Results 1  10
of
740,622
LucasKanade 20 Years On: A Unifying Framework: Part 3
 International Journal of Computer Vision
, 2002
"... Since the LucasKanade algorithm was proposed in 1981 image alignment has become one of the most widely used techniques in computer vision. Applications range from optical flow, tracking, and layered motion, to mosaic construction, medical image registration, and face coding. Numerous algorithms hav ..."
Abstract

Cited by 698 (30 self)
 Add to MetaCart
first consider linear appearance variation when the error function is the Euclidean L2 norm. We describe three different algorithms, the simultaneous, project out, and normalization inverse compositional algorithms, and empirically compare them. Afterwards we consider the combination of linear
For Most Large Underdetermined Systems of Linear Equations the Minimal ℓ1norm Solution is also the Sparsest Solution
 Comm. Pure Appl. Math
, 2004
"... We consider linear equations y = Φα where y is a given vector in R n, Φ is a given n by m matrix with n < m ≤ An, and we wish to solve for α ∈ R m. We suppose that the columns of Φ are normalized to unit ℓ 2 norm 1 and we place uniform measure on such Φ. We prove the existence of ρ = ρ(A) so that ..."
Abstract

Cited by 560 (10 self)
 Add to MetaCart
We consider linear equations y = Φα where y is a given vector in R n, Φ is a given n by m matrix with n < m ≤ An, and we wish to solve for α ∈ R m. We suppose that the columns of Φ are normalized to unit ℓ 2 norm 1 and we place uniform measure on such Φ. We prove the existence of ρ = ρ(A) so
Guaranteed minimumrank solutions of linear matrix equations via nuclear norm minimization
, 2007
"... The affine rank minimization problem consists of finding a matrix of minimum rank that satisfies a given system of linear equality constraints. Such problems have appeared in the literature of a diverse set of fields including system identification and control, Euclidean embedding, and collaborative ..."
Abstract

Cited by 568 (23 self)
 Add to MetaCart
The affine rank minimization problem consists of finding a matrix of minimum rank that satisfies a given system of linear equality constraints. Such problems have appeared in the literature of a diverse set of fields including system identification and control, Euclidean embedding
Polynomial time approximation schemes for Euclidean TSP and other geometric problems
 In Proceedings of the 37th IEEE Symposium on Foundations of Computer Science (FOCS’96
, 1996
"... Abstract. We present a polynomial time approximation scheme for Euclidean TSP in fixed dimensions. For every fixed c � 1 and given any n nodes in � 2, a randomized version of the scheme finds a (1 � 1/c)approximation to the optimum traveling salesman tour in O(n(log n) O(c) ) time. When the nodes a ..."
Abstract

Cited by 399 (3 self)
 Add to MetaCart
Abstract. We present a polynomial time approximation scheme for Euclidean TSP in fixed dimensions. For every fixed c � 1 and given any n nodes in � 2, a randomized version of the scheme finds a (1 � 1/c)approximation to the optimum traveling salesman tour in O(n(log n) O(c) ) time. When the nodes
Approximate Nearest Neighbors: Towards Removing the Curse of Dimensionality
, 1998
"... The nearest neighbor problem is the following: Given a set of n points P = fp 1 ; : : : ; png in some metric space X, preprocess P so as to efficiently answer queries which require finding the point in P closest to a query point q 2 X. We focus on the particularly interesting case of the ddimens ..."
Abstract

Cited by 1017 (40 self)
 Add to MetaCart
dimensional Euclidean space where X = ! d under some l p norm. Despite decades of effort, the current solutions are far from satisfactory; in fact, for large d, in theory or in practice, they provide little improvement over the bruteforce algorithm which compares the query point to each data point. Of late
The selfduality equations on a Riemann surface
 Proc. Lond. Math. Soc., III. Ser
, 1987
"... In this paper we shall study a special class of solutions of the selfdual YangMills equations. The original selfduality equations which arose in mathematical physics were defined on Euclidean 4space. The physically relevant solutions were the ones with finite action—the socalled 'instanton ..."
Abstract

Cited by 524 (6 self)
 Add to MetaCart
In this paper we shall study a special class of solutions of the selfdual YangMills equations. The original selfduality equations which arose in mathematical physics were defined on Euclidean 4space. The physically relevant solutions were the ones with finite action—the socalled &apos
GMRES: A generalized minimal residual algorithm for solving nonsymmetric linear systems
 SIAM J. SCI. STAT. COMPUT
, 1986
"... We present an iterative method for solving linear systems, which has the property ofminimizing at every step the norm of the residual vector over a Krylov subspace. The algorithm is derived from the Arnoldi process for constructing an l2orthogonal basis of Krylov subspaces. It can be considered a ..."
Abstract

Cited by 2046 (40 self)
 Add to MetaCart
We present an iterative method for solving linear systems, which has the property ofminimizing at every step the norm of the residual vector over a Krylov subspace. The algorithm is derived from the Arnoldi process for constructing an l2orthogonal basis of Krylov subspaces. It can be considered
Learning the Kernel Matrix with SemiDefinite Programming
, 2002
"... Kernelbased learning algorithms work by embedding the data into a Euclidean space, and then searching for linear relations among the embedded data points. The embedding is performed implicitly, by specifying the inner products between each pair of points in the embedding space. This information ..."
Abstract

Cited by 780 (22 self)
 Add to MetaCart
Kernelbased learning algorithms work by embedding the data into a Euclidean space, and then searching for linear relations among the embedded data points. The embedding is performed implicitly, by specifying the inner products between each pair of points in the embedding space. This information
Directional Statistics and Shape Analysis
, 1995
"... There have been various developments in shape analysis in the last decade. We describe here some relationships of shape analysis with directional statistics. For shape, rotations are to be integrated out or to be optimized over whilst they are the basis for directional statistics. However, various c ..."
Abstract

Cited by 775 (31 self)
 Add to MetaCart
to shape analysis. Note that the idea of using tangent space for analysis is common to both manifold as well. 1 Introduction Consider shapes of configurations of points in Euclidean space. There are various contexts in which k labelled points (or "landmarks") x 1 ; :::; x k in IR m are given
Ktheory for operator algebras
 Mathematical Sciences Research Institute Publications
, 1998
"... p. XII line5: since p. 12: I blew this simple formula: should be α = −〈ξ, η〉/〈η, η〉. p. 2 I.1.1.4: The RieszFischer Theorem is often stated this way today, but neither Riesz nor Fischer (who worked independently) phrased it in terms of completeness of the orthogonal system {e int}. If [a, b] is a ..."
Abstract

Cited by 559 (0 self)
 Add to MetaCart
space is not σfinite. p. 13: add after I.2.6.16: I.2.6.17. If X is a compact subset of C not containing 0, and k ∈ N, there is in general no bound on the norm of T −1 as T ranges over all operators with ‖T ‖ ≤ k and σ(T) ⊆ X. For example, let Sn ∈ L(l 2) be the truncated shift: Sn(α1, α2,...) = (0
Results 1  10
of
740,622