• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations
Advanced Search Include Citations | Disambiguate

Low-rank solutions of linear matrix equations via procrustes flow. arXiv preprint arXiv:1507.03566, (2015)

by Stephen Tu, Ross Boczar, Mahdi Soltanolkotabi, Benjamin Recht
Add To MetaCart

Tools

Sorted by:
Results 1 - 4 of 4

A convergent gradient descent algorithm for rank minimization and semidefinite programming from random linear measurements. arXiv preprint arXiv:1506.06081

by Qinqing Zheng , John Lafferty , 2015
"... Abstract We propose a simple, scalable, and fast gradient descent algorithm to optimize a nonconvex objective for the rank minimization problem and a closely related family of semidefinite programs. With O(r 3 κ 2 n log n) random measurements of a positive semidefinite n×n matrix of rank r and cond ..."
Abstract - Cited by 4 (0 self) - Add to MetaCart
Abstract We propose a simple, scalable, and fast gradient descent algorithm to optimize a nonconvex objective for the rank minimization problem and a closely related family of semidefinite programs. With O(r 3 κ 2 n log n) random measurements of a positive semidefinite n×n matrix of rank r and condition number κ, our method is guaranteed to converge linearly to the global optimum.
(Show Context)

Citation Context

... AltMinSense converge to X? linearly under a RIP condition. However, the least squares problems are often ill-conditioned, it is difficult to observe AltMinSense converging to X? in practice. As described above, considerable progress has been made on algorithms for rank minimization and certain semidefinite programming problems. Yet truly efficient, scalable and provably convergent 3 algorithms have not yet been obtained. In the specific setting that X? is positive semidefinite, our algorithm exploits this structure to achieve these goals. We note that recent and independent work of Tu et al. [21] proposes a hybrid algorithm called Procrustes Flow (PF), which uses a few iterations of SVP as initialization, and then applies gradient descent. 4 A Gradient Descent Algorithm for Rank Minimization Our method is described in Algorithm 1. It is parallel to the Wirtinger Flow (WF) algorithm for phase retrieval [6], to recover a complex vector x ∈ Cn given the squared magnitudes of its linear measurements bi = |〈ai, x〉|2, i ∈ [m], where a1, . . . , am ∈ Cn. Candes et al. [6] propose a first-order method to minimize the sum of squared residuals fWF(z) = n∑ i=1 ( |〈ai, z〉|2 − bi )2 . (5) The aut...

Fast Algorithms for Robust PCA via Gradient Descent

by Xinyang Yi , Dohyung Park , Yudong Chen , Constantine Caramanis
"... Abstract We consider the problem of Robust PCA in the fully and partially observed settings. Without corruptions, this is the well-known matrix completion problem. From a statistical standpoint this problem has been recently well-studied, and conditions on when recovery is possible (how many observ ..."
Abstract - Add to MetaCart
Abstract We consider the problem of Robust PCA in the fully and partially observed settings. Without corruptions, this is the well-known matrix completion problem. From a statistical standpoint this problem has been recently well-studied, and conditions on when recovery is possible (how many observations do we need, how many corruptions can we tolerate) via polynomial-time algorithms is by now understood. This paper presents and analyzes a non-convex optimization approach that greatly reduces the computational complexity of the above problems, compared to the best available algorithms. In particular, in the fully observed case, with r denoting rank and d dimension, we reduce the complexity from O(r 2 d 2 log(1/ε)) to O(rd 2 log(1/ε)) -a big savings when the rank is big. For the partially observed case, we show the complexity of our algorithm is no more than O(r 4 d log d log(1/ε)). Not only is this the best-known run-time for a provable algorithm under partial observation, but in the setting where r is small compared to d, it also allows for near-linear-in-d run-time that can be exploited in the fully-observed case as well, by simply running our algorithm on a subset of the observations.
(Show Context)

Citation Context

... d. Work in [21] considers the deterministic corruption model, and improves this running time without sacrificing the robustness guarantee on α. They propose an alternating projection (AltProj) method to estimate the low rank and sparse structures iteratively and simultaneously, and show their algorithm has complexity O(r2d2 log(1/ε)), which is faster than the convex approach but still slower than SVD. Non-convex approaches have recently seen numerous developments for applications in low-rank estimation, including alternating minimization (see e.g. [19, 17, 16]) and gradient descent (see e.g. [4, 12, 23, 24, 29, 30]). These works have fast running times, yet do not provide robustness guarantees. One exception is [12], where the authors analyze a row-wise `1 projection method for recovering S∗. Their analysis hinges on positive semidefinite M∗, and the algorithm requires prior knowledge of the `1 norm of every row of S∗ and is thus prohibitive in practice. Another exception is work [16], which analyzes alternating minimization plus an overall sparse projection. Their algorithm is shown to tolerate at most a fraction of α = O(1/(µ2/3r2/3d)) corruptions. As we discuss in Section 1.2, we can allow S∗ to have...

A Geometric Analysis of Phase Retrieval

by Ju Sun , Qing Qu , John Wright
"... Abstract Can we recover a complex signal from its Fourier magnitudes? More generally, given a set of m measurements, y k = |a * k x| for k = 1, . . . , m, is it possible to recover x ∈ C n (i.e., length-n complex vector)? This generalized phase retrieval (GPR) problem is a fundamental task in vario ..."
Abstract - Add to MetaCart
Abstract Can we recover a complex signal from its Fourier magnitudes? More generally, given a set of m measurements, y k = |a * k x| for k = 1, . . . , m, is it possible to recover x ∈ C n (i.e., length-n complex vector)? This generalized phase retrieval (GPR) problem is a fundamental task in various disciplines, and has been the subject of much recent investigation. Natural nonconvex heuristics often work remarkably well for GPR in practice, but lack clear theoretical explanations. In this paper, we take a step towards bridging this gap. We prove that when the measurement vectors a k 's are generic (i.i.d. complex Gaussian) and the number of measurements is large enough (m ≥ Cn log 3 n), with high probability, a natural least-squares formulation for GPR has the following benign geometric structure: (1) there are no spurious local minimizers, and all global minimizers are equal to the target signal x, up to a global phase; and (2) the objective function has a negative curvature around each saddle point. This structure allows a number of iterative optimization methods to efficiently find a global minimizer, without special initialization. To corroborate the claim, we describe and analyze a second-order trust-region algorithm.
(Show Context)

Citation Context

...rieval problem has a natural generalization to recovering low-rank positive semidefinite matrices. Consider the problem of recovering an unknown rank-r matrix M 0 in Rn×n from linear measurement of the form zk = tr(AkM) with symmetricAk for k = 1, . . . ,m. One can solve the problem by considering the “factorized” version: recoveringX ∈ Rn×r (up to right invertible transform) from measurements zk = tr(X ∗AkX). This is a natural generalization of GPR, as one can write the GPR measurements as y2k = |a∗kx| 2 = x∗(aka ∗ k)x. This generalization and related problems have recently been studied in [SRO15, ZL15, TBSR15, CW15]. 1.5 Notations, Organization, and Reproducible Research Basic notations and facts. Throughout the paper, we define complex inner product as: 〈a, b〉 .= a∗b for any a, b ∈ Cn. We use CSn−1 for the complex unit sphere in Cn. CSn−1(λ) with λ > 0 denotes the centered complex sphere with radius λ in Cn. Similarly, we use CBn(λ) to denote the centered complex ball of radius λ. We use CN (k) for a standard complex Gaussian vector of length k defined in (1.2). We reserve C and c, and their indexed versions to denote absolute constants. Their value vary with the context. Let < (z) ∈ Rn and =(z) ∈ Rn de...

Nonconvex Low Rank Matrix Factorization via Inexact First Order Oracle

by Tuo Zhao, Zhaoran Wang, Han Liu
"... We study the low rank matrix factorization problem via nonconvex optimization. Com-pared with the convex relaxation approach, nonconvex optimization exhibits superior empirical performance for large scale low rank matrix estimation. However, the understanding of its theo-retical guarantees is limite ..."
Abstract - Add to MetaCart
We study the low rank matrix factorization problem via nonconvex optimization. Com-pared with the convex relaxation approach, nonconvex optimization exhibits superior empirical performance for large scale low rank matrix estimation. However, the understanding of its theo-retical guarantees is limited. To bridge this gap, we exploit the notion of inexact first order oracle, which naturally appears in low rank matrix factorization problems such as matrix sensing and completion. Particularly, our analysis shows that a broad class of nonconvex optimization algo-rithms, including alternating minimization and gradient-type methods, can be treated as solving two sequences of convex optimization algorithms using inexact first order oracle. Thus we can show that these algorithms converge geometrically to the global optima and recover the true low rank matrices under suitable conditions. Numerical results are provided to support our theory. 1
Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University