• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations
Advanced Search Include Citations

Convergence of stochastic proximal gradient algorithm, (2014)

by L Rosasco, S Villa, B C Vu
Add To MetaCart

Tools

Sorted by:
Results 1 - 5 of 5

A STOCHASTIC FORWARD-BACKWARD SPLITTING METHOD FOR SOLVING MONOTONE INCLUSIONS IN HILBERT SPACES

by Lorenzo Rosasco, Silvia Villa, B, Ang Công Vu ̃
"... ar ..."
Abstract - Cited by 1 (0 self) - Add to MetaCart
Abstract not found
(Show Context)

Citation Context

...ties, we obtain an additional convergence result without imposing stronger monotonicity properties on B, which requires averaging of the iterates. The present paper extends a short conference version =-=[47]-=- restricted to the minimization case. The paper is organized as follows. We first review related work in Section 2. Section 3 collects some preliminaries and Section 4 contains the main results of the...

using

by Pascal Bianchi, Walid Hachem , 2015
"... behavior of a stochastic forward-backward algorithm ..."
Abstract - Add to MetaCart
behavior of a stochastic forward-backward algorithm
(Show Context)

Citation Context

...eded, and the Lipschitz property is essentially replaced with the condition ‖f(ξ, x)‖2 ≤ M(ξ)(1 + ‖x‖) where M(ξ) satisfies a moment condition. Regarding the convergence rate analysis, let us mention =-=[3, 35]-=- which investigate the performance of the algorithm xn+1 = proxγn+1g(xn − γn+1Hn+1) where Hn+1 is a noisy estimate of the gradient ∇f(xn). The same algorithm is addressed in [36] where the proximity o...

Stochastic Approximations and Perturbations in Forward-Backward Splitting for Monotone Operators *

by Patrick L Combettes , Jean-Christophe Pesquet
"... Abstract We investigate the asymptotic behavior of a stochastic version of the forward-backward splitting algorithm for finding a zero of the sum of a maximally monotone set-valued operator and a cocoercive operator in Hilbert spaces. Our general setting features stochastic approximations of the co ..."
Abstract - Add to MetaCart
Abstract We investigate the asymptotic behavior of a stochastic version of the forward-backward splitting algorithm for finding a zero of the sum of a maximally monotone set-valued operator and a cocoercive operator in Hilbert spaces. Our general setting features stochastic approximations of the cocoercive operator and stochastic perturbations in the evaluation of the resolvents of the set-valued operator. In addition, relaxations and not necessarily vanishing proximal parameters are allowed. Weak and strong almost sure convergence properties of the iterates is established under mild conditions on the underlying stochastic processes. Leveraging these results, we also establish the almost sure convergence of the iterates of a stochastic variant of a primal-dual proximal splitting method for composite minimization problems.
(Show Context)

Citation Context

...1.5) The first instances of the stochastic iteration (1.5) can be traced back to [44] in the context of the gradient method, i.e., when A = 0 and B is the gradient of a convex function. Stochastic approximations in the gradient method were then investigated in the Russian literature of the late 1960s and early 1970s [27, 28, 29, 33, 42, 49]. Stochastic gradient methods have also been used extensively in adaptive signal processing, in control, and in machine learning, e.g., [3, 36, 54]. More generally, proximal stochastic gradient methods have been applied to various problems; see for instance [1, 26, 45, 48, 55]. The objective of the present paper is to provide an analysis of the stochastic forward-backward method in the context of Algorithm 1.3. Almost sure convergence of the iterates (xn)n∈N to a solution to Problem 1.1 will be established under general conditions on the sequences (un)n∈N, (an)n∈N, (γn)n∈N, and (λn)n∈N. In particular, a feature of our analysis is that it allows for relaxation parameters and it does not require that the proximal parameter sequence (γn)n∈N be vanishing. Our proofs are based on properties of stochastic quasi-Fejer iterations [18], for which we provide a novel converg...

Stochastic Optimization for Kernel PCA∗

by Lijun Zhang, Tianbao Yang, Jinfeng Yi, Rong Jin, Zhi-hua Zhou
"... Kernel Principal Component Analysis (PCA) is a popular ex-tension of PCA which is able to find nonlinear patterns from data. However, the application of kernel PCA to large-scale problems remains a big challenge, due to its quadratic space complexity and cubic time complexity in the number of ex-amp ..."
Abstract - Add to MetaCart
Kernel Principal Component Analysis (PCA) is a popular ex-tension of PCA which is able to find nonlinear patterns from data. However, the application of kernel PCA to large-scale problems remains a big challenge, due to its quadratic space complexity and cubic time complexity in the number of ex-amples. To address this limitation, we utilize techniques from stochastic optimization to solve kernel PCA with linear s-pace and time complexities per iteration. Specifically, we for-mulate it as a stochastic composite optimization problem, where a nuclear norm regularizer is introduced to promote low-rankness, and then develop a simple algorithm based on stochastic proximal gradient descent. During the optimization process, the proposed algorithm always maintains a low-rank factorization of iterates that can be conveniently held in mem-ory. Compared to previous iterative approaches, a remarkable property of our algorithm is that it is equipped with an ex-plicit rate of convergence. Theoretical analysis shows that the solution of our algorithm converges to the optimal one at an O(1/T) rate, where T is the number of iterations.

RAPID: Rapidly Accelerated Proximal Gradient Algorithms for Convex Minimization

by Ziming Zhang, Venkatesh Saligrama
"... In this paper, we propose a new algorithm to speed-up the convergence of accel-erated proximal gradient (APG) methods. In order to minimize a convex function f(x), our algorithm introduces a simple line search step after each proximal gra-dient step in APG so that a biconvex function f(θx) is minimi ..."
Abstract - Add to MetaCart
In this paper, we propose a new algorithm to speed-up the convergence of accel-erated proximal gradient (APG) methods. In order to minimize a convex function f(x), our algorithm introduces a simple line search step after each proximal gra-dient step in APG so that a biconvex function f(θx) is minimized over scalar variable θ> 0 while fixing variable x. We propose two new ways of constructing the auxiliary variables in APG based on the intermediate solutions of the proxi-mal gradient and the line search steps. We prove that at arbitrary iteration step t(t ≥ 1), our algorithm can achieve a smaller upper-bound for the gap between the current and optimal objective values than those in the traditional APG methods such as FISTA [4], making it converge faster in practice. In fact, our algorithm can be potentially applied to many important convex optimization problems, such as sparse linear regression and kernel SVMs. Our experimental results clearly demonstrate that our algorithm converges faster than APG in all of the applica-tions above, even comparable to some sophisticated solvers. 1
(Show Context)

Citation Context

...y. Inexact proximal gradient methods [17] allow to approximate the proximal gradients with controllable errors in a faster way while guaranteeing the convergence. Stochastic proximal gradient methods =-=[1, 15]-=- allow to compute the proximal gradients using a small set of data in a stochastic fashion while guaranteeing the convergence as well. Distributed proximal gradient methods [6] decompose the optimizat...

Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University