Results 1  10
of
127
On the Implementation of an InteriorPoint Filter LineSearch Algorithm for LargeScale Nonlinear Programming
, 2004
"... We present a primaldual interiorpoint algorithm with a filter linesearch method for nonlinear programming. Local and global convergence properties of this method were analyzed in previous work. Here we provide a comprehensive description of the algorithm, including the feasibility restoration ph ..."
Abstract

Cited by 294 (6 self)
 Add to MetaCart
(Show Context)
We present a primaldual interiorpoint algorithm with a filter linesearch method for nonlinear programming. Local and global convergence properties of this method were analyzed in previous work. Here we provide a comprehensive description of the algorithm, including the feasibility restoration phase for the filter method, secondorder corrections, and inertia correction of the KKT matrix. Heuristics are also considered that allow faster performance. This method has been implemented in the IPOPT code, which we demonstrate in a detailed numerical study based on 954 problems from the CUTEr test set. An evaluation is made of several linesearch options, and a comparison is provided with two stateoftheart interiorpoint codes for nonlinear programming.
A Sparse Signal Reconstruction Perspective for Source Localization With Sensor Arrays
, 2005
"... We present a source localization method based on a sparse representation of sensor measurements with an overcomplete basis composed of samples from the array manifold. We enforce sparsity by imposing penalties based on the 1norm. A number of recent theoretical results on sparsifying properties of ..."
Abstract

Cited by 231 (6 self)
 Add to MetaCart
We present a source localization method based on a sparse representation of sensor measurements with an overcomplete basis composed of samples from the array manifold. We enforce sparsity by imposing penalties based on the 1norm. A number of recent theoretical results on sparsifying properties of 1 penalties justify this choice. Explicitly enforcing the sparsity of the representation is motivated by a desire to obtain a sharp estimate of the spatial spectrum that exhibits superresolution. We propose to use the singular value decomposition (SVD) of the data matrix to summarize multiple time or frequency samples. Our formulation leads to an optimization problem, which we solve efficiently in a secondorder cone (SOC) programming framework by an interior point implementation. We propose a grid refinement method to mitigate the effects of limiting estimates to a grid of spatial locations and introduce an automatic selection criterion for the regularization parameter involved in our approach. We demonstrate the effectiveness of the method on simulated data by plots of spatial spectra and by comparing the estimator variance to the Cramér–Rao bound (CRB). We observe that our approach has a number of advantages over other source localization techniques, including increased resolution, improved robustness to noise, limitations in data quantity, and correlation of the sources, as well as not requiring an accurate initialization.
Interior methods for mathematical programs with complementarity constraints
 SIAM J. Optim
, 2004
"... This paper studies theoretical and practical properties of interiorpenalty methods for mathematical programs with complementarity constraints. A framework for implementing these methods is presented, and the need for adaptive penalty update strategies is motivated with examples. The algorithm is sh ..."
Abstract

Cited by 37 (10 self)
 Add to MetaCart
(Show Context)
This paper studies theoretical and practical properties of interiorpenalty methods for mathematical programs with complementarity constraints. A framework for implementing these methods is presented, and the need for adaptive penalty update strategies is motivated with examples. The algorithm is shown to be globally convergent to strongly stationary points, under standard assumptions. These results are then extended to an interiorrelaxation approach. Superlinear convergence to strongly stationary points is also established. Two strategies for updating the penalty parameter are proposed, and their efficiency and robustness are studied on an extensive collection of test problems.
The interiorpoint revolution in optimization: history, recent developments, and lasting consequences
 BULL. AMER. MATH. SOC. (N.S
, 2005
"... Interior methods are a pervasive feature of the optimization landscape today, but it was not always so. Although interiorpoint techniques, primarily in the form of barrier methods, were widely used during the 1960s for problems with nonlinear constraints, their use for the fundamental problem of ..."
Abstract

Cited by 33 (1 self)
 Add to MetaCart
(Show Context)
Interior methods are a pervasive feature of the optimization landscape today, but it was not always so. Although interiorpoint techniques, primarily in the form of barrier methods, were widely used during the 1960s for problems with nonlinear constraints, their use for the fundamental problem of linear programming was unthinkable because of the total dominance of the simplex method. During the 1970s, barrier methods were superseded, nearly to the point of oblivion, by newly emerging and seemingly more efficient alternatives such as augmented Lagrangian and sequential quadratic programming methods. By the early 1980s, barrier methods were almost universally regarded as a closed chapter in the history of optimization. This picture changed dramatically in 1984, when Narendra Karmarkar announced a fast polynomialtime interior method for linear programming; in 1985, a formal connection was established between his method and classical barrier methods. Since then, interior methods have continued to transform both the theory and practice of constrained optimization. We present a condensed,
Optimal experimental design and some related control problems
, 2008
"... This paper traces the strong relations between experimental design and control, such as the use of optimal inputs to obtain precise parameter estimation in dynamical systems and the introduction of suitably designed perturbations in adaptive control. The mathematical background of optimal experiment ..."
Abstract

Cited by 26 (0 self)
 Add to MetaCart
This paper traces the strong relations between experimental design and control, such as the use of optimal inputs to obtain precise parameter estimation in dynamical systems and the introduction of suitably designed perturbations in adaptive control. The mathematical background of optimal experimental design is briefly presented, and the role of experimental design in the asymptotic properties of estimators is emphasized. Although most of the paper concerns parametric models, some results are also presented for statistical learning and prediction with nonparametric models.
Interior point methods in function space
 SIAM J. Control Optim
"... Abstract. A primaldual interior point method for optimal control problems is considered. The algorithm is directly applied to the infinitedimensional problem. Existence and convergence of the central path are analyzed, and linear convergence of a shortstep pathfollowing method is established. Ke ..."
Abstract

Cited by 21 (3 self)
 Add to MetaCart
Abstract. A primaldual interior point method for optimal control problems is considered. The algorithm is directly applied to the infinitedimensional problem. Existence and convergence of the central path are analyzed, and linear convergence of a shortstep pathfollowing method is established. Key words. interior point methods in function space, optimal control, complementarity functions
PathFollowing Methods for a Class of Constrained Minimization Problems in Function Space
, 2004
"... Pathfollowing methods for primaldual active set strategies requiring a regularization parameter are introduced. Existence of a path and its differentiability properties are analyzed. Monotonicity and convexity of the primaldual path value function are investigated. Both feasible and infeasible ap ..."
Abstract

Cited by 21 (5 self)
 Add to MetaCart
(Show Context)
Pathfollowing methods for primaldual active set strategies requiring a regularization parameter are introduced. Existence of a path and its differentiability properties are analyzed. Monotonicity and convexity of the primaldual path value function are investigated. Both feasible and infeasible approximations are considered. Numerical path following strategies are developed and their efficiency is demonstrated by means of examples.
A PRIMALDUAL TRUST REGION ALGORITHM FOR NONLINEAR OPTIMIZATION
, 2003
"... This paper concerns general (nonconvex) nonlinear optimization when first and second derivatives of the objective and constraint functions are available. The proposed method is based on finding an approximate solution of a sequence of unconstrained subproblems parameterized by a scalar parameter. T ..."
Abstract

Cited by 21 (3 self)
 Add to MetaCart
This paper concerns general (nonconvex) nonlinear optimization when first and second derivatives of the objective and constraint functions are available. The proposed method is based on finding an approximate solution of a sequence of unconstrained subproblems parameterized by a scalar parameter. The objective function of each unconstrained subproblem is an augmented penaltybarrier function that involves both primal and dual variables. Each subproblem is solved using a secondderivative Newtontype method that employs a combined trust region and line search strategy to ensure global convergence. It is shown that the trustregion step can be computed by factorizing a sequence of systems with diagonallymodified primaldual structure, where the inertia of these systems can be determined without recourse to a special factorization method. This has the benefit that offtheshelf linear system software can be used at all times, allowing the straightforward extension to largescale problems. Numerical results are given for problems in the COPS test collection.
Iterative solution of augmented systems arising in interior methods
 SIAM JOURNAL ON OPTIMIZATION
, 2007
"... Iterative methods are proposed for certain augmented systems of linear equations that arise in interior methods for general nonlinear optimization. Interior methods define a sequence of KKT equations that represent the symmetrized (but indefinite) equations associated with Newton’s method for a po ..."
Abstract

Cited by 20 (1 self)
 Add to MetaCart
(Show Context)
Iterative methods are proposed for certain augmented systems of linear equations that arise in interior methods for general nonlinear optimization. Interior methods define a sequence of KKT equations that represent the symmetrized (but indefinite) equations associated with Newton’s method for a point satisfying the perturbed optimality conditions. These equations involve both the primal and dual variables and become increasingly illconditioned as the optimization proceeds. In this context, an iterative linear solver must not only handle the illconditioning but also detect the occurrence of KKT matrices with the wrong matrix inertia. A oneparameter family of equivalent linear equations is formulated that includes the KKT system as a special case. The discussion focuses on a particular system from this family, known as the “doubly augmented system, ” that is positive definite with respect to both the primal and dual variables. This property means that a standard preconditioned conjugategradient method involving both primal and dual variables will either terminate successfully or detect if the KKT matrix has the wrong inertia. Constraint preconditioning is a wellknown technique for preconditioning the conjugategradient method on augmented systems. A family of constraint preconditioners is proposed that provably eliminates the inherent illconditioning in the augmented system. A considerable benefit of combining constraint preconditioning with the doubly augmented system is that the preconditioner need not be applied exactly. Two particular “activese ” constraint preconditioners are formulated that involve only a subset of the rows of the augmented system and thereby may be applied with considerably less work. Finally, some numerical experiments illustrate the numerical performance of the proposed preconditioners and highlight some theoretical properties of the preconditioned matrices.
Interiorpoint methods for optimization
, 2008
"... This article describes the current state of the art of interiorpoint methods (IPMs) for convex, conic, and general nonlinear optimization. We discuss the theory, outline the algorithms, and comment on the applicability of this class of methods, which have revolutionized the field over the last twen ..."
Abstract

Cited by 18 (0 self)
 Add to MetaCart
This article describes the current state of the art of interiorpoint methods (IPMs) for convex, conic, and general nonlinear optimization. We discuss the theory, outline the algorithms, and comment on the applicability of this class of methods, which have revolutionized the field over the last twenty years.