Elementary optimality conditions for nonlinear SDPs

The goal of this paper is an easy and self-contained presentation of optimality conditions for nonlinear semidefinite programs concentrating on the differences between nonlinear semidefinite programs and nonlinear programs. CitationTechnical Report, Department of Mathematics, Universit\”at D\”usseldorf.ArticleDownload View PDF

On convex relaxations for quadratically constrained quadratic programming

We consider convex relaxations for the problem of minimizing a (possibly nonconvex) quadratic objective subject to linear and (possibly nonconvex) quadratic constraints. Let F denote the feasible region for the linear constraints. We first show that replacing the quadratic objective and constraint functions with their convex lower envelopes on F is dominated by an alternative … Read more

A Linearly Convergent Algorithm for Solving a Class of Nonconvex/Affine Feasibility Problems

We introduce a class of nonconvex/affine feasibility problems, called (NCF), that consists of finding a point in the intersection of affine constraints with a nonconvex closed set. This class captures some interesting fundamental and NP hard problems arising in various application areas such as sparse recovery of signals and affine rank minimization that we briefly … Read more

Efficient preconditioner updates for shifted linear systems

We present a new technique for building effective and low cost preconditioners for sequences of shifted linear systems (A+aI)x=b, where A is symmetric positive definite and a>0. This technique updates a preconditioner for A, available in the form of an LDL’ factorization, by modifying only the nonzero entries of the L factor in such a … Read more

A Practical Relative Error Criterion for Augmented Lagrangians

This paper develops a new error criterion for the approximate minimization of augmented Lagrangian subproblems. This criterion is practical in the sense that it requires only information that is ordinarily readily available, such as the gradient (or a subgradient) of the augmented Lagrangian. It is also “relative” in the sense of relative error criteria for … Read more

New concave penalty functions for improving the Feasibility Pump

Mixed-Integer optimization represents a powerful tool for modeling many optimization problems arising from real-world applications. The Feasibility pump is a heuristic for finding feasible solutions to mixed integer linear problems. In this work, we propose a new feasibility pump approach using concave non-differentiable penalty functions for measuring solution integrality. We present computational results on binary … Read more

Derivative-free methods for nonlinear programming with general lower-level constraints

Augmented Lagrangian methods for derivative-free continuous optimization with constraints are introduced in this paper. The algorithms inherit the convergence results obtained by Andreani, Birgin, Martínez and Schuverdt for the case in which analytic derivatives exist and are available. In particular, feasible limit points satisfy KKT conditions under the Constant Positive Linear Dependence (CPLD) constraint qualification. … Read more

A Penalty-Interior-Point Algorithm for Nonlinear Constrained Optimization

Penalty and interior-point methods for nonlinear optimization problems have enjoyed great successes for decades. Penalty methods have proved to be effective for a variety of problem classes due to their regularization effects on the constraints. They have also been shown to allow for rapid infeasibility detection. Interior-point methods have become the workhorse in large-scale optimization … Read more

Augmented Lagrangian method with nonmonotone penalty parameters for constrained optimization

At each outer iteration of standard Augmented Lagrangian methods one tries to solve a box-constrained optimization problem with some prescribed tolerance. In the continuous world, using exact arithmetic, this subproblem is always solvable. Therefore, the possibility of finishing the subproblem resolution without satisfying the theoretical stopping conditions is not contemplated in usual convergence theories. However, … Read more

A relaxed constant positive linear dependence constraint qualification and applications

In this work we introduce a relaxed version of the constant positive linear dependence constraint qualification (CPLD) that we call RCPLD. This development is inspired by a recent generalization of the constant rank constraint qualification from Minchenko and Stakhovski that was called RCR. We show that RCPLD is enough to ensure the convergence of an … Read more