Approximate norm descent methods for constrained nonlinear systems

We address the solution of convex-constrained nonlinear systems of equations where the Jacobian matrix is unavailable or its computation/storage is burdensome. In order to efficiently solve such problems, we propose a new class of algorithms which are “derivative-free” both in the computation of the search direction and in the selection of the steplength. Search directions … Read more

Second-order optimality and beyond: characterization and evaluation complexity in convexly-constrained nonlinear optimization

High-order optimality conditions for convexly-constrained nonlinear optimization problems are analyzed. A corresponding (expensive) measure of criticality for arbitrary order is proposed and extended to define high-order $\epsilon$-approximate critical points. This new measure is then used within a conceptual trust-region algorithm to show that, if derivatives of the objective function up to order $q \geq 1$ … Read more

Low-Rank Matrix Completion using Nuclear Norm with Facial Reduction

Minimization of the nuclear norm is often used as a surrogate, convex relaxation, for finding the minimum rank completion (recovery) of a partial matrix. The minimum nuclear norm problem can be solved as a trace minimization semidefinite programming problem (\SDP). The \SDP and its dual are regular in the sense that they both satisfy strict … Read more

On a Practical Notion of Geoffrion Proper Optimality in Multicriteria Optimization

Geoffrion proper optimality is a widely used optimality notion in multicriteria optimization that prevents exact solutions having unbounded trade-offs. As algorithms for multicriteria optimization usually give only approximate solutions, we analyze the notion of approximate Geoffrion proper optimality. We show that in the limit, approximate Geoffrion proper optimality may converge to solutions having unbounded trade-offs. … Read more

A new branch-and-bound algorithm for standard quadratic programming problems

In this paper we propose convex and LP bounds for Standard Quadratic Programming (StQP) problems and employ them within a branch-and-bound approach. We first compare different bounding strategies for StQPs in terms both of the quality of the bound and of the computation times. It turns out that the polyhedral bounding strategy is the best … Read more

Asymptotical Analysis of a SAA Estimator for Optimal Value of a Two Stage Problem with Quadratic Recourse

In this paper, we first consider the stability analysis of a convex quadratic programming problem and its restricted Wolfe dual in which all parameters in the problem are perturbed. We demonstrate the upper semi-continuity of solution mappings for the primal problem and the restricted Wolfe dual problem and establish the Hadamard directionally differentiability of the … Read more

Exploiting Problem Structure in Optimization under Uncertainty via Online Convex Optimization

In this paper, we consider two paradigms that are developed to account for uncertainty in optimization models: robust optimization (RO) and joint estimation-optimization (JEO). We examine recent developments on efficient and scalable iterative first-order methods for these problems, and show that these iterative methods can be viewed through the lens of online convex optimization (OCO). … Read more

An inexact dual logarithmic barrier method for solving sparse semidefinite programs

A dual logarithmic barrier method for solving large, sparse semidefinite programs is proposed in this paper. The method avoids any explicit use of the primal variable X and therefore is well-suited to problems with a sparse dual matrix S. It relies on inexact Newton steps in dual space which are computed by the conjugate gradient … Read more

A Two-Stage Active-Set Algorithm for Bound-Constrained Optimization

In this paper, we describe a two-stage method for solving optimization problems with bound constraints. It combines the active-set estimate described in [Facchinei and Lucidi, 1995] with a modification of the non-monotone line search framework recently proposed in [De Santis et al., 2012]. In the first stage, the algorithm exploits a property of the active-set … Read more

Step lengths in BFGS method for monotone gradients

In this paper, we consider how to directly apply the BFGS method to finding a zero point of any given monotone gradient and thus suggest new conditions to locate the corresponding step lengths. The suggested conditions involve curvature condition and merely use gradients’ computations. Furthermore, they can guarantee convergence without any other restrictions. Finally, preliminary … Read more