Perturbation resilience and superiorization of iterative algorithms

Iterative algorithms aimed at solving some problems are discussed. For certain problems, such as finding a common point in the intersection of a finite number of convex sets, there often exist iterative algorithms that impose very little demand on computer resources. For other problems, such as finding that point in the intersection at which the … Read more

On Equivalence of Semidefinite Relaxations for Quadratic Matrix Programming

In this paper, we analyze two popular semidefinite programming \SDPb relaxations for quadratically constrained quadratic programs \QCQPb with matrix variables. These are based on \emph{vector-lifting} and on \emph{matrix lifting} and are of different size and expense. We prove, under mild assumptions, that these two relaxations provide equivalent bounds. Thus, our results provide a theoretical guideline … Read more

Achieving Higher Frequencies in Large-Scale Nonlinear Model Predictive Control

We present new insights into how to achieve higher frequencies in large-scale nonlinear predictive control using truncated-like schemes. The basic idea is that, instead of solving the full nonlinear programming (NLP) problem at each sampling time, we solve a single, truncated quadratic programming (QP) problem. We present conditions guaranteeing stability of the approximation error derived … Read more

A Gauss-Newton approach for solving constrained optimization problems using differentiable exact penalties

We propose a Gauss-Newton-type method for nonlinear constrained optimization using the exact penalty introduced recently by Andre and Silva for variational inequalities. We extend their penalty function to both equality and inequality constraints using a weak regularity assumption, and as a result, we obtain a continuously differentiable exact penalty function and a new reformulation of … Read more

A concave optimization-based approach for sparse portfolio selection

This paper considers a portfolio selection problem in which portfolios with minimum number of active assets are sought. This problem is motivated by the need of inducing sparsity on the selected portfolio to reduce transaction costs, complexity of portfolio management, and instability of the solution. The resulting problem is a difficult combinatorial problem. We propose … Read more

^phBcnorms, log-barriers and Cramer transform in optimization

We show that the Laplace approximation of a supremum by $L^p$-norms has interesting consequences in optimization. For instance, the logarithmic barrier functions (LBF) of a primal convex problem $P$ and its dual $P^*$ appear naturally when using this simple approximation technique for the value function $g$ of $P$ or its Legendre-Fenchel conjugate $g^*$. In addition, … Read more

A Robust Implementation of a Sequential Quadratic Programming Algorithm with Successive Error Restoration

We consider sequential quadratic programming (SQP) methods for solving constrained nonlinear programming problems. It is generally believed that SQP methods are sensitive to the accuracy by which partial derivatives are provided. One reason is that differences of gradients of the Lagrangian function are used for updating a quasi-Newton matrix, e.g., by the BFGS formula. The … Read more

A Sequential Quadratic Programming Algorithm for Nonconvex, Nonsmooth Constrained Optimization

We consider optimization problems with objective and constraint functions that may be nonconvex and nonsmooth. Problems of this type arise in important applications, many having solutions at points of nondifferentiability of the problem functions. We present a line search algorithm for situations when the objective and constraint functions are locally Lipschitz and continuously differentiable on … Read more

A sufficiently exact inexact Newton step based on reusing matrix information

Newton’s method is a classical method for solving a nonlinear equation $F(z)=0$. We derive inexact Newton steps that lead to an inexact Newton method, applicable near a solution. The method is based on solving for a particular $F'(z_{k’})$ during $p$ consecutive iterations $k=k’,k’+1,\dots,k’+p-1$. One such $p$-cycle requires $2^p-1$ solves with the matrix $F'(z_{k’})$. If matrix … Read more

MathOptimizer: A nonlinear optimization package for Mathematica users

Mathematica is an advanced software system that enables symbolic computing, numerics, program code development, model visualization and professional documentation in a unified framework. Our MathOptimizer software package serves to solve global and local optimization models developed using Mathematica. We introduce MathOptimizer’s key features and discuss its usage options that support a range of operational modes. … Read more