Convex Maximization via Adjustable Robust Optimization

Maximizing a convex function over convex constraints is an NP-hard problem in general. We prove that such a problem can be reformulated as an adjustable robust optimization (ARO) problem where each adjustable variable corresponds to a unique constraint of the original problem. We use ARO techniques to obtain approximate solutions to the convex maximization problem. … Read more

A Nonmonotone Matrix-Free Algorithm for Nonlinear Equality-Constrained Least-Squares Problems

Least squares form one of the most prominent classes of optimization problems, with numerous applications in scientific computing and data fitting. When such formulations aim at modeling complex systems, the optimization process must account for nonlinear dynamics by incorporating constraints. In addition, these systems often incorporate a large number of variables, which increases the difficulty … Read more

Iteratively Reweighted Group Lasso based on Log-composite Regularization

This paper addresses supervised learning problems with structured sparsity, where subsets of model coefficients form distinct groups. We introduce a novel log-composite regularizer in a bi-criteria optimization problem together with a loss (e.g., least squares) in order to reconstruct the desired group sparsity structure. We develop an iteratively reweighted algorithm that solves the group LASSO … Read more

Riemannian Optimization on the Symplectic Stiefel Manifold

The symplectic Stiefel manifold, denoted by $\mathrm{Sp}(2p,2n)$, is the set of linear symplectic maps between the standard symplectic spaces $\mathbb{R}^{2p}$ and $\mathbb{R}^{2n}$. When $p=n$, it reduces to the well-known set of $2n\times 2n$ symplectic matrices. Optimization problems on $\mathrm{Sp}(2p,2n)$ find applications in various areas, such as optics, quantum physics, numerical linear algebra and model order … Read more

Inexact Derivative-Free Optimization for Bilevel Learning

Variational regularization techniques are dominant in the field of mathematical imaging. A drawback of these techniques is that they are dependent on a number of parameters which have to be set by the user. A by now common strategy to resolve this issue is to learn these parameters from data. While mathematically appealing this strategy … Read more

Mixed-Integer Optimal Control for Multimodal Chromatography

Multimodal chromatography is a powerful tool in the downstream processing of biopharmaceuticals. To fully benefit from this technology, an efficient process strategy must be determined beforehand. To facilitate this task, we employ a recent mechanistic model for multimodal chromatography, which takes salt concentration and pH into account, and we present a mathematical framework for the … Read more

Cycle-based formulations in Distance Geometry

The distance geometry problem asks to find a realization of a given simple edge-weighted graph in a Euclidean space of given dimension K, where the edges are realized as straight segments of lengths equal (or as close as possible) to the edge weights. The problem is often modelled as a mathematical programming formulation involving decision … Read more

Optimality conditions in discrete-continuous nonlinear optimization

This paper presents necessary and sufficient optimality conditions for discrete-continuous nonlinear optimization problems including mixed-integer nonlinear problems. This theory does not utilize an extension of the Lagrange theory of continuous optimization but it works with certain max functionals for a separation of two sets where one of them is nonconvex. These functionals have the advantage … Read more

A simplified treatment of Ramana’s exact dual for semidefinite programming

In semidefinite programming the dual may fail to attain its optimal value and there could be a duality gap, i.e., the primal and dual optimal values may differ. In a striking paper, Ramana proposed a polynomial size extended dual that does not have these deficiencies and yields a number of fundamental results in complexity theory. … Read more

A Line-Search Descent Algorithm for Strict Saddle Functions with Complexity Guarantees

We describe a line-search algorithm which achieves the best-known worst-case complexity results for problems with a certain “strict saddle” property that has been observed to hold in low-rank matrix optimization problems. Our algorithm is adaptive, in the sense that it makes use of backtracking line searches and does not require prior knowledge of the parameters … Read more