Copositivity for second-order optimality conditions in general smooth optimization problems

Second-order local optimality conditions involving copositivity of the Hessian of the Lagrangian on the reduced linearization cone have the advantage that there is only a small gap between sufficient (the Hessian is strictly copositive) and necessary (the Hessian is copositive) conditions. In this respect, this is a proper generalization of convexity of the Lagrangian. We … Read more

On the application of the spectral projected gradient method in image segmentation

We investigate the application of the nonmonotone spectral projected gradient (SPG) method to a region-based variational model for image segmentation. We consider a “discretize-then-optimize” approach and solve the resulting nonlinear optimization problem by an alternating minimization procedure that exploits the SPG2 algorithm by Birgin, Martì­nez and Raydan (SIAM J. Optim., 10(4), 2000). We provide a … Read more

A cone-continuity constraint qualification and algorithmic consequences

Every local minimizer of a smooth constrained optimization problem satisfies the sequential Approximate Karush-Kuhn-Tucker (AKKT) condition. This optimality condition is used to define the stopping criteria of many practical nonlinear programming algorithms. It is natural to ask for conditions on the constraints under which AKKT implies KKT. These conditions will be called Strict Constraint Qualifications … Read more

Alternating direction methods for non convex optimization with applications to second-order least-squares and risk parity portfolio selection

In this paper we mainly focus on optimization of sums of squares of quadratic functions, which we refer to as second-order least-squares problems, subject to convex constraints. Our motivation arises from applications in risk parity portfolio selection. We generalize the setting further by considering a class of nonlinear, non convex functions which admit a (non … Read more

A Fast Branch-and-Bound Algorithm for Non-convex Quadratic Integer Optimization Subject To Linear Constraints Using Ellipsoidal Relaxations

We propose two exact approaches for non-convex quadratic integer minimization subject to linear constraints where lower bounds are computed by considering ellipsoidal relaxations of the feasible set. In the first approach, we intersect the ellipsoids with the feasible linear subspace. In the second approach we penalize exactly the linear constraints. We investigate the connection between … Read more

On the Quadratic Shortest Path Problem

Finding the shortest path in a directed graph is one of the most important combinatorial optimization problems, having applications in a wide range of fields. In its basic version, however, the problem fails to represent situations in which the value of the objective function is determined not only by the choice of each single arc, … Read more

An external penalty-type method for multicriteria

We propose an extension of the classical real-valued external penalty method to the multicriteria optimization setting. As its single objective counterpart, it also requires an external penalty function for the constraint set, as well as an exogenous divergent sequence of nonnegative real numbers, the so-called penalty parameters, but, differently from the scalar procedure, the vector-valued … Read more

Lower bounding procedure for the Asymmetric Quadratic Traveling Salesman Problem

In this paper we consider the Asymmetric Quadratic Traveling Salesman Problem. Given a directed graph and a function that maps every pair of consecutive arcs to a cost, the problem consists in finding a cycle that visits every vertex exactly once and such that the sum of the costs is minimum. We propose an extended … Read more

On the Performance of SQP Methods for Nonlinear Optimization

This paper concerns some practical issues associated with the formulation of sequential quadratic programming (SQP) methods for large-scale nonlinear optimization. SQP methods find an approximate solution of a sequence of quadratic programming (QP) subproblems in which a quadratic model of the objective function is minimized subject to the linearized constraints. Extensive numerical results are given … Read more

Regret Analysis of Block Coordinate Gradient Methods for Online Convex Programming

In this paper, we propose two block coordinate gradient (BCG) methods for the online convex programming: the BCG method with the cyclic rule and the BCG method with the random rule. The proposed methods solve a low dimensional problem at each iteration, and hence they are efficient for large scale problems. For the proposed methods, … Read more