Relaxing the Optimality Conditions of Box QP

We present semidefinite relaxations of nonconvex, box-constrained quadratic programming, which incorporate the first- and second-order necessary optimality conditions. We compare these relaxations with a basic semidefinite relaxation due to Shor, particularly in the context of branch-and-bound to determine a global optimal solution, where it is shown empirically that the new relaxations are significantly stronger. We … Read more

Classification problems with imprecise data through separating hyperplanes

We consider a supervised classification problem in which the elements to be classified are sets with certain geometrical properties. In particular, this model can be applied to deal with data affected by some kind of noise and in the case of interval-valued data. Two classification rules, a fuzzy one and a crisp one, are defined … Read more

Optimal solutions for unrelated parallel machines scheduling problems using convex quadratic reformulations

In this work, we take advantage of the powerful quadratic programming theory to obtain optimal solutions of scheduling problems. We apply a methodology that starts, in contrast to more classical approaches, by formulating three unrelated parallel machine scheduling problems as 0–1 quadratic programs under linear constraints. By construction, these quadratic programs are non-convex. Therefore, before … Read more

Exploiting separability in large-scale linear support vector machine training

Linear support vector machine training can be represented as a large quadratic program. We present an efficient and numerically stable algorithm for this problem using interior point methods, which requires only O(n) operations per iteration. Through exploiting the separability of the Hessian, we provide a unified approach, from an optimization perspective, to 1-norm classification, 2-norm … Read more

A conic duality Frank–Wolfe type theorem via exact penalization in quadratic optimization

The famous Frank–Wolfe theorem ensures attainability of the optimal value for quadratic objective functions over a (possibly unbounded) polyhedron if the feasible values are bounded. This theorem does not hold in general for conic programs where linear constraints are replaced by more general convex constraints like positive-semidefiniteness or copositivity conditions, despite the fact that the … Read more

A New Unblocking Technique to Warmstart Interior Point Methods based on Sensitivity Analysis

One of the main drawbacks associated with Interior Point Methods (IPM) is the perceived lack of an efficient warmstarting scheme which would enable the use of information from a previous solution of a similar problem. Recently there has been renewed interest in the subject. A common problem with warmstarting for IPM is that an advanced … Read more

Large Scale Portfolio Optimization with Piecewise Linear Transaction Costs

We consider the fundamental problem of computing an optimal portfolio based on a quadratic mean-variance model of the objective function and a given polyhedral representation of the constraints. The main departure from the classical quadratic programming formulation is the inclusion in the objective function of piecewise linear, separable functions representing the transaction costs. We handle … Read more

An Adaptive Primal-Dual Warm-Start Technique for Quadratic Multiobjective Optimization

We present a new primal-dual algorithm for convex quadratic multicriteria optimization. The algorithm is able to adaptively refine the approximation to the set of efficient points by way of a warm-start interior-point scalarization approach. Results of this algorithm when applied on a three-criteria real-world power plant optimization problem are reported, thereby illustrating the feasibility of … Read more

Global and finite termination of a two-phase augmented Lagrangian filter method for general quadratic programs

We present a two-phase algorithm for solving large-scale quadratic programs (QPs). In the first phase, gradient-projection iterations approximately minimize an augmented Lagrangian function and provide an estimate of the optimal active set. In the second phase, an equality-constrained QP defined by the current inactive variables is approximately minimized in order to generate a second-order search … Read more

On the behavior of the conjugate-gradient method on ill-conditioned problems

We study the behavior of the conjugate-gradient method for solving a set of linear equations, where the matrix is symmetric and positive definite with one set of eigenvalues that are large and the remaining are small. We characterize the behavior of the residuals associated with the large eigenvalues throughout the iterations, and also characterize the … Read more