A Quadratic Programming Bibliography

The following is a list of all of the published and unpublished works on quadratic programming that we are aware of. Some are general references to background material, while others are central to the development of the quadratic programming methods and to the applications we intend to cover in our evolving book on the subject. … Read more

Feasible Interior Methods Using Slacks for Nonlinear Optimization

A slack-based feasible interior point method is described which can be derived as a modification of infeasible methods. The modification is minor for most line search methods, but trust region methods require special attention. It is shown how the Cauchy point, which is often computed in trust region methods, must be modified so that the … Read more

A BFGS-IP algorithm for solving strongly convex optimization problems with feasibility enforced by an exact penalty approach

This paper introduces and analyses a new algorithm for minimizing a convex function subject to a finite number of convex inequality constraints. It is assumed that the Lagrangian of the problem is strongly convex. The algorithm combines interior point methods for dealing with the inequality constraints and quasi-Newton techniques for accelerating the convergence. Feasibility of … Read more

Reducing the number of AD passes for computing a sparse Jacobian matrix

A reduction in the computational work is possible if we do not require that the nonzeros of a Jacobian matrix be determined directly. If a column or row partition is available, the proposed substitution technique can be used to reduce the number of groups in the partition further. In this chapter, we present a substitution … Read more

Constraint Identification and Algorithm Stabilization for Degenerate Nonlinear Programs

In the vicinity of a solution of a nonlinear programming problem at which both strict complementarity and linear independence of the active constraints may fail to hold, we describe a technique for distinguishing weakly active from strongly active constraints. We show that this information can be used to modify the sequential quadratic programming algorithm so … Read more

iNEOS : An Interactive Environment for Nonlinear Optimization

In this paper we describe iNEOS, an Internet-based environment which facilitates the solution of complex nonlinear optimization problems. It enables a user to easily invoke a remote optimization code without having to supply the model to be optimized. An interactive communication between client and server is established and maintainted using CORBA. We test the system … Read more

Benchmarking Optimization Software with COPS

We describe version 2.0 of the COPS set of nonlinearly constrained optimization problems. We have added new problems, as well as streamlined and improved most of the problems. We also provide a comparison of the LANCELOT, LOQO, MINOS, and SNOPT solvers on these problems. Citation Technical Report ANL/MCS-246 Mathematics and Computer Science Division Argonne National … Read more

Feasibility Control in Nonlinear Optimization

We analyze the properties that optimization algorithms must possess in order to prevent convergence to non-stationary points for the merit function. We show that demanding the exact satisfaction of constraint linearizations results in difficulties in a wide range of optimization algorithms. Feasibility control is a mechanism that prevents convergence to spurious solutions by ensuring that … Read more

An infeasible active set method for convex problems with simple bounds

A primal-dual active set method for convex quadratic problems with bound constraints is presented. Based on a guess on the active set, a primal-dual pair $(x,s)$ is computed that satisfies the first order optimality condition and the complementarity condition. If $(x,s)$ is not feasible, a new active set is determined, and the process is iterated. … Read more

On the global convergence of an SLP-filter algorithm

A mechanism for proving global convergence infilter-type methods for nonlinear programming is described. Such methods are characterized by their use of the dominance concept of multi objective optimization, instead of a penalty parameter whose adjustment can be problematic. The main point of interest is to demonstrate how convergence for NLP can be induced without forcing … Read more