Optimization problems with value function objectives

The family of optimization problems with value function objectives includes the minmax programming problem and the bilevel optimization problem. In this paper, we derive necessary optimality conditions for this class of problems. The main focus is on the case where the functions involved are nonsmooth and the constraints are the very general operator constraints. CitationsubmittedArticleDownload … Read more

Regularized Sequential Quadratic Programming

We present the formulation and analysis of a new sequential quadratic programming (\SQP) method for general nonlinearly constrained optimization. The method pairs a primal-dual generalized augmented Lagrangian merit function with a \emph{flexible} line search to obtain a sequence of improving estimates of the solution. This function is a primal-dual variant of the augmented Lagrangian proposed … Read more

A SIMPLE APPROACH TO OPTIMALITY CONDITIONS IN MINMAX PROGRAMMING

Considering the minmax programming problem, lower and upper subdi erential optimality conditions, in the sense of Mordukhovich, are derived. The approach here, mainly based on the nonsmooth dual objects of Mordukhovich, is completely di erent from that of most of the previous works where generalizations of the alternative theorem of Farkas have been applied. The results obtained … Read more

Differentiable exact penalty functions for nonlinear second-order cone programs

We propose a method to solve nonlinear second-order cone programs (SOCPs), based on a continuously differentiable exact penalty function. The construction of the penalty function is given by incorporating a multipliers estimate in the augmented Lagrangian for SOCPs. Under the nondegeneracy assumption and the strong second-order sufficient condition, we show that a generalized Newton method … Read more

Higher-Order Confidence Intervals for Stochastic Programming using Bootstrapping

We study the problem of constructing confidence intervals for the optimal value of a stochastic programming problem by using bootstrapping. Bootstrapping is a resampling method used in the statistical inference of unknown parameters for which only a small number of samples can be obtained. One such parameter is the optimal value of a stochastic optimization … Read more

A robust Kantorovich’s theorem on inexact Newton method with relative residual error tolerance

We prove that under semi-local assumptions, the inexact Newton method with a fixed relative residual error tolerance converges Q-linearly to a zero of the non-linear operator under consideration. Using this result we show that Newton method for minimizing a self-concordant function or to find a zero of an analytic function can be implemented with a … Read more

Approximate spectral factorization for design of efficient sub-filter sequences

A well-known approach to the design of computationally efficient filters is to use spectral factorization, i.e. a decomposition of a filter into a sequence of sub-filters. Due to the sparsity of the sub-filters, the typical processing speedup factor is within the range 1-10 in 2D, and for 3D it achieves 10-100. The design of such … Read more

Optimization over the Efficient Set of a Bicriteria Convex Programming Problem

The problem of optimizing a real function over the efficient set of a multiple objective programming problem arises in a variety of applications. In this article, we propose an outer approximation algorithm for maximizing a function $h(x) = \varphi(f(x))$ over the efficient set $X_E$ of the bi-criteria convex programming problem $ {\rm Vmin} \{f(x)=(f_1(x), f_2(x))^T … Read more

Derivative-free methods for constrained mixed-integer optimization

We consider the problem of minimizing a continuously di erentiable function of several variables subject to simple bound and general nonlinear inequality constraints, where some of the variables are restricted to take integer values. We assume that the rst order derivatives of the objective and constraint functions can be neither calculated nor approximated explicitly. This class … Read more

Conjugate gradient methods based on secant conditions that generate descent search directions for unconstrained optimization

Conjugate gradient methods have been paid attention to, because they can be directly applied to large-scale unconstrained optimization problems. In order to incorporate second order information of the objective function into conjugate gradient methods, Dai and Liao (2001) proposed a conjugate gradient method based on the secant condition. However, their method does not necessarily generate … Read more