Block stochastic gradient iteration for convex and nonconvex optimization

The stochastic gradient (SG) method can minimize an objective function composed of a large number of differentiable functions, or solve a stochastic optimization problem, to a moderate accuracy. The block coordinate descent/update (BCD) method, on the other hand, handles problems with multiple blocks of variables by updating them one at a time; when the blocks … Read more

An improved algorithm for L2-Lp minimization problem

In this paper we consider a class of non-Lipschitz and non-convex minimization problems which generalize the L2−Lp minimization problem. We propose an iterative algorithm that decides the next iteration based on the local convexity/concavity/sparsity of its current position. We show that our algorithm finds an epsilon-KKT point within O(log(1/epsilon)) iterations. The same result is also … Read more

Coercive polynomials and their Newton polytopes

Many interesting properties of polynomials are closely related to the geometry of their Newton polytopes. In this article we analyze the coercivity on $\mathbb{R}^n$ of multivariate polynomials $f\in \mathbb{R}[x]$ in terms of their Newton polytopes. In fact, we introduce the broad class of so-called gem regular polynomials and characterize their coercivity via conditions imposed on … Read more

Global Optimization via Slack Variables

This paper presents a method for finding global optima to constrained nonlinear programs via slack variables. The method only applies if all functions involved are of class C1 but without any further qualification on the types of constraints allowed; it proceeds by reformulating the given program into a bi-objective program that is then solved for … Read more

Mathematical Programs with Cardinality Constraints: Reformulation by Complementarity-type Constraints and a Regularization Method

Optimization problems with cardinality constraints are very dicult mathematical programs which are typically solved by global techniques from discrete optimization. Here we introduce a mixed-integer formulation whose standard relaxation still has the same solutions (in the sense of global minima) as the underlying cardinality-constrained problem; the relation between the local minima is also discussed in … Read more

An error analysis for polynomial optimization over the simplex based on the multivariate hypergeometric distribution

We study the minimization of fixed-degree polynomials over the simplex. This problem is well-known to be NP-hard, as it contains the maximum stable set problem in graph theory as a special case. In this paper, we consider a rational approximation by taking the minimum over the regular grid, which consists of rational points with denominator … Read more

On Global Optimization

This paper presents a relatively “unfettered” method for finding global optima to constrained nonlinear programs. The method reformulates the given program into a bi-objective mixed-integer program that is then solved for the Nash equilibrium. A numerical example (whose solution provides a new benchmark against which other algorithms may be assessed) is included to illustrate the … Read more

Directed modified Cholesky factorizations and convex quadratic relaxations

A directed Cholesky factorization of a symmetric interval matrix \A consists of a permuted upper triangular matrix R such that for all symmetric A \in \A, the residual matrix A – R^T R is positive semidefinite with tiny entries. This must holds with full mathematical rigor, although the computations are done in floating-point arithmetic. Similarly, … Read more

Constraint aggregation for rigorous global optimization

In rigorous constrained global optimization, upper bounds on the objective function help to reduce the search space. Their construction requires finding a narrow box around an approximately feasible solution, verified to contain a feasible point. Approximations are easily found by local optimization, but the verification often fails. In this paper we show that even if … Read more

How Difficult is Nonlinear Optimization? A Practical Solver Tuning Approach, with Illustrative Results

Nonlinear optimization (NLO) per definitionem covers a vast range of problems, from trivial to practically intractable. For this reason, it is impossible to offer “guaranteed” advice to NLO software users. This fact becomes especially obvious, when facing unusually hard and/or previously unexplored NLO challenges. In the present study we offer some related practical observations, propose … Read more