Computational Experiments with Cross and Crooked Cross Cuts

In a recent paper, Dash, Dey and Gunluk (2010) showed that many families of inequalities for the two-row continuous group relaxation and variants of this relaxation are cross cuts or crooked cross cuts, both of which generalize split cuts. Li and Richard (2008) recently studied t-branch split cuts for mixed-integer programs for integers t >= … Read more

A lower bound on the optimal self-concordance parameter of convex cones

Let $K \subset \mathbb R^n$ be a regular convex cone, let $e_1,\dots,e_n \in \partial K$ be linearly independent points on the boundary of a compact affine section of the cone, and let $x^* \in K^o$ be a point in the relative interior of this section. For $k = 1,\dots,n$, let $l_k$ be the line through … Read more

On Penalty and Gap Function Methods for Bilevel Equilibrium Problems

We consider bilevel pseudomonotone equilibrium problems. We use a penalty function to convert a bilevel problem into one-level ones. We generalize a pseudo $\nabla$-monotonicity concept from $\nabla$-monotonicity and prove that under pseudo $\nabla$-monotonicity property any stationary point of a regularized gap function is a solution of the penalized equilibrium problem. As an application, we discuss … Read more

Unharnessing the power of Schrijver’s permanental inequality

Let $A \in \Omega_n$ be doubly-stochastic $n \times n$ matrix. Alexander Schrijver proved in 1998 the following remarkable inequality \begin{equation} \label{le} per(\widetilde{A}) \geq \prod_{1 \leq i,j \leq n} (1- A(i,j)); \widetilde{A}(i,j) =: A(i,j)(1-A(i,j)), 1 \leq i,j \leq n \end{equation} We prove in this paper the following generalization (or just clever reformulation) of (\ref{le}):\\ For all … Read more

On DC. optimization algorithms for solving minmax flow problems

We formulate minmax flow problems as a DC. optimization problem. We then apply a DC primal-dual algorithm to solve the resulting problem.The obtained computational results show that the proposed algorithm is efficient thanks to particular structures of the minmax flow problems. Citation1. An L. T. H. and Tao P. D., The DC (Difference of convex … Read more

Inexact projected gradient method for vector optimization

In this work, we propose an inexact projected gradient-like method for solving smooth constrained vector optimization problems. In the unconstrained case, we retrieve the steepest descent method introduced by Graña Drummond and Svaiter. In the constrained setting, the method we present extends the exact one proposed by Graña Drummond and Iusem, since it admits relative … Read more

Sufficient Conditions for Low-rank Matrix Recovery,Translated from Sparse Signal Recovery

The low-rank matrix recovery (LMR) is a rank minimization problem subject to linear equality constraints, and it arises in many fields such as signal and image processing, statistics, computer vision, system identification and control. This class of optimization problems is $NP$-hard and a popular approach replaces the rank function with the nuclear norm of the … Read more

Solving Two-stage Robust Optimization Problems by A Constraint-and-Column Generation Method

We present a constraint-and-column generation algorithm to solve two-stage robust optimization problems. Compared with existing Benders style cutting plane methods, it is a general procedure with a unified approach to deal with optimality and feasibility. A computational study on a two-stage robust location-transportation problem shows that it performs an order of magnitude faster. Also, it … Read more

Immunizing conic quadratic optimization problems against implementation errors

We show that the robust counterpart of a convex quadratic constraint with ellipsoidal implementation error is equivalent to a system of conic quadratic constraints. To prove this result we first derive a sharper result for the S-lemma in case the two matrices involved can be simultaneously diagonalized. This extension of the S-lemma may also be … Read more

Computing the Grothendieck constant of some graph classes

Given a graph $G=([n],E)$ and $w\in\R^E$, consider the integer program ${\max}_{x\in \{\pm 1\}^n} \sum_{ij \in E} w_{ij}x_ix_j$ and its canonical semidefinite programming relaxation ${\max} \sum_{ij \in E} w_{ij}v_i^Tv_j$, where the maximum is taken over all unit vectors $v_i\in\R^n$. The integrality gap of this relaxation is known as the Grothendieck constant $\ka(G)$ of $G$. We present … Read more