Faster approximation algorithms for packing and covering problems

We adapt a method due to Nesterov so as to obtain an algorithm for solving block-angular fractional packing or covering problems to relative tolerance epsilon, while using a number of iterations that grows polynomially in the size of the problem and whose dependency on epsilon is proportional to 1/epsilon. CitationCORC report TR-2004-09, Computational Optimization Research … Read more

Interior point methods for large-scale linear programming

We discuss interior point methods for large-scale linear programming, with an emphasis on methods that are useful for problems arising in telecommunications. We give the basic framework of a primal-dual interior point method, and consider the numerical issues involved in calculating the search direction in each iteration, including the use of factorization methods and/or preconditioned … Read more

A NEW SELF-CONCORDANT BARRIER FOR THE HYPERCUBE

In this paper we introduce a new barrier function $\sum\limits_{i=1}^n(2x_i-1)[\ln{x_i}-\ln(1-x_i)]$ to solve the following optimization problem: $\min\,\, f(x)$ subject to: $Ax=b;\;\;0\leq x\leq e$. We show that this function is a $(3/2)n$-self-concordant barrier on the hypercube $[0,1]^n$. We prove that the central path is well defined and that under an additional assumption on the objective function, … Read more

Sensitivity analysis in linear optimization: Invariant support set intervals

Sensitivity analysis is one of the most nteresting and preoccupying areas in optimization. Many attempts are made to investigate the problem’s behavior when the input data changes. Usually variation occurs in the right hand side of the constraints and/or the objective function coefficients. Degeneracy of optimal solutions causes considerable difficulties in sensitivity analysis. In this … Read more

Symmetry Points of Convex Set: Basic Properties and Computational Complexity

Given a convex body S and a point x \in S, let sym(x,S) denote the symmetry value of x in S: sym(x,S):= max{t : x + t(x – y) \in S for every y \in S}, which essentially measures how symmetric S is about the point x, and define sym(S):=\max{sym(x,S) : x \in S }. … Read more

Pointillism via Linear Programming

Pointillism is a painting technique in which the painter places dots of paint on the canvas in such a way that they blend together into desired forms when viewed from a distance. In this brief note, we describe how to use linear programming to construct a pointillist portrait. CitationDept. of Mathematics, Oberlin College, Oberlin, OH … Read more

Sensitivity analysis for linear optimization problem with fuzzy data in the objective function

Linear programming problems with fuzzy coefficients in the objective function are considered. Emphasis is on the dependence of the optimal solution from linear perturbations of the membership functions of the objective function coefficients as well as on the computation of a robust solution of the fuzzy linear problem if the membership functions are not surely … Read more

On Implementing Self-Regular Proximity Based Feasible IPMs

Self-regular based interior point methods present a unified novel approach for solving linear optimization and conic optimization problems. So far it was not known if the new Self-Regular IPMs can lead to similar advances in computational practice as shown in the theoretical analysis. In this paper, we present our experiences in developing the software package … Read more

Dual Convergence of the Proximal Point Method with Bregman Distances for Linear Programming

In this paper we consider the proximal point method with Bregman distance applied to linear programming problems, and study the dual sequence obtained from the optimal multipliers of the linear constraints of each subproblem. We establish the convergence of this dual sequence, as well as convergence rate results for the primal sequence, for a suitable … Read more

Boundedness Theorems for the Relaxation Method

A classical theorem by Block and Levin says that certain variants of the relaxation method for solving systems of linear inequalities produce bounded sequences of intermediate solutions even when running on inconsistent input data. Using a new approach, we prove a more general version of this result and answer an old open problem of quantifying … Read more