Sums of Random Symmetric Matrices and Applications

Let B_i be deterministic symmetric m\times m matrices, and \xi_i be independent random scalars with zero mean and “of order of one” (e.g., \xi_i are Gaussian with zero mean and unit standard deviation). We are interested in conditions for the “typical norm” of the random matrix S_N = \xi_1B_1+…+\xi_NB_N to be of order of 1. … Read more

Convergent relaxations of polynomial matrix inequalities and static output feedback

Using a moment interpretation of recent results on sum-of-squares decompositions of non-negative polynomial matrices, we propose a hierarchy of convex linear matrix inequality (LMI) relaxations to solve non-convex polynomial matrix inequality (PMI) optimization problems, including bilinear matrix inequality (BMI) problems. This hierarchy of LMI relaxations generates a monotone sequence of lower bounds that converges to … Read more

Large-Scale Semidefinite Programming via Saddle Point Mirror-Prox Algorithm

In this paper, we first develop “economical” representations for positive semidefinitness of well-structured sparse symmetric matrix. Using the representations, we then reformulate well-structured large-scale semidefinite problems into smooth convex-concave saddle point problems, which can be solved by a Prox-method with efficiency ${\cal O}(\epsilon^{-1})$ developed in \cite{Nem}. Some numerical implementations for large-scale Lovasz capacity and MAXCUT … Read more

The Q Method for Second-order Cone Programming

Based on the Q method for SDP, we develop the Q method for SOCP. A modified Q method is also introduced. Properties of the algorithms are discussed. Convergence proofs are given. Finally, we present numerical results. CitationAdvOl-Report#2004/15 McMaster University, Advanced Optimization LaboratoryArticleDownload View PDF

On the Behavior of the Homogeneous Self-Dual Model for Conic Convex Optimization

There is a natural norm associated with a starting point of the homogeneous self-dual (HSD) embedding model for conic convex optimization. In this norm two measures of the HSD model’s behavior are precisely controlled independent of the problem instance: (i) the sizes of epsilon-optimal solutions, and (ii) the maximum distance of epsilon-optimal solutions to the … Read more

Complex Quadratic Optimization and Semidefinite Programming

In this paper we study the approximation algorithms for a class of discrete quadratic optimization problems in the Hermitian complex form. A special case of the problem that we study corresponds to the max-3-cut model used in a recent paper of Goemans and Williamson. We first develop a closed-form formula to compute the probability of … Read more

Recovering Risk-Neutral Probability Density Functions from Options Prices using Cubic Splines

We present a new approach to estimate the risk-neutral probability density function (pdf) of the future prices of an underlying asset from the prices of options written on the asset. The estimation is carried out in the space of cubic spline functions, yielding appropriate smoothness. The resulting optimization problem, used to invert the data and … Read more

Computational Enhancements in Low-Rank Semidefinite Programming

We discuss computational enhancements for the low-rank semidefinite programming algorithm, including the extension to block semidefinite programs, an exact linesearch procedure, and a dynamic rank reduction scheme. A truncated Newton method is also introduced, and several preconditioning strategies are proposed. Numerical experiments illustrating these enhancements are provided. CitationManuscript, Department of Mangagement Sciences, University of Iowa, … Read more

A direct formulation for sparse PCA using semidefinite programming

We examine the problem of approximating, in the Frobenius-norm sense, a positive, semidefinite symmetric matrix by a rank-one matrix, with an upper bound on the cardinality of its eigenvector. The problem arises in the decomposition of a covariance matrix into sparse factors, and has wide applications ranging from biology to finance. We use a modification … Read more

Primal-Dual Interior-Point Algorithms for Semidefinite Optimization Based on a Simple Kernel Function

Interior-point methods (IPMs) for semidefinite optimization (SDO) have been studied intensively, due to their polynomial complexity and practical efficiency. Recently, J.Peng et al. introduced so-called self-regular kernel (and barrier) functions and designed primal-dual interior-point algorithms based on self-regular proximity for linear optimization (LO) problems. They have also extended the approach for LO to SDO. In … Read more