Forward-Backward and Tseng’s Type Penalty Schemes for Monotone Inclusion Problems

We deal with monotone inclusion problems of the form $0\in Ax+Dx+N_C(x)$ in real Hilbert spaces, where $A$ is a maximally monotone operator, $D$ a cocoercive operator and $C$ the nonempty set of zeros of another cocoercive operator. We propose a forward-backward penalty algorithm for solving this problem which extends the one proposed by H. Attouch, … Read more

Facially exposed cones are not always nice

We address the conjecture proposed by Gabor Pataki that every facially exposed cone is nice. We show that the conjecture is true in the three-dimensional case, however, there exists a four-dimensional counterexample of a cone that is facially exposed but is not nice. CitationCRN, University of BallaratArticleDownload View PDF

New and Improved Conditions for Uniqueness of Sparsest Solutions of Underdetermined Linear Systems

The uniqueness of sparsest solutions of underdetermined linear systems plays a fundamental role in the newly developed compressed sensing theory. Several new algebraic concepts, including the sub-mutual coherence, scaled mutual coherence, coherence rank, and sub-coherence rank, are introduced in this paper in order to develop new and improved sufficient conditions for the uniqueness of sparsest … Read more

The Complexity of Large-scale Convex Programming under a Linear Optimization Oracle

This paper considers a general class of iterative optimization algorithms, referred to as linear-optimization-based convex programming (LCP) methods, for solving large-scale convex programming (CP) problems. The LCP methods, covering the classic conditional gradient (CG) method (a.k.a., Frank-Wolfe method) as a special case, can only solve a linear optimization subproblem at each iteration. In this paper, … Read more

On full Jacobian decomposition of the augmented Lagrangian method for separable convex programming

The augmented Lagrangian method (ALM) is a benchmark for solving the convex minimization problem with linear constraints. We consider the special case where the objective is in form of the sum of m functions without coupled variables. For solving this separable convex programming model, it is usually required to decompose the ALM subproblem at each … Read more

Level Bundle Methods for Constrained Convex Optimization with Various Oracles

We propose restricted memory level bundle methods for minimizing constrained convex nonsmooth optimization problems whose objective and constraint functions are known through oracles (black-boxes) that might provide inexact information. Our approach is general and covers many instances of inexact oracles, such as upper, lower and on-demand oracles. We show that the proposed level bundle methods … Read more

Nonsmooth Optimization Using Uncontrolled Inexact Information

We consider convex nonsmooth optimization problems whose objective function is known through a (fine) oracle together with some additional (cheap but poor) information – formalized as a second coarse oracle with uncontrolled inexactness. It is the case when the objective function is itself the output of an optimization solver, using a branch-and-bound procedure, or decomposing … Read more

Robust convex relaxation for the planted clique and densest k-subgraph problems

We consider the problem of identifying the densest k-node subgraph in a given graph. We write this problem as an instance of rank-constrained cardinality minimization and then relax using the nuclear and l1 norms. Although the original combinatorial problem is NP-hard, we show that the densest k-subgraph can be recovered from the solution of our … Read more

On the Complexity Analysis of Randomized Block-Coordinate Descent Methods

In this paper we analyze the randomized block-coordinate descent (RBCD) methods for minimizing the sum of a smooth convex function and a block-separable convex function. In particular, we extend Nesterov’s technique (SIOPT 2012) for analyzing the RBCD method for minimizing a smooth convex function over a block-separable closed convex set to the aforementioned more general … Read more

A proximal technique for computing the Karcher mean of symmetric positive definite matrices

This paper presents a proximal point approach for computing the Riemannian or intrinsic Karcher mean of symmetric positive definite matrices. Our method derives from proximal point algorithm with Schur decomposition developed to compute minimum points of convex functions on symmetric positive definite matrices set when it is seen as a Hadamard manifold. The main idea … Read more