Fast Multiple Splitting Algorithms for Convex Optimization

We present in this paper two different classes of general $K$-splitting algorithms for solving finite-dimensional convex optimization problems. Under the assumption that the function being minimized has a Lipschitz continuous gradient, we prove that the number of iterations needed by the first class of algorithms to obtain an $\epsilon$-optimal solution is $O(1/\epsilon)$. The algorithms in … Read more

Fast Alternating Linearization Methods for Minimizing the Sum of Two Convex Functions

We present in this paper first-order alternating linearization algorithms based on an alternating direction augmented Lagrangian approach for minimizing the sum of two convex functions. Our basic methods require at most $O(1/\epsilon)$ iterations to obtain an $\epsilon$-optimal solution, while our accelerated (i.e., fast) versions of them require at most $O(1/\sqrt{\epsilon})$ iterations, with little change in … Read more

Solving log-determinant optimization problems by a Newton-CG primal proximal point algorithm

We propose a Newton-CG primal proximal point algorithm for solving large scale log-determinant optimization problems. Our algorithm employs the essential ideas of the proximal point algorithm, the Newton method and the preconditioned conjugate gradient solver. When applying the Newton method to solve the inner sub-problem, we find that the log-determinant term plays the role of … Read more

An Implementable Proximal Point Algorithmic Framework for Nuclear Norm Minimization

The nuclear norm minimization problem is to find a matrix with the minimum nuclear norm subject to linear and second order cone constraints. Such a problem often arises from the convex relaxation of a rank minimization problem with noisy data, and arises in many fields of engineering and science. In this paper, we study inexact … Read more

A Redistributed Proximal Bundle Method for Nonconvex Optimization

Proximal bundle methods have been shown to be highly successful optimization methods for unconstrained convex problems with discontinuous first derivatives. This naturally leads to the question of whether proximal variants of bundle methods can be extended to a nonconvex setting. This work proposes an approach based on generating cutting-planes models, not of the objective function … Read more

A Logarithmic-Quadratic Proximal Point Scalarization Method for Multiobjective Programming

We present a proximal point method to solve multiobjective problems based on the scalarization for maps. We build a family of a convex scalar strict representation of a convex map F with respect to the lexicographic order on Rm and we add a variant of the logarithmquadratic regularization of Auslender, where the unconstrained variables in … Read more

Proximal Point Methods for Functions Involving Lojasiewicz, Quasiconvex and Convex Properties on Hadamard Manifolds

This paper extends the full convergence of the proximal point method with Riemannian, Semi-Bregman and Bregman distances to solve minimization problems on Hadamard manifolds. For the unconstrained problem, under the assumptions that the optimal set is nonempty and the objective function is continuous and either quasiconvex or satisfies a generalized Lojasiewicz property, we prove the … Read more


The notion of soft thresholding plays a central role in problems from various areas of applied mathematics, in which the ideal solution is known to possess a sparse decomposition in some orthonormal basis. Using convex-analytical tools, we extend this notion to that of proximal thresholding and investigate its properties, providing in particular several characterizations of … Read more

A Proximal Point Algorithm with phi-Divergence to Quasiconvex Programming

We use the proximal point method with the phi-divergence given by phi(t) = t – log t – 1 for the minimization of quasiconvex functions subject to nonnegativity constraints. We establish that the sequence generated by our algorithm is well-defined in the sense that it exists and it is not cyclical. Without any assumption of … Read more

An Extension of the Proximal Point Method for Quasiconvex Minimization

In this paper we propose an extension of the proximal point method to solve minimization problems with quasiconvex objective functions on the Euclidean space and the nonnegative orthant. For the unconstrained minimization problem, assumming that the function is bounded from below and lower semicontinuous we prove that iterations fxkg given by 0 2 b@(f(:)+(k=2)jj:􀀀xk􀀀1jj2)(xk) are … Read more