Political districting to optimize the Polsby-Popper compactness score

\(\)In the academic literature and in expert testimony, the Polsby-Popper score is the most popular way to measure the compactness of a political district. Given a district with area \(A\) and perimeter \(P\), its Polsby-Popper score is given by \( (4 \pi A)/P^2\). This score takes values between zero and one, with circular districts achieving … Read more

On the equilibrium prices of a regular locally Lipschitz exchange economy

We extend classical results by Debreu and Dierker about equilibrium prices of a regular economy with continuously differentiable demand functions/excess demand function to a regular exchange economy with these functions being locally Lipschitz. Our concept of a regular economy is based on Clarke’s concept of regular value and we show that such a regular economy … Read more

Asynchronous Iterations in Optimization: New Sequence Results and Sharper Algorithmic Guarantees

We introduce novel convergence results for asynchronous iterations that appear in the analysis of parallel and distributed optimization algorithms. The results are simple to apply and give explicit estimates for how the degree of asynchrony impacts the convergence rates of the iterates. Our results shorten, streamline and strengthen existing convergence proofs for several asynchronous optimization … Read more

A Moment-SOS Hierarchy for Robust Polynomial Matrix Inequality Optimization with SOS-Convexity

We study a class of polynomial optimization problems with a robust polynomial matrix inequality constraint for which the uncertainty set is defined also by a polynomial matrix inequality (including robust polynomial semidefinite programs as a special case). Under certain SOS-convexity assumptions, we construct a hierarchy of moment-SOS relaxations for this problem to obtain convergent upper … Read more

Robust optimal design of a tree-based water distribution network with intermittent demand

This paper discusses the design of a tree-shaped water distribution system for small, dispersed rural communities. It revisits the topic that was discussed in [Babonneau et al., 2019] and is nowadays implemented in the field. It proposes a new approach to pipe selection based on robust optimization to account for the uncertainty inherent in intermittent … Read more

Stochastic Ratios Tracking Algorithm for Large Scale Machine Learning Problems

Many machine learning applications and tasks rely on the stochastic gradient descent (SGD) algorithm and its variants. Effective step length selection is crucial for the success of these algorithms, which has motivated the development of algorithms such as ADAM or AdaGrad. In this paper, we propose a novel algorithm for adaptive step length selection in … Read more

Finding Regions of Counterfactual Explanations via Robust Optimization

Counterfactual explanations play an important role in detecting bias and improving the explainability of data-driven classification models. A counterfactual explanation (CE) is a minimal perturbed data point for which the decision of the model changes. Most of the existing methods can only provide one CE, which may not be achievable for the user. In this … Read more

Maximum Likelihood Probability Measures over Sets and Applications to Data-Driven Optimization

\(\) Motivated by data-driven approaches to sequential decision-making under uncertainty, we study maximum likelihood estimation of a distribution over a general measurable space when, unlike traditional setups, realizations of the underlying uncertainty are not directly observable but instead are known to lie within observable sets. While extant work studied the special cases when the observed … Read more

Optimal Low-Rank Matrix Completion: Semidefinite Relaxations and Eigenvector Disjunctions

Low-rank matrix completion consists of computing a matrix of minimal complexity that recovers a given set of observations as accurately as possible. Unfortunately, existing methods for matrix completion are heuristics that, while highly scalable and often identifying high-quality solutions, do not possess any optimality guarantees. We reexamine matrix completion with an optimality-oriented eye. We reformulate … Read more

Learning in Inverse Optimization: Incenter Cost, Augmented Suboptimality Loss, and Algorithms

In Inverse Optimization (IO), an expert agent solves an optimization problem parametric in an exogenous signal. From a learning perspective, the goal is to learn the expert’s cost function given a dataset of signals and corresponding optimal actions. Motivated by the geometry of the IO set of consistent cost vectors, we introduce the “incenter” concept, … Read more