Accessible Theoretical Complexity of the Restarted Primal-Dual Hybrid Gradient Method for Linear Programs with Unique Optima
\(\)
The restarted primal-dual hybrid gradient method (rPDHG) has recently emerged as an important tool for solving large-scale linear programs (LPs). For LPs with unique optima, we present an iteration bound of \(\widetilde{O}\left(\kappa\Phi\cdot\ln\left(\frac{\|w^*\|}{\varepsilon}\right)\right)\), where \(\varepsilon\) is the target tolerance, \(\kappa\) is the standard matrix condition number, \(\|w^*\|\) is the norm of the optimal solution, and \(\Phi\) is a geometric condition number of the LP sublevel sets. This iteration bound is “accessible” in the sense that computing it is no more difficult than computing the optimal solution itself. Indeed, we present a closed-form and tractably computable expression for \(\Phi\). This enables an analysis of the “two-stage performance” of rPDHG: we show that the first stage identifies the optimal basis in \(\widetilde{O}\left(\kappa\Phi\right)\) iterations, and the second stage computes an \(\varepsilon\)-optimal solution in \(O\left(\|B^{-1}\|\|A\|\cdot\ln\left(\frac{\xi}{\varepsilon}\right)\right)\) additional iterations, where \(A\) is the constraint matrix, \(B\) is the optimal basis and \(\xi\) is the smallest nonzero in the optimal solution. Furthermore, computational tests mostly confirm the tightness of our iterations bounds. We also show a reciprocal relation between the iteration bound and three equivalent types of condition measures: (i) stability under data perturbation, (ii) proximity to multiple optima, and (iii) the LP sharpness of the instance. Finally, we analyze an “optimized” primal-dual reweighting which offers some intuition concerning the step-size heuristics used in practice.