The Gradient Sampling method is a recently developed tool for solving unconstrained nonsmooth optimization problems. Using just first order information about the objective function, it generalizes the steepest descent method, one of the most classical methods to minimize a smooth function. This manuscript aims at determining under which circumstances one can expect the same local convergence result of the Cauchy method for the Gradient Sampling algorithm. Additionally, at the end of this study, we show how to practically accomplish the required hypotheses during the execution of the algorithm.
View On the local convergence analysis of the Gradient Sampling method