Many optimization algorithms require gradients of the model functions, but computing accurate gradients can be computationally expensive. We study the implications of using inexact gradients in the context of the multilevel optimization algorithm MGOpt. MGOpt recursively uses (typically cheaper) coarse models to obtain search directions for finer-level models. However, MGOpt requires the gradient on the fine level to define the recursion. Our primary focus here is the impact of the gradient errors on the multilevel recursion. We analyze, partly through model problems, how MGOpt is affected under various assumptions about the source of the error in the gradients, and demonstrate that in many cases the effect of the errors is benign. Computational experiments are included.

## Citation

Computational Optimization with Applications, volume 56 (2013), pp. 39-61.

## Article

View Using Inexact Gradients in a Multilevel Optimization Algorithm