Item

On the Worst-Case Complexity of the Gradient Method with Exact Line Search for Smooth Strongly Convex Functions

de Klerk,Etienne
Glineur,Francois
Taylor,Adrien
Abstract
We consider the gradient (or steepest) descent method with exact line search applied to a strongly convex function with Lipschitz continuous gradient. We establish the exact worst-case rate of convergence of this scheme, and show that this worst-case behavior is exhibited by a certain convex quadratic function. We also extend the result to a noisy variant of gradient descent method, where exact line-search is performed in a search direction that differs from negative gradient by at most a prescribed relative tolerance. The proof is computer-assisted, and relies on the resolution of semidefinite programming performance estimation problems as introduced in the paper [Y. Drori and M. Teboulle. Performance of first-order methods for smooth convex minimization: a novel approach. Mathematical Programming, 145(1-2):451-482, 2014].
Description
Date
2016-06-30
Journal Title
Journal ISSN
Volume Title
Publisher
Cornell University Library
Research Projects
Organizational Units
Journal Issue
Keywords
Citation
de Klerk, E, Glineur, F & Taylor, A 2016 'On the Worst-Case Complexity of the Gradient Method with Exact Line Search for Smooth Strongly Convex Functions' arXiv, vol. arXiv:1606.09365, Cornell University Library, Itacha. < http://arxiv.org/abs/1606.09365 >
Embedded videos