# Backtracking Gradient Descent allowing unbounded learning rates

7 Jan 2020Tuyen Trung Truong

In unconstrained optimisation on an Euclidean space, to prove convergence in Gradient Descent processes (GD) $x_{n+1}=x_n-\delta _n \nabla f(x_n)$ it usually is required that the learning rates $\delta _n$'s are bounded: $\delta _n\leq \delta$ for some positive $\delta$. Under this assumption, if the sequence $x_n$ converges to a critical point $z$, then with large values of $n$ the update will be small because $||x_{n+1}-x_n||\lesssim ||\nabla f(x_n)||$... (read more)

PDF Abstract

No code implementations yet. Submit your code now