Gradient descent is used in many machine learning algorithms , It's an optimization tool . But gradient descent is essentially a greedy algorithm , Easy to fall into the local most . References in this paper [1] Of Momentum Methods help the gradient descent method to jump out of the local optimal solution .
For various optimization methods, see [2] Detailed explanation .
———————- I am the elegant dividing line , Wash and sleep first —————–