How to skip local optimization by gradient descent （ Detailed explanation of various optimization methods in depth learning ）【 Continuous update 】
Gradient descent is used in many machine learning algorithms , It's an optimization tool . But gradient descent is essentially a greedy algorithm , Easy to fall into the local most . References in this paper  Of Momentum Methods help the gradient descent method to jump out of the local optimal solution .
For various optimization methods, see  Detailed explanation .
———————- I am the elegant dividing line , Wash and sleep first —————–