Linear Algebra for Data Science
In the context of optimization algorithms, momentum is a technique used to accelerate the convergence of gradient descent by incorporating past gradients into the current update. This method helps to smooth out the updates and allows for faster convergence, particularly in areas of the optimization landscape with high curvature. By combining the current gradient with a fraction of the previous update, momentum can help to overcome local minima and oscillations in the cost function.
congrats on reading the definition of momentum. now let's actually learn it.