Vanishing Gradient Problem
Meaning ⎊ Training issue where gradients shrink to near zero, preventing deep network layers from updating their weights.
Momentum-Based Optimization
Meaning ⎊ Optimization technique using moving averages of past gradients to accelerate convergence and smooth out noise.
Stochastic Gradient Descent
Meaning ⎊ Gradient optimization method using random data subsets to improve computational speed and escape local minima.
Gradient Descent Optimization
Meaning ⎊ Mathematical technique to find the minimum of a function by iteratively moving against the gradient of the loss.
