Glossary

Stochastic Gradient Descent

Stochastic Gradient Descent (SGD) is an optimization algorithm that updates model parameters using the gradient from a single (or small batch of) randomly chosen training sample(s) rather than the full dataset. This makes updates faster but noisier, which can help escape local minima. It iteratively moves parameters in the direction that reduces the loss function until convergence.

by Frank Zickert