Machine Learning Engineering
Stochastic Gradient Descent (SGD) is an optimization algorithm used to minimize the loss function in machine learning models, particularly in training neural networks. Unlike standard gradient descent, which computes the gradient using the entire dataset, SGD updates the model weights using only a single sample or a small batch of samples at each iteration. This approach introduces randomness into the learning process, making it faster and often more effective for large datasets, while also helping to avoid local minima.
congrats on reading the definition of Stochastic Gradient Descent. now let's actually learn it.