gradient-descent

Search IconIcon to open search

Parent: Variations of gradient descent

Source: google-ml-course

Gradient descent

An iterated approach

  1. Labeled data arrives
  2. Gradient of the loss function is computed
  3. Now the direction for updating the model parameters $\mathbf{w}$ is known (negative gradient).
  4. A step is taken in this direction, in the parameter space. The step size is equivalent to the learning rate .
  5. Repeat

This process tunes all model parameters simultaneously.

Notes