View source on GitHub
|
Returns a tff.learning.optimizers.Optimizer for momentum SGD.
tff.learning.optimizers.build_sgdm(
learning_rate: optimizer.Float = 0.01,
momentum: Optional[optimizer.Float] = None
) -> tff.learning.optimizers.Optimizer
Used in the notebooks
| Used in the tutorials |
|---|
This class supports the simple gradient descent and its variant with momentum.
If momentum is not used, the update rule given learning rate lr, weights w
and gradients g is:
w = w - lr * g
If momentum m (a float between 0.0 and 1.0) is used, the update rule is
v = m * v + g
w = w - lr * v
where v is the velocity from previous steps of the optimizer.
Args | |
|---|---|
learning_rate
|
A positive float for learning rate, default to 0.01. |
momentum
|
An optional float between 0.0 and 1.0. If None, no momentum is
used.
|
View source on GitHub