tff.learning.optimizers.build_rmsprop
Stay organized with collections
Save and categorize content based on your preferences.
Returns a tff.learning.optimizers.Optimizer
for RMSprop.
tff.learning.optimizers.build_rmsprop(
learning_rate: optimizer.Float,
decay: optimizer.Float = 0.9,
epsilon: optimizer.Float = 1e-07
) -> tff.learning.optimizers.Optimizer
The RMSprop optimizer is based on Tieleman and Hinton, 2012.
The update rule given learning rate lr
, epsilon eps
, decay d
,
preconditioner s
, weights w
and gradients g
is:
s = d * s + (1 - d) * g**2
w = w - lr * g / (sqrt(s) + eps)
Args |
learning_rate
|
A positive float for learning rate, default to 0.01.
|
decay
|
A float between 0.0 and 1.0 for the decay used to track the magnitude
of previous gradients.
|
epsilon
|
A small non-negative float, used to maintain numerical stability.
|
Except as otherwise noted, the content of this page is licensed under the Creative Commons Attribution 4.0 License, and code samples are licensed under the Apache 2.0 License. For details, see the Google Developers Site Policies. Java is a registered trademark of Oracle and/or its affiliates.
Last updated 2024-09-20 UTC.
[null,null,["Last updated 2024-09-20 UTC."],[],[],null,["# tff.learning.optimizers.build_rmsprop\n\n\u003cbr /\u003e\n\n|-------------------------------------------------------------------------------------------------------------------------------------------------------------|\n| [View source on GitHub](https://github.com/tensorflow/federated/blob/v0.87.0 Version 2.0, January 2004 Licensed under the Apache License, Version 2.0 (the) |\n\nReturns a [`tff.learning.optimizers.Optimizer`](../../../tff/learning/optimizers/Optimizer) for RMSprop. \n\n tff.learning.optimizers.build_rmsprop(\n learning_rate: optimizer.Float,\n decay: optimizer.Float = 0.9,\n epsilon: optimizer.Float = 1e-07\n ) -\u003e ../../../tff/learning/optimizers/Optimizer\n\nThe RMSprop optimizer is based on [Tieleman and Hinton, 2012](http://www.cs.toronto.edu/%7Ehinton/coursera/lecture6/lec6.pdf).\n\nThe update rule given learning rate `lr`, epsilon `eps`, decay `d`,\npreconditioner `s`, weights `w` and gradients `g` is: \n\n s = d * s + (1 - d) * g**2\n w = w - lr * g / (sqrt(s) + eps)\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n| Args ---- ||\n|-----------------|----------------------------------------------------------------------------------------------|\n| `learning_rate` | A positive float for learning rate, default to 0.01. |\n| `decay` | A float between 0.0 and 1.0 for the decay used to track the magnitude of previous gradients. |\n| `epsilon` | A small non-negative float, used to maintain numerical stability. |\n\n\u003cbr /\u003e"]]