tff.learning.optimizers.build_adafactor
Stay organized with collections
Save and categorize content based on your preferences.
Builds an Adafactor optimizer.
tff.learning.optimizers.build_adafactor(
learning_rate: optimizer.Float,
*,
beta_2_decay: optimizer.Float = -0.8,
epsilon_1: optimizer.Float = 1e-30,
epsilon_2: optimizer.Float = 0.001,
clip_threshold: optimizer.Float = 1.0,
relative_step: bool = True
) -> tff.learning.optimizers.Optimizer
An implementation of Adafactor from Shazeer, Noam et al described in
https://arxiv.org/abs/1804.04235
Args |
learning_rate
|
Initial value of the learning rate.
|
beta_2_decay
|
The decay rate of beta_2 .
|
epsilon_1
|
A small offset to keep denomiantor away from zero.
|
epsilon_2
|
A small offset to avoid learning rate becoming two small over
time.
|
clip_threshold
|
The clipping threshold of the Adafactor algorithm.
|
relative_step
|
If True , learning rate is adjusted based on number of
iterations. This is the default Adafactor learning rate decay.
|
Except as otherwise noted, the content of this page is licensed under the Creative Commons Attribution 4.0 License, and code samples are licensed under the Apache 2.0 License. For details, see the Google Developers Site Policies. Java is a registered trademark of Oracle and/or its affiliates.
Last updated 2024-09-20 UTC.
[null,null,["Last updated 2024-09-20 UTC."],[],[],null,["# tff.learning.optimizers.build_adafactor\n\n\u003cbr /\u003e\n\n|-------------------------------------------------------------------------------------------------------------------------------------------------------------|\n| [View source on GitHub](https://github.com/tensorflow/federated/blob/v0.87.0 Version 2.0, January 2004 Licensed under the Apache License, Version 2.0 (the) |\n\nBuilds an Adafactor optimizer. \n\n tff.learning.optimizers.build_adafactor(\n learning_rate: optimizer.Float,\n *,\n beta_2_decay: optimizer.Float = -0.8,\n epsilon_1: optimizer.Float = 1e-30,\n epsilon_2: optimizer.Float = 0.001,\n clip_threshold: optimizer.Float = 1.0,\n relative_step: bool = True\n ) -\u003e ../../../tff/learning/optimizers/Optimizer\n\nAn implementation of Adafactor from Shazeer, Noam et al described in\n\u003chttps://arxiv.org/abs/1804.04235\u003e\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n| Args ---- ||\n|------------------|------------------------------------------------------------------------------------------------------------------------|\n| `learning_rate` | Initial value of the learning rate. |\n| `beta_2_decay` | The decay rate of `beta_2`. |\n| `epsilon_1` | A small offset to keep denomiantor away from zero. |\n| `epsilon_2` | A small offset to avoid learning rate becoming two small over time. |\n| `clip_threshold` | The clipping threshold of the Adafactor algorithm. |\n| `relative_step` | If `True`, learning rate is adjusted based on number of iterations. This is the default Adafactor learning rate decay. |\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n| Returns ------- ||\n|---|---|\n| A [`tff.learning.optimizers.Optimizer`](../../../tff/learning/optimizers/Optimizer) that implements the Adafactor optimizer. ||\n\n\u003cbr /\u003e"]]