Stay organized with collections
Save and categorize content based on your preferences.
tensorflow::ops::ApplyAdagrad
#include <training_ops.h>
Update '*var' according to the adagrad scheme.
Summary
accum += grad * grad var -= lr * grad * (1 / sqrt(accum))
Arguments:
- scope: A Scope object
- var: Should be from a Variable().
- accum: Should be from a Variable().
- lr: Scaling factor. Must be a scalar.
- grad: The gradient.
Optional attributes (see Attrs
):
- use_locking: If
True
, updating of the var and accum tensors will be protected by a lock; otherwise the behavior is undefined, but may exhibit less contention.
Returns:
Public attributes
Public functions
node
::tensorflow::Node * node() const
operator::tensorflow::Input() const
operator::tensorflow::Output
operator::tensorflow::Output() const
Public static functions
UpdateSlots
Attrs UpdateSlots(
bool x
)
UseLocking
Attrs UseLocking(
bool x
)
Except as otherwise noted, the content of this page is licensed under the Creative Commons Attribution 4.0 License, and code samples are licensed under the Apache 2.0 License. For details, see the Google Developers Site Policies. Java is a registered trademark of Oracle and/or its affiliates.
Last updated 2020-04-20 UTC.
[null,null,["Last updated 2020-04-20 UTC."],[],[],null,["# tensorflow::ops::ApplyAdagrad Class Reference\n\ntensorflow::ops::ApplyAdagrad\n=============================\n\n`#include \u003ctraining_ops.h\u003e`\n\nUpdate '\\*var' according to the adagrad scheme.\n\nSummary\n-------\n\naccum += grad \\* grad var -= lr \\* grad \\* (1 / sqrt(accum))\n\nArguments:\n\n- scope: A [Scope](/versions/r1.15/api_docs/cc/class/tensorflow/scope#classtensorflow_1_1_scope) object\n- var: Should be from a Variable().\n- accum: Should be from a Variable().\n- lr: Scaling factor. Must be a scalar.\n- grad: The gradient.\n\n\u003cbr /\u003e\n\nOptional attributes (see [Attrs](/versions/r1.15/api_docs/cc/struct/tensorflow/ops/apply-adagrad/attrs#structtensorflow_1_1ops_1_1_apply_adagrad_1_1_attrs)):\n\n- use_locking: If `True`, updating of the var and accum tensors will be protected by a lock; otherwise the behavior is undefined, but may exhibit less contention.\n\n\u003cbr /\u003e\n\nReturns:\n\n- [Output](/versions/r1.15/api_docs/cc/class/tensorflow/output#classtensorflow_1_1_output): Same as \"var\".\n\n\u003cbr /\u003e\n\n| ### Constructors and Destructors ||\n|---|---|\n| [ApplyAdagrad](#classtensorflow_1_1ops_1_1_apply_adagrad_1a119cb4675749a5746ce8598f8793a671)`(const ::`[tensorflow::Scope](/versions/r1.15/api_docs/cc/class/tensorflow/scope#classtensorflow_1_1_scope)` & scope, ::`[tensorflow::Input](/versions/r1.15/api_docs/cc/class/tensorflow/input#classtensorflow_1_1_input)` var, ::`[tensorflow::Input](/versions/r1.15/api_docs/cc/class/tensorflow/input#classtensorflow_1_1_input)` accum, ::`[tensorflow::Input](/versions/r1.15/api_docs/cc/class/tensorflow/input#classtensorflow_1_1_input)` lr, ::`[tensorflow::Input](/versions/r1.15/api_docs/cc/class/tensorflow/input#classtensorflow_1_1_input)` grad)` ||\n| [ApplyAdagrad](#classtensorflow_1_1ops_1_1_apply_adagrad_1a251be3569dad2f58e5820e423d32f28c)`(const ::`[tensorflow::Scope](/versions/r1.15/api_docs/cc/class/tensorflow/scope#classtensorflow_1_1_scope)` & scope, ::`[tensorflow::Input](/versions/r1.15/api_docs/cc/class/tensorflow/input#classtensorflow_1_1_input)` var, ::`[tensorflow::Input](/versions/r1.15/api_docs/cc/class/tensorflow/input#classtensorflow_1_1_input)` accum, ::`[tensorflow::Input](/versions/r1.15/api_docs/cc/class/tensorflow/input#classtensorflow_1_1_input)` lr, ::`[tensorflow::Input](/versions/r1.15/api_docs/cc/class/tensorflow/input#classtensorflow_1_1_input)` grad, const `[ApplyAdagrad::Attrs](/versions/r1.15/api_docs/cc/struct/tensorflow/ops/apply-adagrad/attrs#structtensorflow_1_1ops_1_1_apply_adagrad_1_1_attrs)` & attrs)` ||\n\n| ### Public attributes ||\n|-------------------------------------------------------------------------------------------|----------------------------------------------------------------------------------------------------------|\n| [operation](#classtensorflow_1_1ops_1_1_apply_adagrad_1a7e34dfd358db60a1c864ec24b278de1e) | [Operation](/versions/r1.15/api_docs/cc/class/tensorflow/operation#classtensorflow_1_1_operation) |\n| [out](#classtensorflow_1_1ops_1_1_apply_adagrad_1a3fe05f28ae9aa48b8351a53fc562e03d) | `::`[tensorflow::Output](/versions/r1.15/api_docs/cc/class/tensorflow/output#classtensorflow_1_1_output) |\n\n| ### Public functions ||\n|-------------------------------------------------------------------------------------------------------------------------|------------------------|\n| [node](#classtensorflow_1_1ops_1_1_apply_adagrad_1ac3e8e9b94978a64092adee8ae4836249)`() const ` | `::tensorflow::Node *` |\n| [operator::tensorflow::Input](#classtensorflow_1_1ops_1_1_apply_adagrad_1a881551beab3939a976b1a859acc77e89)`() const ` | ` ` ` ` |\n| [operator::tensorflow::Output](#classtensorflow_1_1ops_1_1_apply_adagrad_1a97a0d1b59f2d6bf45ed2281809bfba96)`() const ` | ` ` ` ` |\n\n| ### Public static functions ||\n|-------------------------------------------------------------------------------------------------------|------------------------------------------------------------------------------------------------------------------------------------|\n| [UpdateSlots](#classtensorflow_1_1ops_1_1_apply_adagrad_1aebe5e4875477a6e8f3634d4c5c33dba0)`(bool x)` | [Attrs](/versions/r1.15/api_docs/cc/struct/tensorflow/ops/apply-adagrad/attrs#structtensorflow_1_1ops_1_1_apply_adagrad_1_1_attrs) |\n| [UseLocking](#classtensorflow_1_1ops_1_1_apply_adagrad_1a412d43f4d6cbb6c10d0e49edf9f256f2)`(bool x)` | [Attrs](/versions/r1.15/api_docs/cc/struct/tensorflow/ops/apply-adagrad/attrs#structtensorflow_1_1ops_1_1_apply_adagrad_1_1_attrs) |\n\n| ### Structs ||\n|---------------------------------------------------------------------------------------------------------------|---------------------------------------------------------------------------------------------------------------------------------------------------------|\n| [tensorflow::ops::ApplyAdagrad::Attrs](/versions/r1.15/api_docs/cc/struct/tensorflow/ops/apply-adagrad/attrs) | Optional attribute setters for [ApplyAdagrad](/versions/r1.15/api_docs/cc/class/tensorflow/ops/apply-adagrad#classtensorflow_1_1ops_1_1_apply_adagrad). |\n\nPublic attributes\n-----------------\n\n### operation\n\n```text\nOperation operation\n``` \n\n### out\n\n```text\n::tensorflow::Output out\n``` \n\nPublic functions\n----------------\n\n### ApplyAdagrad\n\n```gdscript\n ApplyAdagrad(\n const ::tensorflow::Scope & scope,\n ::tensorflow::Input var,\n ::tensorflow::Input accum,\n ::tensorflow::Input lr,\n ::tensorflow::Input grad\n)\n``` \n\n### ApplyAdagrad\n\n```gdscript\n ApplyAdagrad(\n const ::tensorflow::Scope & scope,\n ::tensorflow::Input var,\n ::tensorflow::Input accum,\n ::tensorflow::Input lr,\n ::tensorflow::Input grad,\n const ApplyAdagrad::Attrs & attrs\n)\n``` \n\n### node\n\n```gdscript\n::tensorflow::Node * node() const \n``` \n\n### operator::tensorflow::Input\n\n```gdscript\n operator::tensorflow::Input() const \n``` \n\n### operator::tensorflow::Output\n\n```gdscript\n operator::tensorflow::Output() const \n``` \n\nPublic static functions\n-----------------------\n\n### UpdateSlots\n\n```text\nAttrs UpdateSlots(\n bool x\n)\n``` \n\n### UseLocking\n\n```text\nAttrs UseLocking(\n bool x\n)\n```"]]