Training Ops
Summary
Classes |
|
---|---|
tensorflow::
|
Update '*var' according to the adadelta scheme. |
tensorflow::
|
Update '*var' according to the adagrad scheme. |
tensorflow::
|
Update '*var' according to the proximal adagrad scheme. |
tensorflow::
|
Update '*var' according to the Adam algorithm. |
tensorflow::
|
Update '*var' according to the AddSign update. |
tensorflow::
|
Update '*var' according to the centered RMSProp algorithm. |
tensorflow::
|
Update '*var' according to the Ftrl-proximal scheme. |
tensorflow::
|
Update '*var' according to the Ftrl-proximal scheme. |
tensorflow::
|
Update '*var' by subtracting 'alpha' * 'delta' from it. |
tensorflow::
|
Update '*var' according to the momentum scheme. |
tensorflow::
|
Update '*var' according to the AddSign update. |
tensorflow::
|
Update '*var' and '*accum' according to FOBOS with Adagrad learning rate. |
tensorflow::
|
Update '*var' as FOBOS algorithm with fixed learning rate. |
tensorflow::
|
Update '*var' according to the RMSProp algorithm. |
tensorflow::
|
Update '*var' according to the adadelta scheme. |
tensorflow::
|
Update '*var' according to the adagrad scheme. |
tensorflow::
|
Update '*var' according to the proximal adagrad scheme. |
tensorflow::
|
Update '*var' according to the Adam algorithm. |
tensorflow::
|
Update '*var' according to the Adam algorithm. |
tensorflow::
|
Update '*var' according to the AddSign update. |
tensorflow::
|
Update '*var' according to the centered RMSProp algorithm. |
tensorflow::
|
Update '*var' according to the Ftrl-proximal scheme. |
tensorflow::
|
Update '*var' according to the Ftrl-proximal scheme. |
tensorflow::
|
Update '*var' by subtracting 'alpha' * 'delta' from it. |
tensorflow::
|
Update '*var' according to the momentum scheme. |
tensorflow::
|
Update '*var' according to the momentum scheme. |
tensorflow::
|
Update '*var' according to the AddSign update. |
tensorflow::
|
Update '*var' and '*accum' according to FOBOS with Adagrad learning rate. |
tensorflow::
|
Update '*var' as FOBOS algorithm with fixed learning rate. |
tensorflow::
|
Update '*var' according to the RMSProp algorithm. |
tensorflow::
|
var: Should be from a Variable(). |
tensorflow::
|
Update relevant entries in '*var' and '*accum' according to the adagrad scheme. |
tensorflow::
|
Update entries in '*var' and '*accum' according to the proximal adagrad scheme. |
tensorflow::
|
Update '*var' according to the centered RMSProp algorithm. |
tensorflow::
|
Update relevant entries in '*var' according to the Ftrl-proximal scheme. |
tensorflow::
|
Update relevant entries in '*var' according to the Ftrl-proximal scheme. |
tensorflow::
|
Update relevant entries in '*var' and '*accum' according to the momentum scheme. |
tensorflow::
|
Update relevant entries in '*var' and '*accum' according to the momentum scheme. |
tensorflow::
|
Sparse update entries in '*var' and '*accum' according to FOBOS algorithm. |
tensorflow::
|
Sparse update '*var' as FOBOS algorithm with fixed learning rate. |
tensorflow::
|
Update '*var' according to the RMSProp algorithm. |
tensorflow::
|
var: Should be from a Variable(). |
tensorflow::
|
Update relevant entries in '*var' and '*accum' according to the adagrad scheme. |
tensorflow::
|
Update entries in '*var' and '*accum' according to the proximal adagrad scheme. |
tensorflow::
|
Update '*var' according to the centered RMSProp algorithm. |
tensorflow::
|
Update relevant entries in '*var' according to the Ftrl-proximal scheme. |
tensorflow::
|
Update relevant entries in '*var' according to the Ftrl-proximal scheme. |
tensorflow::
|
Update relevant entries in '*var' and '*accum' according to the momentum scheme. |
tensorflow::
|
Sparse update entries in '*var' and '*accum' according to FOBOS algorithm. |
tensorflow::
|
Sparse update '*var' as FOBOS algorithm with fixed learning rate. |
tensorflow::
|
Update '*var' according to the RMSProp algorithm. |