#include <training_ops.h>
Update '*var' according to the proximal adagrad scheme.
Arguments:
Optional attributes (see Attrs
):
Returns:
Operation
Constructors and Destructors | |
---|---|
ResourceApplyAdagradDA(const ::tensorflow::Scope & scope, ::tensorflow::Input var, ::tensorflow::Input gradient_accumulator, ::tensorflow::Input gradient_squared_accumulator, ::tensorflow::Input grad, ::tensorflow::Input lr, ::tensorflow::Input l1, ::tensorflow::Input l2, ::tensorflow::Input global_step) | |
ResourceApplyAdagradDA(const ::tensorflow::Scope & scope, ::tensorflow::Input var, ::tensorflow::Input gradient_accumulator, ::tensorflow::Input gradient_squared_accumulator, ::tensorflow::Input grad, ::tensorflow::Input lr, ::tensorflow::Input l1, ::tensorflow::Input l2, ::tensorflow::Input global_step, const ResourceApplyAdagradDA::Attrs & attrs) |
Public attributes | |
---|---|
operation |
Public functions | |
---|---|
operator::tensorflow::Operation() const |
Public static functions | |
---|---|
UseLocking(bool x) |
Structs | |
---|---|
tensorflow::ops::ResourceApplyAdagradDA::Attrs | Optional attribute setters for ResourceApplyAdagradDA. |
Operation operation
ResourceApplyAdagradDA( const ::tensorflow::Scope & scope, ::tensorflow::Input var, ::tensorflow::Input gradient_accumulator, ::tensorflow::Input gradient_squared_accumulator, ::tensorflow::Input grad, ::tensorflow::Input lr, ::tensorflow::Input l1, ::tensorflow::Input l2, ::tensorflow::Input global_step )
ResourceApplyAdagradDA( const ::tensorflow::Scope & scope, ::tensorflow::Input var, ::tensorflow::Input gradient_accumulator, ::tensorflow::Input gradient_squared_accumulator, ::tensorflow::Input grad, ::tensorflow::Input lr, ::tensorflow::Input l1, ::tensorflow::Input l2, ::tensorflow::Input global_step, const ResourceApplyAdagradDA::Attrs & attrs )
operator::tensorflow::Operation() const
Attrs UseLocking( bool x )
© 2017 The TensorFlow Authors. All rights reserved.
Licensed under the Creative Commons Attribution License 3.0.
Code samples licensed under the Apache 2.0 License.
https://www.tensorflow.org/api_docs/cc/class/tensorflow/ops/resource-apply-adagrad-d-a.html