aboutsummaryrefslogtreecommitdiffhomepage
path: root/tensorflow/tools/api/golden/tensorflow.train.-proximal-adagrad-optimizer.pbtxt
diff options
context:
space:
mode:
authorGravatar Martin Wicke <wicke@google.com>2017-04-07 11:19:20 -0800
committerGravatar TensorFlower Gardener <gardener@tensorflow.org>2017-04-07 12:29:19 -0700
commit2238fd708c191e3939a9caee0de7cf74a4b391f6 (patch)
treef952cca2b36be6ca297da9f948cdee6a48954c55 /tensorflow/tools/api/golden/tensorflow.train.-proximal-adagrad-optimizer.pbtxt
parent9a1b76bc77c06d65799c1f0352eb5dc1dd5b932f (diff)
Automated rollback of change 152310869
Change: 152528732
Diffstat (limited to 'tensorflow/tools/api/golden/tensorflow.train.-proximal-adagrad-optimizer.pbtxt')
-rw-r--r--tensorflow/tools/api/golden/tensorflow.train.-proximal-adagrad-optimizer.pbtxt46
1 files changed, 0 insertions, 46 deletions
diff --git a/tensorflow/tools/api/golden/tensorflow.train.-proximal-adagrad-optimizer.pbtxt b/tensorflow/tools/api/golden/tensorflow.train.-proximal-adagrad-optimizer.pbtxt
deleted file mode 100644
index 571d846b6c..0000000000
--- a/tensorflow/tools/api/golden/tensorflow.train.-proximal-adagrad-optimizer.pbtxt
+++ /dev/null
@@ -1,46 +0,0 @@
-path: "tensorflow.train.ProximalAdagradOptimizer"
-tf_class {
- is_instance: "<class \'tensorflow.python.training.proximal_adagrad.ProximalAdagradOptimizer\'>"
- is_instance: "<class \'tensorflow.python.training.optimizer.Optimizer\'>"
- is_instance: "<type \'object\'>"
- member {
- name: "GATE_GRAPH"
- mtype: "<type \'int\'>"
- }
- member {
- name: "GATE_NONE"
- mtype: "<type \'int\'>"
- }
- member {
- name: "GATE_OP"
- mtype: "<type \'int\'>"
- }
- member_method {
- name: "__init__"
- argspec: "args=[\'self\', \'learning_rate\', \'initial_accumulator_value\', \'l1_regularization_strength\', \'l2_regularization_strength\', \'use_locking\', \'name\'], varargs=None, keywords=None, defaults=[\'0.1\', \'0.0\', \'0.0\', \'False\', \'ProximalAdagrad\'], "
- }
- member_method {
- name: "apply_gradients"
- argspec: "args=[\'self\', \'grads_and_vars\', \'global_step\', \'name\'], varargs=None, keywords=None, defaults=[\'None\', \'None\'], "
- }
- member_method {
- name: "compute_gradients"
- argspec: "args=[\'self\', \'loss\', \'var_list\', \'gate_gradients\', \'aggregation_method\', \'colocate_gradients_with_ops\', \'grad_loss\'], varargs=None, keywords=None, defaults=[\'None\', \'1\', \'None\', \'False\', \'None\'], "
- }
- member_method {
- name: "get_name"
- argspec: "args=[\'self\'], varargs=None, keywords=None, defaults=None"
- }
- member_method {
- name: "get_slot"
- argspec: "args=[\'self\', \'var\', \'name\'], varargs=None, keywords=None, defaults=None"
- }
- member_method {
- name: "get_slot_names"
- argspec: "args=[\'self\'], varargs=None, keywords=None, defaults=None"
- }
- member_method {
- name: "minimize"
- argspec: "args=[\'self\', \'loss\', \'global_step\', \'var_list\', \'gate_gradients\', \'aggregation_method\', \'colocate_gradients_with_ops\', \'name\', \'grad_loss\'], varargs=None, keywords=None, defaults=[\'None\', \'None\', \'1\', \'None\', \'False\', \'None\', \'None\'], "
- }
-}