diff options
author | Mark Daoust <markdaoust@google.com> | 2018-09-28 11:12:42 -0700 |
---|---|---|
committer | TensorFlower Gardener <gardener@tensorflow.org> | 2018-09-28 11:21:49 -0700 |
commit | 1a834d3aa84ba47afe39c22fffd60d03ca151d30 (patch) | |
tree | eed1e3c7bab04b64c9e505ddf0ec82e69f80dc29 /tensorflow/python/distribute | |
parent | d644fa0542a5a9995512674c7ac708468941fe28 (diff) |
Remove @{} api_links and ban "@{}" from python and md files.
PiperOrigin-RevId: 214964988
Diffstat (limited to 'tensorflow/python/distribute')
-rw-r--r-- | tensorflow/python/distribute/distribute_coordinator.py | 4 | ||||
-rw-r--r-- | tensorflow/python/distribute/estimator_training.py | 2 |
2 files changed, 3 insertions, 3 deletions
diff --git a/tensorflow/python/distribute/distribute_coordinator.py b/tensorflow/python/distribute/distribute_coordinator.py index bd3562f1ff..b9b77d4a5b 100644 --- a/tensorflow/python/distribute/distribute_coordinator.py +++ b/tensorflow/python/distribute/distribute_coordinator.py @@ -126,7 +126,7 @@ class _WorkerContext(object): replicated training. task_id: an integer indicating id of the corresponding task. It can be None if it is local training or in-graph replicated training. - session_config: an optional @{tf.ConfigProto} object. + session_config: an optional `tf.ConfigProto` object. rpc_layer: optional string specifying the RPC protocol for communication with worker masters. If None or empty, hosts in the `cluster_spec` will be used directly. @@ -685,7 +685,7 @@ def run_distribute_coordinator(worker_fn, in a cluster. If not set or empty, fall back to local training. task_type: the current task type, optional if this is a client. task_id: the current task id, optional if this is a client. - session_config: an optional @{tf.ConfigProto} object which will be passed + session_config: an optional `tf.ConfigProto` object which will be passed to `strategy`'s `configure` method and used to create a session. rpc_layer: optional string, the protocol for RPC, e.g. "grpc". diff --git a/tensorflow/python/distribute/estimator_training.py b/tensorflow/python/distribute/estimator_training.py index 8daa34c885..0289689134 100644 --- a/tensorflow/python/distribute/estimator_training.py +++ b/tensorflow/python/distribute/estimator_training.py @@ -62,7 +62,7 @@ def _get_global_id(cluster_spec, task_type, task_id, chief_task_type): # Sort task names in cluster by "chief"/"master", "evaluator", "worker" # and "ps". More details can be found at the documentation of - # @{tf.estimator.RunConfig.global_id_in_cluster}. + # `tf.estimator.RunConfig.global_id_in_cluster`. task_type_ordered_list = [] if chief_task_type in cluster_spec.jobs: task_type_ordered_list = [chief_task_type] |