aboutsummaryrefslogtreecommitdiffhomepage
path: root/tensorflow/docs_src/api_guides/python/input_dataset.md
diff options
context:
space:
mode:
authorGravatar Mark Daoust <markdaoust@google.com>2018-08-07 14:28:32 -0700
committerGravatar TensorFlower Gardener <gardener@tensorflow.org>2018-08-07 14:32:57 -0700
commit02df0f46d562a0c48b6f24803eba6330d13d7213 (patch)
tree3a39933e12b7300ddcefb5afb90db053f295d824 /tensorflow/docs_src/api_guides/python/input_dataset.md
parent452f995e2c23cbd67c14b15b678bb3a352212633 (diff)
Remove usage of magic-api-link syntax from docs.
Back-ticks are now converted to links in the api_docs generator. With the new docs repo we're moving to simplify the docs pipeline, and make everything more readable. By doing this we no longer get test failures for symbols that don't exist (`tf.does_not_exist` will not get a link). There is also no way to set custom link text now. That's okay. This is the result of the following regex replacement (+ a couple of manual edits.): re: @\{([^$].*?)(\$.+?)?} sub: `\1` Which does the following replacements: "@{tf.symbol}" --> "`tf.symbol`" "@{tf.symbol$link_text}" --> "`tf.symbol`" PiperOrigin-RevId: 207780049
Diffstat (limited to 'tensorflow/docs_src/api_guides/python/input_dataset.md')
-rw-r--r--tensorflow/docs_src/api_guides/python/input_dataset.md96
1 files changed, 48 insertions, 48 deletions
diff --git a/tensorflow/docs_src/api_guides/python/input_dataset.md b/tensorflow/docs_src/api_guides/python/input_dataset.md
index a6612d1bf7..ab572e53d4 100644
--- a/tensorflow/docs_src/api_guides/python/input_dataset.md
+++ b/tensorflow/docs_src/api_guides/python/input_dataset.md
@@ -1,27 +1,27 @@
# Dataset Input Pipeline
[TOC]
-@{tf.data.Dataset} allows you to build complex input pipelines. See the
+`tf.data.Dataset` allows you to build complex input pipelines. See the
@{$guide/datasets} for an in-depth explanation of how to use this API.
## Reader classes
Classes that create a dataset from input files.
-* @{tf.data.FixedLengthRecordDataset}
-* @{tf.data.TextLineDataset}
-* @{tf.data.TFRecordDataset}
+* `tf.data.FixedLengthRecordDataset`
+* `tf.data.TextLineDataset`
+* `tf.data.TFRecordDataset`
## Creating new datasets
Static methods in `Dataset` that create new datasets.
-* @{tf.data.Dataset.from_generator}
-* @{tf.data.Dataset.from_tensor_slices}
-* @{tf.data.Dataset.from_tensors}
-* @{tf.data.Dataset.list_files}
-* @{tf.data.Dataset.range}
-* @{tf.data.Dataset.zip}
+* `tf.data.Dataset.from_generator`
+* `tf.data.Dataset.from_tensor_slices`
+* `tf.data.Dataset.from_tensors`
+* `tf.data.Dataset.list_files`
+* `tf.data.Dataset.range`
+* `tf.data.Dataset.zip`
## Transformations on existing datasets
@@ -32,54 +32,54 @@ can be chained together, as shown in the example below:
train_data = train_data.batch(100).shuffle().repeat()
```
-* @{tf.data.Dataset.apply}
-* @{tf.data.Dataset.batch}
-* @{tf.data.Dataset.cache}
-* @{tf.data.Dataset.concatenate}
-* @{tf.data.Dataset.filter}
-* @{tf.data.Dataset.flat_map}
-* @{tf.data.Dataset.interleave}
-* @{tf.data.Dataset.map}
-* @{tf.data.Dataset.padded_batch}
-* @{tf.data.Dataset.prefetch}
-* @{tf.data.Dataset.repeat}
-* @{tf.data.Dataset.shard}
-* @{tf.data.Dataset.shuffle}
-* @{tf.data.Dataset.skip}
-* @{tf.data.Dataset.take}
+* `tf.data.Dataset.apply`
+* `tf.data.Dataset.batch`
+* `tf.data.Dataset.cache`
+* `tf.data.Dataset.concatenate`
+* `tf.data.Dataset.filter`
+* `tf.data.Dataset.flat_map`
+* `tf.data.Dataset.interleave`
+* `tf.data.Dataset.map`
+* `tf.data.Dataset.padded_batch`
+* `tf.data.Dataset.prefetch`
+* `tf.data.Dataset.repeat`
+* `tf.data.Dataset.shard`
+* `tf.data.Dataset.shuffle`
+* `tf.data.Dataset.skip`
+* `tf.data.Dataset.take`
### Custom transformation functions
-Custom transformation functions can be applied to a `Dataset` using @{tf.data.Dataset.apply}. Below are custom transformation functions from `tf.contrib.data`:
-
-* @{tf.contrib.data.batch_and_drop_remainder}
-* @{tf.contrib.data.dense_to_sparse_batch}
-* @{tf.contrib.data.enumerate_dataset}
-* @{tf.contrib.data.group_by_window}
-* @{tf.contrib.data.ignore_errors}
-* @{tf.contrib.data.map_and_batch}
-* @{tf.contrib.data.padded_batch_and_drop_remainder}
-* @{tf.contrib.data.parallel_interleave}
-* @{tf.contrib.data.rejection_resample}
-* @{tf.contrib.data.scan}
-* @{tf.contrib.data.shuffle_and_repeat}
-* @{tf.contrib.data.unbatch}
+Custom transformation functions can be applied to a `Dataset` using `tf.data.Dataset.apply`. Below are custom transformation functions from `tf.contrib.data`:
+
+* `tf.contrib.data.batch_and_drop_remainder`
+* `tf.contrib.data.dense_to_sparse_batch`
+* `tf.contrib.data.enumerate_dataset`
+* `tf.contrib.data.group_by_window`
+* `tf.contrib.data.ignore_errors`
+* `tf.contrib.data.map_and_batch`
+* `tf.contrib.data.padded_batch_and_drop_remainder`
+* `tf.contrib.data.parallel_interleave`
+* `tf.contrib.data.rejection_resample`
+* `tf.contrib.data.scan`
+* `tf.contrib.data.shuffle_and_repeat`
+* `tf.contrib.data.unbatch`
## Iterating over datasets
-These functions make a @{tf.data.Iterator} from a `Dataset`.
+These functions make a `tf.data.Iterator` from a `Dataset`.
-* @{tf.data.Dataset.make_initializable_iterator}
-* @{tf.data.Dataset.make_one_shot_iterator}
+* `tf.data.Dataset.make_initializable_iterator`
+* `tf.data.Dataset.make_one_shot_iterator`
-The `Iterator` class also contains static methods that create a @{tf.data.Iterator} that can be used with multiple `Dataset` objects.
+The `Iterator` class also contains static methods that create a `tf.data.Iterator` that can be used with multiple `Dataset` objects.
-* @{tf.data.Iterator.from_structure}
-* @{tf.data.Iterator.from_string_handle}
+* `tf.data.Iterator.from_structure`
+* `tf.data.Iterator.from_string_handle`
## Extra functions from `tf.contrib.data`
-* @{tf.contrib.data.get_single_element}
-* @{tf.contrib.data.make_saveable_from_iterator}
-* @{tf.contrib.data.read_batch_features}
+* `tf.contrib.data.get_single_element`
+* `tf.contrib.data.make_saveable_from_iterator`
+* `tf.contrib.data.read_batch_features`