diff options
author | Cao Zongyan <zongyan.cao@alibaba-inc.com> | 2018-09-26 11:54:30 +0800 |
---|---|---|
committer | Cao Zongyan <zongyan.cao@alibaba-inc.com> | 2018-09-26 11:54:30 +0800 |
commit | 35174f46b973c66a2e6894a12b3018d60e8414ec (patch) | |
tree | 5bdae0172159bc02ec3a470722bf959b14dd47ba /tensorflow/core/api_def | |
parent | f0886f7269de900d226455d4831722f6fc94a71b (diff) | |
parent | 6666516f390f125ed70ddbd4e6f89b83d953c408 (diff) |
Merge remote-tracking branch 'origin'
Diffstat (limited to 'tensorflow/core/api_def')
26 files changed, 625 insertions, 7 deletions
diff --git a/tensorflow/core/api_def/base_api/api_def_BoostedTreesBucketize.pbtxt b/tensorflow/core/api_def/base_api/api_def_BoostedTreesBucketize.pbtxt new file mode 100644 index 0000000000..cdaeb5091c --- /dev/null +++ b/tensorflow/core/api_def/base_api/api_def_BoostedTreesBucketize.pbtxt @@ -0,0 +1,34 @@ +op { + graph_op_name: "BoostedTreesBucketize" + visibility: HIDDEN + in_arg { + name: "float_values" + description: <<END +float; List of Rank 2 Tensor each containing float values for a single feature. +END + } + in_arg { + name: "bucket_boundaries" + description: <<END +float; List of Rank 1 Tensors each containing the bucket boundaries for a single +feature. +END + } + out_arg { + name: "buckets" + description: <<END +int; List of Rank 2 Tensors each containing the bucketized values for a single feature. +END + } + attr { + name: "num_features" + description: <<END +inferred int; number of features. +END + } + summary: "Bucketize each feature based on bucket boundaries." + description: <<END +An op that returns a list of float tensors, where each tensor represents the +bucketized values for a single feature. +END +} diff --git a/tensorflow/core/api_def/base_api/api_def_BoostedTreesCreateQuantileStreamResource.pbtxt b/tensorflow/core/api_def/base_api/api_def_BoostedTreesCreateQuantileStreamResource.pbtxt new file mode 100644 index 0000000000..20da1295f6 --- /dev/null +++ b/tensorflow/core/api_def/base_api/api_def_BoostedTreesCreateQuantileStreamResource.pbtxt @@ -0,0 +1,29 @@ +op { + graph_op_name: "BoostedTreesCreateQuantileStreamResource" + visibility: HIDDEN + in_arg { + name: "quantile_stream_resource_handle" + description: <<END +resource; Handle to quantile stream resource. +END + } + in_arg { + name: "epsilon" + description: <<END +float; The required approximation error of the stream resource. +END + } + in_arg { + name: "num_streams" + description: <<END +int; The number of streams managed by the resource that shares the same epsilon. +END + } + attr { + name: "max_elements" + description : <<END +int; The maximum number of data points that can be fed to the stream. +END + } + summary: "Create the Resource for Quantile Streams." +} diff --git a/tensorflow/core/api_def/base_api/api_def_BoostedTreesMakeQuantileSummaries.pbtxt b/tensorflow/core/api_def/base_api/api_def_BoostedTreesMakeQuantileSummaries.pbtxt new file mode 100644 index 0000000000..ca111af312 --- /dev/null +++ b/tensorflow/core/api_def/base_api/api_def_BoostedTreesMakeQuantileSummaries.pbtxt @@ -0,0 +1,40 @@ +op { + graph_op_name: "BoostedTreesMakeQuantileSummaries" + visibility: HIDDEN + in_arg { + name: "float_values" + description: <<END +float; List of Rank 2 Tensors each containing values for a single feature. +END + } + in_arg { + name: "example_weights" + description: <<END +float; Rank 1 Tensor with weights per instance. +END + } + in_arg { + name: "epsilon" + description: <<END +float; The required maximum approximation error. +END + } + out_arg { + name: "summaries" + description: <<END +float; List of Rank 2 Tensors each containing the quantile summary (value, weight, +min_rank, max_rank) of a single feature. +END + } + attr { + name: "num_features" + description: <<END +int; Inferred from the size of float_values. +The number of float features. +END + } + summary: "Makes the summary of quantiles for the batch." + description: <<END +An op that takes a list of tensors and outputs the quantile summaries for each tensor. +END +} diff --git a/tensorflow/core/api_def/base_api/api_def_BoostedTreesQuantileStreamResourceAddSummaries.pbtxt b/tensorflow/core/api_def/base_api/api_def_BoostedTreesQuantileStreamResourceAddSummaries.pbtxt new file mode 100644 index 0000000000..bbeecbf32b --- /dev/null +++ b/tensorflow/core/api_def/base_api/api_def_BoostedTreesQuantileStreamResourceAddSummaries.pbtxt @@ -0,0 +1,22 @@ +op { + graph_op_name: "BoostedTreesQuantileStreamResourceAddSummaries" + visibility: HIDDEN + in_arg { + name: "quantile_stream_resource_handle" + description: <<END +resource handle referring to a QuantileStreamResource. +END + } + in_arg { + name: "summaries" + description: <<END +string; List of Rank 2 Tensor each containing the summaries for a single feature. +END + } + summary: "Add the quantile summaries to each quantile stream resource." + description: <<END +An op that adds a list of quantile summaries to a quantile stream resource. Each +summary Tensor is rank 2, containing summaries (value, weight, min_rank, max_rank) +for a single feature. +END +} diff --git a/tensorflow/core/api_def/base_api/api_def_BoostedTreesQuantileStreamResourceFlush.pbtxt b/tensorflow/core/api_def/base_api/api_def_BoostedTreesQuantileStreamResourceFlush.pbtxt new file mode 100644 index 0000000000..2fd94efa10 --- /dev/null +++ b/tensorflow/core/api_def/base_api/api_def_BoostedTreesQuantileStreamResourceFlush.pbtxt @@ -0,0 +1,31 @@ +op { + graph_op_name: "BoostedTreesQuantileStreamResourceFlush" + visibility: HIDDEN + in_arg { + name: "quantile_stream_resource_handle" + description: <<END +resource handle referring to a QuantileStreamResource. +END + } + in_arg { + name: "num_buckets", + description: <<END +int; approximate number of buckets unless using generate_quantiles. +END + } + attr { + name: "generate_quantiles" + description: <<END +bool; If True, the output will be the num_quantiles for each stream where the ith +entry is the ith quantile of the input with an approximation error of epsilon. +Duplicate values may be present. +If False, the output will be the points in the histogram that we got which roughly +translates to 1/epsilon boundaries and without any duplicates. +Default to False. +END + } + summary: "Flush the summaries for a quantile stream resource." + description: <<END +An op that flushes the summaries for a quantile stream resource. +END +} diff --git a/tensorflow/core/api_def/base_api/api_def_BoostedTreesQuantileStreamResourceGetBucketBoundaries.pbtxt b/tensorflow/core/api_def/base_api/api_def_BoostedTreesQuantileStreamResourceGetBucketBoundaries.pbtxt new file mode 100644 index 0000000000..206672802f --- /dev/null +++ b/tensorflow/core/api_def/base_api/api_def_BoostedTreesQuantileStreamResourceGetBucketBoundaries.pbtxt @@ -0,0 +1,27 @@ +op { + graph_op_name: "BoostedTreesQuantileStreamResourceGetBucketBoundaries" + visibility: HIDDEN + in_arg { + name: "quantile_stream_resource_handle" + description: <<END +resource handle referring to a QuantileStreamResource. +END + } + out_arg { + name: "bucket_boundaries" + description: <<END +float; List of Rank 1 Tensors each containing the bucket boundaries for a feature. +END + } + attr { + name: "num_features" + description: <<END +inferred int; number of features to get bucket boundaries for. +END + } + summary: "Generate the bucket boundaries for each feature based on accumulated summaries." + description: <<END +An op that returns a list of float tensors for a quantile stream resource. Each +tensor is Rank 1 containing bucket boundaries for a single feature. +END +} diff --git a/tensorflow/core/api_def/base_api/api_def_BoostedTreesQuantileStreamResourceHandleOp.pbtxt b/tensorflow/core/api_def/base_api/api_def_BoostedTreesQuantileStreamResourceHandleOp.pbtxt new file mode 100644 index 0000000000..cb7786c051 --- /dev/null +++ b/tensorflow/core/api_def/base_api/api_def_BoostedTreesQuantileStreamResourceHandleOp.pbtxt @@ -0,0 +1,5 @@ +op { + graph_op_name: "BoostedTreesQuantileStreamResourceHandleOp" + visibility: HIDDEN + summary: "Creates a handle to a BoostedTreesQuantileStreamResource." +} diff --git a/tensorflow/core/api_def/base_api/api_def_DecodeCSV.pbtxt b/tensorflow/core/api_def/base_api/api_def_DecodeCSV.pbtxt index e39213cbc7..440800704e 100644 --- a/tensorflow/core/api_def/base_api/api_def_DecodeCSV.pbtxt +++ b/tensorflow/core/api_def/base_api/api_def_DecodeCSV.pbtxt @@ -11,7 +11,8 @@ END name: "record_defaults" description: <<END One tensor per column of the input record, with either a -scalar default value for that column or empty if the column is required. +scalar default value for that column or an empty vector if the column is +required. END } out_arg { diff --git a/tensorflow/core/api_def/base_api/api_def_ExtractVolumePatches.pbtxt b/tensorflow/core/api_def/base_api/api_def_ExtractVolumePatches.pbtxt new file mode 100644 index 0000000000..3c8a455983 --- /dev/null +++ b/tensorflow/core/api_def/base_api/api_def_ExtractVolumePatches.pbtxt @@ -0,0 +1,49 @@ +op { + graph_op_name: "ExtractVolumePatches" + in_arg { + name: "input" + description: <<END +5-D Tensor with shape `[batch, in_planes, in_rows, in_cols, depth]`. +END + } + out_arg { + name: "patches" + description: <<END +5-D Tensor with shape `[batch, out_planes, out_rows, out_cols, +ksize_planes * ksize_rows * ksize_cols * depth]` containing patches +with size `ksize_planes x ksize_rows x ksize_cols x depth` vectorized +in the "depth" dimension. Note `out_planes`, `out_rows` and `out_cols` +are the dimensions of the output patches. +END + } + attr { + name: "ksizes" + description: <<END +The size of the sliding window for each dimension of `input`. +END + } + attr { + name: "strides" + description: <<END +1-D of length 5. How far the centers of two consecutive patches are in +`input`. Must be: `[1, stride_planes, stride_rows, stride_cols, 1]`. +END + } + attr { + name: "padding" + description: <<END +The type of padding algorithm to use. + +We specify the size-related attributes as: + +```python + ksizes = [1, ksize_planes, ksize_rows, ksize_cols, 1] + strides = [1, stride_planes, strides_rows, strides_cols, 1] +``` +END + } + summary: <<END +Extract `patches` from `input` and put them in the "depth" output +dimension. 3D extension of `extract_image_patches`. +END +} diff --git a/tensorflow/core/api_def/base_api/api_def_IsBoostedTreesQuantileStreamResourceInitialized.pbtxt b/tensorflow/core/api_def/base_api/api_def_IsBoostedTreesQuantileStreamResourceInitialized.pbtxt new file mode 100644 index 0000000000..758eeb96f0 --- /dev/null +++ b/tensorflow/core/api_def/base_api/api_def_IsBoostedTreesQuantileStreamResourceInitialized.pbtxt @@ -0,0 +1,20 @@ +op { + graph_op_name: "IsBoostedTreesQuantileStreamResourceInitialized" + visibility: HIDDEN + in_arg { + name: "quantile_stream_resource_handle" + description: <<END +resource; The reference to quantile stream resource handle. +END + } + out_arg { + name: "is_initialized" + description: <<END +bool; True if the resource is initialized, False otherwise. +END + } + summary: "Checks whether a quantile stream has been initialized." + description: <<END +An Op that checks if quantile stream resource is initialized. +END +} diff --git a/tensorflow/core/api_def/base_api/api_def_LowerBound.pbtxt b/tensorflow/core/api_def/base_api/api_def_LowerBound.pbtxt new file mode 100644 index 0000000000..5ce825ae04 --- /dev/null +++ b/tensorflow/core/api_def/base_api/api_def_LowerBound.pbtxt @@ -0,0 +1,45 @@ +op { + graph_op_name: "LowerBound" + visibility: HIDDEN + in_arg { + name: "sorted_inputs" + description: <<END +2-D Tensor where each row is ordered. +END + } + in_arg { + name: "values" + description: <<END +2-D Tensor with the same numbers of rows as `sorted_search_values`. Contains +the values that will be searched for in `sorted_search_values`. +END + } + out_arg { + name: "output" + description: <<END +A `Tensor` with the same shape as `values`. It contains the first scalar index +into the last dimension where values can be inserted without changing the +ordered property. +END + } + summary: "Applies lower_bound(sorted_search_values, values) along each row." + description: <<END +Each set of rows with the same index in (sorted_inputs, values) is treated +independently. The resulting row is the equivalent of calling +`np.searchsorted(sorted_inputs, values, side='left')`. + +The result is not a global index to the entire +`Tensor`, but rather just the index in the last dimension. + +A 2-D example: + sorted_sequence = [[0, 3, 9, 9, 10], + [1, 2, 3, 4, 5]] + values = [[2, 4, 9], + [0, 2, 6]] + + result = LowerBound(sorted_sequence, values) + + result == [[1, 2, 2], + [0, 1, 5]] +END +} diff --git a/tensorflow/core/api_def/base_api/api_def_ModelDataset.pbtxt b/tensorflow/core/api_def/base_api/api_def_ModelDataset.pbtxt new file mode 100644 index 0000000000..171add16d4 --- /dev/null +++ b/tensorflow/core/api_def/base_api/api_def_ModelDataset.pbtxt @@ -0,0 +1,14 @@ +op { + graph_op_name: "ModelDataset" + visibility: HIDDEN + in_arg { + name: "input_dataset" + description: <<END +A variant tensor representing the input dataset. +END + } + summary: "Identity transformation that models performance." + description: <<END +Identity transformation that models performance. +END +} diff --git a/tensorflow/core/api_def/base_api/api_def_MultiDeviceIterator.pbtxt b/tensorflow/core/api_def/base_api/api_def_MultiDeviceIterator.pbtxt new file mode 100644 index 0000000000..4b0a5d8f65 --- /dev/null +++ b/tensorflow/core/api_def/base_api/api_def_MultiDeviceIterator.pbtxt @@ -0,0 +1,43 @@ +op { + graph_op_name: "MultiDeviceIterator" + out_arg { + name: "handle" + description: <<END +Handle to the resource created. +END + } + attr { + name: "devices" + description: <<END +A list of devices the iterator works across. +END + } + attr { + name: "shared_name" + description: <<END +If non-empty, this resource will be shared under the given name +across multiple sessions. +END + } + attr { + name: "container" + description: <<END +If non-empty, this resource is placed in the given container. +Otherwise, a default container is used. +END + } + attr { + name: "output_types" + description: <<END +The type list for the return values. +END + } + attr { + name: "output_shapes" + description: <<END +The list of shapes being produced. +END + } + summary: "Creates a MultiDeviceIterator resource." + visibility: HIDDEN +} diff --git a/tensorflow/core/api_def/base_api/api_def_MultiDeviceIteratorFromStringHandle.pbtxt b/tensorflow/core/api_def/base_api/api_def_MultiDeviceIteratorFromStringHandle.pbtxt new file mode 100644 index 0000000000..adaacd8ab7 --- /dev/null +++ b/tensorflow/core/api_def/base_api/api_def_MultiDeviceIteratorFromStringHandle.pbtxt @@ -0,0 +1,29 @@ +op { + graph_op_name: "MultiDeviceIteratorFromStringHandle" + in_arg { + name: "string_handle" + description: <<END +String representing the resource. +END + } + out_arg { + name: "multi_device_iterator" + description: <<END +A MultiDeviceIterator resource. +END + } + attr { + name: "output_types" + description: <<END +The type list for the return values. +END + } + attr { + name: "output_shapes" + description: <<END +The list of shapes being produced. +END + } + summary: "Generates a MultiDeviceIterator resource from its provided string handle." + visibility: HIDDEN +} diff --git a/tensorflow/core/api_def/base_api/api_def_MultiDeviceIteratorGetNextFromShard.pbtxt b/tensorflow/core/api_def/base_api/api_def_MultiDeviceIteratorGetNextFromShard.pbtxt new file mode 100644 index 0000000000..f9be9188cc --- /dev/null +++ b/tensorflow/core/api_def/base_api/api_def_MultiDeviceIteratorGetNextFromShard.pbtxt @@ -0,0 +1,41 @@ +op { + graph_op_name: "MultiDeviceIteratorGetNextFromShard" + in_arg { + name: "multi_device_iterator" + description: <<END +A MultiDeviceIterator resource. +END + } + in_arg { + name: "shard_num" + description: <<END +Integer representing which shard to fetch data for. +END + } + in_arg { + name: "incarnation_id" + description: <<END +Which incarnation of the MultiDeviceIterator is running. +END + } + out_arg { + name: "components" + description: <<END +Result of the get_next on the dataset. +END + } + attr { + name: "output_types" + description: <<END +The type list for the return values. +END + } + attr { + name: "output_shapes" + description: <<END +The list of shapes being produced. +END + } + summary: "Gets next element for the provided shard number." + visibility: HIDDEN +} diff --git a/tensorflow/core/api_def/base_api/api_def_MultiDeviceIteratorInit.pbtxt b/tensorflow/core/api_def/base_api/api_def_MultiDeviceIteratorInit.pbtxt new file mode 100644 index 0000000000..6b54fa1307 --- /dev/null +++ b/tensorflow/core/api_def/base_api/api_def_MultiDeviceIteratorInit.pbtxt @@ -0,0 +1,30 @@ +op { + graph_op_name: "MultiDeviceIteratorInit" + in_arg { + name: "dataset" + description: <<END +Dataset to be iterated upon. +END + } + in_arg { + name: "multi_device_iterator" + description: <<END +A MultiDeviceIteratorResource. +END + } + in_arg { + name: "max_buffer_size" + description: <<END +The maximum size of the host side per device buffer to keep. +END + } + out_arg { + name: "incarnation_id" + description: <<END +An int64 indicating which incarnation of the MultiDeviceIterator +is running. +END + } + summary: "Initializes the multi device iterator with the given dataset." + visibility: HIDDEN +} diff --git a/tensorflow/core/api_def/base_api/api_def_MultiDeviceIteratorToStringHandle.pbtxt b/tensorflow/core/api_def/base_api/api_def_MultiDeviceIteratorToStringHandle.pbtxt new file mode 100644 index 0000000000..1f1fdf99b4 --- /dev/null +++ b/tensorflow/core/api_def/base_api/api_def_MultiDeviceIteratorToStringHandle.pbtxt @@ -0,0 +1,17 @@ +op { + graph_op_name: "MultiDeviceIteratorToStringHandle" + in_arg { + name: "multi_device_iterator" + description: <<END +A MultiDeviceIterator resource. +END + } + out_arg { + name: "string_handle" + description: <<END +A string representing the resource. +END + } + summary: "Produces a string handle for the given MultiDeviceIterator." + visibility: HIDDEN +} diff --git a/tensorflow/core/api_def/base_api/api_def_PrintV2.pbtxt b/tensorflow/core/api_def/base_api/api_def_PrintV2.pbtxt new file mode 100644 index 0000000000..4cb8955dcb --- /dev/null +++ b/tensorflow/core/api_def/base_api/api_def_PrintV2.pbtxt @@ -0,0 +1,19 @@ +op { + graph_op_name: "PrintV2" + in_arg { + name: "input" + description: <<END +The string scalar to print. +END + } + attr { + name: "output_stream" + description: <<END +A string specifying the output stream or logging level to print to. +END + } + summary: "Prints a string scalar." + description: <<END +Prints a string scalar to the desired output_stream. +END +} diff --git a/tensorflow/core/api_def/base_api/api_def_StringFormat.pbtxt b/tensorflow/core/api_def/base_api/api_def_StringFormat.pbtxt new file mode 100644 index 0000000000..a82dae9e48 --- /dev/null +++ b/tensorflow/core/api_def/base_api/api_def_StringFormat.pbtxt @@ -0,0 +1,38 @@ +op { + graph_op_name: "StringFormat" + in_arg { + name: "inputs" + description: <<END +The list of tensors to format into the placeholder string. +END + } + + out_arg { + name: "output" + description: <<END += The resulting string scalar. +END + } + attr { + name: "template" + description: <<END +A string, the template to format tensor summaries into. +END + } + attr { + name: "placeholder" + description: <<END +A string, at each placeholder in the template a subsequent tensor summary will be inserted. +END + } + attr { + name: "summarize" + description: <<END +When formatting the tensor summaries print the first and last summarize entries of each tensor dimension. +END + } + summary: "Formats a string template using a list of tensors." + description: <<END +Formats a string template using a list of tensors, pretty-printing tensor summaries. +END +} diff --git a/tensorflow/core/api_def/base_api/api_def_StringLength.pbtxt b/tensorflow/core/api_def/base_api/api_def_StringLength.pbtxt index cc21ddc815..7d2fbcd00b 100644 --- a/tensorflow/core/api_def/base_api/api_def_StringLength.pbtxt +++ b/tensorflow/core/api_def/base_api/api_def_StringLength.pbtxt @@ -1,5 +1,15 @@ op { graph_op_name: "StringLength" + attr { + name: "unit" + description: <<END +The unit that is counted to compute string length. One of: `"BYTE"` (for +the number of bytes in each string) or `"UTF8_CHAR"` (for the number of UTF-8 +encoded Unicode code points in each string). Results are undefined +if `unit=UTF8_CHAR` and the `input` strings do not contain structurally +valid UTF-8. +END + } in_arg { name: "input" description: <<END diff --git a/tensorflow/core/api_def/base_api/api_def_Substr.pbtxt b/tensorflow/core/api_def/base_api/api_def_Substr.pbtxt index 8fc1e5cba3..5246090ab3 100644 --- a/tensorflow/core/api_def/base_api/api_def_Substr.pbtxt +++ b/tensorflow/core/api_def/base_api/api_def_Substr.pbtxt @@ -32,8 +32,10 @@ For each string in the input `Tensor`, creates a substring starting at index If `len` defines a substring that would extend beyond the length of the input string, then as many characters as possible are used. -If `pos` is negative or specifies a character index larger than any of the input -strings, then an `InvalidArgumentError` is thrown. +A negative `pos` indicates distance within the string backwards from the end. + +If `pos` specifies an index which is out of range for any of the input strings, +then an `InvalidArgumentError` is thrown. `pos` and `len` must have the same shape, otherwise a `ValueError` is thrown on Op creation. diff --git a/tensorflow/core/api_def/base_api/api_def_UpperBound.pbtxt b/tensorflow/core/api_def/base_api/api_def_UpperBound.pbtxt new file mode 100644 index 0000000000..0630f6e82a --- /dev/null +++ b/tensorflow/core/api_def/base_api/api_def_UpperBound.pbtxt @@ -0,0 +1,45 @@ +op { + graph_op_name: "UpperBound" + visibility: HIDDEN + in_arg { + name: "sorted_inputs" + description: <<END +2-D Tensor where each row is ordered. +END + } + in_arg { + name: "values" + description: <<END +2-D Tensor with the same numbers of rows as `sorted_search_values`. Contains +the values that will be searched for in `sorted_search_values`. +END + } + out_arg { + name: "output" + description: <<END +A `Tensor` with the same shape as `values`. It contains the last scalar index +into the last dimension where values can be inserted without changing the +ordered property. +END + } + summary: "Applies upper_bound(sorted_search_values, values) along each row." + description: <<END +Each set of rows with the same index in (sorted_inputs, values) is treated +independently. The resulting row is the equivalent of calling +`np.searchsorted(sorted_inputs, values, side='right')`. + +The result is not a global index to the entire +`Tensor`, but rather just the index in the last dimension. + +A 2-D example: + sorted_sequence = [[0, 3, 9, 9, 10], + [1, 2, 3, 4, 5]] + values = [[2, 4, 9], + [0, 2, 6]] + + result = UpperBound(sorted_sequence, values) + + result == [[1, 2, 4], + [0, 2, 5]] +END +} diff --git a/tensorflow/core/api_def/base_api/api_def_WindowDataset.pbtxt b/tensorflow/core/api_def/base_api/api_def_WindowDataset.pbtxt index 1bc3660479..01387b7527 100644 --- a/tensorflow/core/api_def/base_api/api_def_WindowDataset.pbtxt +++ b/tensorflow/core/api_def/base_api/api_def_WindowDataset.pbtxt @@ -2,10 +2,31 @@ op { visibility: HIDDEN graph_op_name: "WindowDataset" in_arg { - name: "window_size" + name: "size" description: <<END A scalar representing the number of elements to accumulate in a window. END } + in_arg { + name: "shift" + description: <<END +A scalar representing the steps moving the sliding window forward in one +iteration. It must be positive. +END + } + in_arg { + name: "stride" + description: <<END +A scalar representing the stride of the input elements of the sliding window. +It must be positive. +END + } + in_arg { + name: "drop_remainder" + description: <<END +A scalar representing whether a window should be dropped in case its size is +smaller than desired. +END + } summary: "A dataset that creates window datasets from the input dataset." } diff --git a/tensorflow/core/api_def/python_api/api_def_PrintV2.pbtxt b/tensorflow/core/api_def/python_api/api_def_PrintV2.pbtxt new file mode 100644 index 0000000000..e22d980424 --- /dev/null +++ b/tensorflow/core/api_def/python_api/api_def_PrintV2.pbtxt @@ -0,0 +1,4 @@ +op { + graph_op_name: "PrintV2" + visibility: HIDDEN +} diff --git a/tensorflow/core/api_def/python_api/api_def_StringFormat.pbtxt b/tensorflow/core/api_def/python_api/api_def_StringFormat.pbtxt new file mode 100644 index 0000000000..8f0b1db45d --- /dev/null +++ b/tensorflow/core/api_def/python_api/api_def_StringFormat.pbtxt @@ -0,0 +1,4 @@ +op { + graph_op_name: "StringFormat" + visibility: HIDDEN +} diff --git a/tensorflow/core/api_def/python_api/api_def_StringLength.pbtxt b/tensorflow/core/api_def/python_api/api_def_StringLength.pbtxt index 01c02e1f70..df012414e3 100644 --- a/tensorflow/core/api_def/python_api/api_def_StringLength.pbtxt +++ b/tensorflow/core/api_def/python_api/api_def_StringLength.pbtxt @@ -1,6 +1,4 @@ op { graph_op_name: "StringLength" - endpoint { - name: "strings.length" - } + visibility: HIDDEN } |