aboutsummaryrefslogtreecommitdiffhomepage
path: root/third_party/googleapis/google/bigtable/v1/bigtable_service_messages.proto
diff options
context:
space:
mode:
Diffstat (limited to 'third_party/googleapis/google/bigtable/v1/bigtable_service_messages.proto')
-rw-r--r--third_party/googleapis/google/bigtable/v1/bigtable_service_messages.proto218
1 files changed, 218 insertions, 0 deletions
diff --git a/third_party/googleapis/google/bigtable/v1/bigtable_service_messages.proto b/third_party/googleapis/google/bigtable/v1/bigtable_service_messages.proto
new file mode 100644
index 0000000000..6d75af78e1
--- /dev/null
+++ b/third_party/googleapis/google/bigtable/v1/bigtable_service_messages.proto
@@ -0,0 +1,218 @@
+// Copyright 2016 Google Inc.
+//
+// Licensed under the Apache License, Version 2.0 (the "License");
+// you may not use this file except in compliance with the License.
+// You may obtain a copy of the License at
+//
+// http://www.apache.org/licenses/LICENSE-2.0
+//
+// Unless required by applicable law or agreed to in writing, software
+// distributed under the License is distributed on an "AS IS" BASIS,
+// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+// See the License for the specific language governing permissions and
+// limitations under the License.
+
+syntax = "proto3";
+
+package google.bigtable.v1;
+
+import "google/bigtable/v1/bigtable_data.proto";
+import "google/rpc/status.proto";
+
+option go_package = "google.golang.org/genproto/googleapis/bigtable/v1;bigtable";
+option java_multiple_files = true;
+option java_outer_classname = "BigtableServiceMessagesProto";
+option java_package = "com.google.bigtable.v1";
+
+
+// Request message for BigtableServer.ReadRows.
+message ReadRowsRequest {
+ // The unique name of the table from which to read.
+ string table_name = 1;
+
+ // If neither row_key nor row_range is set, reads from all rows.
+ oneof target {
+ // The key of a single row from which to read.
+ bytes row_key = 2;
+
+ // A range of rows from which to read.
+ RowRange row_range = 3;
+
+ // A set of rows from which to read. Entries need not be in order, and will
+ // be deduplicated before reading.
+ // The total serialized size of the set must not exceed 1MB.
+ RowSet row_set = 8;
+ }
+
+ // The filter to apply to the contents of the specified row(s). If unset,
+ // reads the entire table.
+ RowFilter filter = 5;
+
+ // By default, rows are read sequentially, producing results which are
+ // guaranteed to arrive in increasing row order. Setting
+ // "allow_row_interleaving" to true allows multiple rows to be interleaved in
+ // the response stream, which increases throughput but breaks this guarantee,
+ // and may force the client to use more memory to buffer partially-received
+ // rows. Cannot be set to true when specifying "num_rows_limit".
+ bool allow_row_interleaving = 6;
+
+ // The read will terminate after committing to N rows' worth of results. The
+ // default (zero) is to return all results.
+ // Note that "allow_row_interleaving" cannot be set to true when this is set.
+ int64 num_rows_limit = 7;
+}
+
+// Response message for BigtableService.ReadRows.
+message ReadRowsResponse {
+ // Specifies a piece of a row's contents returned as part of the read
+ // response stream.
+ message Chunk {
+ oneof chunk {
+ // A subset of the data from a particular row. As long as no "reset_row"
+ // is received in between, multiple "row_contents" from the same row are
+ // from the same atomic view of that row, and will be received in the
+ // expected family/column/timestamp order.
+ Family row_contents = 1;
+
+ // Indicates that the client should drop all previous chunks for
+ // "row_key", as it will be re-read from the beginning.
+ bool reset_row = 2;
+
+ // Indicates that the client can safely process all previous chunks for
+ // "row_key", as its data has been fully read.
+ bool commit_row = 3;
+ }
+ }
+
+ // The key of the row for which we're receiving data.
+ // Results will be received in increasing row key order, unless
+ // "allow_row_interleaving" was specified in the request.
+ bytes row_key = 1;
+
+ // One or more chunks of the row specified by "row_key".
+ repeated Chunk chunks = 2;
+}
+
+// Request message for BigtableService.SampleRowKeys.
+message SampleRowKeysRequest {
+ // The unique name of the table from which to sample row keys.
+ string table_name = 1;
+}
+
+// Response message for BigtableService.SampleRowKeys.
+message SampleRowKeysResponse {
+ // Sorted streamed sequence of sample row keys in the table. The table might
+ // have contents before the first row key in the list and after the last one,
+ // but a key containing the empty string indicates "end of table" and will be
+ // the last response given, if present.
+ // Note that row keys in this list may not have ever been written to or read
+ // from, and users should therefore not make any assumptions about the row key
+ // structure that are specific to their use case.
+ bytes row_key = 1;
+
+ // Approximate total storage space used by all rows in the table which precede
+ // "row_key". Buffering the contents of all rows between two subsequent
+ // samples would require space roughly equal to the difference in their
+ // "offset_bytes" fields.
+ int64 offset_bytes = 2;
+}
+
+// Request message for BigtableService.MutateRow.
+message MutateRowRequest {
+ // The unique name of the table to which the mutation should be applied.
+ string table_name = 1;
+
+ // The key of the row to which the mutation should be applied.
+ bytes row_key = 2;
+
+ // Changes to be atomically applied to the specified row. Entries are applied
+ // in order, meaning that earlier mutations can be masked by later ones.
+ // Must contain at least one entry and at most 100000.
+ repeated Mutation mutations = 3;
+}
+
+// Request message for BigtableService.MutateRows.
+message MutateRowsRequest {
+ message Entry {
+ // The key of the row to which the `mutations` should be applied.
+ bytes row_key = 1;
+
+ // Changes to be atomically applied to the specified row. Mutations are
+ // applied in order, meaning that earlier mutations can be masked by
+ // later ones.
+ // At least one mutation must be specified.
+ repeated Mutation mutations = 2;
+ }
+
+ // The unique name of the table to which the mutations should be applied.
+ string table_name = 1;
+
+ // The row keys/mutations to be applied in bulk.
+ // Each entry is applied as an atomic mutation, but the entries may be
+ // applied in arbitrary order (even between entries for the same row).
+ // At least one entry must be specified, and in total the entries may
+ // contain at most 100000 mutations.
+ repeated Entry entries = 2;
+}
+
+// Response message for BigtableService.MutateRows.
+message MutateRowsResponse {
+ // The results for each Entry from the request, presented in the order
+ // in which the entries were originally given.
+ // Depending on how requests are batched during execution, it is possible
+ // for one Entry to fail due to an error with another Entry. In the event
+ // that this occurs, the same error will be reported for both entries.
+ repeated google.rpc.Status statuses = 1;
+}
+
+// Request message for BigtableService.CheckAndMutateRowRequest
+message CheckAndMutateRowRequest {
+ // The unique name of the table to which the conditional mutation should be
+ // applied.
+ string table_name = 1;
+
+ // The key of the row to which the conditional mutation should be applied.
+ bytes row_key = 2;
+
+ // The filter to be applied to the contents of the specified row. Depending
+ // on whether or not any results are yielded, either "true_mutations" or
+ // "false_mutations" will be executed. If unset, checks that the row contains
+ // any values at all.
+ RowFilter predicate_filter = 6;
+
+ // Changes to be atomically applied to the specified row if "predicate_filter"
+ // yields at least one cell when applied to "row_key". Entries are applied in
+ // order, meaning that earlier mutations can be masked by later ones.
+ // Must contain at least one entry if "false_mutations" is empty, and at most
+ // 100000.
+ repeated Mutation true_mutations = 4;
+
+ // Changes to be atomically applied to the specified row if "predicate_filter"
+ // does not yield any cells when applied to "row_key". Entries are applied in
+ // order, meaning that earlier mutations can be masked by later ones.
+ // Must contain at least one entry if "true_mutations" is empty, and at most
+ // 100000.
+ repeated Mutation false_mutations = 5;
+}
+
+// Response message for BigtableService.CheckAndMutateRowRequest.
+message CheckAndMutateRowResponse {
+ // Whether or not the request's "predicate_filter" yielded any results for
+ // the specified row.
+ bool predicate_matched = 1;
+}
+
+// Request message for BigtableService.ReadModifyWriteRowRequest.
+message ReadModifyWriteRowRequest {
+ // The unique name of the table to which the read/modify/write rules should be
+ // applied.
+ string table_name = 1;
+
+ // The key of the row to which the read/modify/write rules should be applied.
+ bytes row_key = 2;
+
+ // Rules specifying how the specified row's contents are to be transformed
+ // into writes. Entries are applied in order, meaning that earlier rules will
+ // affect the results of later ones.
+ repeated ReadModifyWriteRule rules = 3;
+}