aboutsummaryrefslogtreecommitdiffhomepage
path: root/test/core/end2end/fuzzers/api_fuzzer.c
diff options
context:
space:
mode:
Diffstat (limited to 'test/core/end2end/fuzzers/api_fuzzer.c')
-rw-r--r--test/core/end2end/fuzzers/api_fuzzer.c896
1 files changed, 896 insertions, 0 deletions
diff --git a/test/core/end2end/fuzzers/api_fuzzer.c b/test/core/end2end/fuzzers/api_fuzzer.c
new file mode 100644
index 0000000000..b133a948ee
--- /dev/null
+++ b/test/core/end2end/fuzzers/api_fuzzer.c
@@ -0,0 +1,896 @@
+/*
+ *
+ * Copyright 2016, Google Inc.
+ * All rights reserved.
+ *
+ * Redistribution and use in source and binary forms, with or without
+ * modification, are permitted provided that the following conditions are
+ * met:
+ *
+ * * Redistributions of source code must retain the above copyright
+ * notice, this list of conditions and the following disclaimer.
+ * * Redistributions in binary form must reproduce the above
+ * copyright notice, this list of conditions and the following disclaimer
+ * in the documentation and/or other materials provided with the
+ * distribution.
+ * * Neither the name of Google Inc. nor the names of its
+ * contributors may be used to endorse or promote products derived from
+ * this software without specific prior written permission.
+ *
+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
+ * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
+ * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
+ * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
+ * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
+ * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
+ * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
+ * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
+ * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
+ * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
+ *
+ */
+
+#include <string.h>
+
+#include <grpc/grpc.h>
+#include <grpc/support/alloc.h>
+#include <grpc/support/log.h>
+#include <grpc/support/string_util.h>
+
+#include "src/core/ext/transport/chttp2/transport/chttp2_transport.h"
+#include "src/core/lib/channel/channel_args.h"
+#include "src/core/lib/iomgr/resolve_address.h"
+#include "src/core/lib/iomgr/tcp_client.h"
+#include "src/core/lib/iomgr/timer.h"
+#include "src/core/lib/surface/server.h"
+#include "src/core/lib/transport/metadata.h"
+#include "test/core/util/passthru_endpoint.h"
+
+////////////////////////////////////////////////////////////////////////////////
+// logging
+
+static const bool squelch = true;
+
+static void dont_log(gpr_log_func_args *args) {}
+
+////////////////////////////////////////////////////////////////////////////////
+// input_stream: allows easy access to input bytes, and allows reading a little
+// past the end (avoiding needing to check everywhere)
+
+typedef struct {
+ const uint8_t *cur;
+ const uint8_t *end;
+} input_stream;
+
+static uint8_t next_byte(input_stream *inp) {
+ if (inp->cur == inp->end) {
+ return 0;
+ }
+ return *inp->cur++;
+}
+
+static void end(input_stream *inp) { inp->cur = inp->end; }
+
+static char *read_string(input_stream *inp) {
+ char *str = NULL;
+ size_t cap = 0;
+ size_t sz = 0;
+ char c;
+ do {
+ if (cap == sz) {
+ cap = GPR_MAX(3 * cap / 2, cap + 8);
+ str = gpr_realloc(str, cap);
+ }
+ c = (char)next_byte(inp);
+ str[sz++] = c;
+ } while (c != 0);
+ return str;
+}
+
+static void read_buffer(input_stream *inp, char **buffer, size_t *length) {
+ *length = next_byte(inp);
+ *buffer = gpr_malloc(*length);
+ for (size_t i = 0; i < *length; i++) {
+ (*buffer)[i] = (char)next_byte(inp);
+ }
+}
+
+static uint32_t read_uint22(input_stream *inp) {
+ uint8_t b = next_byte(inp);
+ uint32_t x = b & 0x7f;
+ if (b & 0x80) {
+ x <<= 7;
+ b = next_byte(inp);
+ x |= b & 0x7f;
+ if (b & 0x80) {
+ x <<= 8;
+ x |= next_byte(inp);
+ }
+ }
+ return x;
+}
+
+static uint32_t read_uint32(input_stream *inp) {
+ uint8_t b = next_byte(inp);
+ uint32_t x = b & 0x7f;
+ if (b & 0x80) {
+ x <<= 7;
+ b = next_byte(inp);
+ x |= b & 0x7f;
+ if (b & 0x80) {
+ x <<= 7;
+ b = next_byte(inp);
+ x |= b & 0x7f;
+ if (b & 0x80) {
+ x <<= 7;
+ b = next_byte(inp);
+ x |= b & 0x7f;
+ if (b & 0x80) {
+ x = (x << 4) | (next_byte(inp) & 0x0f);
+ }
+ }
+ }
+ }
+ return x;
+}
+
+static grpc_byte_buffer *read_message(input_stream *inp) {
+ gpr_slice slice = gpr_slice_malloc(read_uint22(inp));
+ memset(GPR_SLICE_START_PTR(slice), 0, GPR_SLICE_LENGTH(slice));
+ grpc_byte_buffer *out = grpc_raw_byte_buffer_create(&slice, 1);
+ gpr_slice_unref(slice);
+ return out;
+}
+
+static int read_int(input_stream *inp) { return (int)read_uint32(inp); }
+
+static grpc_channel_args *read_args(input_stream *inp) {
+ size_t n = next_byte(inp);
+ grpc_arg *args = gpr_malloc(sizeof(*args) * n);
+ for (size_t i = 0; i < n; i++) {
+ bool is_string = next_byte(inp) & 1;
+ args[i].type = is_string ? GRPC_ARG_STRING : GRPC_ARG_INTEGER;
+ args[i].key = read_string(inp);
+ if (is_string) {
+ args[i].value.string = read_string(inp);
+ } else {
+ args[i].value.integer = read_int(inp);
+ }
+ }
+ grpc_channel_args *a = gpr_malloc(sizeof(*a));
+ a->args = args;
+ a->num_args = n;
+ return a;
+}
+
+static bool is_eof(input_stream *inp) { return inp->cur == inp->end; }
+
+////////////////////////////////////////////////////////////////////////////////
+// global state
+
+static gpr_timespec g_now;
+static grpc_server *g_server;
+static grpc_channel *g_channel;
+
+extern gpr_timespec (*gpr_now_impl)(gpr_clock_type clock_type);
+
+static gpr_timespec now_impl(gpr_clock_type clock_type) {
+ GPR_ASSERT(clock_type != GPR_TIMESPAN);
+ return g_now;
+}
+
+////////////////////////////////////////////////////////////////////////////////
+// dns resolution
+
+typedef struct addr_req {
+ grpc_timer timer;
+ char *addr;
+ grpc_resolve_cb cb;
+ void *arg;
+} addr_req;
+
+static void finish_resolve(grpc_exec_ctx *exec_ctx, void *arg, bool success) {
+ addr_req *r = arg;
+
+ if (success && 0 == strcmp(r->addr, "server")) {
+ grpc_resolved_addresses *addrs = gpr_malloc(sizeof(*addrs));
+ addrs->naddrs = 1;
+ addrs->addrs = gpr_malloc(sizeof(*addrs->addrs));
+ addrs->addrs[0].len = 0;
+ r->cb(exec_ctx, r->arg, addrs);
+ } else {
+ r->cb(exec_ctx, r->arg, NULL);
+ }
+
+ gpr_free(r->addr);
+ gpr_free(r);
+}
+
+void my_resolve_address(grpc_exec_ctx *exec_ctx, const char *addr,
+ const char *default_port, grpc_resolve_cb cb,
+ void *arg) {
+ addr_req *r = gpr_malloc(sizeof(*r));
+ r->addr = gpr_strdup(addr);
+ r->cb = cb;
+ r->arg = arg;
+ grpc_timer_init(exec_ctx, &r->timer,
+ gpr_time_add(gpr_now(GPR_CLOCK_MONOTONIC),
+ gpr_time_from_seconds(1, GPR_TIMESPAN)),
+ finish_resolve, r, gpr_now(GPR_CLOCK_MONOTONIC));
+}
+
+////////////////////////////////////////////////////////////////////////////////
+// client connection
+
+// defined in tcp_client_posix.c
+extern void (*grpc_tcp_client_connect_impl)(
+ grpc_exec_ctx *exec_ctx, grpc_closure *closure, grpc_endpoint **ep,
+ grpc_pollset_set *interested_parties, const struct sockaddr *addr,
+ size_t addr_len, gpr_timespec deadline);
+
+static void sched_connect(grpc_exec_ctx *exec_ctx, grpc_closure *closure,
+ grpc_endpoint **ep, gpr_timespec deadline);
+
+typedef struct {
+ grpc_timer timer;
+ grpc_closure *closure;
+ grpc_endpoint **ep;
+ gpr_timespec deadline;
+} future_connect;
+
+static void do_connect(grpc_exec_ctx *exec_ctx, void *arg, bool success) {
+ future_connect *fc = arg;
+ if (!success) {
+ *fc->ep = NULL;
+ grpc_exec_ctx_enqueue(exec_ctx, fc->closure, false, NULL);
+ } else if (g_server != NULL) {
+ grpc_endpoint *client;
+ grpc_endpoint *server;
+ grpc_passthru_endpoint_create(&client, &server);
+ *fc->ep = client;
+
+ grpc_transport *transport =
+ grpc_create_chttp2_transport(exec_ctx, NULL, server, 0);
+ grpc_server_setup_transport(exec_ctx, g_server, transport, NULL);
+ grpc_chttp2_transport_start_reading(exec_ctx, transport, NULL, 0);
+
+ grpc_exec_ctx_enqueue(exec_ctx, fc->closure, false, NULL);
+ } else {
+ sched_connect(exec_ctx, fc->closure, fc->ep, fc->deadline);
+ }
+ gpr_free(fc);
+}
+
+static void sched_connect(grpc_exec_ctx *exec_ctx, grpc_closure *closure,
+ grpc_endpoint **ep, gpr_timespec deadline) {
+ if (gpr_time_cmp(deadline, gpr_now(deadline.clock_type)) < 0) {
+ *ep = NULL;
+ grpc_exec_ctx_enqueue(exec_ctx, closure, false, NULL);
+ return;
+ }
+
+ future_connect *fc = gpr_malloc(sizeof(*fc));
+ fc->closure = closure;
+ fc->ep = ep;
+ fc->deadline = deadline;
+ grpc_timer_init(exec_ctx, &fc->timer,
+ gpr_time_add(gpr_now(GPR_CLOCK_MONOTONIC),
+ gpr_time_from_millis(1, GPR_TIMESPAN)),
+ do_connect, fc, gpr_now(GPR_CLOCK_MONOTONIC));
+}
+
+static void my_tcp_client_connect(grpc_exec_ctx *exec_ctx,
+ grpc_closure *closure, grpc_endpoint **ep,
+ grpc_pollset_set *interested_parties,
+ const struct sockaddr *addr, size_t addr_len,
+ gpr_timespec deadline) {
+ sched_connect(exec_ctx, closure, ep, deadline);
+}
+
+////////////////////////////////////////////////////////////////////////////////
+// test driver
+
+typedef struct validator {
+ void (*validate)(void *arg, bool success);
+ void *arg;
+} validator;
+
+static validator *create_validator(void (*validate)(void *arg, bool success),
+ void *arg) {
+ validator *v = gpr_malloc(sizeof(*v));
+ v->validate = validate;
+ v->arg = arg;
+ return v;
+}
+
+static void assert_success_and_decrement(void *counter, bool success) {
+ GPR_ASSERT(success);
+ --*(int *)counter;
+}
+
+static void decrement(void *counter, bool success) { --*(int *)counter; }
+
+typedef struct connectivity_watch {
+ int *counter;
+ gpr_timespec deadline;
+} connectivity_watch;
+
+static connectivity_watch *make_connectivity_watch(gpr_timespec s,
+ int *counter) {
+ connectivity_watch *o = gpr_malloc(sizeof(*o));
+ o->deadline = s;
+ o->counter = counter;
+ return o;
+}
+
+static void validate_connectivity_watch(void *p, bool success) {
+ connectivity_watch *w = p;
+ if (!success) {
+ GPR_ASSERT(gpr_time_cmp(gpr_now(w->deadline.clock_type), w->deadline) >= 0);
+ }
+ --*w->counter;
+ gpr_free(w);
+}
+
+static void free_non_null(void *p) {
+ GPR_ASSERT(p != NULL);
+ gpr_free(p);
+}
+
+typedef enum { ROOT, CLIENT, SERVER, PENDING_SERVER } call_state_type;
+
+typedef struct call_state {
+ call_state_type type;
+ grpc_call *call;
+ grpc_byte_buffer *recv_message;
+ grpc_status_code status;
+ grpc_metadata_array recv_initial_metadata;
+ grpc_metadata_array recv_trailing_metadata;
+ char *recv_status_details;
+ size_t recv_status_details_capacity;
+ int cancelled;
+ int pending_ops;
+ grpc_call_details call_details;
+
+ // array of pointers to free later
+ size_t num_to_free;
+ size_t cap_to_free;
+ void **to_free;
+
+ struct call_state *next;
+ struct call_state *prev;
+} call_state;
+
+static call_state *g_active_call;
+
+static call_state *new_call(call_state *sibling, call_state_type type) {
+ call_state *c = gpr_malloc(sizeof(*c));
+ memset(c, 0, sizeof(*c));
+ if (sibling != NULL) {
+ c->next = sibling;
+ c->prev = sibling->prev;
+ c->next->prev = c->prev->next = c;
+ } else {
+ c->next = c->prev = c;
+ }
+ c->type = type;
+ return c;
+}
+
+static call_state *maybe_delete_call_state(call_state *call) {
+ call_state *next = call->next;
+
+ if (call->call != NULL) return next;
+ if (call->pending_ops != 0) return next;
+
+ if (call == g_active_call) {
+ g_active_call = call->next;
+ GPR_ASSERT(call != g_active_call);
+ }
+
+ call->prev->next = call->next;
+ call->next->prev = call->prev;
+ grpc_metadata_array_destroy(&call->recv_initial_metadata);
+ grpc_metadata_array_destroy(&call->recv_trailing_metadata);
+ gpr_free(call->recv_status_details);
+ grpc_call_details_destroy(&call->call_details);
+
+ for (size_t i = 0; i < call->num_to_free; i++) {
+ gpr_free(call->to_free[i]);
+ }
+ gpr_free(call->to_free);
+
+ gpr_free(call);
+
+ return next;
+}
+
+static void add_to_free(call_state *call, void *p) {
+ if (call->num_to_free == call->cap_to_free) {
+ call->cap_to_free = GPR_MAX(8, 2 * call->cap_to_free);
+ call->to_free =
+ gpr_realloc(call->to_free, sizeof(*call->to_free) * call->cap_to_free);
+ }
+ call->to_free[call->num_to_free++] = p;
+}
+
+static void read_metadata(input_stream *inp, size_t *count,
+ grpc_metadata **metadata, call_state *cs) {
+ *count = next_byte(inp);
+ *metadata = gpr_malloc(*count * sizeof(**metadata));
+ memset(*metadata, 0, *count * sizeof(**metadata));
+ for (size_t i = 0; i < *count; i++) {
+ (*metadata)[i].key = read_string(inp);
+ read_buffer(inp, (char **)&(*metadata)[i].value,
+ &(*metadata)[i].value_length);
+ (*metadata)[i].flags = read_uint32(inp);
+ add_to_free(cs, (void *)(*metadata)[i].key);
+ add_to_free(cs, (void *)(*metadata)[i].value);
+ }
+ add_to_free(cs, *metadata);
+}
+
+static call_state *destroy_call(call_state *call) {
+ grpc_call_destroy(call->call);
+ call->call = NULL;
+ return maybe_delete_call_state(call);
+}
+
+static void finished_request_call(void *csp, bool success) {
+ call_state *cs = csp;
+ GPR_ASSERT(cs->pending_ops > 0);
+ --cs->pending_ops;
+ if (success) {
+ GPR_ASSERT(cs->call != NULL);
+ cs->type = SERVER;
+ } else {
+ maybe_delete_call_state(cs);
+ }
+}
+
+static void finished_batch(void *csp, bool success) {
+ call_state *cs = csp;
+ --cs->pending_ops;
+ maybe_delete_call_state(cs);
+}
+
+int LLVMFuzzerTestOneInput(const uint8_t *data, size_t size) {
+ grpc_test_only_set_metadata_hash_seed(0);
+ if (squelch) gpr_set_log_function(dont_log);
+ input_stream inp = {data, data + size};
+ grpc_resolve_address = my_resolve_address;
+ grpc_tcp_client_connect_impl = my_tcp_client_connect;
+ gpr_now_impl = now_impl;
+ grpc_init();
+
+ GPR_ASSERT(g_channel == NULL);
+ GPR_ASSERT(g_server == NULL);
+
+ bool server_shutdown = false;
+ int pending_server_shutdowns = 0;
+ int pending_channel_watches = 0;
+ int pending_pings = 0;
+
+ g_active_call = new_call(NULL, ROOT);
+
+ grpc_completion_queue *cq = grpc_completion_queue_create(NULL);
+
+ while (!is_eof(&inp) || g_channel != NULL || g_server != NULL ||
+ pending_channel_watches > 0 || pending_pings > 0 ||
+ g_active_call->type != ROOT || g_active_call->next != g_active_call) {
+ if (is_eof(&inp)) {
+ if (g_channel != NULL) {
+ grpc_channel_destroy(g_channel);
+ g_channel = NULL;
+ }
+ if (g_server != NULL) {
+ if (!server_shutdown) {
+ grpc_server_shutdown_and_notify(
+ g_server, cq, create_validator(assert_success_and_decrement,
+ &pending_server_shutdowns));
+ server_shutdown = true;
+ pending_server_shutdowns++;
+ } else if (pending_server_shutdowns == 0) {
+ grpc_server_destroy(g_server);
+ g_server = NULL;
+ }
+ }
+ call_state *s = g_active_call;
+ do {
+ if (s->type != PENDING_SERVER && s->call != NULL) {
+ s = destroy_call(s);
+ } else {
+ s = s->next;
+ }
+ } while (s != g_active_call);
+
+ g_now = gpr_time_add(g_now, gpr_time_from_seconds(1, GPR_TIMESPAN));
+ }
+
+ switch (next_byte(&inp)) {
+ // terminate on bad bytes
+ default:
+ end(&inp);
+ break;
+ // tickle completion queue
+ case 0: {
+ grpc_event ev = grpc_completion_queue_next(
+ cq, gpr_inf_past(GPR_CLOCK_REALTIME), NULL);
+ switch (ev.type) {
+ case GRPC_OP_COMPLETE: {
+ validator *v = ev.tag;
+ v->validate(v->arg, ev.success);
+ gpr_free(v);
+ break;
+ }
+ case GRPC_QUEUE_TIMEOUT:
+ break;
+ case GRPC_QUEUE_SHUTDOWN:
+ abort();
+ break;
+ }
+ break;
+ }
+ // increment global time
+ case 1: {
+ g_now = gpr_time_add(
+ g_now, gpr_time_from_micros(read_uint32(&inp), GPR_TIMESPAN));
+ break;
+ }
+ // create an insecure channel
+ case 2: {
+ if (g_channel == NULL) {
+ char *target = read_string(&inp);
+ char *target_uri;
+ gpr_asprintf(&target_uri, "dns:%s", target);
+ grpc_channel_args *args = read_args(&inp);
+ g_channel = grpc_insecure_channel_create(target_uri, args, NULL);
+ GPR_ASSERT(g_channel != NULL);
+ grpc_channel_args_destroy(args);
+ gpr_free(target_uri);
+ gpr_free(target);
+ } else {
+ end(&inp);
+ }
+ break;
+ }
+ // destroy a channel
+ case 3: {
+ if (g_channel != NULL) {
+ grpc_channel_destroy(g_channel);
+ g_channel = NULL;
+ } else {
+ end(&inp);
+ }
+ break;
+ }
+ // bring up a server
+ case 4: {
+ if (g_server == NULL) {
+ grpc_channel_args *args = read_args(&inp);
+ g_server = grpc_server_create(args, NULL);
+ GPR_ASSERT(g_server != NULL);
+ grpc_channel_args_destroy(args);
+ grpc_server_register_completion_queue(g_server, cq, NULL);
+ grpc_server_start(g_server);
+ server_shutdown = false;
+ GPR_ASSERT(pending_server_shutdowns == 0);
+ } else {
+ end(&inp);
+ }
+ }
+ // begin server shutdown
+ case 5: {
+ if (g_server != NULL) {
+ grpc_server_shutdown_and_notify(
+ g_server, cq, create_validator(assert_success_and_decrement,
+ &pending_server_shutdowns));
+ pending_server_shutdowns++;
+ server_shutdown = true;
+ } else {
+ end(&inp);
+ }
+ break;
+ }
+ // cancel all calls if shutdown
+ case 6: {
+ if (g_server != NULL && server_shutdown) {
+ grpc_server_cancel_all_calls(g_server);
+ } else {
+ end(&inp);
+ }
+ break;
+ }
+ // destroy server
+ case 7: {
+ if (g_server != NULL && server_shutdown &&
+ pending_server_shutdowns == 0) {
+ grpc_server_destroy(g_server);
+ g_server = NULL;
+ } else {
+ end(&inp);
+ }
+ break;
+ }
+ // check connectivity
+ case 8: {
+ if (g_channel != NULL) {
+ uint8_t try_to_connect = next_byte(&inp);
+ if (try_to_connect == 0 || try_to_connect == 1) {
+ grpc_channel_check_connectivity_state(g_channel, try_to_connect);
+ } else {
+ end(&inp);
+ }
+ } else {
+ end(&inp);
+ }
+ break;
+ }
+ // watch connectivity
+ case 9: {
+ if (g_channel != NULL) {
+ grpc_connectivity_state st =
+ grpc_channel_check_connectivity_state(g_channel, 0);
+ if (st != GRPC_CHANNEL_FATAL_FAILURE) {
+ gpr_timespec deadline = gpr_time_add(
+ gpr_now(GPR_CLOCK_REALTIME),
+ gpr_time_from_micros(read_uint32(&inp), GPR_TIMESPAN));
+ grpc_channel_watch_connectivity_state(
+ g_channel, st, deadline, cq,
+ create_validator(validate_connectivity_watch,
+ make_connectivity_watch(
+ deadline, &pending_channel_watches)));
+ pending_channel_watches++;
+ }
+ } else {
+ end(&inp);
+ }
+ break;
+ }
+ // create a call
+ case 10: {
+ bool ok = true;
+ if (g_channel == NULL) ok = false;
+ grpc_call *parent_call = NULL;
+ if (g_active_call->type != ROOT) {
+ if (g_active_call->call == NULL || g_active_call->type == CLIENT) {
+ end(&inp);
+ break;
+ }
+ parent_call = g_active_call->call;
+ }
+ uint32_t propagation_mask = read_uint32(&inp);
+ char *method = read_string(&inp);
+ char *host = read_string(&inp);
+ gpr_timespec deadline =
+ gpr_time_add(gpr_now(GPR_CLOCK_REALTIME),
+ gpr_time_from_micros(read_uint32(&inp), GPR_TIMESPAN));
+
+ if (ok) {
+ call_state *cs = new_call(g_active_call, CLIENT);
+ cs->call =
+ grpc_channel_create_call(g_channel, parent_call, propagation_mask,
+ cq, method, host, deadline, NULL);
+ } else {
+ end(&inp);
+ }
+ gpr_free(method);
+ gpr_free(host);
+ break;
+ }
+ // switch the 'current' call
+ case 11: {
+ g_active_call = g_active_call->next;
+ break;
+ }
+ // queue some ops on a call
+ case 12: {
+ if (g_active_call->type == PENDING_SERVER ||
+ g_active_call->type == ROOT || g_active_call->call == NULL) {
+ end(&inp);
+ break;
+ }
+ size_t num_ops = next_byte(&inp);
+ if (num_ops > 6) {
+ end(&inp);
+ break;
+ }
+ grpc_op *ops = gpr_malloc(sizeof(grpc_op) * num_ops);
+ bool ok = true;
+ size_t i;
+ grpc_op *op;
+ for (i = 0; i < num_ops; i++) {
+ op = &ops[i];
+ switch (next_byte(&inp)) {
+ default:
+ /* invalid value */
+ op->op = (grpc_op_type)-1;
+ ok = false;
+ break;
+ case GRPC_OP_SEND_INITIAL_METADATA:
+ op->op = GRPC_OP_SEND_INITIAL_METADATA;
+ read_metadata(&inp, &op->data.send_initial_metadata.count,
+ &op->data.send_initial_metadata.metadata,
+ g_active_call);
+ break;
+ case GRPC_OP_SEND_MESSAGE:
+ op->op = GRPC_OP_SEND_MESSAGE;
+ op->data.send_message = read_message(&inp);
+ break;
+ case GRPC_OP_SEND_CLOSE_FROM_CLIENT:
+ op->op = GRPC_OP_SEND_CLOSE_FROM_CLIENT;
+ break;
+ case GRPC_OP_SEND_STATUS_FROM_SERVER:
+ op->op = GRPC_OP_SEND_STATUS_FROM_SERVER;
+ read_metadata(
+ &inp,
+ &op->data.send_status_from_server.trailing_metadata_count,
+ &op->data.send_status_from_server.trailing_metadata,
+ g_active_call);
+ op->data.send_status_from_server.status = next_byte(&inp);
+ op->data.send_status_from_server.status_details =
+ read_string(&inp);
+ break;
+ case GRPC_OP_RECV_INITIAL_METADATA:
+ op->op = GRPC_OP_RECV_INITIAL_METADATA;
+ op->data.recv_initial_metadata =
+ &g_active_call->recv_initial_metadata;
+ break;
+ case GRPC_OP_RECV_MESSAGE:
+ op->op = GRPC_OP_RECV_MESSAGE;
+ op->data.recv_message = &g_active_call->recv_message;
+ break;
+ case GRPC_OP_RECV_STATUS_ON_CLIENT:
+ op->op = GRPC_OP_RECV_STATUS_ON_CLIENT;
+ op->data.recv_status_on_client.status = &g_active_call->status;
+ op->data.recv_status_on_client.trailing_metadata =
+ &g_active_call->recv_trailing_metadata;
+ op->data.recv_status_on_client.status_details =
+ &g_active_call->recv_status_details;
+ op->data.recv_status_on_client.status_details_capacity =
+ &g_active_call->recv_status_details_capacity;
+ break;
+ case GRPC_OP_RECV_CLOSE_ON_SERVER:
+ op->op = GRPC_OP_RECV_CLOSE_ON_SERVER;
+ op->data.recv_close_on_server.cancelled =
+ &g_active_call->cancelled;
+ break;
+ }
+ op->reserved = NULL;
+ op->flags = read_uint32(&inp);
+ }
+ if (ok) {
+ validator *v = create_validator(finished_batch, g_active_call);
+ g_active_call->pending_ops++;
+ grpc_call_error error =
+ grpc_call_start_batch(g_active_call->call, ops, num_ops, v, NULL);
+ if (error != GRPC_CALL_OK) {
+ v->validate(v->arg, false);
+ gpr_free(v);
+ }
+ } else {
+ end(&inp);
+ }
+ for (i = 0; i < num_ops; i++) {
+ op = &ops[i];
+ switch (op->op) {
+ case GRPC_OP_SEND_INITIAL_METADATA:
+ break;
+ case GRPC_OP_SEND_MESSAGE:
+ grpc_byte_buffer_destroy(op->data.send_message);
+ break;
+ case GRPC_OP_SEND_STATUS_FROM_SERVER:
+ gpr_free((void *)op->data.send_status_from_server.status_details);
+ break;
+ case GRPC_OP_SEND_CLOSE_FROM_CLIENT:
+ case GRPC_OP_RECV_INITIAL_METADATA:
+ case GRPC_OP_RECV_MESSAGE:
+ case GRPC_OP_RECV_STATUS_ON_CLIENT:
+ case GRPC_OP_RECV_CLOSE_ON_SERVER:
+ break;
+ }
+ }
+ gpr_free(ops);
+
+ break;
+ }
+ // cancel current call
+ case 13: {
+ if (g_active_call->type != ROOT && g_active_call->call != NULL) {
+ grpc_call_cancel(g_active_call->call, NULL);
+ } else {
+ end(&inp);
+ }
+ break;
+ }
+ // get a calls peer
+ case 14: {
+ if (g_active_call->type != ROOT && g_active_call->call != NULL) {
+ free_non_null(grpc_call_get_peer(g_active_call->call));
+ } else {
+ end(&inp);
+ }
+ break;
+ }
+ // get a channels target
+ case 15: {
+ if (g_channel != NULL) {
+ free_non_null(grpc_channel_get_target(g_channel));
+ } else {
+ end(&inp);
+ }
+ break;
+ }
+ // send a ping on a channel
+ case 16: {
+ if (g_channel != NULL) {
+ pending_pings++;
+ grpc_channel_ping(g_channel, cq,
+ create_validator(decrement, &pending_pings), NULL);
+ } else {
+ end(&inp);
+ }
+ break;
+ }
+ // enable a tracer
+ case 17: {
+ char *tracer = read_string(&inp);
+ grpc_tracer_set_enabled(tracer, 1);
+ gpr_free(tracer);
+ break;
+ }
+ // disable a tracer
+ case 18: {
+ char *tracer = read_string(&inp);
+ grpc_tracer_set_enabled(tracer, 0);
+ gpr_free(tracer);
+ break;
+ }
+ // request a server call
+ case 19: {
+ if (g_server == NULL) {
+ end(&inp);
+ break;
+ }
+ call_state *cs = new_call(g_active_call, PENDING_SERVER);
+ cs->pending_ops++;
+ validator *v = create_validator(finished_request_call, cs);
+ grpc_call_error error =
+ grpc_server_request_call(g_server, &cs->call, &cs->call_details,
+ &cs->recv_initial_metadata, cq, cq, v);
+ if (error != GRPC_CALL_OK) {
+ v->validate(v->arg, false);
+ gpr_free(v);
+ }
+ break;
+ }
+ // destroy a call
+ case 20: {
+ if (g_active_call->type != ROOT &&
+ g_active_call->type != PENDING_SERVER &&
+ g_active_call->call != NULL) {
+ destroy_call(g_active_call);
+ } else {
+ end(&inp);
+ }
+ break;
+ }
+ }
+ }
+
+ GPR_ASSERT(g_channel == NULL);
+ GPR_ASSERT(g_server == NULL);
+ GPR_ASSERT(g_active_call->type == ROOT);
+ GPR_ASSERT(g_active_call->next == g_active_call);
+ gpr_free(g_active_call);
+
+ grpc_completion_queue_shutdown(cq);
+ GPR_ASSERT(
+ grpc_completion_queue_next(cq, gpr_inf_past(GPR_CLOCK_REALTIME), NULL)
+ .type == GRPC_QUEUE_SHUTDOWN);
+ grpc_completion_queue_destroy(cq);
+
+ grpc_shutdown();
+ return 0;
+}