From b7ebd3b8c6fe39f99c40b10c1b563e4adb607b6c Mon Sep 17 00:00:00 2001 From: Nicolas Noble Date: Wed, 26 Nov 2014 16:33:03 -0800 Subject: Initial import. --- test/core/surface/completion_queue_test.c | 435 ++++++++++++++++++++++++++++++ 1 file changed, 435 insertions(+) create mode 100644 test/core/surface/completion_queue_test.c (limited to 'test/core/surface/completion_queue_test.c') diff --git a/test/core/surface/completion_queue_test.c b/test/core/surface/completion_queue_test.c new file mode 100644 index 0000000000..6df159f697 --- /dev/null +++ b/test/core/surface/completion_queue_test.c @@ -0,0 +1,435 @@ +/* + * + * Copyright 2014, Google Inc. + * All rights reserved. + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions are + * met: + * + * * Redistributions of source code must retain the above copyright + * notice, this list of conditions and the following disclaimer. + * * Redistributions in binary form must reproduce the above + * copyright notice, this list of conditions and the following disclaimer + * in the documentation and/or other materials provided with the + * distribution. + * * Neither the name of Google Inc. nor the names of its + * contributors may be used to endorse or promote products derived from + * this software without specific prior written permission. + * + * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS + * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT + * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR + * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT + * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, + * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT + * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, + * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY + * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT + * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE + * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. + * + */ + +#include "src/core/surface/completion_queue.h" + +#include +#include +#include +#include +#include +#include "src/core/surface/surface_em.h" +#include "test/core/util/test_config.h" + +#define LOG_TEST() gpr_log(GPR_INFO, "%s", __FUNCTION__) + +static void increment_int_on_finish(void *user_data, grpc_op_error error) { + ++*(int *)user_data; +} + +static void *create_test_tag() { + static gpr_intptr i = 0; + return (void *)(++i); +} + +/* helper for tests to shutdown correctly and tersely */ +static void shutdown_and_destroy(grpc_completion_queue *cc) { + grpc_event *ev; + grpc_completion_queue_shutdown(cc); + ev = grpc_completion_queue_next(cc, gpr_inf_past); + GPR_ASSERT(ev != NULL); + GPR_ASSERT(ev->type == GRPC_QUEUE_SHUTDOWN); + grpc_event_finish(ev); + grpc_completion_queue_destroy(cc); +} + +/* ensure we can create and destroy a completion channel */ +static void test_no_op() { + LOG_TEST(); + shutdown_and_destroy(grpc_completion_queue_create()); +} + +static void test_wait_empty() { + grpc_completion_queue *cc; + + LOG_TEST(); + + cc = grpc_completion_queue_create(); + GPR_ASSERT(grpc_completion_queue_next(cc, gpr_now()) == NULL); + shutdown_and_destroy(cc); +} + +static void test_cq_end_read() { + grpc_event *ev; + grpc_completion_queue *cc; + int on_finish_called = 0; + void *tag = create_test_tag(); + + LOG_TEST(); + + cc = grpc_completion_queue_create(); + + grpc_cq_begin_op(cc, NULL, GRPC_READ); + grpc_cq_end_read(cc, tag, NULL, increment_int_on_finish, &on_finish_called, + NULL); + + ev = grpc_completion_queue_next(cc, gpr_inf_past); + GPR_ASSERT(ev != NULL); + GPR_ASSERT(ev->type == GRPC_READ); + GPR_ASSERT(ev->tag == tag); + GPR_ASSERT(ev->data.read == NULL); + GPR_ASSERT(on_finish_called == 0); + grpc_event_finish(ev); + GPR_ASSERT(on_finish_called == 1); + + shutdown_and_destroy(cc); +} + +static void test_cq_end_invoke_accepted() { + grpc_event *ev; + grpc_completion_queue *cc; + int on_finish_called = 0; + void *tag = create_test_tag(); + + LOG_TEST(); + + cc = grpc_completion_queue_create(); + + grpc_cq_begin_op(cc, NULL, GRPC_INVOKE_ACCEPTED); + grpc_cq_end_invoke_accepted(cc, tag, NULL, increment_int_on_finish, + &on_finish_called, GRPC_OP_OK); + + ev = grpc_completion_queue_next(cc, gpr_inf_past); + GPR_ASSERT(ev != NULL); + GPR_ASSERT(ev->type == GRPC_INVOKE_ACCEPTED); + GPR_ASSERT(ev->tag == tag); + GPR_ASSERT(ev->data.invoke_accepted == GRPC_OP_OK); + GPR_ASSERT(on_finish_called == 0); + grpc_event_finish(ev); + GPR_ASSERT(on_finish_called == 1); + + shutdown_and_destroy(cc); +} + +static void test_cq_end_write_accepted() { + grpc_event *ev; + grpc_completion_queue *cc; + int on_finish_called = 0; + void *tag = create_test_tag(); + + LOG_TEST(); + + cc = grpc_completion_queue_create(); + + grpc_cq_begin_op(cc, NULL, GRPC_WRITE_ACCEPTED); + grpc_cq_end_write_accepted(cc, tag, NULL, increment_int_on_finish, + &on_finish_called, GRPC_OP_OK); + + ev = grpc_completion_queue_next(cc, gpr_inf_past); + GPR_ASSERT(ev != NULL); + GPR_ASSERT(ev->type == GRPC_WRITE_ACCEPTED); + GPR_ASSERT(ev->tag == tag); + GPR_ASSERT(ev->data.write_accepted == GRPC_OP_OK); + GPR_ASSERT(on_finish_called == 0); + grpc_event_finish(ev); + GPR_ASSERT(on_finish_called == 1); + + shutdown_and_destroy(cc); +} + +static void test_cq_end_finish_accepted() { + grpc_event *ev; + grpc_completion_queue *cc; + int on_finish_called = 0; + void *tag = create_test_tag(); + + LOG_TEST(); + + cc = grpc_completion_queue_create(); + + grpc_cq_begin_op(cc, NULL, GRPC_FINISH_ACCEPTED); + grpc_cq_end_finish_accepted(cc, tag, NULL, increment_int_on_finish, + &on_finish_called, GRPC_OP_OK); + + ev = grpc_completion_queue_next(cc, gpr_inf_past); + GPR_ASSERT(ev != NULL); + GPR_ASSERT(ev->type == GRPC_FINISH_ACCEPTED); + GPR_ASSERT(ev->tag == tag); + GPR_ASSERT(ev->data.finish_accepted == GRPC_OP_OK); + GPR_ASSERT(on_finish_called == 0); + grpc_event_finish(ev); + GPR_ASSERT(on_finish_called == 1); + + shutdown_and_destroy(cc); +} + +static void test_cq_end_client_metadata_read() { + grpc_event *ev; + grpc_completion_queue *cc; + int on_finish_called = 0; + void *tag = create_test_tag(); + + LOG_TEST(); + + cc = grpc_completion_queue_create(); + + grpc_cq_begin_op(cc, NULL, GRPC_CLIENT_METADATA_READ); + grpc_cq_end_client_metadata_read(cc, tag, NULL, increment_int_on_finish, + &on_finish_called, 0, NULL); + + ev = grpc_completion_queue_next(cc, gpr_inf_past); + GPR_ASSERT(ev != NULL); + GPR_ASSERT(ev->type == GRPC_CLIENT_METADATA_READ); + GPR_ASSERT(ev->tag == tag); + GPR_ASSERT(ev->data.client_metadata_read.count == 0); + GPR_ASSERT(ev->data.client_metadata_read.elements == NULL); + GPR_ASSERT(on_finish_called == 0); + grpc_event_finish(ev); + GPR_ASSERT(on_finish_called == 1); + + shutdown_and_destroy(cc); +} + +static void test_pluck() { + grpc_event *ev; + grpc_completion_queue *cc; + void *tags[128]; + int i, j; + int on_finish_called = 0; + + LOG_TEST(); + + for (i = 0; i < GPR_ARRAY_SIZE(tags); i++) { + tags[i] = create_test_tag(); + for (j = 0; j < i; j++) { + GPR_ASSERT(tags[i] != tags[j]); + } + } + + cc = grpc_completion_queue_create(); + + for (i = 0; i < GPR_ARRAY_SIZE(tags); i++) { + grpc_cq_begin_op(cc, NULL, GRPC_WRITE_ACCEPTED); + grpc_cq_end_write_accepted(cc, tags[i], NULL, increment_int_on_finish, + &on_finish_called, GRPC_OP_OK); + } + + for (i = 0; i < GPR_ARRAY_SIZE(tags); i++) { + ev = grpc_completion_queue_pluck(cc, tags[i], gpr_inf_past); + GPR_ASSERT(ev->tag == tags[i]); + grpc_event_finish(ev); + } + + GPR_ASSERT(on_finish_called == GPR_ARRAY_SIZE(tags)); + + for (i = 0; i < GPR_ARRAY_SIZE(tags); i++) { + grpc_cq_begin_op(cc, NULL, GRPC_WRITE_ACCEPTED); + grpc_cq_end_write_accepted(cc, tags[i], NULL, increment_int_on_finish, + &on_finish_called, GRPC_OP_OK); + } + + for (i = 0; i < GPR_ARRAY_SIZE(tags); i++) { + ev = grpc_completion_queue_pluck(cc, tags[GPR_ARRAY_SIZE(tags) - i - 1], + gpr_inf_past); + GPR_ASSERT(ev->tag == tags[GPR_ARRAY_SIZE(tags) - i - 1]); + grpc_event_finish(ev); + } + + GPR_ASSERT(on_finish_called == 2 * GPR_ARRAY_SIZE(tags)); + + shutdown_and_destroy(cc); +} + +#define TEST_THREAD_EVENTS 10000 + +typedef struct test_thread_options { + gpr_event on_started; + gpr_event *phase1; + gpr_event on_phase1_done; + gpr_event *phase2; + gpr_event on_finished; + int events_triggered; + int id; + grpc_completion_queue *cc; +} test_thread_options; + +gpr_timespec ten_seconds_time() { + return gpr_time_add(gpr_now(), gpr_time_from_micros(10 * 1000000)); +} + +static void producer_thread(void *arg) { + test_thread_options *opt = arg; + int i; + + gpr_log(GPR_INFO, "producer %d started", opt->id); + gpr_event_set(&opt->on_started, (void *)(gpr_intptr) 1); + GPR_ASSERT(gpr_event_wait(opt->phase1, ten_seconds_time())); + + gpr_log(GPR_INFO, "producer %d phase 1", opt->id); + for (i = 0; i < TEST_THREAD_EVENTS; i++) { + grpc_cq_begin_op(opt->cc, NULL, GRPC_WRITE_ACCEPTED); + } + + gpr_log(GPR_INFO, "producer %d phase 1 done", opt->id); + gpr_event_set(&opt->on_phase1_done, (void *)(gpr_intptr) 1); + GPR_ASSERT(gpr_event_wait(opt->phase2, ten_seconds_time())); + + gpr_log(GPR_INFO, "producer %d phase 2", opt->id); + for (i = 0; i < TEST_THREAD_EVENTS; i++) { + grpc_cq_end_write_accepted(opt->cc, (void *)(gpr_intptr) 1, NULL, NULL, + NULL, GRPC_OP_OK); + opt->events_triggered++; + } + + gpr_log(GPR_INFO, "producer %d phase 2 done", opt->id); + gpr_event_set(&opt->on_finished, (void *)(gpr_intptr) 1); +} + +static void consumer_thread(void *arg) { + test_thread_options *opt = arg; + grpc_event *ev; + + gpr_log(GPR_INFO, "consumer %d started", opt->id); + gpr_event_set(&opt->on_started, (void *)(gpr_intptr) 1); + GPR_ASSERT(gpr_event_wait(opt->phase1, ten_seconds_time())); + + gpr_log(GPR_INFO, "consumer %d phase 1", opt->id); + + gpr_log(GPR_INFO, "consumer %d phase 1 done", opt->id); + gpr_event_set(&opt->on_phase1_done, (void *)(gpr_intptr) 1); + GPR_ASSERT(gpr_event_wait(opt->phase2, ten_seconds_time())); + + gpr_log(GPR_INFO, "consumer %d phase 2", opt->id); + for (;;) { + ev = grpc_completion_queue_next(opt->cc, ten_seconds_time()); + GPR_ASSERT(ev); + switch (ev->type) { + case GRPC_WRITE_ACCEPTED: + GPR_ASSERT(ev->data.write_accepted == GRPC_OP_OK); + opt->events_triggered++; + grpc_event_finish(ev); + break; + case GRPC_QUEUE_SHUTDOWN: + gpr_log(GPR_INFO, "consumer %d phase 2 done", opt->id); + gpr_event_set(&opt->on_finished, (void *)(gpr_intptr) 1); + grpc_event_finish(ev); + return; + default: + gpr_log(GPR_ERROR, "Invalid event received: %d", ev->type); + abort(); + } + } +} + +static void test_threading(int producers, int consumers) { + test_thread_options *options = + gpr_malloc((producers + consumers) * sizeof(test_thread_options)); + gpr_event phase1 = GPR_EVENT_INIT; + gpr_event phase2 = GPR_EVENT_INIT; + grpc_completion_queue *cc = grpc_completion_queue_create(); + int i; + int total_consumed = 0; + static int optid = 101; + + gpr_log(GPR_INFO, "%s: %d producers, %d consumers", __FUNCTION__, producers, + consumers); + + grpc_completion_queue_dont_poll_test_only(cc); + + /* start all threads: they will wait for phase1 */ + for (i = 0; i < producers + consumers; i++) { + gpr_thd_id id; + gpr_event_init(&options[i].on_started); + gpr_event_init(&options[i].on_phase1_done); + gpr_event_init(&options[i].on_finished); + options[i].phase1 = &phase1; + options[i].phase2 = &phase2; + options[i].events_triggered = 0; + options[i].cc = cc; + options[i].id = optid++; + GPR_ASSERT(gpr_thd_new(&id, + i < producers ? producer_thread : consumer_thread, + options + i, NULL)); + gpr_event_wait(&options[i].on_started, ten_seconds_time()); + } + + /* start phase1: producers will pre-declare all operations they will + complete */ + gpr_log(GPR_INFO, "start phase 1"); + gpr_event_set(&phase1, (void *)(gpr_intptr) 1); + + gpr_log(GPR_INFO, "wait phase 1"); + for (i = 0; i < producers + consumers; i++) { + GPR_ASSERT(gpr_event_wait(&options[i].on_phase1_done, ten_seconds_time())); + } + gpr_log(GPR_INFO, "done phase 1"); + + /* start phase2: operations will complete, and consumers will consume them */ + gpr_log(GPR_INFO, "start phase 2"); + gpr_event_set(&phase2, (void *)(gpr_intptr) 1); + + /* in parallel, we shutdown the completion channel - all events should still + be consumed */ + grpc_completion_queue_shutdown(cc); + + /* join all threads */ + gpr_log(GPR_INFO, "wait phase 2"); + for (i = 0; i < producers + consumers; i++) { + GPR_ASSERT(gpr_event_wait(&options[i].on_finished, ten_seconds_time())); + } + gpr_log(GPR_INFO, "done phase 2"); + + /* destroy the completion channel */ + grpc_completion_queue_destroy(cc); + + /* verify that everything was produced and consumed */ + for (i = 0; i < producers + consumers; i++) { + if (i < producers) { + GPR_ASSERT(options[i].events_triggered == TEST_THREAD_EVENTS); + } else { + total_consumed += options[i].events_triggered; + } + } + GPR_ASSERT(total_consumed == producers * TEST_THREAD_EVENTS); + + gpr_free(options); +} + +int main(int argc, char **argv) { + grpc_test_init(argc, argv); + grpc_surface_em_init(); + test_no_op(); + test_wait_empty(); + test_cq_end_read(); + test_cq_end_invoke_accepted(); + test_cq_end_write_accepted(); + test_cq_end_finish_accepted(); + test_cq_end_client_metadata_read(); + test_pluck(); + test_threading(1, 1); + test_threading(1, 10); + test_threading(10, 1); + test_threading(10, 10); + grpc_surface_em_shutdown(); + return 0; +} -- cgit v1.2.3