diff options
Diffstat (limited to 'src/core/statistics/census_log.c')
-rw-r--r-- | src/core/statistics/census_log.c | 29 |
1 files changed, 14 insertions, 15 deletions
diff --git a/src/core/statistics/census_log.c b/src/core/statistics/census_log.c index 88e338038d..257ba586a3 100644 --- a/src/core/statistics/census_log.c +++ b/src/core/statistics/census_log.c @@ -116,7 +116,7 @@ typedef struct census_log_block { simultaneously by reader and writer. */ gpr_atm bytes_committed; /* Bytes already read */ - gpr_int32 bytes_read; + int32_t bytes_read; /* Links for list */ cl_block_list_struct link; /* We want this structure to be cacheline aligned. We assume the following @@ -124,7 +124,7 @@ typedef struct census_log_block { type 32b size 64b size char* 4 8 3x gpr_atm 12 24 - gpr_int32 4 8 (assumes padding) + int32_t 4 8 (assumes padding) cl_block_list_struct 12 24 TOTAL 32 64 @@ -147,7 +147,7 @@ typedef struct census_log_block { /* A list of cl_blocks, doubly-linked through cl_block::link. */ typedef struct census_log_block_list { - gpr_int32 count; /* Number of items in list. */ + int32_t count; /* Number of items in list. */ cl_block_list_struct ht; /* head/tail of linked list. */ } cl_block_list; @@ -175,7 +175,7 @@ struct census_log { /* Number of cores (aka hardware-contexts) */ unsigned num_cores; /* number of CENSUS_LOG_2_MAX_RECORD_SIZE blocks in log */ - gpr_int32 num_blocks; + int32_t num_blocks; cl_block *blocks; /* Block metadata. */ cl_core_local_block *core_local_blocks; /* Keeps core to block mappings. */ gpr_mu lock; @@ -183,7 +183,7 @@ struct census_log { /* Keeps the state of the reader iterator. A value of 0 indicates that iterator has reached the end. census_log_init_reader() resets the value to num_core to restart iteration. */ - gpr_uint32 read_iterator_state; + uint32_t read_iterator_state; /* Points to the block being read. If non-NULL, the block is locked for reading (block_being_read_->reader_lock is held). */ cl_block *block_being_read; @@ -276,11 +276,11 @@ static void cl_block_initialize(cl_block *block, char *buffer) { /* Guards against exposing partially written buffer to the reader. */ static void cl_block_set_bytes_committed(cl_block *block, - gpr_int32 bytes_committed) { + int32_t bytes_committed) { gpr_atm_rel_store(&block->bytes_committed, bytes_committed); } -static gpr_int32 cl_block_get_bytes_committed(cl_block *block) { +static int32_t cl_block_get_bytes_committed(cl_block *block) { return gpr_atm_acq_load(&block->bytes_committed); } @@ -317,7 +317,7 @@ static void cl_block_enable_access(cl_block *block) { /* Returns with writer_lock held. */ static void *cl_block_start_write(cl_block *block, size_t size) { - gpr_int32 bytes_committed; + int32_t bytes_committed; if (!cl_try_lock(&block->writer_lock)) { return NULL; } @@ -395,8 +395,7 @@ static cl_block *cl_allocate_block(void) { - allocated a new block OR - 'core_id' => 'old_block' mapping changed (another thread allocated a block before lock was acquired). */ -static int cl_allocate_core_local_block(gpr_int32 core_id, - cl_block *old_block) { +static int cl_allocate_core_local_block(int32_t core_id, cl_block *old_block) { /* Now that we have the lock, check if core-local mapping has changed. */ cl_core_local_block *core_local_block = &g_log.core_local_blocks[core_id]; cl_block *block = cl_core_local_block_get_block(core_local_block); @@ -418,8 +417,8 @@ static int cl_allocate_core_local_block(gpr_int32 core_id, } static cl_block *cl_get_block(void *record) { - gpr_uintptr p = (gpr_uintptr)((char *)record - g_log.buffer); - gpr_uintptr index = p >> CENSUS_LOG_2_MAX_RECORD_SIZE; + uintptr_t p = (uintptr_t)((char *)record - g_log.buffer); + uintptr_t index = p >> CENSUS_LOG_2_MAX_RECORD_SIZE; return &g_log.blocks[index]; } @@ -460,7 +459,7 @@ static cl_block *cl_next_block_to_read(cl_block *prev) { /* External functions: primary stats_log interface */ void census_log_initialize(size_t size_in_mb, int discard_old_records) { - gpr_int32 ix; + int32_t ix; /* Check cacheline alignment. */ GPR_ASSERT(sizeof(cl_block) % GPR_CACHELINE_SIZE == 0); GPR_ASSERT(sizeof(cl_core_local_block) % GPR_CACHELINE_SIZE == 0); @@ -510,9 +509,9 @@ void census_log_shutdown(void) { void *census_log_start_write(size_t size) { /* Used to bound number of times block allocation is attempted. */ - gpr_int32 attempts_remaining = g_log.num_blocks; + int32_t attempts_remaining = g_log.num_blocks; /* TODO(aveitch): move this inside the do loop when current_cpu is fixed */ - gpr_int32 core_id = gpr_cpu_current_cpu(); + int32_t core_id = gpr_cpu_current_cpu(); GPR_ASSERT(g_log.initialized); if (size > CENSUS_LOG_MAX_RECORD_SIZE) { return NULL; |