// Protocol Buffers - Google's data interchange format // Copyright 2008 Google Inc. All rights reserved. // https://developers.google.com/protocol-buffers/ // // Redistribution and use in source and binary forms, with or without // modification, are permitted provided that the following conditions are // met: // // * Redistributions of source code must retain the above copyright // notice, this list of conditions and the following disclaimer. // * Redistributions in binary form must reproduce the above // copyright notice, this list of conditions and the following disclaimer // in the documentation and/or other materials provided with the // distribution. // * Neither the name of Google Inc. nor the names of its // contributors may be used to endorse or promote products derived from // this software without specific prior written permission. // // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS // "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT // LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR // A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT // OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, // SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT // LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, // DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. #include #include #include #ifdef ADDRESS_SANITIZER #include #endif // ADDRESS_SANITIZER namespace google { namespace protobuf { google::protobuf::internal::SequenceNumber Arena::lifecycle_id_generator_; #if defined(GOOGLE_PROTOBUF_NO_THREADLOCAL) Arena::ThreadCache& Arena::thread_cache() { static internal::ThreadLocalStorage* thread_cache_ = new internal::ThreadLocalStorage(); return *thread_cache_->Get(); } #elif defined(PROTOBUF_USE_DLLS) Arena::ThreadCache& Arena::thread_cache() { static GOOGLE_THREAD_LOCAL ThreadCache thread_cache_ = { -1, NULL }; return thread_cache_; } #else GOOGLE_THREAD_LOCAL Arena::ThreadCache Arena::thread_cache_ = { -1, NULL }; #endif void Arena::Init() { lifecycle_id_ = lifecycle_id_generator_.GetNext(); blocks_ = 0; hint_ = 0; space_allocated_ = 0; owns_first_block_ = true; cleanup_list_ = 0; if (options_.initial_block != NULL && options_.initial_block_size > 0) { GOOGLE_CHECK_GE(options_.initial_block_size, sizeof(Block)) << ": Initial block size too small for header."; // Add first unowned block to list. Block* first_block = reinterpret_cast(options_.initial_block); first_block->size = options_.initial_block_size; first_block->pos = kHeaderSize; first_block->next = NULL; // Thread which calls Init() owns the first block. This allows the // single-threaded case to allocate on the first block without taking any // locks. first_block->owner = &thread_cache(); SetThreadCacheBlock(first_block); AddBlockInternal(first_block); owns_first_block_ = false; } // Call the initialization hook if (options_.on_arena_init != NULL) { hooks_cookie_ = options_.on_arena_init(this); } else { hooks_cookie_ = NULL; } } Arena::~Arena() { uint64 space_allocated = ResetInternal(); // Call the destruction hook if (options_.on_arena_destruction != NULL) { options_.on_arena_destruction(this, hooks_cookie_, space_allocated); } } uint64 Arena::Reset() { // Invalidate any ThreadCaches pointing to any blocks we just destroyed. lifecycle_id_ = lifecycle_id_generator_.GetNext(); return ResetInternal(); } uint64 Arena::ResetInternal() { CleanupList(); uint64 space_allocated = FreeBlocks(); // Call the reset hook if (options_.on_arena_reset != NULL) { options_.on_arena_reset(this, hooks_cookie_, space_allocated); } return space_allocated; } Arena::Block* Arena::NewBlock(void* me, Block* my_last_block, size_t n, size_t start_block_size, size_t max_block_size) { size_t size; if (my_last_block != NULL) { // Double the current block size, up to a limit. size = 2 * (my_last_block->size); if (size > max_block_size) size = max_block_size; } else { size = start_block_size; } // Verify that n + kHeaderSize won't overflow. GOOGLE_CHECK_LE(n, std::numeric_limits::max() - kHeaderSize); size = std::max(size, kHeaderSize + n); Block* b = reinterpret_cast(options_.block_alloc(size)); b->pos = kHeaderSize + n; b->size = size; b->owner = me; #ifdef ADDRESS_SANITIZER // Poison the rest of the block for ASAN. It was unpoisoned by the underlying // malloc but it's not yet usable until we return it as part of an allocation. ASAN_POISON_MEMORY_REGION( reinterpret_cast(b) + b->pos, b->size - b->pos); #endif // ADDRESS_SANITIZER return b; } void Arena::AddBlock(Block* b) { MutexLock l(&blocks_lock_); AddBlockInternal(b); } void Arena::AddBlockInternal(Block* b) { b->next = reinterpret_cast(google::protobuf::internal::NoBarrier_Load(&blocks_)); google::protobuf::internal::Release_Store(&blocks_, reinterpret_cast(b)); if (b->avail() != 0) { // Direct future allocations to this block. google::protobuf::internal::Release_Store(&hint_, reinterpret_cast(b)); } space_allocated_ += b->size; } void Arena::AddListNode(void* elem, void (*cleanup)(void*)) { Node* node = reinterpret_cast(AllocateAligned(sizeof(Node))); node->elem = elem; node->cleanup = cleanup; node->next = reinterpret_cast( google::protobuf::internal::NoBarrier_AtomicExchange(&cleanup_list_, reinterpret_cast(node))); } void* Arena::AllocateAligned(const std::type_info* allocated, size_t n) { // Align n to next multiple of 8 (from Hacker's Delight, Chapter 3.) n = (n + 7) & -8; // Monitor allocation if needed. if (GOOGLE_PREDICT_FALSE(hooks_cookie_ != NULL) && options_.on_arena_allocation != NULL) { options_.on_arena_allocation(allocated, n, hooks_cookie_); } // If this thread already owns a block in this arena then try to use that. // This fast path optimizes the case where multiple threads allocate from the // same arena. if (thread_cache().last_lifecycle_id_seen == lifecycle_id_ && thread_cache().last_block_used_ != NULL) { if (thread_cache().last_block_used_->avail() < n) { return SlowAlloc(n); } return AllocFromBlock(thread_cache().last_block_used_, n); } // Check whether we own the last accessed block on this arena. // This fast path optimizes the case where a single thread uses multiple // arenas. void* me = &thread_cache(); Block* b = reinterpret_cast(google::protobuf::internal::Acquire_Load(&hint_)); if (!b || b->owner != me || b->avail() < n) { return SlowAlloc(n); } return AllocFromBlock(b, n); } void* Arena::AllocFromBlock(Block* b, size_t n) { size_t p = b->pos; b->pos = p + n; #ifdef ADDRESS_SANITIZER ASAN_UNPOISON_MEMORY_REGION(reinterpret_cast(b) + p, n); #endif // ADDRESS_SANITIZER return reinterpret_cast(b) + p; } void* Arena::SlowAlloc(size_t n) { void* me = &thread_cache(); Block* b = FindBlock(me); // Find block owned by me. // See if allocation fits in my latest block. if (b != NULL && b->avail() >= n) { SetThreadCacheBlock(b); google::protobuf::internal::NoBarrier_Store(&hint_, reinterpret_cast(b)); return AllocFromBlock(b, n); } b = NewBlock(me, b, n, options_.start_block_size, options_.max_block_size); AddBlock(b); SetThreadCacheBlock(b); return reinterpret_cast(b) + kHeaderSize; } uint64 Arena::SpaceAllocated() const { MutexLock l(&blocks_lock_); return space_allocated_; } uint64 Arena::SpaceUsed() const { uint64 space_used = 0; Block* b = reinterpret_cast(google::protobuf::internal::NoBarrier_Load(&blocks_)); while (b != NULL) { space_used += (b->pos - kHeaderSize); b = b->next; } return space_used; } std::pair Arena::SpaceAllocatedAndUsed() const { return std::make_pair(SpaceAllocated(), SpaceUsed()); } uint64 Arena::FreeBlocks() { uint64 space_allocated = 0; Block* b = reinterpret_cast(google::protobuf::internal::NoBarrier_Load(&blocks_)); Block* first_block = NULL; while (b != NULL) { space_allocated += (b->size); Block* next = b->next; if (next != NULL) { #ifdef ADDRESS_SANITIZER // This memory was provided by the underlying allocator as unpoisoned, so // return it in an unpoisoned state. ASAN_UNPOISON_MEMORY_REGION(reinterpret_cast(b), b->size); #endif // ADDRESS_SANITIZER options_.block_dealloc(b, b->size); } else { if (owns_first_block_) { #ifdef ADDRESS_SANITIZER // This memory was provided by the underlying allocator as unpoisoned, // so return it in an unpoisoned state. ASAN_UNPOISON_MEMORY_REGION(reinterpret_cast(b), b->size); #endif // ADDRESS_SANITIZER options_.block_dealloc(b, b->size); } else { // User passed in the first block, skip free'ing the memory. first_block = b; } } b = next; } blocks_ = 0; hint_ = 0; space_allocated_ = 0; if (!owns_first_block_) { // Make the first block that was passed in through ArenaOptions // available for reuse. first_block->pos = kHeaderSize; // Thread which calls Reset() owns the first block. This allows the // single-threaded case to allocate on the first block without taking any // locks. first_block->owner = &thread_cache(); SetThreadCacheBlock(first_block); AddBlockInternal(first_block); } return space_allocated; } void Arena::CleanupList() { Node* head = reinterpret_cast(google::protobuf::internal::NoBarrier_Load(&cleanup_list_)); while (head != NULL) { head->cleanup(head->elem); head = head->next; } cleanup_list_ = 0; } Arena::Block* Arena::FindBlock(void* me) { // TODO(sanjay): We might want to keep a separate list with one // entry per thread. Block* b = reinterpret_cast(google::protobuf::internal::Acquire_Load(&blocks_)); while (b != NULL && b->owner != me) { b = b->next; } return b; } } // namespace protobuf } // namespace google