/* * Copyright 2015 Google Inc. * * Use of this source code is governed by a BSD-style license that can be * found in the LICENSE file. */ #include "SkRWBuffer.h" #include "SkAtomics.h" #include "SkMakeUnique.h" #include "SkMalloc.h" #include "SkStream.h" #include "SkTo.h" #include #include // Force small chunks to be a page's worth static const size_t kMinAllocSize = 4096; struct SkBufferBlock { SkBufferBlock* fNext; // updated by the writer size_t fUsed; // updated by the writer const size_t fCapacity; SkBufferBlock(size_t capacity) : fNext(nullptr), fUsed(0), fCapacity(capacity) {} const void* startData() const { return this + 1; } size_t avail() const { return fCapacity - fUsed; } void* availData() { return (char*)this->startData() + fUsed; } static SkBufferBlock* Alloc(size_t length) { size_t capacity = LengthToCapacity(length); void* buffer = sk_malloc_throw(sizeof(SkBufferBlock) + capacity); return new (buffer) SkBufferBlock(capacity); } // Return number of bytes actually appended. Important that we always completely this block // before spilling into the next, since the reader uses fCapacity to know how many it can read. // size_t append(const void* src, size_t length) { this->validate(); size_t amount = SkTMin(this->avail(), length); memcpy(this->availData(), src, amount); fUsed += amount; this->validate(); return amount; } // Do not call in the reader thread, since the writer may be updating fUsed. // (The assertion is still true, but TSAN still may complain about its raciness.) void validate() const { #ifdef SK_DEBUG SkASSERT(fCapacity > 0); SkASSERT(fUsed <= fCapacity); #endif } private: static size_t LengthToCapacity(size_t length) { const size_t minSize = kMinAllocSize - sizeof(SkBufferBlock); return SkTMax(length, minSize); } }; struct SkBufferHead { mutable std::atomic fRefCnt; SkBufferBlock fBlock; SkBufferHead(size_t capacity) : fRefCnt(1), fBlock(capacity) {} static size_t LengthToCapacity(size_t length) { const size_t minSize = kMinAllocSize - sizeof(SkBufferHead); return SkTMax(length, minSize); } static SkBufferHead* Alloc(size_t length) { size_t capacity = LengthToCapacity(length); size_t size = sizeof(SkBufferHead) + capacity; void* buffer = sk_malloc_throw(size); return new (buffer) SkBufferHead(capacity); } void ref() const { SkAssertResult(fRefCnt.fetch_add(+1, std::memory_order_relaxed)); } void unref() const { // A release here acts in place of all releases we "should" have been doing in ref(). int32_t oldRefCnt = fRefCnt.fetch_add(-1, std::memory_order_acq_rel); SkASSERT(oldRefCnt); if (1 == oldRefCnt) { // Like unique(), the acquire is only needed on success. SkBufferBlock* block = fBlock.fNext; sk_free((void*)this); while (block) { SkBufferBlock* next = block->fNext; sk_free(block); block = next; } } } void validate(size_t minUsed, const SkBufferBlock* tail = nullptr) const { #ifdef SK_DEBUG SkASSERT(fRefCnt.load(std::memory_order_relaxed) > 0); size_t totalUsed = 0; const SkBufferBlock* block = &fBlock; const SkBufferBlock* lastBlock = block; while (block) { block->validate(); totalUsed += block->fUsed; lastBlock = block; block = block->fNext; } SkASSERT(minUsed <= totalUsed); if (tail) { SkASSERT(tail == lastBlock); } #endif } }; /////////////////////////////////////////////////////////////////////////////////////////////////// // The reader can only access block.fCapacity (which never changes), and cannot access // block.fUsed, which may be updated by the writer. // SkROBuffer::SkROBuffer(const SkBufferHead* head, size_t available, const SkBufferBlock* tail) : fHead(head), fAvailable(available), fTail(tail) { if (head) { fHead->ref(); SkASSERT(available > 0); head->validate(available, tail); } else { SkASSERT(0 == available); SkASSERT(!tail); } } SkROBuffer::~SkROBuffer() { if (fHead) { fHead->unref(); } } SkROBuffer::Iter::Iter(const SkROBuffer* buffer) { this->reset(buffer); } SkROBuffer::Iter::Iter(const sk_sp& buffer) { this->reset(buffer.get()); } void SkROBuffer::Iter::reset(const SkROBuffer* buffer) { fBuffer = buffer; if (buffer && buffer->fHead) { fBlock = &buffer->fHead->fBlock; fRemaining = buffer->fAvailable; } else { fBlock = nullptr; fRemaining = 0; } } const void* SkROBuffer::Iter::data() const { return fRemaining ? fBlock->startData() : nullptr; } size_t SkROBuffer::Iter::size() const { if (!fBlock) { return 0; } return SkTMin(fBlock->fCapacity, fRemaining); } bool SkROBuffer::Iter::next() { if (fRemaining) { fRemaining -= this->size(); if (fBuffer->fTail == fBlock) { // There are more blocks, but fBuffer does not know about them. SkASSERT(0 == fRemaining); fBlock = nullptr; } else { fBlock = fBlock->fNext; } } return fRemaining != 0; } /////////////////////////////////////////////////////////////////////////////////////////////////// SkRWBuffer::SkRWBuffer(size_t initialCapacity) : fHead(nullptr), fTail(nullptr), fTotalUsed(0) { if (initialCapacity) { fHead = SkBufferHead::Alloc(initialCapacity); fTail = &fHead->fBlock; } } SkRWBuffer::~SkRWBuffer() { this->validate(); if (fHead) { fHead->unref(); } } // It is important that we always completely fill the current block before spilling over to the // next, since our reader will be using fCapacity (min'd against its total available) to know how // many bytes to read from a given block. // void SkRWBuffer::append(const void* src, size_t length, size_t reserve) { this->validate(); if (0 == length) { return; } fTotalUsed += length; if (nullptr == fHead) { fHead = SkBufferHead::Alloc(length + reserve); fTail = &fHead->fBlock; } size_t written = fTail->append(src, length); SkASSERT(written <= length); src = (const char*)src + written; length -= written; if (length) { SkBufferBlock* block = SkBufferBlock::Alloc(length + reserve); fTail->fNext = block; fTail = block; written = fTail->append(src, length); SkASSERT(written == length); } this->validate(); } #ifdef SK_DEBUG void SkRWBuffer::validate() const { if (fHead) { fHead->validate(fTotalUsed, fTail); } else { SkASSERT(nullptr == fTail); SkASSERT(0 == fTotalUsed); } } #endif /////////////////////////////////////////////////////////////////////////////////////////////////// class SkROBufferStreamAsset : public SkStreamAsset { void validate() const { #ifdef SK_DEBUG SkASSERT(fGlobalOffset <= fBuffer->size()); SkASSERT(fLocalOffset <= fIter.size()); SkASSERT(fLocalOffset <= fGlobalOffset); #endif } #ifdef SK_DEBUG class AutoValidate { SkROBufferStreamAsset* fStream; public: AutoValidate(SkROBufferStreamAsset* stream) : fStream(stream) { stream->validate(); } ~AutoValidate() { fStream->validate(); } }; #define AUTO_VALIDATE AutoValidate av(this); #else #define AUTO_VALIDATE #endif public: SkROBufferStreamAsset(sk_sp buffer) : fBuffer(std::move(buffer)), fIter(fBuffer) { fGlobalOffset = fLocalOffset = 0; } size_t getLength() const override { return fBuffer->size(); } bool rewind() override { AUTO_VALIDATE fIter.reset(fBuffer.get()); fGlobalOffset = fLocalOffset = 0; return true; } size_t read(void* dst, size_t request) override { AUTO_VALIDATE size_t bytesRead = 0; for (;;) { size_t size = fIter.size(); SkASSERT(fLocalOffset <= size); size_t avail = SkTMin(size - fLocalOffset, request - bytesRead); if (dst) { memcpy(dst, (const char*)fIter.data() + fLocalOffset, avail); dst = (char*)dst + avail; } bytesRead += avail; fLocalOffset += avail; SkASSERT(bytesRead <= request); if (bytesRead == request) { break; } // If we get here, we've exhausted the current iter SkASSERT(fLocalOffset == size); fLocalOffset = 0; if (!fIter.next()) { break; // ran out of data } } fGlobalOffset += bytesRead; SkASSERT(fGlobalOffset <= fBuffer->size()); return bytesRead; } bool isAtEnd() const override { return fBuffer->size() == fGlobalOffset; } size_t getPosition() const override { return fGlobalOffset; } bool seek(size_t position) override { AUTO_VALIDATE if (position < fGlobalOffset) { this->rewind(); } (void)this->skip(position - fGlobalOffset); return true; } bool move(long offset) override{ AUTO_VALIDATE offset += fGlobalOffset; if (offset <= 0) { this->rewind(); } else { (void)this->seek(SkToSizeT(offset)); } return true; } private: SkStreamAsset* onDuplicate() const override { return new SkROBufferStreamAsset(fBuffer); } SkStreamAsset* onFork() const override { auto clone = this->duplicate(); clone->seek(this->getPosition()); return clone.release(); } sk_sp fBuffer; SkROBuffer::Iter fIter; size_t fLocalOffset; size_t fGlobalOffset; }; std::unique_ptr SkRWBuffer::makeStreamSnapshot() const { return skstd::make_unique(this->makeROBufferSnapshot()); }