/* * Copyright 2010 Google Inc. * * Use of this source code is governed by a BSD-style license that can be * found in the LICENSE file. */ #include "GrBufferAllocPool.h" #include "GrDrawTargetCaps.h" #include "GrGpu.h" #include "GrIndexBuffer.h" #include "GrTypes.h" #include "GrVertexBuffer.h" #if GR_DEBUG #define VALIDATE validate #else static void VALIDATE(bool = false) {} #endif // page size #define GrBufferAllocPool_MIN_BLOCK_SIZE ((size_t)1 << 12) GrBufferAllocPool::GrBufferAllocPool(GrGpu* gpu, BufferType bufferType, bool frequentResetHint, size_t blockSize, int preallocBufferCnt) : fBlocks(GrMax(8, 2*preallocBufferCnt)) { SkASSERT(NULL != gpu); fGpu = gpu; fGpu->ref(); fGpuIsReffed = true; fBufferType = bufferType; fFrequentResetHint = frequentResetHint; fBufferPtr = NULL; fMinBlockSize = GrMax(GrBufferAllocPool_MIN_BLOCK_SIZE, blockSize); fBytesInUse = 0; fPreallocBuffersInUse = 0; fPreallocBufferStartIdx = 0; for (int i = 0; i < preallocBufferCnt; ++i) { GrGeometryBuffer* buffer = this->createBuffer(fMinBlockSize); if (NULL != buffer) { *fPreallocBuffers.append() = buffer; } } } GrBufferAllocPool::~GrBufferAllocPool() { VALIDATE(); if (fBlocks.count()) { GrGeometryBuffer* buffer = fBlocks.back().fBuffer; if (buffer->isLocked()) { buffer->unlock(); } } while (!fBlocks.empty()) { destroyBlock(); } fPreallocBuffers.unrefAll(); releaseGpuRef(); } void GrBufferAllocPool::releaseGpuRef() { if (fGpuIsReffed) { fGpu->unref(); fGpuIsReffed = false; } } void GrBufferAllocPool::reset() { VALIDATE(); fBytesInUse = 0; if (fBlocks.count()) { GrGeometryBuffer* buffer = fBlocks.back().fBuffer; if (buffer->isLocked()) { buffer->unlock(); } } // fPreallocBuffersInUse will be decremented down to zero in the while loop int preallocBuffersInUse = fPreallocBuffersInUse; while (!fBlocks.empty()) { this->destroyBlock(); } if (fPreallocBuffers.count()) { // must set this after above loop. fPreallocBufferStartIdx = (fPreallocBufferStartIdx + preallocBuffersInUse) % fPreallocBuffers.count(); } // we may have created a large cpu mirror of a large VB. Reset the size // to match our pre-allocated VBs. fCpuData.reset(fMinBlockSize); SkASSERT(0 == fPreallocBuffersInUse); VALIDATE(); } void GrBufferAllocPool::unlock() { VALIDATE(); if (NULL != fBufferPtr) { BufferBlock& block = fBlocks.back(); if (block.fBuffer->isLocked()) { block.fBuffer->unlock(); } else { size_t flushSize = block.fBuffer->sizeInBytes() - block.fBytesFree; flushCpuData(fBlocks.back().fBuffer, flushSize); } fBufferPtr = NULL; } VALIDATE(); } #if GR_DEBUG void GrBufferAllocPool::validate(bool unusedBlockAllowed) const { if (NULL != fBufferPtr) { SkASSERT(!fBlocks.empty()); if (fBlocks.back().fBuffer->isLocked()) { GrGeometryBuffer* buf = fBlocks.back().fBuffer; SkASSERT(buf->lockPtr() == fBufferPtr); } else { SkASSERT(fCpuData.get() == fBufferPtr); } } else { SkASSERT(fBlocks.empty() || !fBlocks.back().fBuffer->isLocked()); } size_t bytesInUse = 0; for (int i = 0; i < fBlocks.count() - 1; ++i) { SkASSERT(!fBlocks[i].fBuffer->isLocked()); } for (int i = 0; i < fBlocks.count(); ++i) { size_t bytes = fBlocks[i].fBuffer->sizeInBytes() - fBlocks[i].fBytesFree; bytesInUse += bytes; SkASSERT(bytes || unusedBlockAllowed); } SkASSERT(bytesInUse == fBytesInUse); if (unusedBlockAllowed) { SkASSERT((fBytesInUse && !fBlocks.empty()) || (!fBytesInUse && (fBlocks.count() < 2))); } else { SkASSERT((0 == fBytesInUse) == fBlocks.empty()); } } #endif void* GrBufferAllocPool::makeSpace(size_t size, size_t alignment, const GrGeometryBuffer** buffer, size_t* offset) { VALIDATE(); SkASSERT(NULL != buffer); SkASSERT(NULL != offset); if (NULL != fBufferPtr) { BufferBlock& back = fBlocks.back(); size_t usedBytes = back.fBuffer->sizeInBytes() - back.fBytesFree; size_t pad = GrSizeAlignUpPad(usedBytes, alignment); if ((size + pad) <= back.fBytesFree) { usedBytes += pad; *offset = usedBytes; *buffer = back.fBuffer; back.fBytesFree -= size + pad; fBytesInUse += size + pad; VALIDATE(); return (void*)(reinterpret_cast(fBufferPtr) + usedBytes); } } // We could honor the space request using by a partial update of the current // VB (if there is room). But we don't currently use draw calls to GL that // allow the driver to know that previously issued draws won't read from // the part of the buffer we update. Also, the GL buffer implementation // may be cheating on the actual buffer size by shrinking the buffer on // updateData() if the amount of data passed is less than the full buffer // size. if (!createBlock(size)) { return NULL; } SkASSERT(NULL != fBufferPtr); *offset = 0; BufferBlock& back = fBlocks.back(); *buffer = back.fBuffer; back.fBytesFree -= size; fBytesInUse += size; VALIDATE(); return fBufferPtr; } int GrBufferAllocPool::currentBufferItems(size_t itemSize) const { VALIDATE(); if (NULL != fBufferPtr) { const BufferBlock& back = fBlocks.back(); size_t usedBytes = back.fBuffer->sizeInBytes() - back.fBytesFree; size_t pad = GrSizeAlignUpPad(usedBytes, itemSize); return (back.fBytesFree - pad) / itemSize; } else if (fPreallocBuffersInUse < fPreallocBuffers.count()) { return fMinBlockSize / itemSize; } return 0; } int GrBufferAllocPool::preallocatedBuffersRemaining() const { return fPreallocBuffers.count() - fPreallocBuffersInUse; } int GrBufferAllocPool::preallocatedBufferCount() const { return fPreallocBuffers.count(); } void GrBufferAllocPool::putBack(size_t bytes) { VALIDATE(); // if the putBack unwinds all the preallocated buffers then we will // advance the starting index. As blocks are destroyed fPreallocBuffersInUse // will be decremented. I will reach zero if all blocks using preallocated // buffers are released. int preallocBuffersInUse = fPreallocBuffersInUse; while (bytes) { // caller shouldnt try to put back more than they've taken SkASSERT(!fBlocks.empty()); BufferBlock& block = fBlocks.back(); size_t bytesUsed = block.fBuffer->sizeInBytes() - block.fBytesFree; if (bytes >= bytesUsed) { bytes -= bytesUsed; fBytesInUse -= bytesUsed; // if we locked a vb to satisfy the make space and we're releasing // beyond it, then unlock it. if (block.fBuffer->isLocked()) { block.fBuffer->unlock(); } this->destroyBlock(); } else { block.fBytesFree += bytes; fBytesInUse -= bytes; bytes = 0; break; } } if (!fPreallocBuffersInUse && fPreallocBuffers.count()) { fPreallocBufferStartIdx = (fPreallocBufferStartIdx + preallocBuffersInUse) % fPreallocBuffers.count(); } VALIDATE(); } bool GrBufferAllocPool::createBlock(size_t requestSize) { size_t size = GrMax(requestSize, fMinBlockSize); SkASSERT(size >= GrBufferAllocPool_MIN_BLOCK_SIZE); VALIDATE(); BufferBlock& block = fBlocks.push_back(); if (size == fMinBlockSize && fPreallocBuffersInUse < fPreallocBuffers.count()) { uint32_t nextBuffer = (fPreallocBuffersInUse + fPreallocBufferStartIdx) % fPreallocBuffers.count(); block.fBuffer = fPreallocBuffers[nextBuffer]; block.fBuffer->ref(); ++fPreallocBuffersInUse; } else { block.fBuffer = this->createBuffer(size); if (NULL == block.fBuffer) { fBlocks.pop_back(); return false; } } block.fBytesFree = size; if (NULL != fBufferPtr) { SkASSERT(fBlocks.count() > 1); BufferBlock& prev = fBlocks.fromBack(1); if (prev.fBuffer->isLocked()) { prev.fBuffer->unlock(); } else { flushCpuData(prev.fBuffer, prev.fBuffer->sizeInBytes() - prev.fBytesFree); } fBufferPtr = NULL; } SkASSERT(NULL == fBufferPtr); // If the buffer is CPU-backed we lock it because it is free to do so and saves a copy. // Otherwise when buffer locking is supported: // a) If the frequently reset hint is set we only lock when the requested size meets a // threshold (since we don't expect it is likely that we will see more vertex data) // b) If the hint is not set we lock if the buffer size is greater than the threshold. bool attemptLock = block.fBuffer->isCPUBacked(); if (!attemptLock && fGpu->caps()->bufferLockSupport()) { if (fFrequentResetHint) { attemptLock = requestSize > GR_GEOM_BUFFER_LOCK_THRESHOLD; } else { attemptLock = size > GR_GEOM_BUFFER_LOCK_THRESHOLD; } } if (attemptLock) { fBufferPtr = block.fBuffer->lock(); } if (NULL == fBufferPtr) { fBufferPtr = fCpuData.reset(size); } VALIDATE(true); return true; } void GrBufferAllocPool::destroyBlock() { SkASSERT(!fBlocks.empty()); BufferBlock& block = fBlocks.back(); if (fPreallocBuffersInUse > 0) { uint32_t prevPreallocBuffer = (fPreallocBuffersInUse + fPreallocBufferStartIdx + (fPreallocBuffers.count() - 1)) % fPreallocBuffers.count(); if (block.fBuffer == fPreallocBuffers[prevPreallocBuffer]) { --fPreallocBuffersInUse; } } SkASSERT(!block.fBuffer->isLocked()); block.fBuffer->unref(); fBlocks.pop_back(); fBufferPtr = NULL; } void GrBufferAllocPool::flushCpuData(GrGeometryBuffer* buffer, size_t flushSize) { SkASSERT(NULL != buffer); SkASSERT(!buffer->isLocked()); SkASSERT(fCpuData.get() == fBufferPtr); SkASSERT(flushSize <= buffer->sizeInBytes()); VALIDATE(true); if (fGpu->caps()->bufferLockSupport() && flushSize > GR_GEOM_BUFFER_LOCK_THRESHOLD) { void* data = buffer->lock(); if (NULL != data) { memcpy(data, fBufferPtr, flushSize); buffer->unlock(); return; } } buffer->updateData(fBufferPtr, flushSize); VALIDATE(true); } GrGeometryBuffer* GrBufferAllocPool::createBuffer(size_t size) { if (kIndex_BufferType == fBufferType) { return fGpu->createIndexBuffer(size, true); } else { SkASSERT(kVertex_BufferType == fBufferType); return fGpu->createVertexBuffer(size, true); } } //////////////////////////////////////////////////////////////////////////////// GrVertexBufferAllocPool::GrVertexBufferAllocPool(GrGpu* gpu, bool frequentResetHint, size_t bufferSize, int preallocBufferCnt) : GrBufferAllocPool(gpu, kVertex_BufferType, frequentResetHint, bufferSize, preallocBufferCnt) { } void* GrVertexBufferAllocPool::makeSpace(size_t vertexSize, int vertexCount, const GrVertexBuffer** buffer, int* startVertex) { SkASSERT(vertexCount >= 0); SkASSERT(NULL != buffer); SkASSERT(NULL != startVertex); size_t offset = 0; // assign to suppress warning const GrGeometryBuffer* geomBuffer = NULL; // assign to suppress warning void* ptr = INHERITED::makeSpace(vertexSize * vertexCount, vertexSize, &geomBuffer, &offset); *buffer = (const GrVertexBuffer*) geomBuffer; SkASSERT(0 == offset % vertexSize); *startVertex = offset / vertexSize; return ptr; } bool GrVertexBufferAllocPool::appendVertices(size_t vertexSize, int vertexCount, const void* vertices, const GrVertexBuffer** buffer, int* startVertex) { void* space = makeSpace(vertexSize, vertexCount, buffer, startVertex); if (NULL != space) { memcpy(space, vertices, vertexSize * vertexCount); return true; } else { return false; } } int GrVertexBufferAllocPool::preallocatedBufferVertices(size_t vertexSize) const { return INHERITED::preallocatedBufferSize() / vertexSize; } int GrVertexBufferAllocPool::currentBufferVertices(size_t vertexSize) const { return currentBufferItems(vertexSize); } //////////////////////////////////////////////////////////////////////////////// GrIndexBufferAllocPool::GrIndexBufferAllocPool(GrGpu* gpu, bool frequentResetHint, size_t bufferSize, int preallocBufferCnt) : GrBufferAllocPool(gpu, kIndex_BufferType, frequentResetHint, bufferSize, preallocBufferCnt) { } void* GrIndexBufferAllocPool::makeSpace(int indexCount, const GrIndexBuffer** buffer, int* startIndex) { SkASSERT(indexCount >= 0); SkASSERT(NULL != buffer); SkASSERT(NULL != startIndex); size_t offset = 0; // assign to suppress warning const GrGeometryBuffer* geomBuffer = NULL; // assign to suppress warning void* ptr = INHERITED::makeSpace(indexCount * sizeof(uint16_t), sizeof(uint16_t), &geomBuffer, &offset); *buffer = (const GrIndexBuffer*) geomBuffer; SkASSERT(0 == offset % sizeof(uint16_t)); *startIndex = offset / sizeof(uint16_t); return ptr; } bool GrIndexBufferAllocPool::appendIndices(int indexCount, const void* indices, const GrIndexBuffer** buffer, int* startIndex) { void* space = makeSpace(indexCount, buffer, startIndex); if (NULL != space) { memcpy(space, indices, sizeof(uint16_t) * indexCount); return true; } else { return false; } } int GrIndexBufferAllocPool::preallocatedBufferIndices() const { return INHERITED::preallocatedBufferSize() / sizeof(uint16_t); } int GrIndexBufferAllocPool::currentBufferIndices() const { return currentBufferItems(sizeof(uint16_t)); }