/* * Copyright 2010 Google Inc. * * Use of this source code is governed by a BSD-style license that can be * found in the LICENSE file. */ #include "GrAtlas.h" #include "GrContext.h" #include "GrGpu.h" #include "GrRectanizer.h" /////////////////////////////////////////////////////////////////////////////// // for testing #define FONT_CACHE_STATS 0 #if FONT_CACHE_STATS static int g_UploadCount = 0; #endif GrPlot::GrPlot() : fDrawToken(NULL, 0) , fTexture(NULL) , fRects(NULL) , fAtlasMgr(NULL) , fBytesPerPixel(1) , fDirty(false) , fBatchUploads(false) { fOffset.set(0, 0); } GrPlot::~GrPlot() { SkDELETE_ARRAY(fPlotData); fPlotData = NULL; delete fRects; } void GrPlot::init(GrAtlasMgr* mgr, int offX, int offY, int width, int height, size_t bpp, bool batchUploads) { fRects = GrRectanizer::Factory(width, height); fAtlasMgr = mgr; fOffset.set(offX * width, offY * height); fBytesPerPixel = bpp; fPlotData = NULL; fDirtyRect.setEmpty(); fDirty = false; fBatchUploads = batchUploads; } static inline void adjust_for_offset(SkIPoint16* loc, const SkIPoint16& offset) { loc->fX += offset.fX; loc->fY += offset.fY; } bool GrPlot::addSubImage(int width, int height, const void* image, SkIPoint16* loc) { float percentFull = fRects->percentFull(); if (!fRects->addRect(width, height, loc)) { return false; } // if batching uploads, create backing memory on first use // once the plot is nearly full we will revert to uploading each subimage individually int plotWidth = fRects->width(); int plotHeight = fRects->height(); if (fBatchUploads && NULL == fPlotData && 0.0f == percentFull) { fPlotData = SkNEW_ARRAY(unsigned char, fBytesPerPixel*plotWidth*plotHeight); memset(fPlotData, 0, fBytesPerPixel*plotWidth*plotHeight); } // if we have backing memory, copy to the memory and set for future upload if (NULL != fPlotData) { const unsigned char* imagePtr = (const unsigned char*) image; // point ourselves at the right starting spot unsigned char* dataPtr = fPlotData; dataPtr += fBytesPerPixel*plotWidth*loc->fY; dataPtr += fBytesPerPixel*loc->fX; // copy into the data buffer for (int i = 0; i < height; ++i) { memcpy(dataPtr, imagePtr, fBytesPerPixel*width); dataPtr += fBytesPerPixel*plotWidth; imagePtr += fBytesPerPixel*width; } fDirtyRect.join(loc->fX, loc->fY, loc->fX + width, loc->fY + height); adjust_for_offset(loc, fOffset); fDirty = true; // otherwise, just upload the image directly } else { adjust_for_offset(loc, fOffset); GrContext* context = fTexture->getContext(); context->writeTexturePixels(fTexture, loc->fX, loc->fY, width, height, fTexture->config(), image, 0, GrContext::kDontFlush_PixelOpsFlag); } #if FONT_CACHE_STATS ++g_UploadCount; #endif return true; } void GrPlot::uploadToTexture() { static const float kNearlyFullTolerance = 0.85f; // should only do this if batching is enabled SkASSERT(fBatchUploads); if (fDirty) { SkASSERT(NULL != fTexture); GrContext* context = fTexture->getContext(); // We pass the flag that does not force a flush. We assume our caller is // smart and hasn't referenced the part of the texture we're about to update // since the last flush. int rowBytes = fBytesPerPixel*fRects->width(); const unsigned char* dataPtr = fPlotData; dataPtr += rowBytes*fDirtyRect.fTop; dataPtr += fBytesPerPixel*fDirtyRect.fLeft; context->writeTexturePixels(fTexture, fOffset.fX + fDirtyRect.fLeft, fOffset.fY + fDirtyRect.fTop, fDirtyRect.width(), fDirtyRect.height(), fTexture->config(), dataPtr, rowBytes, GrContext::kDontFlush_PixelOpsFlag); fDirtyRect.setEmpty(); fDirty = false; // If the Plot is nearly full, anything else we add will probably be small and one // at a time, so free up the memory and after this upload any new images directly. if (fRects->percentFull() > kNearlyFullTolerance) { SkDELETE_ARRAY(fPlotData); fPlotData = NULL; } } } void GrPlot::resetRects() { SkASSERT(NULL != fRects); fRects->reset(); } /////////////////////////////////////////////////////////////////////////////// GrAtlasMgr::GrAtlasMgr(GrGpu* gpu, GrPixelConfig config, const SkISize& backingTextureSize, int numPlotsX, int numPlotsY, bool batchUploads) { fGpu = SkRef(gpu); fPixelConfig = config; fBackingTextureSize = backingTextureSize; fNumPlotsX = numPlotsX; fNumPlotsY = numPlotsY; fBatchUploads = batchUploads; fTexture = NULL; int textureWidth = fBackingTextureSize.width(); int textureHeight = fBackingTextureSize.height(); int plotWidth = textureWidth / fNumPlotsX; int plotHeight = textureHeight / fNumPlotsY; SkASSERT(plotWidth * fNumPlotsX == textureWidth); SkASSERT(plotHeight * fNumPlotsY == textureHeight); // We currently do not support compressed atlases... SkASSERT(!GrPixelConfigIsCompressed(config)); // set up allocated plots size_t bpp = GrBytesPerPixel(fPixelConfig); fPlotArray = SkNEW_ARRAY(GrPlot, (fNumPlotsX*fNumPlotsY)); GrPlot* currPlot = fPlotArray; for (int y = numPlotsY-1; y >= 0; --y) { for (int x = numPlotsX-1; x >= 0; --x) { currPlot->init(this, x, y, plotWidth, plotHeight, bpp, batchUploads); // build LRU list fPlotList.addToHead(currPlot); ++currPlot; } } } GrAtlasMgr::~GrAtlasMgr() { SkSafeUnref(fTexture); SkDELETE_ARRAY(fPlotArray); fGpu->unref(); #if FONT_CACHE_STATS GrPrintf("Num uploads: %d\n", g_UploadCount); #endif } void GrAtlasMgr::moveToHead(GrPlot* plot) { if (fPlotList.head() == plot) { return; } fPlotList.remove(plot); fPlotList.addToHead(plot); }; GrPlot* GrAtlasMgr::addToAtlas(GrAtlas* atlas, int width, int height, const void* image, SkIPoint16* loc) { // iterate through entire plot list for this atlas, see if we can find a hole // last one was most recently added and probably most empty for (int i = atlas->fPlots.count()-1; i >= 0; --i) { GrPlot* plot = atlas->fPlots[i]; if (plot->addSubImage(width, height, image, loc)) { this->moveToHead(plot); return plot; } } // before we get a new plot, make sure we have a backing texture if (NULL == fTexture) { // TODO: Update this to use the cache rather than directly creating a texture. GrTextureDesc desc; desc.fFlags = kDynamicUpdate_GrTextureFlagBit; desc.fWidth = fBackingTextureSize.width(); desc.fHeight = fBackingTextureSize.height(); desc.fConfig = fPixelConfig; fTexture = fGpu->createTexture(desc, NULL, 0); if (NULL == fTexture) { return NULL; } } // now look through all allocated plots for one we can share, in MRU order GrPlotList::Iter plotIter; plotIter.init(fPlotList, GrPlotList::Iter::kHead_IterStart); GrPlot* plot; while (NULL != (plot = plotIter.get())) { // make sure texture is set for quick lookup plot->fTexture = fTexture; if (plot->addSubImage(width, height, image, loc)) { this->moveToHead(plot); // new plot for atlas, put at end of array *(atlas->fPlots.append()) = plot; return plot; } plotIter.next(); } // If the above fails, then the current plot list has no room return NULL; } bool GrAtlasMgr::removePlot(GrAtlas* atlas, const GrPlot* plot) { // iterate through plot list for this atlas int count = atlas->fPlots.count(); for (int i = 0; i < count; ++i) { if (plot == atlas->fPlots[i]) { atlas->fPlots.remove(i); return true; } } return false; } // get a plot that's not being used by the current draw GrPlot* GrAtlasMgr::getUnusedPlot() { GrPlotList::Iter plotIter; plotIter.init(fPlotList, GrPlotList::Iter::kTail_IterStart); GrPlot* plot; while (NULL != (plot = plotIter.get())) { if (plot->drawToken().isIssued()) { return plot; } plotIter.prev(); } return NULL; } void GrAtlasMgr::uploadPlotsToTexture() { if (fBatchUploads) { GrPlotList::Iter plotIter; plotIter.init(fPlotList, GrPlotList::Iter::kHead_IterStart); GrPlot* plot; while (NULL != (plot = plotIter.get())) { plot->uploadToTexture(); plotIter.next(); } } }