/* * Copyright 2014 Google Inc. * * Use of this source code is governed by a BSD-style license that can be * found in the LICENSE file. */ #ifndef GrPathRange_DEFINED #define GrPathRange_DEFINED #include "GrGpuResource.h" #include "SkPath.h" #include "SkRefCnt.h" #include "SkTArray.h" class SkDescriptor; /** * Represents a contiguous range of GPU path objects. * This object is immutable with the exception that individual paths may be * initialized lazily. */ class GrPathRange : public GrGpuResource { public: enum PathIndexType { kU8_PathIndexType, //!< uint8_t kU16_PathIndexType, //!< uint16_t kU32_PathIndexType, //!< uint32_t kLast_PathIndexType = kU32_PathIndexType }; static inline int PathIndexSizeInBytes(PathIndexType type) { GR_STATIC_ASSERT(0 == kU8_PathIndexType); GR_STATIC_ASSERT(1 == kU16_PathIndexType); GR_STATIC_ASSERT(2 == kU32_PathIndexType); GR_STATIC_ASSERT(kU32_PathIndexType == kLast_PathIndexType); return 1 << type; } /** * Class that generates the paths for a specific range. */ class PathGenerator : public SkRefCnt { public: virtual int getNumPaths() = 0; virtual void generatePath(int index, SkPath* out) = 0; #ifdef SK_DEBUG virtual bool isEqualTo(const SkDescriptor&) const { return false; } #endif virtual ~PathGenerator() {} }; /** * Initialize a lazy-loaded path range. This class will generate an SkPath and call * onInitPath() for each path within the range before it is drawn for the first time. */ GrPathRange(GrGpu*, PathGenerator*); /** * Initialize an eager-loaded path range. The subclass is responsible for ensuring all * the paths are initialized up front. */ GrPathRange(GrGpu*, int numPaths); int getNumPaths() const { return fNumPaths; } const PathGenerator* getPathGenerator() const { return fPathGenerator.get(); } void loadPathsIfNeeded(const void* indices, PathIndexType, int count) const; template void loadPathsIfNeeded(const IndexType* indices, int count) const { if (!fPathGenerator) { return; } bool didLoadPaths = false; for (int i = 0; i < count; ++i) { SkASSERT(indices[i] < static_cast(fNumPaths)); const int groupIndex = indices[i] / kPathsPerGroup; const int groupByte = groupIndex / 8; const uint8_t groupBit = 1 << (groupIndex % 8); const bool hasPath = SkToBool(fGeneratedPaths[groupByte] & groupBit); if (!hasPath) { // We track which paths are loaded in groups of kPathsPerGroup. To // mark a path as loaded we need to load the entire group. const int groupFirstPath = groupIndex * kPathsPerGroup; const int groupLastPath = SkTMin(groupFirstPath + kPathsPerGroup, fNumPaths) - 1; SkPath path; for (int pathIdx = groupFirstPath; pathIdx <= groupLastPath; ++pathIdx) { fPathGenerator->generatePath(pathIdx, &path); this->onInitPath(pathIdx, path); } fGeneratedPaths[groupByte] |= groupBit; didLoadPaths = true; } } if (didLoadPaths) { this->didChangeGpuMemorySize(); } } #ifdef SK_DEBUG void assertPathsLoaded(const void* indices, PathIndexType, int count) const; template void assertPathsLoaded(const IndexType* indices, int count) const { if (!fPathGenerator) { return; } for (int i = 0; i < count; ++i) { SkASSERT(indices[i] < static_cast(fNumPaths)); const int groupIndex = indices[i] / kPathsPerGroup; const int groupByte = groupIndex / 8; const uint8_t groupBit = 1 << (groupIndex % 8); SkASSERT(fGeneratedPaths[groupByte] & groupBit); } } virtual bool isEqualTo(const SkDescriptor& desc) const { return nullptr != fPathGenerator.get() && fPathGenerator->isEqualTo(desc); } #endif protected: // Initialize a path in the range before drawing. This is only called when // fPathGenerator is non-null. The child class need not call didChangeGpuMemorySize(), // GrPathRange will take care of that after the call is complete. virtual void onInitPath(int index, const SkPath&) const = 0; private: enum { kPathsPerGroup = 16 // Paths get tracked in groups of 16 for lazy loading. }; mutable sk_sp fPathGenerator; mutable SkTArray fGeneratedPaths; const int fNumPaths; typedef GrGpuResource INHERITED; }; #endif