diff options
-rw-r--r-- | src/gpu/vk/GrVkBuffer.cpp | 4 | ||||
-rw-r--r-- | src/gpu/vk/GrVkMemory.cpp | 80 | ||||
-rw-r--r-- | src/gpu/vk/GrVkMemory.h | 22 | ||||
-rwxr-xr-x | tests/VkHeapTests.cpp | 64 |
4 files changed, 49 insertions, 121 deletions
diff --git a/src/gpu/vk/GrVkBuffer.cpp b/src/gpu/vk/GrVkBuffer.cpp index 82674b4cfb..5d2b5659f0 100644 --- a/src/gpu/vk/GrVkBuffer.cpp +++ b/src/gpu/vk/GrVkBuffer.cpp @@ -45,9 +45,7 @@ const GrVkBuffer::Resource* GrVkBuffer::Create(const GrVkGpu* gpu, const Desc& d bufInfo.usage = VK_BUFFER_USAGE_TRANSFER_DST_BIT; break; } - if (!desc.fDynamic) { - bufInfo.usage |= VK_BUFFER_USAGE_TRANSFER_DST_BIT; - } + bufInfo.usage |= VK_BUFFER_USAGE_TRANSFER_DST_BIT; bufInfo.sharingMode = VK_SHARING_MODE_EXCLUSIVE; bufInfo.queueFamilyIndexCount = 0; diff --git a/src/gpu/vk/GrVkMemory.cpp b/src/gpu/vk/GrVkMemory.cpp index 4f96fcdba0..98b2f89e24 100644 --- a/src/gpu/vk/GrVkMemory.cpp +++ b/src/gpu/vk/GrVkMemory.cpp @@ -10,24 +10,16 @@ #include "GrVkGpu.h" #include "GrVkUtil.h" -#ifdef SK_DEBUG -// for simple tracking of how much we're using in each heap -// last counter is for non-subheap allocations -VkDeviceSize gHeapUsage[VK_MAX_MEMORY_HEAPS+1] = { 0 }; -#endif - static bool get_valid_memory_type_index(const VkPhysicalDeviceMemoryProperties& physDevMemProps, uint32_t typeBits, VkMemoryPropertyFlags requestedMemFlags, - uint32_t* typeIndex, - uint32_t* heapIndex) { + uint32_t* typeIndex) { for (uint32_t i = 0; i < physDevMemProps.memoryTypeCount; ++i) { if (typeBits & (1 << i)) { uint32_t supportedFlags = physDevMemProps.memoryTypes[i].propertyFlags & requestedMemFlags; if (supportedFlags == requestedMemFlags) { *typeIndex = i; - *heapIndex = physDevMemProps.memoryTypes[i].heapIndex; return true; } } @@ -64,7 +56,6 @@ bool GrVkMemory::AllocAndBindBufferMemory(const GrVkGpu* gpu, GR_VK_CALL(iface, GetBufferMemoryRequirements(device, buffer, &memReqs)); uint32_t typeIndex = 0; - uint32_t heapIndex = 0; const VkPhysicalDeviceMemoryProperties& phDevMemProps = gpu->physicalDeviceMemoryProperties(); if (dynamic) { // try to get cached and ideally non-coherent memory first @@ -72,14 +63,12 @@ bool GrVkMemory::AllocAndBindBufferMemory(const GrVkGpu* gpu, memReqs.memoryTypeBits, VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT | VK_MEMORY_PROPERTY_HOST_CACHED_BIT, - &typeIndex, - &heapIndex)) { + &typeIndex)) { // some sort of host-visible memory type should always be available for dynamic buffers SkASSERT_RELEASE(get_valid_memory_type_index(phDevMemProps, memReqs.memoryTypeBits, VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT, - &typeIndex, - &heapIndex)); + &typeIndex)); } VkMemoryPropertyFlags mpf = phDevMemProps.memoryTypes[typeIndex].propertyFlags; @@ -90,22 +79,15 @@ bool GrVkMemory::AllocAndBindBufferMemory(const GrVkGpu* gpu, SkASSERT_RELEASE(get_valid_memory_type_index(phDevMemProps, memReqs.memoryTypeBits, VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT, - &typeIndex, - &heapIndex)); + &typeIndex)); alloc->fFlags = 0x0; } GrVkHeap* heap = gpu->getHeap(buffer_type_to_heap(type)); - if (!heap->alloc(memReqs.size, memReqs.alignment, typeIndex, heapIndex, alloc)) { - // if static, try to allocate from non-host-visible non-device-local memory instead - if (dynamic || - !get_valid_memory_type_index(phDevMemProps, memReqs.memoryTypeBits, - 0, &typeIndex, &heapIndex) || - !heap->alloc(memReqs.size, memReqs.alignment, typeIndex, heapIndex, alloc)) { - SkDebugf("Failed to alloc buffer\n"); - return false; - } + if (!heap->alloc(memReqs.size, memReqs.alignment, typeIndex, alloc)) { + SkDebugf("Failed to alloc buffer\n"); + return false; } // Bind buffer @@ -148,7 +130,6 @@ bool GrVkMemory::AllocAndBindImageMemory(const GrVkGpu* gpu, GR_VK_CALL(iface, GetImageMemoryRequirements(device, image, &memReqs)); uint32_t typeIndex = 0; - uint32_t heapIndex = 0; GrVkHeap* heap; const VkPhysicalDeviceMemoryProperties& phDevMemProps = gpu->physicalDeviceMemoryProperties(); if (linearTiling) { @@ -157,14 +138,12 @@ bool GrVkMemory::AllocAndBindImageMemory(const GrVkGpu* gpu, if (!get_valid_memory_type_index(phDevMemProps, memReqs.memoryTypeBits, desiredMemProps, - &typeIndex, - &heapIndex)) { + &typeIndex)) { // some sort of host-visible memory type should always be available SkASSERT_RELEASE(get_valid_memory_type_index(phDevMemProps, memReqs.memoryTypeBits, VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT, - &typeIndex, - &heapIndex)); + &typeIndex)); } heap = gpu->getHeap(GrVkGpu::kLinearImage_Heap); VkMemoryPropertyFlags mpf = phDevMemProps.memoryTypes[typeIndex].propertyFlags; @@ -175,8 +154,7 @@ bool GrVkMemory::AllocAndBindImageMemory(const GrVkGpu* gpu, SkASSERT_RELEASE(get_valid_memory_type_index(phDevMemProps, memReqs.memoryTypeBits, VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT, - &typeIndex, - &heapIndex)); + &typeIndex)); if (memReqs.size <= kMaxSmallImageSize) { heap = gpu->getHeap(GrVkGpu::kSmallOptimalImage_Heap); } else { @@ -185,15 +163,9 @@ bool GrVkMemory::AllocAndBindImageMemory(const GrVkGpu* gpu, alloc->fFlags = 0x0; } - if (!heap->alloc(memReqs.size, memReqs.alignment, typeIndex, heapIndex, alloc)) { - // if optimal, try to allocate from non-host-visible non-device-local memory instead - if (linearTiling || - !get_valid_memory_type_index(phDevMemProps, memReqs.memoryTypeBits, - 0, &typeIndex, &heapIndex) || - !heap->alloc(memReqs.size, memReqs.alignment, typeIndex, heapIndex, alloc)) { - SkDebugf("Failed to alloc image\n"); - return false; - } + if (!heap->alloc(memReqs.size, memReqs.alignment, typeIndex, alloc)) { + SkDebugf("Failed to alloc image\n"); + return false; } // Bind image @@ -459,12 +431,11 @@ void GrVkFreeListAlloc::free(VkDeviceSize allocOffset, VkDeviceSize allocSize) { #endif } -GrVkSubHeap::GrVkSubHeap(const GrVkGpu* gpu, uint32_t memoryTypeIndex, uint32_t heapIndex, +GrVkSubHeap::GrVkSubHeap(const GrVkGpu* gpu, uint32_t memoryTypeIndex, VkDeviceSize size, VkDeviceSize alignment) : INHERITED(size, alignment) , fGpu(gpu) - , fMemoryTypeIndex(memoryTypeIndex) - , fHeapIndex(heapIndex) { + , fMemoryTypeIndex(memoryTypeIndex) { VkMemoryAllocateInfo allocInfo = { VK_STRUCTURE_TYPE_MEMORY_ALLOCATE_INFO, // sType @@ -479,20 +450,12 @@ GrVkSubHeap::GrVkSubHeap(const GrVkGpu* gpu, uint32_t memoryTypeIndex, uint32_t &fAlloc)); if (VK_SUCCESS != err) { this->reset(); - } -#ifdef SK_DEBUG - else { - gHeapUsage[heapIndex] += size; } -#endif } GrVkSubHeap::~GrVkSubHeap() { const GrVkInterface* iface = fGpu->vkInterface(); GR_VK_CALL(iface, FreeMemory(fGpu->device(), fAlloc, nullptr)); -#ifdef SK_DEBUG - gHeapUsage[fHeapIndex] -= fSize; -#endif } bool GrVkSubHeap::alloc(VkDeviceSize size, GrVkAlloc* alloc) { @@ -507,7 +470,7 @@ void GrVkSubHeap::free(const GrVkAlloc& alloc) { } bool GrVkHeap::subAlloc(VkDeviceSize size, VkDeviceSize alignment, - uint32_t memoryTypeIndex, uint32_t heapIndex, GrVkAlloc* alloc) { + uint32_t memoryTypeIndex, GrVkAlloc* alloc) { VkDeviceSize alignedSize = align_size(size, alignment); // if requested is larger than our subheap allocation, just alloc directly @@ -528,9 +491,6 @@ bool GrVkHeap::subAlloc(VkDeviceSize size, VkDeviceSize alignment, } alloc->fOffset = 0; alloc->fSize = 0; // hint that this is not a subheap allocation -#ifdef SK_DEBUG - gHeapUsage[VK_MAX_MEMORY_HEAPS] += alignedSize; -#endif return true; } @@ -560,11 +520,11 @@ bool GrVkHeap::subAlloc(VkDeviceSize size, VkDeviceSize alignment, // need to allocate a new subheap SkAutoTDelete<GrVkSubHeap>& subHeap = fSubHeaps.push_back(); - subHeap.reset(new GrVkSubHeap(fGpu, memoryTypeIndex, heapIndex, fSubHeapSize, alignment)); + subHeap.reset(new GrVkSubHeap(fGpu, memoryTypeIndex, fSubHeapSize, alignment)); // try to recover from failed allocation by only allocating what we need if (subHeap->size() == 0) { VkDeviceSize alignedSize = align_size(size, alignment); - subHeap.reset(new GrVkSubHeap(fGpu, memoryTypeIndex, heapIndex, alignedSize, alignment)); + subHeap.reset(new GrVkSubHeap(fGpu, memoryTypeIndex, alignedSize, alignment)); if (subHeap->size() == 0) { return false; } @@ -579,7 +539,7 @@ bool GrVkHeap::subAlloc(VkDeviceSize size, VkDeviceSize alignment, } bool GrVkHeap::singleAlloc(VkDeviceSize size, VkDeviceSize alignment, - uint32_t memoryTypeIndex, uint32_t heapIndex, GrVkAlloc* alloc) { + uint32_t memoryTypeIndex, GrVkAlloc* alloc) { VkDeviceSize alignedSize = align_size(size, alignment); // first try to find an unallocated subheap that fits our allocation request @@ -608,7 +568,7 @@ bool GrVkHeap::singleAlloc(VkDeviceSize size, VkDeviceSize alignment, // need to allocate a new subheap SkAutoTDelete<GrVkSubHeap>& subHeap = fSubHeaps.push_back(); - subHeap.reset(new GrVkSubHeap(fGpu, memoryTypeIndex, heapIndex, alignedSize, alignment)); + subHeap.reset(new GrVkSubHeap(fGpu, memoryTypeIndex, alignedSize, alignment)); fAllocSize += alignedSize; if (subHeap->alloc(size, alloc)) { fUsedSize += alloc->fSize; diff --git a/src/gpu/vk/GrVkMemory.h b/src/gpu/vk/GrVkMemory.h index 34850dba06..f8d5fdf91c 100644 --- a/src/gpu/vk/GrVkMemory.h +++ b/src/gpu/vk/GrVkMemory.h @@ -93,7 +93,7 @@ protected: class GrVkSubHeap : public GrVkFreeListAlloc { public: - GrVkSubHeap(const GrVkGpu* gpu, uint32_t memoryTypeIndex, uint32_t heapIndex, + GrVkSubHeap(const GrVkGpu* gpu, uint32_t memoryTypeIndex, VkDeviceSize size, VkDeviceSize alignment); ~GrVkSubHeap(); @@ -106,7 +106,6 @@ public: private: const GrVkGpu* fGpu; uint32_t fMemoryTypeIndex; - uint32_t fHeapIndex; VkDeviceMemory fAlloc; typedef GrVkFreeListAlloc INHERITED; @@ -136,24 +135,21 @@ public: VkDeviceSize allocSize() const { return fAllocSize; } VkDeviceSize usedSize() const { return fUsedSize; } - bool alloc(VkDeviceSize size, VkDeviceSize alignment, uint32_t memoryTypeIndex, - uint32_t heapIndex, GrVkAlloc* alloc) { + bool alloc(VkDeviceSize size, VkDeviceSize alignment, uint32_t memoryTypeIndex, + GrVkAlloc* alloc) { SkASSERT(size > 0); - return (*this.*fAllocFunc)(size, alignment, memoryTypeIndex, heapIndex, alloc); + return (*this.*fAllocFunc)(size, alignment, memoryTypeIndex, alloc); } bool free(const GrVkAlloc& alloc); private: - typedef bool (GrVkHeap::*AllocFunc)(VkDeviceSize size, VkDeviceSize alignment, - uint32_t memoryTypeIndex, uint32_t heapIndex, - GrVkAlloc* alloc); + typedef bool (GrVkHeap::*AllocFunc)(VkDeviceSize size, VkDeviceSize alignment, + uint32_t memoryTypeIndex, GrVkAlloc* alloc); - bool subAlloc(VkDeviceSize size, VkDeviceSize alignment, - uint32_t memoryTypeIndex, uint32_t heapIndex, - GrVkAlloc* alloc); + bool subAlloc(VkDeviceSize size, VkDeviceSize alignment, + uint32_t memoryTypeIndex, GrVkAlloc* alloc); bool singleAlloc(VkDeviceSize size, VkDeviceSize alignment, - uint32_t memoryTypeIndex, uint32_t heapIndex, - GrVkAlloc* alloc); + uint32_t memoryTypeIndex, GrVkAlloc* alloc); const GrVkGpu* fGpu; VkDeviceSize fSubHeapSize; diff --git a/tests/VkHeapTests.cpp b/tests/VkHeapTests.cpp index 4561c90803..c4a9beb5f2 100755 --- a/tests/VkHeapTests.cpp +++ b/tests/VkHeapTests.cpp @@ -21,8 +21,8 @@ using sk_gpu_test::GrContextFactory; void subheap_test(skiatest::Reporter* reporter, GrContext* context) { GrVkGpu* gpu = static_cast<GrVkGpu*>(context->getGpu()); - // memtype doesn't matter, we're just testing the suballocation algorithm so we'll use 0 - GrVkSubHeap heap(gpu, 0, 0, 64 * 1024, 32); + // heap index doesn't matter, we're just testing the suballocation algorithm so we'll use 0 + GrVkSubHeap heap(gpu, 0, 64 * 1024, 32); GrVkAlloc alloc0, alloc1, alloc2, alloc3; // test full allocation and free REPORTER_ASSERT(reporter, heap.alloc(64 * 1024, &alloc0)); @@ -118,35 +118,34 @@ void subheap_test(skiatest::Reporter* reporter, GrContext* context) { void suballoc_test(skiatest::Reporter* reporter, GrContext* context) { GrVkGpu* gpu = static_cast<GrVkGpu*>(context->getGpu()); - // memtype/heap index don't matter, we're just testing the allocation algorithm so we'll use 0 + // heap index doesn't matter, we're just testing the allocation algorithm so we'll use 0 GrVkHeap heap(gpu, GrVkHeap::kSubAlloc_Strategy, 64 * 1024); GrVkAlloc alloc0, alloc1, alloc2, alloc3; const VkDeviceSize kAlignment = 16; - const uint32_t kMemType = 0; const uint32_t kHeapIndex = 0; REPORTER_ASSERT(reporter, heap.allocSize() == 0 && heap.usedSize() == 0); // fragment allocations so we need to grow heap - REPORTER_ASSERT(reporter, heap.alloc(19 * 1024 - 3, kAlignment, kMemType, kHeapIndex, &alloc0)); - REPORTER_ASSERT(reporter, heap.alloc(5 * 1024 - 9, kAlignment, kMemType, kHeapIndex, &alloc1)); - REPORTER_ASSERT(reporter, heap.alloc(15 * 1024 - 15, kAlignment, kMemType, kHeapIndex, &alloc2)); - REPORTER_ASSERT(reporter, heap.alloc(3 * 1024 - 6, kAlignment, kMemType, kHeapIndex, &alloc3)); + REPORTER_ASSERT(reporter, heap.alloc(19 * 1024 - 3, kAlignment, kHeapIndex, &alloc0)); + REPORTER_ASSERT(reporter, heap.alloc(5 * 1024 - 9, kAlignment, kHeapIndex, &alloc1)); + REPORTER_ASSERT(reporter, heap.alloc(15 * 1024 - 15, kAlignment, kHeapIndex, &alloc2)); + REPORTER_ASSERT(reporter, heap.alloc(3 * 1024 - 6, kAlignment, kHeapIndex, &alloc3)); REPORTER_ASSERT(reporter, heap.allocSize() == 64 * 1024 && heap.usedSize() == 42 * 1024); heap.free(alloc0); REPORTER_ASSERT(reporter, heap.allocSize() == 64 * 1024 && heap.usedSize() == 23 * 1024); heap.free(alloc2); REPORTER_ASSERT(reporter, heap.allocSize() == 64 * 1024 && heap.usedSize() == 8 * 1024); // we expect the heap to grow here - REPORTER_ASSERT(reporter, heap.alloc(40 * 1024, kAlignment, kMemType, kHeapIndex, &alloc0)); + REPORTER_ASSERT(reporter, heap.alloc(40 * 1024, kAlignment, kHeapIndex, &alloc0)); REPORTER_ASSERT(reporter, heap.allocSize() == 128 * 1024 && heap.usedSize() == 48 * 1024); heap.free(alloc3); REPORTER_ASSERT(reporter, heap.allocSize() == 128 * 1024 && heap.usedSize() == 45 * 1024); // heap should not grow here (first subheap has exactly enough room) - REPORTER_ASSERT(reporter, heap.alloc(40 * 1024, kAlignment, kMemType, kHeapIndex, &alloc3)); + REPORTER_ASSERT(reporter, heap.alloc(40 * 1024, kAlignment, kHeapIndex, &alloc3)); REPORTER_ASSERT(reporter, heap.allocSize() == 128 * 1024 && heap.usedSize() == 85 * 1024); // heap should not grow here (second subheap has room) - REPORTER_ASSERT(reporter, heap.alloc(22 * 1024, kAlignment, kMemType, kHeapIndex, &alloc2)); + REPORTER_ASSERT(reporter, heap.alloc(22 * 1024, kAlignment, kHeapIndex, &alloc2)); REPORTER_ASSERT(reporter, heap.allocSize() == 128 * 1024 && heap.usedSize() == 107 * 1024); heap.free(alloc1); REPORTER_ASSERT(reporter, heap.allocSize() == 128 * 1024 && heap.usedSize() == 102 * 1024); @@ -157,58 +156,45 @@ void suballoc_test(skiatest::Reporter* reporter, GrContext* context) { heap.free(alloc3); REPORTER_ASSERT(reporter, heap.allocSize() == 128 * 1024 && heap.usedSize() == 0 * 1024); // heap should not grow here (allocating more than subheap size) - REPORTER_ASSERT(reporter, heap.alloc(128 * 1024, kAlignment, kMemType, kHeapIndex, &alloc0)); + REPORTER_ASSERT(reporter, heap.alloc(128 * 1024, kAlignment, kHeapIndex, &alloc0)); REPORTER_ASSERT(reporter, 0 == alloc0.fSize); REPORTER_ASSERT(reporter, heap.allocSize() == 128 * 1024 && heap.usedSize() == 0 * 1024); heap.free(alloc0); - REPORTER_ASSERT(reporter, heap.alloc(24 * 1024, kAlignment, kMemType, kHeapIndex, &alloc0)); - REPORTER_ASSERT(reporter, heap.allocSize() == 128 * 1024 && heap.usedSize() == 24 * 1024); - // heap should alloc a new subheap because the memory type is different - REPORTER_ASSERT(reporter, heap.alloc(24 * 1024, kAlignment, kMemType+1, kHeapIndex, &alloc1)); - REPORTER_ASSERT(reporter, heap.allocSize() == 192 * 1024 && heap.usedSize() == 48 * 1024); - // heap should alloc a new subheap because the alignment is different - REPORTER_ASSERT(reporter, heap.alloc(24 * 1024, 128, kMemType, kHeapIndex, &alloc2)); - REPORTER_ASSERT(reporter, heap.allocSize() == 256 * 1024 && heap.usedSize() == 72 * 1024); - heap.free(alloc2); - heap.free(alloc0); - heap.free(alloc1); - REPORTER_ASSERT(reporter, heap.allocSize() == 256 * 1024 && heap.usedSize() == 0 * 1024); } void singlealloc_test(skiatest::Reporter* reporter, GrContext* context) { GrVkGpu* gpu = static_cast<GrVkGpu*>(context->getGpu()); - // memtype/heap index don't matter, we're just testing the allocation algorithm so we'll use 0 + // heap index doesn't matter, we're just testing the allocation algorithm so we'll use 0 GrVkHeap heap(gpu, GrVkHeap::kSingleAlloc_Strategy, 64 * 1024); GrVkAlloc alloc0, alloc1, alloc2, alloc3; const VkDeviceSize kAlignment = 64; - const uint32_t kMemType = 0; const uint32_t kHeapIndex = 0; REPORTER_ASSERT(reporter, heap.allocSize() == 0 && heap.usedSize() == 0); // make a few allocations - REPORTER_ASSERT(reporter, heap.alloc(49 * 1024 - 3, kAlignment, kMemType, kHeapIndex, &alloc0)); - REPORTER_ASSERT(reporter, heap.alloc(5 * 1024 - 37, kAlignment, kMemType, kHeapIndex, &alloc1)); - REPORTER_ASSERT(reporter, heap.alloc(15 * 1024 - 11, kAlignment, kMemType, kHeapIndex, &alloc2)); - REPORTER_ASSERT(reporter, heap.alloc(3 * 1024 - 29, kAlignment, kMemType, kHeapIndex, &alloc3)); + REPORTER_ASSERT(reporter, heap.alloc(49 * 1024 - 3, kAlignment, kHeapIndex, &alloc0)); + REPORTER_ASSERT(reporter, heap.alloc(5 * 1024 - 37, kAlignment, kHeapIndex, &alloc1)); + REPORTER_ASSERT(reporter, heap.alloc(15 * 1024 - 11, kAlignment, kHeapIndex, &alloc2)); + REPORTER_ASSERT(reporter, heap.alloc(3 * 1024 - 29, kAlignment, kHeapIndex, &alloc3)); REPORTER_ASSERT(reporter, heap.allocSize() == 72 * 1024 && heap.usedSize() == 72 * 1024); heap.free(alloc0); REPORTER_ASSERT(reporter, heap.allocSize() == 72 * 1024 && heap.usedSize() == 23 * 1024); heap.free(alloc2); REPORTER_ASSERT(reporter, heap.allocSize() == 72 * 1024 && heap.usedSize() == 8 * 1024); // heap should not grow here (first subheap has room) - REPORTER_ASSERT(reporter, heap.alloc(40 * 1024, kAlignment, kMemType, kHeapIndex, &alloc0)); + REPORTER_ASSERT(reporter, heap.alloc(40 * 1024, kAlignment, kHeapIndex, &alloc0)); REPORTER_ASSERT(reporter, heap.allocSize() == 72 * 1024 && heap.usedSize() == 48 * 1024); heap.free(alloc3); REPORTER_ASSERT(reporter, heap.allocSize() == 72 * 1024 && heap.usedSize() == 45 * 1024); // check for exact fit -- heap should not grow here (third subheap has room) - REPORTER_ASSERT(reporter, heap.alloc(15 * 1024 - 63, kAlignment, kMemType, kHeapIndex, &alloc2)); + REPORTER_ASSERT(reporter, heap.alloc(15 * 1024 - 63, kAlignment, kHeapIndex, &alloc2)); REPORTER_ASSERT(reporter, heap.allocSize() == 72 * 1024 && heap.usedSize() == 60 * 1024); heap.free(alloc2); REPORTER_ASSERT(reporter, heap.allocSize() == 72 * 1024 && heap.usedSize() == 45 * 1024); // heap should grow here (no subheap has room) - REPORTER_ASSERT(reporter, heap.alloc(40 * 1024, kAlignment, kMemType, kHeapIndex, &alloc3)); + REPORTER_ASSERT(reporter, heap.alloc(40 * 1024, kAlignment, kHeapIndex, &alloc3)); REPORTER_ASSERT(reporter, heap.allocSize() == 112 * 1024 && heap.usedSize() == 85 * 1024); heap.free(alloc1); REPORTER_ASSERT(reporter, heap.allocSize() == 112 * 1024 && heap.usedSize() == 80 * 1024); @@ -216,18 +202,6 @@ void singlealloc_test(skiatest::Reporter* reporter, GrContext* context) { REPORTER_ASSERT(reporter, heap.allocSize() == 112 * 1024 && heap.usedSize() == 40 * 1024); heap.free(alloc3); REPORTER_ASSERT(reporter, heap.allocSize() == 112 * 1024 && heap.usedSize() == 0 * 1024); - REPORTER_ASSERT(reporter, heap.alloc(24 * 1024, kAlignment, kMemType, kHeapIndex, &alloc0)); - REPORTER_ASSERT(reporter, heap.allocSize() == 112 * 1024 && heap.usedSize() == 24 * 1024); - // heap should alloc a new subheap because the memory type is different - REPORTER_ASSERT(reporter, heap.alloc(24 * 1024, kAlignment, kMemType + 1, kHeapIndex, &alloc1)); - REPORTER_ASSERT(reporter, heap.allocSize() == 136 * 1024 && heap.usedSize() == 48 * 1024); - // heap should alloc a new subheap because the alignment is different - REPORTER_ASSERT(reporter, heap.alloc(24 * 1024, 128, kMemType, kHeapIndex, &alloc2)); - REPORTER_ASSERT(reporter, heap.allocSize() == 160 * 1024 && heap.usedSize() == 72 * 1024); - heap.free(alloc1); - heap.free(alloc2); - heap.free(alloc0); - REPORTER_ASSERT(reporter, heap.allocSize() == 160 * 1024 && heap.usedSize() == 0 * 1024); } DEF_GPUTEST_FOR_VULKAN_CONTEXT(VkHeapTests, reporter, ctxInfo) { |