| /* |
| * Copyright 2020 Google LLC |
| * |
| * Use of this source code is governed by a BSD-style license that can be |
| * found in the LICENSE file. |
| */ |
| |
| #include "src/gpu/GrBlockAllocator.h" |
| #include "tests/Test.h" |
| |
| using Block = GrBlockAllocator::Block; |
| using GrowthPolicy = GrBlockAllocator::GrowthPolicy; |
| |
| // Helper functions for modifying the allocator in a controlled manner |
| template<size_t N> |
| static int block_count(const GrSBlockAllocator<N>& pool) { |
| int ct = 0; |
| for (const Block* b : pool->blocks()) { |
| (void) b; |
| ct++; |
| } |
| return ct; |
| } |
| |
| template<size_t N> |
| static Block* get_block(GrSBlockAllocator<N>& pool, int blockIndex) { |
| Block* found = nullptr; |
| int i = 0; |
| for (Block* b: pool->blocks()) { |
| if (i == blockIndex) { |
| found = b; |
| break; |
| } |
| i++; |
| } |
| |
| SkASSERT(found != nullptr); |
| return found; |
| } |
| |
| template<size_t N> |
| static size_t add_block(GrSBlockAllocator<N>& pool) { |
| size_t currentSize = pool->totalSize(); |
| while(pool->totalSize() == currentSize) { |
| pool->template allocate<4>(pool->preallocSize() / 2); |
| } |
| return pool->totalSize() - currentSize; |
| } |
| |
| template<size_t N> |
| static void* alloc_byte(GrSBlockAllocator<N>& pool) { |
| auto br = pool->template allocate<1>(1); |
| return br.fBlock->ptr(br.fAlignedOffset); |
| } |
| |
| DEF_TEST(GrBlockAllocatorPreallocSize, r) { |
| // Tests stack/member initialization, option #1 described in doc |
| GrBlockAllocator stack{GrowthPolicy::kFixed, 2048}; |
| SkDEBUGCODE(stack.validate();) |
| |
| REPORTER_ASSERT(r, stack.preallocSize() == sizeof(GrBlockAllocator)); |
| REPORTER_ASSERT(r, stack.preallocUsableSpace() == (size_t) stack.currentBlock()->avail()); |
| |
| // Tests placement new initialization to increase head block size, option #2 |
| void* mem = operator new(1024); |
| GrBlockAllocator* placement = new (mem) GrBlockAllocator(GrowthPolicy::kLinear, 1024, |
| 1024 - sizeof(GrBlockAllocator)); |
| REPORTER_ASSERT(r, placement->preallocSize() == 1024); |
| REPORTER_ASSERT(r, placement->preallocUsableSpace() < 1024 && |
| placement->preallocUsableSpace() >= (1024 - sizeof(GrBlockAllocator))); |
| delete placement; |
| |
| // Tests inline increased preallocation, option #3 |
| GrSBlockAllocator<2048> inlined{}; |
| SkDEBUGCODE(inlined->validate();) |
| REPORTER_ASSERT(r, inlined->preallocSize() == 2048); |
| REPORTER_ASSERT(r, inlined->preallocUsableSpace() < 2048 && |
| inlined->preallocUsableSpace() >= (2048 - sizeof(GrBlockAllocator))); |
| } |
| |
| DEF_TEST(GrBlockAllocatorAlloc, r) { |
| GrSBlockAllocator<1024> pool{}; |
| SkDEBUGCODE(pool->validate();) |
| |
| // Assumes the previous pointer was in the same block |
| auto validate_ptr = [&](int align, int size, |
| GrBlockAllocator::ByteRange br, |
| GrBlockAllocator::ByteRange* prevBR) { |
| uintptr_t pt = reinterpret_cast<uintptr_t>(br.fBlock->ptr(br.fAlignedOffset)); |
| // Matches the requested align |
| REPORTER_ASSERT(r, pt % align == 0); |
| // And large enough |
| REPORTER_ASSERT(r, br.fEnd - br.fAlignedOffset >= size); |
| // And has enough padding for alignment |
| REPORTER_ASSERT(r, br.fAlignedOffset - br.fStart >= 0); |
| REPORTER_ASSERT(r, br.fAlignedOffset - br.fStart <= align - 1); |
| // And block of the returned struct is the current block of the allocator |
| REPORTER_ASSERT(r, pool->currentBlock() == br.fBlock); |
| |
| // And make sure that we're past the required end of the previous allocation |
| if (prevBR) { |
| uintptr_t prevEnd = |
| reinterpret_cast<uintptr_t>(prevBR->fBlock->ptr(prevBR->fEnd - 1)); |
| REPORTER_ASSERT(r, pt > prevEnd); |
| } |
| }; |
| |
| auto p1 = pool->allocate<1>(14); |
| validate_ptr(1, 14, p1, nullptr); |
| |
| auto p2 = pool->allocate<2>(24); |
| validate_ptr(2, 24, p2, &p1); |
| |
| auto p4 = pool->allocate<4>(28); |
| validate_ptr(4, 28, p4, &p2); |
| |
| auto p8 = pool->allocate<8>(40); |
| validate_ptr(8, 40, p8, &p4); |
| |
| auto p16 = pool->allocate<16>(64); |
| validate_ptr(16, 64, p16, &p8); |
| |
| auto p32 = pool->allocate<32>(96); |
| validate_ptr(32, 96, p32, &p16); |
| |
| // All of these allocations should be in the head block |
| REPORTER_ASSERT(r, pool->totalSize() == pool->preallocSize()); |
| SkDEBUGCODE(pool->validate();) |
| |
| // Requesting an allocation of avail() should not make a new block |
| size_t avail = pool->currentBlock()->avail<4>(); |
| auto pAvail = pool->allocate<4>(avail); |
| validate_ptr(4, avail, pAvail, &p32); |
| |
| // Remaining should be less than the alignment that was requested, and then |
| // the next allocation will make a new block |
| REPORTER_ASSERT(r, pool->currentBlock()->avail<4>() < 4); |
| auto pNextBlock = pool->allocate<4>(4); |
| validate_ptr(4, 4, pNextBlock, nullptr); |
| REPORTER_ASSERT(r, pool->totalSize() > pool->preallocSize()); |
| |
| // Allocating more than avail() makes an another block |
| size_t currentSize = pool->totalSize(); |
| size_t bigRequest = pool->currentBlock()->avail<4>() * 2; |
| auto pTooBig = pool->allocate<4>(bigRequest); |
| validate_ptr(4, bigRequest, pTooBig, nullptr); |
| REPORTER_ASSERT(r, pool->totalSize() > currentSize); |
| |
| // Allocating more than the default growth policy (1024 in this case), will fulfill the request |
| REPORTER_ASSERT(r, pool->totalSize() - currentSize < 4096); |
| currentSize = pool->totalSize(); |
| auto pReallyTooBig = pool->allocate<4>(4096); |
| validate_ptr(4, 4096, pReallyTooBig, nullptr); |
| REPORTER_ASSERT(r, pool->totalSize() >= currentSize + 4096); |
| SkDEBUGCODE(pool->validate();) |
| } |
| |
| DEF_TEST(GrBlockAllocatorResize, r) { |
| GrSBlockAllocator<1024> pool{}; |
| SkDEBUGCODE(pool->validate();) |
| |
| // Fixed resize from 16 to 32 |
| auto p = pool->allocate<4>(16); |
| REPORTER_ASSERT(r, p.fBlock->avail<4>() > 16); |
| REPORTER_ASSERT(r, p.fBlock->resize(p.fStart, p.fEnd, 16)); |
| p.fEnd += 16; |
| |
| // Subsequent allocation is 32 bytes ahead of 'p' now, and 'p' cannot be resized further. |
| auto pNext = pool->allocate<4>(16); |
| REPORTER_ASSERT(r, reinterpret_cast<uintptr_t>(pNext.fBlock->ptr(pNext.fAlignedOffset)) - |
| reinterpret_cast<uintptr_t>(pNext.fBlock->ptr(p.fAlignedOffset)) == 32); |
| REPORTER_ASSERT(r, p.fBlock == pNext.fBlock); |
| REPORTER_ASSERT(r, !p.fBlock->resize(p.fStart, p.fEnd, 48)); |
| |
| // Confirm that releasing pNext allows 'p' to be resized, and that it can be resized up to avail |
| REPORTER_ASSERT(r, p.fBlock->release(pNext.fStart, pNext.fEnd)); |
| int fillBlock = p.fBlock->avail<4>(); |
| REPORTER_ASSERT(r, p.fBlock->resize(p.fStart, p.fEnd, fillBlock)); |
| p.fEnd += fillBlock; |
| |
| // Confirm that resizing when there's not enough room fails |
| REPORTER_ASSERT(r, p.fBlock->avail<4>() < fillBlock); |
| REPORTER_ASSERT(r, !p.fBlock->resize(p.fStart, p.fEnd, fillBlock)); |
| |
| // Confirm that we can shrink 'p' back to 32 bytes and then further allocate again |
| int shrinkTo32 = p.fStart - p.fEnd + 32; |
| REPORTER_ASSERT(r, p.fBlock->resize(p.fStart, p.fEnd, shrinkTo32)); |
| p.fEnd += shrinkTo32; |
| REPORTER_ASSERT(r, p.fEnd - p.fStart == 32); |
| |
| pNext = pool->allocate<4>(16); |
| REPORTER_ASSERT(r, reinterpret_cast<uintptr_t>(pNext.fBlock->ptr(pNext.fAlignedOffset)) - |
| reinterpret_cast<uintptr_t>(pNext.fBlock->ptr(p.fAlignedOffset)) == 32); |
| SkDEBUGCODE(pool->validate();) |
| |
| // Confirm that we can't shrink past the start of the allocation, but we can shrink it to 0 |
| int shrinkTo0 = pNext.fStart - pNext.fEnd; |
| #ifndef SK_DEBUG |
| // Only test for false on release builds; a negative size should assert on debug builds |
| REPORTER_ASSERT(r, !pNext.fBlock->resize(pNext.fStart, pNext.fEnd, shrinkTo0 - 1)); |
| #endif |
| REPORTER_ASSERT(r, pNext.fBlock->resize(pNext.fStart, pNext.fEnd, shrinkTo0)); |
| } |
| |
| DEF_TEST(GrBlockAllocatorRelease, r) { |
| GrSBlockAllocator<1024> pool{}; |
| SkDEBUGCODE(pool->validate();) |
| |
| // Successful allocate and release |
| auto p = pool->allocate<8>(32); |
| REPORTER_ASSERT(r, pool->currentBlock()->release(p.fStart, p.fEnd)); |
| // Ensure the above release actually means the next allocation reuses the same space |
| auto p2 = pool->allocate<8>(32); |
| REPORTER_ASSERT(r, p.fStart == p2.fStart); |
| |
| // Confirm that 'p2' cannot be released if another allocation came after it |
| auto p3 = pool->allocate<8>(64); |
| (void) p3; |
| REPORTER_ASSERT(r, !p2.fBlock->release(p2.fStart, p2.fEnd)); |
| |
| // Confirm that 'p4' can be released if 'p5' is released first, and confirm that 'p2' and 'p3' |
| // can be released simultaneously (equivalent to 'p3' then 'p2'). |
| auto p4 = pool->allocate<8>(16); |
| auto p5 = pool->allocate<8>(96); |
| REPORTER_ASSERT(r, p5.fBlock->release(p5.fStart, p5.fEnd)); |
| REPORTER_ASSERT(r, p4.fBlock->release(p4.fStart, p4.fEnd)); |
| REPORTER_ASSERT(r, p2.fBlock->release(p2.fStart, p3.fEnd)); |
| |
| // And confirm that passing in the wrong size for the allocation fails |
| p = pool->allocate<8>(32); |
| REPORTER_ASSERT(r, !p.fBlock->release(p.fStart, p.fEnd - 16)); |
| REPORTER_ASSERT(r, !p.fBlock->release(p.fStart, p.fEnd + 16)); |
| REPORTER_ASSERT(r, p.fBlock->release(p.fStart, p.fEnd)); |
| SkDEBUGCODE(pool->validate();) |
| } |
| |
| DEF_TEST(GrBlockAllocatorRewind, r) { |
| // Confirm that a bunch of allocations and then releases in stack order fully goes back to the |
| // start of the block (i.e. unwinds the entire stack, and not just the last cursor position) |
| GrSBlockAllocator<1024> pool{}; |
| SkDEBUGCODE(pool->validate();) |
| |
| std::vector<GrBlockAllocator::ByteRange> ptrs; |
| for (int i = 0; i < 32; ++i) { |
| ptrs.push_back(pool->allocate<4>(16)); |
| } |
| |
| // Release everything in reverse order |
| SkDEBUGCODE(pool->validate();) |
| for (int i = 31; i >= 0; --i) { |
| auto br = ptrs[i]; |
| REPORTER_ASSERT(r, br.fBlock->release(br.fStart, br.fEnd)); |
| } |
| |
| // If correct, we've rewound all the way back to the start of the block, so a new allocation |
| // will have the same location as ptrs[0] |
| SkDEBUGCODE(pool->validate();) |
| REPORTER_ASSERT(r, pool->allocate<4>(16).fStart == ptrs[0].fStart); |
| } |
| |
| DEF_TEST(GrBlockAllocatorGrowthPolicy, r) { |
| static constexpr int kInitSize = 128; |
| static constexpr int kBlockCount = 5; |
| static constexpr size_t kExpectedSizes[GrBlockAllocator::kGrowthPolicyCount][kBlockCount] = { |
| // kFixed -> kInitSize per block |
| { kInitSize, kInitSize, kInitSize, kInitSize, kInitSize }, |
| // kLinear -> (block ct + 1) * kInitSize for next block |
| { kInitSize, 2 * kInitSize, 3 * kInitSize, 4 * kInitSize, 5 * kInitSize }, |
| // kFibonacci -> 1, 1, 2, 3, 5 * kInitSize for the blocks |
| { kInitSize, kInitSize, 2 * kInitSize, 3 * kInitSize, 5 * kInitSize }, |
| // kExponential -> 1, 2, 4, 8, 16 * kInitSize for the blocks |
| { kInitSize, 2 * kInitSize, 4 * kInitSize, 8 * kInitSize, 16 * kInitSize }, |
| }; |
| |
| for (int gp = 0; gp < GrBlockAllocator::kGrowthPolicyCount; ++gp) { |
| GrSBlockAllocator<kInitSize> pool{(GrowthPolicy) gp}; |
| SkDEBUGCODE(pool->validate();) |
| |
| REPORTER_ASSERT(r, kExpectedSizes[gp][0] == pool->totalSize()); |
| for (int i = 1; i < kBlockCount; ++i) { |
| REPORTER_ASSERT(r, kExpectedSizes[gp][i] == add_block(pool)); |
| } |
| |
| SkDEBUGCODE(pool->validate();) |
| } |
| } |
| |
| DEF_TEST(GrBlockAllocatorReset, r) { |
| static constexpr int kBlockIncrement = 1024; |
| |
| GrSBlockAllocator<kBlockIncrement> pool{GrowthPolicy::kLinear}; |
| SkDEBUGCODE(pool->validate();) |
| |
| void* firstAlloc = alloc_byte(pool); |
| |
| // Add several blocks |
| add_block(pool); |
| add_block(pool); |
| add_block(pool); |
| SkDEBUGCODE(pool->validate();) |
| |
| REPORTER_ASSERT(r, block_count(pool) == 4); // 3 added plus the implicit head |
| |
| get_block(pool, 0)->setMetadata(2); |
| |
| // Reset and confirm that there's only one block, a new allocation matches 'firstAlloc' again, |
| // and new blocks are sized based on a reset growth policy. |
| pool->reset(); |
| SkDEBUGCODE(pool->validate();) |
| |
| REPORTER_ASSERT(r,block_count(pool) == 1); |
| REPORTER_ASSERT(r, pool->preallocSize() == pool->totalSize()); |
| REPORTER_ASSERT(r, get_block(pool, 0)->metadata() == 0); |
| |
| REPORTER_ASSERT(r, firstAlloc == alloc_byte(pool)); |
| REPORTER_ASSERT(r, 2 * kBlockIncrement == add_block(pool)); |
| REPORTER_ASSERT(r, 3 * kBlockIncrement == add_block(pool)); |
| SkDEBUGCODE(pool->validate();) |
| } |
| |
| DEF_TEST(GrBlockAllocatorReleaseBlock, r) { |
| // This loops over all growth policies to make sure that the incremental releases update the |
| // sequence correctly for each policy. |
| for (int gp = 0; gp < GrBlockAllocator::kGrowthPolicyCount; ++gp) { |
| GrSBlockAllocator<1024> pool{(GrowthPolicy) gp}; |
| SkDEBUGCODE(pool->validate();) |
| |
| void* firstAlloc = alloc_byte(pool); |
| |
| size_t b1Size = pool->totalSize(); |
| size_t b2Size = add_block(pool); |
| size_t b3Size = add_block(pool); |
| size_t b4Size = add_block(pool); |
| SkDEBUGCODE(pool->validate();) |
| |
| get_block(pool, 0)->setMetadata(1); |
| get_block(pool, 1)->setMetadata(2); |
| get_block(pool, 2)->setMetadata(3); |
| get_block(pool, 3)->setMetadata(4); |
| |
| // Remove the 3 added blocks, but always remove the i = 1 to test intermediate removal (and |
| // on the last iteration, will test tail removal). |
| REPORTER_ASSERT(r, pool->totalSize() == b1Size + b2Size + b3Size + b4Size); |
| pool->releaseBlock(get_block(pool, 1)); |
| REPORTER_ASSERT(r, block_count(pool) == 3); |
| REPORTER_ASSERT(r, get_block(pool, 1)->metadata() == 3); |
| REPORTER_ASSERT(r, pool->totalSize() == b1Size + b3Size + b4Size); |
| |
| pool->releaseBlock(get_block(pool, 1)); |
| REPORTER_ASSERT(r, block_count(pool) == 2); |
| REPORTER_ASSERT(r, get_block(pool, 1)->metadata() == 4); |
| REPORTER_ASSERT(r, pool->totalSize() == b1Size + b4Size); |
| |
| pool->releaseBlock(get_block(pool, 1)); |
| REPORTER_ASSERT(r, block_count(pool) == 1); |
| REPORTER_ASSERT(r, pool->totalSize() == b1Size); |
| |
| // Since we're back to just the head block, if we add a new block, the growth policy should |
| // match the original sequence instead of continuing with "b5Size'" |
| size_t size = add_block(pool); |
| REPORTER_ASSERT(r, size == b2Size); |
| pool->releaseBlock(get_block(pool, 1)); |
| |
| // Explicitly release the head block and confirm it's reset |
| pool->releaseBlock(get_block(pool, 0)); |
| REPORTER_ASSERT(r, pool->totalSize() == pool->preallocSize()); |
| REPORTER_ASSERT(r, block_count(pool) == 1); |
| REPORTER_ASSERT(r, firstAlloc == alloc_byte(pool)); |
| REPORTER_ASSERT(r, get_block(pool, 0)->metadata() == 0); // metadata reset too |
| |
| // Confirm that if we have > 1 block, but release the head block we can still access the |
| // others |
| add_block(pool); |
| add_block(pool); |
| pool->releaseBlock(get_block(pool, 0)); |
| REPORTER_ASSERT(r, block_count(pool) == 3); |
| SkDEBUGCODE(pool->validate();) |
| } |
| } |
| |
| // These tests ensure that the allocation padding mechanism works as intended |
| struct TestMeta { |
| int fX1; |
| int fX2; |
| }; |
| struct alignas(32) TestMetaBig { |
| int fX1; |
| int fX2; |
| }; |
| |
| DEF_TEST(GrBlockAllocatorMetadata, r) { |
| GrSBlockAllocator<1024> pool{}; |
| SkDEBUGCODE(pool->validate();) |
| |
| // Allocation where alignment of user data > alignment of metadata |
| SkASSERT(alignof(TestMeta) < 16); |
| auto p1 = pool->allocate<16, sizeof(TestMeta)>(16); |
| SkDEBUGCODE(pool->validate();) |
| |
| REPORTER_ASSERT(r, p1.fAlignedOffset - p1.fStart >= (int) sizeof(TestMeta)); |
| TestMeta* meta = static_cast<TestMeta*>(p1.fBlock->ptr(p1.fAlignedOffset - sizeof(TestMeta))); |
| // Confirm alignment for both pointers |
| REPORTER_ASSERT(r, reinterpret_cast<uintptr_t>(meta) % alignof(TestMeta) == 0); |
| REPORTER_ASSERT(r, reinterpret_cast<uintptr_t>(p1.fBlock->ptr(p1.fAlignedOffset)) % 16 == 0); |
| // Access fields to make sure 'meta' matches compilers expectations... |
| meta->fX1 = 2; |
| meta->fX2 = 5; |
| |
| // Repeat, but for metadata that has a larger alignment than the allocation |
| SkASSERT(alignof(TestMetaBig) == 32); |
| auto p2 = pool->allocate<alignof(TestMetaBig), sizeof(TestMetaBig)>(16); |
| SkDEBUGCODE(pool->validate();) |
| |
| REPORTER_ASSERT(r, p2.fAlignedOffset - p2.fStart >= (int) sizeof(TestMetaBig)); |
| TestMetaBig* metaBig = static_cast<TestMetaBig*>( |
| p2.fBlock->ptr(p2.fAlignedOffset - sizeof(TestMetaBig))); |
| // Confirm alignment for both pointers |
| REPORTER_ASSERT(r, reinterpret_cast<uintptr_t>(metaBig) % alignof(TestMetaBig) == 0); |
| REPORTER_ASSERT(r, reinterpret_cast<uintptr_t>(p2.fBlock->ptr(p2.fAlignedOffset)) % 16 == 0); |
| // Access fields |
| metaBig->fX1 = 3; |
| metaBig->fX2 = 6; |
| |
| // Ensure metadata values persist after allocations |
| REPORTER_ASSERT(r, meta->fX1 == 2 && meta->fX2 == 5); |
| REPORTER_ASSERT(r, metaBig->fX1 == 3 && metaBig->fX2 == 6); |
| } |
| |
| template<size_t Align, size_t Padding> |
| static void run_owning_block_test(skiatest::Reporter* r, GrBlockAllocator* pool) { |
| auto br = pool->allocate<Align, Padding>(1); |
| |
| void* userPtr = br.fBlock->ptr(br.fAlignedOffset); |
| void* metaPtr = br.fBlock->ptr(br.fAlignedOffset - Padding); |
| |
| Block* block = pool->owningBlock<Align, Padding>(userPtr, br.fStart); |
| REPORTER_ASSERT(r, block == br.fBlock); |
| |
| block = pool->owningBlock<Align>(metaPtr, br.fStart); |
| REPORTER_ASSERT(r, block == br.fBlock); |
| |
| block = reinterpret_cast<Block*>(reinterpret_cast<uintptr_t>(userPtr) - br.fAlignedOffset); |
| REPORTER_ASSERT(r, block == br.fBlock); |
| } |
| |
| template<size_t Padding> |
| static void run_owning_block_tests(skiatest::Reporter* r, GrBlockAllocator* pool) { |
| run_owning_block_test<1, Padding>(r, pool); |
| run_owning_block_test<2, Padding>(r, pool); |
| run_owning_block_test<4, Padding>(r, pool); |
| run_owning_block_test<8, Padding>(r, pool); |
| run_owning_block_test<16, Padding>(r, pool); |
| run_owning_block_test<32, Padding>(r, pool); |
| run_owning_block_test<64, Padding>(r, pool); |
| run_owning_block_test<128, Padding>(r, pool); |
| } |
| |
| DEF_TEST(GrBlockAllocatorOwningBlock, r) { |
| GrSBlockAllocator<1024> pool{}; |
| SkDEBUGCODE(pool->validate();) |
| |
| run_owning_block_tests<1>(r, pool.allocator()); |
| run_owning_block_tests<2>(r, pool.allocator()); |
| run_owning_block_tests<4>(r, pool.allocator()); |
| run_owning_block_tests<8>(r, pool.allocator()); |
| run_owning_block_tests<16>(r, pool.allocator()); |
| run_owning_block_tests<32>(r, pool.allocator()); |
| |
| // And some weird numbers |
| run_owning_block_tests<3>(r, pool.allocator()); |
| run_owning_block_tests<9>(r, pool.allocator()); |
| run_owning_block_tests<17>(r, pool.allocator()); |
| } |