blob: 0ebe164c48d92e4a0846feaf3201f31eba61ec06 [file] [log] [blame]
Michael Ludwigcd019792020-03-17 10:14:48 -04001/*
2 * Copyright 2020 Google LLC
3 *
4 * Use of this source code is governed by a BSD-style license that can be
5 * found in the LICENSE file.
6 */
7
8#include "src/gpu/GrBlockAllocator.h"
9#include "tests/Test.h"
10
John Stiles7990e032020-11-06 11:22:16 -050011#include <cstring>
12
Michael Ludwigcd019792020-03-17 10:14:48 -040013using Block = GrBlockAllocator::Block;
14using GrowthPolicy = GrBlockAllocator::GrowthPolicy;
15
16// Helper functions for modifying the allocator in a controlled manner
17template<size_t N>
18static int block_count(const GrSBlockAllocator<N>& pool) {
19 int ct = 0;
20 for (const Block* b : pool->blocks()) {
21 (void) b;
22 ct++;
23 }
24 return ct;
25}
26
27template<size_t N>
28static Block* get_block(GrSBlockAllocator<N>& pool, int blockIndex) {
29 Block* found = nullptr;
30 int i = 0;
31 for (Block* b: pool->blocks()) {
32 if (i == blockIndex) {
33 found = b;
34 break;
35 }
36 i++;
37 }
38
39 SkASSERT(found != nullptr);
40 return found;
41}
42
Michael Ludwig68e5f292020-07-20 16:17:14 -040043// GrBlockAllocator holds on to the largest last-released block to reuse for new allocations,
44// and this is still counted in its totalSize(). However, it's easier to reason about size - scratch
45// in many of these tests.
46template<size_t N>
47static size_t total_size(GrSBlockAllocator<N>& pool) {
48 return pool->totalSize() - pool->testingOnly_scratchBlockSize();
49}
50
Michael Ludwigcd019792020-03-17 10:14:48 -040051template<size_t N>
52static size_t add_block(GrSBlockAllocator<N>& pool) {
Michael Ludwig68e5f292020-07-20 16:17:14 -040053 size_t currentSize = total_size(pool);
54 GrBlockAllocator::Block* current = pool->currentBlock();
55 while(pool->currentBlock() == current) {
Michael Ludwigcd019792020-03-17 10:14:48 -040056 pool->template allocate<4>(pool->preallocSize() / 2);
57 }
Michael Ludwig68e5f292020-07-20 16:17:14 -040058 return total_size(pool) - currentSize;
Michael Ludwigcd019792020-03-17 10:14:48 -040059}
60
61template<size_t N>
62static void* alloc_byte(GrSBlockAllocator<N>& pool) {
63 auto br = pool->template allocate<1>(1);
64 return br.fBlock->ptr(br.fAlignedOffset);
65}
66
67DEF_TEST(GrBlockAllocatorPreallocSize, r) {
68 // Tests stack/member initialization, option #1 described in doc
69 GrBlockAllocator stack{GrowthPolicy::kFixed, 2048};
70 SkDEBUGCODE(stack.validate();)
71
72 REPORTER_ASSERT(r, stack.preallocSize() == sizeof(GrBlockAllocator));
73 REPORTER_ASSERT(r, stack.preallocUsableSpace() == (size_t) stack.currentBlock()->avail());
74
75 // Tests placement new initialization to increase head block size, option #2
76 void* mem = operator new(1024);
77 GrBlockAllocator* placement = new (mem) GrBlockAllocator(GrowthPolicy::kLinear, 1024,
78 1024 - sizeof(GrBlockAllocator));
79 REPORTER_ASSERT(r, placement->preallocSize() == 1024);
80 REPORTER_ASSERT(r, placement->preallocUsableSpace() < 1024 &&
81 placement->preallocUsableSpace() >= (1024 - sizeof(GrBlockAllocator)));
82 delete placement;
83
84 // Tests inline increased preallocation, option #3
85 GrSBlockAllocator<2048> inlined{};
86 SkDEBUGCODE(inlined->validate();)
87 REPORTER_ASSERT(r, inlined->preallocSize() == 2048);
88 REPORTER_ASSERT(r, inlined->preallocUsableSpace() < 2048 &&
89 inlined->preallocUsableSpace() >= (2048 - sizeof(GrBlockAllocator)));
90}
91
92DEF_TEST(GrBlockAllocatorAlloc, r) {
93 GrSBlockAllocator<1024> pool{};
94 SkDEBUGCODE(pool->validate();)
95
96 // Assumes the previous pointer was in the same block
97 auto validate_ptr = [&](int align, int size,
98 GrBlockAllocator::ByteRange br,
99 GrBlockAllocator::ByteRange* prevBR) {
John Stiles96c21652020-11-05 16:17:32 +0000100 uintptr_t pt = reinterpret_cast<uintptr_t>(br.fBlock->ptr(br.fAlignedOffset));
Michael Ludwigcd019792020-03-17 10:14:48 -0400101 // Matches the requested align
102 REPORTER_ASSERT(r, pt % align == 0);
103 // And large enough
104 REPORTER_ASSERT(r, br.fEnd - br.fAlignedOffset >= size);
105 // And has enough padding for alignment
106 REPORTER_ASSERT(r, br.fAlignedOffset - br.fStart >= 0);
107 REPORTER_ASSERT(r, br.fAlignedOffset - br.fStart <= align - 1);
108 // And block of the returned struct is the current block of the allocator
109 REPORTER_ASSERT(r, pool->currentBlock() == br.fBlock);
110
111 // And make sure that we're past the required end of the previous allocation
112 if (prevBR) {
113 uintptr_t prevEnd =
114 reinterpret_cast<uintptr_t>(prevBR->fBlock->ptr(prevBR->fEnd - 1));
115 REPORTER_ASSERT(r, pt > prevEnd);
116 }
John Stiles7990e032020-11-06 11:22:16 -0500117
118 // And make sure that the entire byte range is safe to write into (excluding the dead space
119 // between "start" and "aligned offset," which is just padding and is left poisoned)
120 std::memset(br.fBlock->ptr(br.fAlignedOffset), 0xFF, br.fEnd - br.fAlignedOffset);
Michael Ludwigcd019792020-03-17 10:14:48 -0400121 };
122
123 auto p1 = pool->allocate<1>(14);
124 validate_ptr(1, 14, p1, nullptr);
125
126 auto p2 = pool->allocate<2>(24);
127 validate_ptr(2, 24, p2, &p1);
128
129 auto p4 = pool->allocate<4>(28);
130 validate_ptr(4, 28, p4, &p2);
131
132 auto p8 = pool->allocate<8>(40);
133 validate_ptr(8, 40, p8, &p4);
134
135 auto p16 = pool->allocate<16>(64);
136 validate_ptr(16, 64, p16, &p8);
137
138 auto p32 = pool->allocate<32>(96);
139 validate_ptr(32, 96, p32, &p16);
140
141 // All of these allocations should be in the head block
Michael Ludwig68e5f292020-07-20 16:17:14 -0400142 REPORTER_ASSERT(r, total_size(pool) == pool->preallocSize());
Michael Ludwigcd019792020-03-17 10:14:48 -0400143 SkDEBUGCODE(pool->validate();)
144
145 // Requesting an allocation of avail() should not make a new block
146 size_t avail = pool->currentBlock()->avail<4>();
147 auto pAvail = pool->allocate<4>(avail);
148 validate_ptr(4, avail, pAvail, &p32);
149
150 // Remaining should be less than the alignment that was requested, and then
151 // the next allocation will make a new block
152 REPORTER_ASSERT(r, pool->currentBlock()->avail<4>() < 4);
153 auto pNextBlock = pool->allocate<4>(4);
154 validate_ptr(4, 4, pNextBlock, nullptr);
Michael Ludwig68e5f292020-07-20 16:17:14 -0400155 REPORTER_ASSERT(r, total_size(pool) > pool->preallocSize());
Michael Ludwigcd019792020-03-17 10:14:48 -0400156
157 // Allocating more than avail() makes an another block
Michael Ludwig68e5f292020-07-20 16:17:14 -0400158 size_t currentSize = total_size(pool);
Michael Ludwigcd019792020-03-17 10:14:48 -0400159 size_t bigRequest = pool->currentBlock()->avail<4>() * 2;
160 auto pTooBig = pool->allocate<4>(bigRequest);
161 validate_ptr(4, bigRequest, pTooBig, nullptr);
Michael Ludwig68e5f292020-07-20 16:17:14 -0400162 REPORTER_ASSERT(r, total_size(pool) > currentSize);
Michael Ludwigcd019792020-03-17 10:14:48 -0400163
164 // Allocating more than the default growth policy (1024 in this case), will fulfill the request
Michael Ludwig68e5f292020-07-20 16:17:14 -0400165 REPORTER_ASSERT(r, total_size(pool) - currentSize < 4096);
166 currentSize = total_size(pool);
Michael Ludwigcd019792020-03-17 10:14:48 -0400167 auto pReallyTooBig = pool->allocate<4>(4096);
168 validate_ptr(4, 4096, pReallyTooBig, nullptr);
Michael Ludwig68e5f292020-07-20 16:17:14 -0400169 REPORTER_ASSERT(r, total_size(pool) >= currentSize + 4096);
Michael Ludwigcd019792020-03-17 10:14:48 -0400170 SkDEBUGCODE(pool->validate();)
171}
172
173DEF_TEST(GrBlockAllocatorResize, r) {
174 GrSBlockAllocator<1024> pool{};
175 SkDEBUGCODE(pool->validate();)
176
177 // Fixed resize from 16 to 32
John Stiles7990e032020-11-06 11:22:16 -0500178 GrBlockAllocator::ByteRange p = pool->allocate<4>(16);
Michael Ludwigcd019792020-03-17 10:14:48 -0400179 REPORTER_ASSERT(r, p.fBlock->avail<4>() > 16);
180 REPORTER_ASSERT(r, p.fBlock->resize(p.fStart, p.fEnd, 16));
181 p.fEnd += 16;
182
John Stiles7990e032020-11-06 11:22:16 -0500183 std::memset(p.fBlock->ptr(p.fAlignedOffset), 0x11, p.fEnd - p.fAlignedOffset);
184
Michael Ludwigcd019792020-03-17 10:14:48 -0400185 // Subsequent allocation is 32 bytes ahead of 'p' now, and 'p' cannot be resized further.
186 auto pNext = pool->allocate<4>(16);
187 REPORTER_ASSERT(r, reinterpret_cast<uintptr_t>(pNext.fBlock->ptr(pNext.fAlignedOffset)) -
188 reinterpret_cast<uintptr_t>(pNext.fBlock->ptr(p.fAlignedOffset)) == 32);
189 REPORTER_ASSERT(r, p.fBlock == pNext.fBlock);
190 REPORTER_ASSERT(r, !p.fBlock->resize(p.fStart, p.fEnd, 48));
191
192 // Confirm that releasing pNext allows 'p' to be resized, and that it can be resized up to avail
193 REPORTER_ASSERT(r, p.fBlock->release(pNext.fStart, pNext.fEnd));
194 int fillBlock = p.fBlock->avail<4>();
195 REPORTER_ASSERT(r, p.fBlock->resize(p.fStart, p.fEnd, fillBlock));
196 p.fEnd += fillBlock;
197
John Stiles7990e032020-11-06 11:22:16 -0500198 std::memset(p.fBlock->ptr(p.fAlignedOffset), 0x22, p.fEnd - p.fAlignedOffset);
199
Michael Ludwigcd019792020-03-17 10:14:48 -0400200 // Confirm that resizing when there's not enough room fails
201 REPORTER_ASSERT(r, p.fBlock->avail<4>() < fillBlock);
202 REPORTER_ASSERT(r, !p.fBlock->resize(p.fStart, p.fEnd, fillBlock));
203
204 // Confirm that we can shrink 'p' back to 32 bytes and then further allocate again
205 int shrinkTo32 = p.fStart - p.fEnd + 32;
206 REPORTER_ASSERT(r, p.fBlock->resize(p.fStart, p.fEnd, shrinkTo32));
207 p.fEnd += shrinkTo32;
208 REPORTER_ASSERT(r, p.fEnd - p.fStart == 32);
209
John Stiles7990e032020-11-06 11:22:16 -0500210 std::memset(p.fBlock->ptr(p.fAlignedOffset), 0x33, p.fEnd - p.fAlignedOffset);
211
Michael Ludwigcd019792020-03-17 10:14:48 -0400212 pNext = pool->allocate<4>(16);
213 REPORTER_ASSERT(r, reinterpret_cast<uintptr_t>(pNext.fBlock->ptr(pNext.fAlignedOffset)) -
214 reinterpret_cast<uintptr_t>(pNext.fBlock->ptr(p.fAlignedOffset)) == 32);
215 SkDEBUGCODE(pool->validate();)
216
217 // Confirm that we can't shrink past the start of the allocation, but we can shrink it to 0
218 int shrinkTo0 = pNext.fStart - pNext.fEnd;
219#ifndef SK_DEBUG
220 // Only test for false on release builds; a negative size should assert on debug builds
221 REPORTER_ASSERT(r, !pNext.fBlock->resize(pNext.fStart, pNext.fEnd, shrinkTo0 - 1));
222#endif
223 REPORTER_ASSERT(r, pNext.fBlock->resize(pNext.fStart, pNext.fEnd, shrinkTo0));
224}
225
226DEF_TEST(GrBlockAllocatorRelease, r) {
227 GrSBlockAllocator<1024> pool{};
228 SkDEBUGCODE(pool->validate();)
229
230 // Successful allocate and release
231 auto p = pool->allocate<8>(32);
232 REPORTER_ASSERT(r, pool->currentBlock()->release(p.fStart, p.fEnd));
233 // Ensure the above release actually means the next allocation reuses the same space
234 auto p2 = pool->allocate<8>(32);
235 REPORTER_ASSERT(r, p.fStart == p2.fStart);
236
237 // Confirm that 'p2' cannot be released if another allocation came after it
238 auto p3 = pool->allocate<8>(64);
239 (void) p3;
240 REPORTER_ASSERT(r, !p2.fBlock->release(p2.fStart, p2.fEnd));
241
242 // Confirm that 'p4' can be released if 'p5' is released first, and confirm that 'p2' and 'p3'
243 // can be released simultaneously (equivalent to 'p3' then 'p2').
244 auto p4 = pool->allocate<8>(16);
245 auto p5 = pool->allocate<8>(96);
246 REPORTER_ASSERT(r, p5.fBlock->release(p5.fStart, p5.fEnd));
247 REPORTER_ASSERT(r, p4.fBlock->release(p4.fStart, p4.fEnd));
248 REPORTER_ASSERT(r, p2.fBlock->release(p2.fStart, p3.fEnd));
249
250 // And confirm that passing in the wrong size for the allocation fails
251 p = pool->allocate<8>(32);
252 REPORTER_ASSERT(r, !p.fBlock->release(p.fStart, p.fEnd - 16));
253 REPORTER_ASSERT(r, !p.fBlock->release(p.fStart, p.fEnd + 16));
254 REPORTER_ASSERT(r, p.fBlock->release(p.fStart, p.fEnd));
255 SkDEBUGCODE(pool->validate();)
256}
257
258DEF_TEST(GrBlockAllocatorRewind, r) {
259 // Confirm that a bunch of allocations and then releases in stack order fully goes back to the
260 // start of the block (i.e. unwinds the entire stack, and not just the last cursor position)
261 GrSBlockAllocator<1024> pool{};
262 SkDEBUGCODE(pool->validate();)
263
264 std::vector<GrBlockAllocator::ByteRange> ptrs;
265 for (int i = 0; i < 32; ++i) {
266 ptrs.push_back(pool->allocate<4>(16));
267 }
268
269 // Release everything in reverse order
270 SkDEBUGCODE(pool->validate();)
271 for (int i = 31; i >= 0; --i) {
272 auto br = ptrs[i];
273 REPORTER_ASSERT(r, br.fBlock->release(br.fStart, br.fEnd));
274 }
275
276 // If correct, we've rewound all the way back to the start of the block, so a new allocation
277 // will have the same location as ptrs[0]
278 SkDEBUGCODE(pool->validate();)
279 REPORTER_ASSERT(r, pool->allocate<4>(16).fStart == ptrs[0].fStart);
280}
281
282DEF_TEST(GrBlockAllocatorGrowthPolicy, r) {
283 static constexpr int kInitSize = 128;
284 static constexpr int kBlockCount = 5;
285 static constexpr size_t kExpectedSizes[GrBlockAllocator::kGrowthPolicyCount][kBlockCount] = {
286 // kFixed -> kInitSize per block
287 { kInitSize, kInitSize, kInitSize, kInitSize, kInitSize },
288 // kLinear -> (block ct + 1) * kInitSize for next block
289 { kInitSize, 2 * kInitSize, 3 * kInitSize, 4 * kInitSize, 5 * kInitSize },
290 // kFibonacci -> 1, 1, 2, 3, 5 * kInitSize for the blocks
291 { kInitSize, kInitSize, 2 * kInitSize, 3 * kInitSize, 5 * kInitSize },
292 // kExponential -> 1, 2, 4, 8, 16 * kInitSize for the blocks
293 { kInitSize, 2 * kInitSize, 4 * kInitSize, 8 * kInitSize, 16 * kInitSize },
294 };
295
296 for (int gp = 0; gp < GrBlockAllocator::kGrowthPolicyCount; ++gp) {
297 GrSBlockAllocator<kInitSize> pool{(GrowthPolicy) gp};
298 SkDEBUGCODE(pool->validate();)
299
Michael Ludwig68e5f292020-07-20 16:17:14 -0400300 REPORTER_ASSERT(r, kExpectedSizes[gp][0] == total_size(pool));
Michael Ludwigcd019792020-03-17 10:14:48 -0400301 for (int i = 1; i < kBlockCount; ++i) {
302 REPORTER_ASSERT(r, kExpectedSizes[gp][i] == add_block(pool));
303 }
304
305 SkDEBUGCODE(pool->validate();)
306 }
307}
308
309DEF_TEST(GrBlockAllocatorReset, r) {
310 static constexpr int kBlockIncrement = 1024;
311
312 GrSBlockAllocator<kBlockIncrement> pool{GrowthPolicy::kLinear};
313 SkDEBUGCODE(pool->validate();)
314
315 void* firstAlloc = alloc_byte(pool);
316
317 // Add several blocks
318 add_block(pool);
319 add_block(pool);
320 add_block(pool);
321 SkDEBUGCODE(pool->validate();)
322
323 REPORTER_ASSERT(r, block_count(pool) == 4); // 3 added plus the implicit head
324
325 get_block(pool, 0)->setMetadata(2);
326
327 // Reset and confirm that there's only one block, a new allocation matches 'firstAlloc' again,
328 // and new blocks are sized based on a reset growth policy.
329 pool->reset();
330 SkDEBUGCODE(pool->validate();)
331
332 REPORTER_ASSERT(r,block_count(pool) == 1);
333 REPORTER_ASSERT(r, pool->preallocSize() == pool->totalSize());
334 REPORTER_ASSERT(r, get_block(pool, 0)->metadata() == 0);
335
336 REPORTER_ASSERT(r, firstAlloc == alloc_byte(pool));
337 REPORTER_ASSERT(r, 2 * kBlockIncrement == add_block(pool));
338 REPORTER_ASSERT(r, 3 * kBlockIncrement == add_block(pool));
339 SkDEBUGCODE(pool->validate();)
340}
341
342DEF_TEST(GrBlockAllocatorReleaseBlock, r) {
343 // This loops over all growth policies to make sure that the incremental releases update the
344 // sequence correctly for each policy.
345 for (int gp = 0; gp < GrBlockAllocator::kGrowthPolicyCount; ++gp) {
346 GrSBlockAllocator<1024> pool{(GrowthPolicy) gp};
347 SkDEBUGCODE(pool->validate();)
348
349 void* firstAlloc = alloc_byte(pool);
350
Michael Ludwig68e5f292020-07-20 16:17:14 -0400351 size_t b1Size = total_size(pool);
Michael Ludwigcd019792020-03-17 10:14:48 -0400352 size_t b2Size = add_block(pool);
353 size_t b3Size = add_block(pool);
354 size_t b4Size = add_block(pool);
355 SkDEBUGCODE(pool->validate();)
356
357 get_block(pool, 0)->setMetadata(1);
358 get_block(pool, 1)->setMetadata(2);
359 get_block(pool, 2)->setMetadata(3);
360 get_block(pool, 3)->setMetadata(4);
361
362 // Remove the 3 added blocks, but always remove the i = 1 to test intermediate removal (and
363 // on the last iteration, will test tail removal).
Michael Ludwig68e5f292020-07-20 16:17:14 -0400364 REPORTER_ASSERT(r, total_size(pool) == b1Size + b2Size + b3Size + b4Size);
Michael Ludwigcd019792020-03-17 10:14:48 -0400365 pool->releaseBlock(get_block(pool, 1));
366 REPORTER_ASSERT(r, block_count(pool) == 3);
367 REPORTER_ASSERT(r, get_block(pool, 1)->metadata() == 3);
Michael Ludwig68e5f292020-07-20 16:17:14 -0400368 REPORTER_ASSERT(r, total_size(pool) == b1Size + b3Size + b4Size);
Michael Ludwigcd019792020-03-17 10:14:48 -0400369
370 pool->releaseBlock(get_block(pool, 1));
371 REPORTER_ASSERT(r, block_count(pool) == 2);
372 REPORTER_ASSERT(r, get_block(pool, 1)->metadata() == 4);
Michael Ludwig68e5f292020-07-20 16:17:14 -0400373 REPORTER_ASSERT(r, total_size(pool) == b1Size + b4Size);
Michael Ludwigcd019792020-03-17 10:14:48 -0400374
375 pool->releaseBlock(get_block(pool, 1));
376 REPORTER_ASSERT(r, block_count(pool) == 1);
Michael Ludwig68e5f292020-07-20 16:17:14 -0400377 REPORTER_ASSERT(r, total_size(pool) == b1Size);
Michael Ludwigcd019792020-03-17 10:14:48 -0400378
379 // Since we're back to just the head block, if we add a new block, the growth policy should
380 // match the original sequence instead of continuing with "b5Size'"
Michael Ludwig68e5f292020-07-20 16:17:14 -0400381 pool->resetScratchSpace();
Michael Ludwigcd019792020-03-17 10:14:48 -0400382 size_t size = add_block(pool);
383 REPORTER_ASSERT(r, size == b2Size);
384 pool->releaseBlock(get_block(pool, 1));
385
386 // Explicitly release the head block and confirm it's reset
387 pool->releaseBlock(get_block(pool, 0));
Michael Ludwig68e5f292020-07-20 16:17:14 -0400388 REPORTER_ASSERT(r, total_size(pool) == pool->preallocSize());
Michael Ludwigcd019792020-03-17 10:14:48 -0400389 REPORTER_ASSERT(r, block_count(pool) == 1);
390 REPORTER_ASSERT(r, firstAlloc == alloc_byte(pool));
391 REPORTER_ASSERT(r, get_block(pool, 0)->metadata() == 0); // metadata reset too
392
393 // Confirm that if we have > 1 block, but release the head block we can still access the
394 // others
395 add_block(pool);
396 add_block(pool);
397 pool->releaseBlock(get_block(pool, 0));
398 REPORTER_ASSERT(r, block_count(pool) == 3);
399 SkDEBUGCODE(pool->validate();)
400 }
401}
402
Michael Ludwig0e15cfb2020-07-15 09:09:40 -0400403DEF_TEST(GrBlockAllocatorIterateAndRelease, r) {
404 GrSBlockAllocator<256> pool;
405
406 pool->headBlock()->setMetadata(1);
407 add_block(pool);
408 add_block(pool);
409 add_block(pool);
410
411 // Loop forward and release the blocks
412 int releaseCount = 0;
413 for (auto* b : pool->blocks()) {
414 pool->releaseBlock(b);
415 releaseCount++;
416 }
417 REPORTER_ASSERT(r, releaseCount == 4);
418 // pool should have just the head block, but was reset
419 REPORTER_ASSERT(r, pool->headBlock()->metadata() == 0);
420 REPORTER_ASSERT(r, block_count(pool) == 1);
421
422 // Add more blocks
423 pool->headBlock()->setMetadata(1);
424 add_block(pool);
425 add_block(pool);
426 add_block(pool);
427
428 // Loop in reverse and release the blocks
429 releaseCount = 0;
430 for (auto* b : pool->rblocks()) {
431 pool->releaseBlock(b);
432 releaseCount++;
433 }
434 REPORTER_ASSERT(r, releaseCount == 4);
435 // pool should have just the head block, but was reset
436 REPORTER_ASSERT(r, pool->headBlock()->metadata() == 0);
437 REPORTER_ASSERT(r, block_count(pool) == 1);
438}
439
Michael Ludwig68e5f292020-07-20 16:17:14 -0400440DEF_TEST(GrBlockAllocatorScratchBlockReserve, r) {
441 GrSBlockAllocator<256> pool;
442
443 size_t added = add_block(pool);
444 REPORTER_ASSERT(r, pool->testingOnly_scratchBlockSize() == 0);
445 size_t total = pool->totalSize();
446 pool->releaseBlock(pool->currentBlock());
447
448 // Total size shouldn't have changed, the released block should become scratch
449 REPORTER_ASSERT(r, pool->totalSize() == total);
450 REPORTER_ASSERT(r, (size_t) pool->testingOnly_scratchBlockSize() == added);
451
452 // But a reset definitely deletes any scratch block
453 pool->reset();
454 REPORTER_ASSERT(r, pool->testingOnly_scratchBlockSize() == 0);
455
456 // Reserving more than what's available adds a scratch block, and current block remains avail.
457 size_t avail = pool->currentBlock()->avail();
458 size_t reserve = avail + 1;
459 pool->reserve(reserve);
460 REPORTER_ASSERT(r, (size_t) pool->currentBlock()->avail() == avail);
461 // And rounds up to the fixed size of this pool's growth policy
462 REPORTER_ASSERT(r, (size_t) pool->testingOnly_scratchBlockSize() >= reserve &&
463 pool->testingOnly_scratchBlockSize() % 256 == 0);
464
465 // Allocating more than avail activates the scratch block (so totalSize doesn't change)
466 size_t preAllocTotalSize = pool->totalSize();
467 pool->allocate<1>(avail + 1);
468 REPORTER_ASSERT(r, (size_t) pool->testingOnly_scratchBlockSize() == 0);
469 REPORTER_ASSERT(r, pool->totalSize() == preAllocTotalSize);
470
471 // When reserving less than what's still available in the current block, no scratch block is
472 // added.
473 pool->reserve(pool->currentBlock()->avail());
474 REPORTER_ASSERT(r, pool->testingOnly_scratchBlockSize() == 0);
475
476 // Unless checking available bytes is disabled
477 pool->reserve(pool->currentBlock()->avail(), GrBlockAllocator::kIgnoreExistingBytes_Flag);
478 REPORTER_ASSERT(r, pool->testingOnly_scratchBlockSize() > 0);
479
480 // If kIgnoreGrowthPolicy is specified, the new scratch block should not have been updated to
481 // follow the size (which in this case is a fixed 256 bytes per block).
482 pool->resetScratchSpace();
483 pool->reserve(32, GrBlockAllocator::kIgnoreGrowthPolicy_Flag);
484 REPORTER_ASSERT(r, pool->testingOnly_scratchBlockSize() > 0 &&
485 pool->testingOnly_scratchBlockSize() < 256);
486
487 // When requesting an allocation larger than the current block and the scratch block, a new
488 // block is added, and the scratch block remains scratch.
489 GrBlockAllocator::Block* oldTail = pool->currentBlock();
490 avail = oldTail->avail();
491 size_t scratchAvail = 2 * avail;
492 pool->reserve(scratchAvail);
Leon Scroggins III982fff22020-07-31 14:09:06 -0400493 REPORTER_ASSERT(r, (size_t) pool->testingOnly_scratchBlockSize() >= scratchAvail);
Michael Ludwig68e5f292020-07-20 16:17:14 -0400494
495 // This allocation request is higher than oldTail's available, and the scratch size so we
496 // should add a new block and scratch size should stay the same.
497 scratchAvail = pool->testingOnly_scratchBlockSize();
498 pool->allocate<1>(scratchAvail + 1);
499 REPORTER_ASSERT(r, pool->currentBlock() != oldTail);
500 REPORTER_ASSERT(r, (size_t) pool->testingOnly_scratchBlockSize() == scratchAvail);
501}
502
Michael Ludwig1d4c08f2020-07-21 13:04:42 -0400503DEF_TEST(GrBlockAllocatorStealBlocks, r) {
504 GrSBlockAllocator<256> poolA;
505 GrSBlockAllocator<128> poolB;
506
507 add_block(poolA);
508 add_block(poolA);
509 add_block(poolA);
510
511 add_block(poolB);
512 add_block(poolB);
513
514 char* bAlloc = (char*) alloc_byte(poolB);
515 *bAlloc = 't';
516
517 const GrBlockAllocator::Block* allocOwner = poolB->findOwningBlock(bAlloc);
518
519 REPORTER_ASSERT(r, block_count(poolA) == 4);
520 REPORTER_ASSERT(r, block_count(poolB) == 3);
521
522 size_t aSize = poolA->totalSize();
523 size_t bSize = poolB->totalSize();
524 size_t theftSize = bSize - poolB->preallocSize();
525
526 // This steal should move B's 2 heap blocks to A, bringing A to 6 and B to just its head
527 poolA->stealHeapBlocks(poolB.allocator());
528 REPORTER_ASSERT(r, block_count(poolA) == 6);
529 REPORTER_ASSERT(r, block_count(poolB) == 1);
530 REPORTER_ASSERT(r, poolB->preallocSize() == poolB->totalSize());
531 REPORTER_ASSERT(r, poolA->totalSize() == aSize + theftSize);
532
533 REPORTER_ASSERT(r, *bAlloc == 't');
534 REPORTER_ASSERT(r, (uintptr_t) poolA->findOwningBlock(bAlloc) == (uintptr_t) allocOwner);
535 REPORTER_ASSERT(r, !poolB->findOwningBlock(bAlloc));
536
537 // Redoing the steal now that B is just a head block should be a no-op
538 poolA->stealHeapBlocks(poolB.allocator());
539 REPORTER_ASSERT(r, block_count(poolA) == 6);
540 REPORTER_ASSERT(r, block_count(poolB) == 1);
541}
542
Michael Ludwigcd019792020-03-17 10:14:48 -0400543// These tests ensure that the allocation padding mechanism works as intended
544struct TestMeta {
545 int fX1;
546 int fX2;
547};
548struct alignas(32) TestMetaBig {
549 int fX1;
550 int fX2;
551};
552
553DEF_TEST(GrBlockAllocatorMetadata, r) {
554 GrSBlockAllocator<1024> pool{};
555 SkDEBUGCODE(pool->validate();)
556
557 // Allocation where alignment of user data > alignment of metadata
558 SkASSERT(alignof(TestMeta) < 16);
559 auto p1 = pool->allocate<16, sizeof(TestMeta)>(16);
560 SkDEBUGCODE(pool->validate();)
561
562 REPORTER_ASSERT(r, p1.fAlignedOffset - p1.fStart >= (int) sizeof(TestMeta));
563 TestMeta* meta = static_cast<TestMeta*>(p1.fBlock->ptr(p1.fAlignedOffset - sizeof(TestMeta)));
564 // Confirm alignment for both pointers
565 REPORTER_ASSERT(r, reinterpret_cast<uintptr_t>(meta) % alignof(TestMeta) == 0);
566 REPORTER_ASSERT(r, reinterpret_cast<uintptr_t>(p1.fBlock->ptr(p1.fAlignedOffset)) % 16 == 0);
567 // Access fields to make sure 'meta' matches compilers expectations...
568 meta->fX1 = 2;
569 meta->fX2 = 5;
570
571 // Repeat, but for metadata that has a larger alignment than the allocation
572 SkASSERT(alignof(TestMetaBig) == 32);
573 auto p2 = pool->allocate<alignof(TestMetaBig), sizeof(TestMetaBig)>(16);
574 SkDEBUGCODE(pool->validate();)
575
576 REPORTER_ASSERT(r, p2.fAlignedOffset - p2.fStart >= (int) sizeof(TestMetaBig));
577 TestMetaBig* metaBig = static_cast<TestMetaBig*>(
578 p2.fBlock->ptr(p2.fAlignedOffset - sizeof(TestMetaBig)));
579 // Confirm alignment for both pointers
580 REPORTER_ASSERT(r, reinterpret_cast<uintptr_t>(metaBig) % alignof(TestMetaBig) == 0);
581 REPORTER_ASSERT(r, reinterpret_cast<uintptr_t>(p2.fBlock->ptr(p2.fAlignedOffset)) % 16 == 0);
582 // Access fields
583 metaBig->fX1 = 3;
584 metaBig->fX2 = 6;
585
586 // Ensure metadata values persist after allocations
587 REPORTER_ASSERT(r, meta->fX1 == 2 && meta->fX2 == 5);
588 REPORTER_ASSERT(r, metaBig->fX1 == 3 && metaBig->fX2 == 6);
589}
590
Michael Ludwigc97ebe02020-07-17 08:39:46 -0400591DEF_TEST(GrBlockAllocatorAllocatorMetadata, r) {
592 GrSBlockAllocator<256> pool{};
593 SkDEBUGCODE(pool->validate();)
594
595 REPORTER_ASSERT(r, pool->metadata() == 0); // initial value
596
597 pool->setMetadata(4);
598 REPORTER_ASSERT(r, pool->metadata() == 4);
599
600 // Releasing the head block doesn't change the allocator's metadata (even though that's where
601 // it is stored).
602 pool->releaseBlock(pool->headBlock());
603 REPORTER_ASSERT(r, pool->metadata() == 4);
604
605 // But resetting the whole allocator brings things back to as if it were newly constructed
606 pool->reset();
607 REPORTER_ASSERT(r, pool->metadata() == 0);
608}
609
Michael Ludwigcd019792020-03-17 10:14:48 -0400610template<size_t Align, size_t Padding>
611static void run_owning_block_test(skiatest::Reporter* r, GrBlockAllocator* pool) {
612 auto br = pool->allocate<Align, Padding>(1);
613
614 void* userPtr = br.fBlock->ptr(br.fAlignedOffset);
615 void* metaPtr = br.fBlock->ptr(br.fAlignedOffset - Padding);
616
617 Block* block = pool->owningBlock<Align, Padding>(userPtr, br.fStart);
618 REPORTER_ASSERT(r, block == br.fBlock);
619
620 block = pool->owningBlock<Align>(metaPtr, br.fStart);
621 REPORTER_ASSERT(r, block == br.fBlock);
622
623 block = reinterpret_cast<Block*>(reinterpret_cast<uintptr_t>(userPtr) - br.fAlignedOffset);
624 REPORTER_ASSERT(r, block == br.fBlock);
625}
626
627template<size_t Padding>
628static void run_owning_block_tests(skiatest::Reporter* r, GrBlockAllocator* pool) {
629 run_owning_block_test<1, Padding>(r, pool);
630 run_owning_block_test<2, Padding>(r, pool);
631 run_owning_block_test<4, Padding>(r, pool);
632 run_owning_block_test<8, Padding>(r, pool);
633 run_owning_block_test<16, Padding>(r, pool);
634 run_owning_block_test<32, Padding>(r, pool);
635 run_owning_block_test<64, Padding>(r, pool);
636 run_owning_block_test<128, Padding>(r, pool);
637}
638
639DEF_TEST(GrBlockAllocatorOwningBlock, r) {
640 GrSBlockAllocator<1024> pool{};
641 SkDEBUGCODE(pool->validate();)
642
643 run_owning_block_tests<1>(r, pool.allocator());
644 run_owning_block_tests<2>(r, pool.allocator());
645 run_owning_block_tests<4>(r, pool.allocator());
646 run_owning_block_tests<8>(r, pool.allocator());
647 run_owning_block_tests<16>(r, pool.allocator());
648 run_owning_block_tests<32>(r, pool.allocator());
649
650 // And some weird numbers
651 run_owning_block_tests<3>(r, pool.allocator());
652 run_owning_block_tests<9>(r, pool.allocator());
653 run_owning_block_tests<17>(r, pool.allocator());
654}