| /* |
| * Copyright 2017 Google Inc. |
| * |
| * Use of this source code is governed by a BSD-style license that can be |
| * found in the LICENSE file. |
| */ |
| |
| #include "include/core/SkCanvas.h" |
| #include "include/core/SkSpan.h" |
| #include "include/core/SkSurface.h" |
| #include "include/gpu/GrDirectContext.h" |
| #include "src/gpu/GrDirectContextPriv.h" |
| #include "src/gpu/GrGpu.h" |
| #include "src/gpu/GrProxyProvider.h" |
| #include "src/gpu/GrResourceAllocator.h" |
| #include "src/gpu/GrResourceProviderPriv.h" |
| #include "src/gpu/GrSurfaceProxyPriv.h" |
| #include "src/gpu/GrTexture.h" |
| #include "src/gpu/GrTextureProxy.h" |
| #include "tests/Test.h" |
| #include "tools/gpu/ManagedBackendTexture.h" |
| |
| namespace { |
| struct ProxyParams { |
| int fSize; |
| GrRenderable fRenderable; |
| GrColorType fColorType; |
| SkBackingFit fFit; |
| int fSampleCnt; |
| SkBudgeted fBudgeted; |
| enum Kind { |
| kDeferred, |
| kBackend, |
| kFullyLazy, |
| kLazy, |
| kInstantiated |
| }; |
| Kind fKind; |
| GrUniqueKey fUniqueKey = GrUniqueKey(); |
| // TODO: do we care about mipmapping |
| }; |
| |
| constexpr GrRenderable kRT = GrRenderable::kYes; |
| constexpr GrRenderable kNotRT = GrRenderable::kNo; |
| |
| constexpr GrColorType kRGBA = GrColorType::kRGBA_8888; |
| constexpr GrColorType kAlpha = GrColorType::kAlpha_8; |
| |
| constexpr SkBackingFit kE = SkBackingFit::kExact; |
| constexpr SkBackingFit kA = SkBackingFit::kApprox; |
| |
| constexpr SkBudgeted kNotB = SkBudgeted::kNo; |
| constexpr SkBudgeted kB = SkBudgeted::kYes; |
| |
| constexpr ProxyParams::Kind kDeferred = ProxyParams::Kind::kDeferred; |
| constexpr ProxyParams::Kind kBackend = ProxyParams::Kind::kBackend; |
| constexpr ProxyParams::Kind kInstantiated = ProxyParams::Kind::kInstantiated; |
| constexpr ProxyParams::Kind kLazy = ProxyParams::Kind::kLazy; |
| constexpr ProxyParams::Kind kFullyLazy = ProxyParams::Kind::kFullyLazy; |
| }; |
| |
| static sk_sp<GrSurfaceProxy> make_deferred(GrProxyProvider* proxyProvider, const GrCaps* caps, |
| const ProxyParams& p) { |
| const GrBackendFormat format = caps->getDefaultBackendFormat(p.fColorType, p.fRenderable); |
| return proxyProvider->createProxy(format, {p.fSize, p.fSize}, p.fRenderable, p.fSampleCnt, |
| GrMipmapped::kNo, p.fFit, p.fBudgeted, GrProtected::kNo); |
| } |
| |
| static sk_sp<GrSurfaceProxy> make_backend(GrDirectContext* dContext, const ProxyParams& p) { |
| GrProxyProvider* proxyProvider = dContext->priv().proxyProvider(); |
| |
| SkColorType skColorType = GrColorTypeToSkColorType(p.fColorType); |
| SkASSERT(SkColorType::kUnknown_SkColorType != skColorType); |
| |
| auto mbet = sk_gpu_test::ManagedBackendTexture::MakeWithoutData( |
| dContext, p.fSize, p.fSize, skColorType, GrMipmapped::kNo, GrRenderable::kNo); |
| |
| if (!mbet) { |
| return nullptr; |
| } |
| |
| return proxyProvider->wrapBackendTexture(mbet->texture(), |
| kBorrow_GrWrapOwnership, |
| GrWrapCacheable::kNo, |
| kRead_GrIOType, |
| mbet->refCountedCallback()); |
| } |
| |
| static sk_sp<GrSurfaceProxy> make_fully_lazy(GrProxyProvider* proxyProvider, const GrCaps* caps, |
| const ProxyParams& p) { |
| const GrBackendFormat format = caps->getDefaultBackendFormat(p.fColorType, p.fRenderable); |
| auto cb = [p](GrResourceProvider* provider, const GrSurfaceProxy::LazySurfaceDesc& desc) { |
| auto tex = provider->createTexture({p.fSize, p.fSize}, desc.fFormat, |
| desc.fTextureType, |
| desc.fRenderable, desc.fSampleCnt, |
| desc.fMipmapped, desc.fBudgeted, |
| desc.fProtected); |
| return GrSurfaceProxy::LazyCallbackResult(std::move(tex)); |
| }; |
| return GrProxyProvider::MakeFullyLazyProxy(std::move(cb), format, p.fRenderable, p.fSampleCnt, |
| GrProtected::kNo, *caps, |
| GrSurfaceProxy::UseAllocator::kYes); |
| } |
| |
| static sk_sp<GrSurfaceProxy> make_lazy(GrProxyProvider* proxyProvider, const GrCaps* caps, |
| const ProxyParams& p) { |
| const GrBackendFormat format = caps->getDefaultBackendFormat(p.fColorType, p.fRenderable); |
| auto cb = [](GrResourceProvider* provider, const GrSurfaceProxy::LazySurfaceDesc& desc) { |
| auto tex = provider->createTexture(desc.fDimensions, desc.fFormat, |
| desc.fTextureType, |
| desc.fRenderable, desc.fSampleCnt, |
| desc.fMipmapped, desc.fBudgeted, |
| desc.fProtected); |
| return GrSurfaceProxy::LazyCallbackResult(std::move(tex)); |
| }; |
| return proxyProvider->createLazyProxy(std::move(cb), format, {p.fSize, p.fSize}, |
| GrMipmapped::kNo, GrMipmapStatus::kNotAllocated, |
| GrInternalSurfaceFlags::kNone, |
| p.fFit, p.fBudgeted, GrProtected::kNo, |
| GrSurfaceProxy::UseAllocator::kYes); |
| } |
| |
| static sk_sp<GrSurfaceProxy> make_proxy(GrDirectContext* dContext, const ProxyParams& p) { |
| GrProxyProvider* proxyProvider = dContext->priv().proxyProvider(); |
| const GrCaps* caps = dContext->priv().caps(); |
| sk_sp<GrSurfaceProxy> proxy; |
| switch (p.fKind) { |
| case ProxyParams::kDeferred: |
| proxy = make_deferred(proxyProvider, caps, p); |
| break; |
| case ProxyParams::kBackend: |
| proxy = make_backend(dContext, p); |
| break; |
| case ProxyParams::kFullyLazy: |
| proxy = make_fully_lazy(proxyProvider, caps, p); |
| break; |
| case ProxyParams::kLazy: |
| proxy = make_lazy(proxyProvider, caps, p); |
| break; |
| case ProxyParams::kInstantiated: |
| proxy = make_deferred(proxyProvider, caps, p); |
| if (proxy) { |
| auto surf = proxy->priv().createSurface(dContext->priv().resourceProvider()); |
| proxy->priv().assign(std::move(surf)); |
| } |
| break; |
| } |
| if (proxy && p.fUniqueKey.isValid()) { |
| SkASSERT(proxy->asTextureProxy()); |
| proxyProvider->assignUniqueKeyToProxy(p.fUniqueKey, proxy->asTextureProxy()); |
| } |
| return proxy; |
| } |
| |
| // Basic test that two proxies with overlapping intervals and compatible descriptors are |
| // assigned different GrSurfaces. |
| static void overlap_test(skiatest::Reporter* reporter, GrDirectContext* dContext, |
| sk_sp<GrSurfaceProxy> p1, sk_sp<GrSurfaceProxy> p2, |
| bool expectedResult) { |
| GrResourceAllocator alloc(dContext); |
| |
| alloc.addInterval(p1.get(), 0, 4, GrResourceAllocator::ActualUse::kYes); |
| alloc.incOps(); |
| alloc.addInterval(p2.get(), 1, 2, GrResourceAllocator::ActualUse::kYes); |
| alloc.incOps(); |
| |
| REPORTER_ASSERT(reporter, alloc.planAssignment()); |
| REPORTER_ASSERT(reporter, alloc.makeBudgetHeadroom()); |
| REPORTER_ASSERT(reporter, alloc.assign()); |
| |
| REPORTER_ASSERT(reporter, p1->peekSurface()); |
| REPORTER_ASSERT(reporter, p2->peekSurface()); |
| bool doTheBackingStoresMatch = p1->underlyingUniqueID() == p2->underlyingUniqueID(); |
| REPORTER_ASSERT(reporter, expectedResult == doTheBackingStoresMatch); |
| } |
| |
| // Test various cases when two proxies do not have overlapping intervals. |
| // This mainly acts as a test of the ResourceAllocator's free pool. |
| static void non_overlap_test(skiatest::Reporter* reporter, GrDirectContext* dContext, |
| sk_sp<GrSurfaceProxy> p1, sk_sp<GrSurfaceProxy> p2, |
| bool expectedResult) { |
| GrResourceAllocator alloc(dContext); |
| |
| alloc.incOps(); |
| alloc.incOps(); |
| alloc.incOps(); |
| alloc.incOps(); |
| alloc.incOps(); |
| alloc.incOps(); |
| |
| alloc.addInterval(p1.get(), 0, 2, GrResourceAllocator::ActualUse::kYes); |
| alloc.addInterval(p2.get(), 3, 5, GrResourceAllocator::ActualUse::kYes); |
| |
| REPORTER_ASSERT(reporter, alloc.planAssignment()); |
| REPORTER_ASSERT(reporter, alloc.makeBudgetHeadroom()); |
| REPORTER_ASSERT(reporter, alloc.assign()); |
| |
| REPORTER_ASSERT(reporter, p1->peekSurface()); |
| REPORTER_ASSERT(reporter, p2->peekSurface()); |
| bool doTheBackingStoresMatch = p1->underlyingUniqueID() == p2->underlyingUniqueID(); |
| REPORTER_ASSERT(reporter, expectedResult == doTheBackingStoresMatch); |
| } |
| |
| DEF_GPUTEST_FOR_RENDERING_CONTEXTS(ResourceAllocatorTest, reporter, ctxInfo) { |
| auto dContext = ctxInfo.directContext(); |
| const GrCaps* caps = dContext->priv().caps(); |
| |
| struct TestCase { |
| ProxyParams fP1; |
| ProxyParams fP2; |
| bool fExpectation; |
| }; |
| |
| constexpr bool kShare = true; |
| constexpr bool kDontShare = false; |
| |
| // Non-RT GrSurfaces are never recycled on some platforms. |
| bool kConditionallyShare = caps->reuseScratchTextures(); |
| |
| static const TestCase overlappingTests[] = { |
| // Two proxies with overlapping intervals and compatible descriptors should never share |
| // RT version |
| {{64, kRT, kRGBA, kA, 1, kNotB, kDeferred}, |
| {64, kRT, kRGBA, kA, 1, kNotB, kDeferred}, |
| kDontShare}, |
| // non-RT version |
| {{64, kNotRT, kRGBA, kA, 1, kNotB, kDeferred}, |
| {64, kNotRT, kRGBA, kA, 1, kNotB, kDeferred}, |
| kDontShare}, |
| }; |
| |
| for (size_t i = 0; i < SK_ARRAY_COUNT(overlappingTests); i++) { |
| const TestCase& test = overlappingTests[i]; |
| sk_sp<GrSurfaceProxy> p1 = make_proxy(dContext, test.fP1); |
| sk_sp<GrSurfaceProxy> p2 = make_proxy(dContext, test.fP2); |
| reporter->push(SkStringPrintf("case %d", SkToInt(i))); |
| overlap_test(reporter, dContext, std::move(p1), std::move(p2), test.fExpectation); |
| reporter->pop(); |
| } |
| |
| auto beFormat = caps->getDefaultBackendFormat(GrColorType::kRGBA_8888, GrRenderable::kYes); |
| int k2 = caps->getRenderTargetSampleCount(2, beFormat); |
| int k4 = caps->getRenderTargetSampleCount(4, beFormat); |
| |
| static const TestCase nonOverlappingTests[] = { |
| // Two non-overlapping intervals w/ compatible proxies should share |
| // both same size & approx |
| {{64, kRT, kRGBA, kA, 1, kNotB, kDeferred}, |
| {64, kRT, kRGBA, kA, 1, kNotB, kDeferred}, |
| kShare}, |
| {{64, kNotRT, kRGBA, kA, 1, kNotB, kDeferred}, |
| {64, kNotRT, kRGBA, kA, 1, kNotB, kDeferred}, |
| kConditionallyShare}, |
| // diffs sizes but still approx |
| {{64, kRT, kRGBA, kA, 1, kNotB, kDeferred}, |
| {50, kRT, kRGBA, kA, 1, kNotB, kDeferred}, |
| kShare}, |
| {{64, kNotRT, kRGBA, kA, 1, kNotB, kDeferred}, |
| {50, kNotRT, kRGBA, kA, 1, kNotB, kDeferred}, |
| kConditionallyShare}, |
| // sames sizes but exact |
| {{64, kRT, kRGBA, kE, 1, kNotB, kDeferred}, |
| {64, kRT, kRGBA, kE, 1, kNotB, kDeferred}, |
| kShare}, |
| {{64, kNotRT, kRGBA, kE, 1, kNotB, kDeferred}, |
| {64, kNotRT, kRGBA, kE, 1, kNotB, kDeferred}, |
| kConditionallyShare}, |
| // Two non-overlapping intervals w/ different exact sizes should not share |
| {{56, kRT, kRGBA, kE, 1, kNotB, kDeferred}, |
| {54, kRT, kRGBA, kE, 1, kNotB, kDeferred}, |
| kDontShare}, |
| // Two non-overlapping intervals w/ _very different_ approx sizes should not share |
| {{255, kRT, kRGBA, kA, 1, kNotB, kDeferred}, |
| {127, kRT, kRGBA, kA, 1, kNotB, kDeferred}, |
| kDontShare}, |
| // Two non-overlapping intervals w/ different MSAA sample counts should not share |
| {{64, kRT, kRGBA, kA, k2, kNotB, kDeferred}, |
| {64, kRT, kRGBA, kA, k4, kNotB, kDeferred}, |
| k2 == k4}, |
| // Two non-overlapping intervals w/ different configs should not share |
| {{64, kRT, kRGBA, kA, 1, kNotB, kDeferred}, |
| {64, kRT, kAlpha, kA, 1, kNotB, kDeferred}, |
| kDontShare}, |
| // Two non-overlapping intervals w/ different RT classifications should never share |
| {{64, kRT, kRGBA, kA, 1, kNotB, kDeferred}, |
| {64, kNotRT, kRGBA, kA, 1, kNotB, kDeferred}, |
| kDontShare}, |
| {{64, kNotRT, kRGBA, kA, 1, kNotB, kDeferred}, |
| {64, kRT, kRGBA, kA, 1, kNotB, kDeferred}, |
| kDontShare}, |
| // Two non-overlapping intervals w/ different origins should share |
| {{64, kRT, kRGBA, kA, 1, kNotB, kDeferred}, |
| {64, kRT, kRGBA, kA, 1, kNotB, kDeferred}, |
| kShare}, |
| // Wrapped backend textures should never be reused |
| {{64, kNotRT, kRGBA, kE, 1, kNotB, kBackend}, |
| {64, kNotRT, kRGBA, kE, 1, kNotB, kDeferred}, |
| kDontShare} |
| }; |
| |
| for (size_t i = 0; i < SK_ARRAY_COUNT(nonOverlappingTests); i++) { |
| const TestCase& test = nonOverlappingTests[i]; |
| sk_sp<GrSurfaceProxy> p1 = make_proxy(dContext, test.fP1); |
| sk_sp<GrSurfaceProxy> p2 = make_proxy(dContext, test.fP2); |
| |
| if (!p1 || !p2) { |
| continue; // creation can fail (e.g., for msaa4 on iOS) |
| } |
| |
| reporter->push(SkStringPrintf("case %d", SkToInt(i))); |
| non_overlap_test(reporter, dContext, std::move(p1), std::move(p2), |
| test.fExpectation); |
| reporter->pop(); |
| } |
| } |
| |
| static void draw(GrRecordingContext* rContext) { |
| SkImageInfo ii = SkImageInfo::Make(1024, 1024, kRGBA_8888_SkColorType, kPremul_SkAlphaType); |
| |
| sk_sp<SkSurface> s = SkSurface::MakeRenderTarget(rContext, SkBudgeted::kYes, |
| ii, 1, kTopLeft_GrSurfaceOrigin, nullptr); |
| |
| SkCanvas* c = s->getCanvas(); |
| |
| c->clear(SK_ColorBLACK); |
| } |
| |
| |
| DEF_GPUTEST_FOR_RENDERING_CONTEXTS(ResourceAllocatorStressTest, reporter, ctxInfo) { |
| auto context = ctxInfo.directContext(); |
| |
| size_t maxBytes = context->getResourceCacheLimit(); |
| |
| context->setResourceCacheLimit(0); // We'll always be overbudget |
| |
| draw(context); |
| draw(context); |
| draw(context); |
| draw(context); |
| context->flushAndSubmit(); |
| |
| context->setResourceCacheLimit(maxBytes); |
| } |
| |
| struct Interval { |
| ProxyParams fParams; |
| int fStart; |
| int fEnd; |
| sk_sp<GrSurfaceProxy> fProxy = nullptr; |
| }; |
| |
| struct TestCase { |
| const char * fName; |
| bool fShouldFit; |
| size_t fBudget; |
| SkTArray<ProxyParams> fPurgeableResourcesInCache = {}; |
| SkTArray<ProxyParams> fUnpurgeableResourcesInCache = {}; |
| SkTArray<Interval> fIntervals; |
| }; |
| |
| static void memory_budget_test(skiatest::Reporter* reporter, |
| GrDirectContext* dContext, |
| const TestCase& test) { |
| // Reset cache. |
| auto cache = dContext->priv().getResourceCache(); |
| cache->releaseAll(); |
| cache->setLimit(test.fBudget); |
| |
| // Add purgeable entries. |
| size_t expectedPurgeableBytes = 0; |
| SkTArray<sk_sp<GrSurface>> purgeableSurfaces; |
| for (auto& params : test.fPurgeableResourcesInCache) { |
| SkASSERT(params.fKind == kInstantiated); |
| sk_sp<GrSurfaceProxy> proxy = make_proxy(dContext, params); |
| REPORTER_ASSERT(reporter, proxy->peekSurface()); |
| expectedPurgeableBytes += proxy->gpuMemorySize(); |
| purgeableSurfaces.push_back(sk_ref_sp(proxy->peekSurface())); |
| } |
| purgeableSurfaces.reset(); |
| REPORTER_ASSERT(reporter, expectedPurgeableBytes == cache->getPurgeableBytes(), |
| "%zu", cache->getPurgeableBytes()); |
| |
| // Add unpurgeable entries. |
| size_t expectedUnpurgeableBytes = 0; |
| SkTArray<sk_sp<GrSurface>> unpurgeableSurfaces; |
| for (auto& params : test.fUnpurgeableResourcesInCache) { |
| SkASSERT(params.fKind == kInstantiated); |
| sk_sp<GrSurfaceProxy> proxy = make_proxy(dContext, params); |
| REPORTER_ASSERT(reporter, proxy->peekSurface()); |
| expectedUnpurgeableBytes += proxy->gpuMemorySize(); |
| unpurgeableSurfaces.push_back(sk_ref_sp(proxy->peekSurface())); |
| } |
| |
| auto unpurgeableBytes = cache->getBudgetedResourceBytes() - cache->getPurgeableBytes(); |
| REPORTER_ASSERT(reporter, expectedUnpurgeableBytes == unpurgeableBytes, |
| "%zu", unpurgeableBytes); |
| |
| // Add intervals and test. |
| GrResourceAllocator alloc(dContext); |
| for (auto& interval : test.fIntervals) { |
| for (int i = interval.fStart; i <= interval.fEnd; i++) { |
| alloc.incOps(); |
| } |
| alloc.addInterval(interval.fProxy.get(), interval.fStart, interval.fEnd, |
| GrResourceAllocator::ActualUse::kYes); |
| } |
| REPORTER_ASSERT(reporter, alloc.planAssignment()); |
| REPORTER_ASSERT(reporter, alloc.makeBudgetHeadroom() == test.fShouldFit); |
| } |
| |
| DEF_GPUTEST_FOR_RENDERING_CONTEXTS(ResourceAllocatorMemoryBudgetTest, reporter, ctxInfo) { |
| auto dContext = ctxInfo.directContext(); |
| |
| constexpr bool kUnder = true; |
| constexpr bool kOver = false; |
| constexpr size_t kRGBA64Bytes = 4 * 64 * 64; |
| const ProxyParams kProxy64 = {64, kRT, kRGBA, kE, 1, kB, kDeferred}; |
| const ProxyParams kProxy64NotBudgeted = {64, kRT, kRGBA, kE, 1, kNotB, kDeferred}; |
| const ProxyParams kProxy64Lazy = {64, kRT, kRGBA, kE, 1, kB, kLazy}; |
| const ProxyParams kProxy64FullyLazy = {64, kRT, kRGBA, kE, 1, kB, kFullyLazy}; |
| const ProxyParams kProxy32Instantiated = {32, kRT, kRGBA, kE, 1, kB, kInstantiated}; |
| const ProxyParams kProxy64Instantiated = {64, kRT, kRGBA, kE, 1, kB, kInstantiated}; |
| |
| TestCase tests[] = { |
| {"empty DAG", kUnder, 0, {}, {}, {}}, |
| {"unbudgeted", kUnder, 0, {}, {}, {{kProxy64NotBudgeted, 0, 2}}}, |
| {"basic", kUnder, kRGBA64Bytes, {}, {}, {{kProxy64, 0, 2}}}, |
| {"basic, over", kOver, kRGBA64Bytes - 1, {}, {}, {{kProxy64, 0, 2}}}, |
| {"shared", kUnder, kRGBA64Bytes, {}, {}, |
| { |
| {kProxy64, 0, 2}, |
| {kProxy64, 3, 5}, |
| }}, |
| {"retrieved from cache", kUnder, kRGBA64Bytes, |
| /* purgeable */{kProxy64Instantiated}, |
| /* unpurgeable */{}, |
| { |
| {kProxy64, 0, 2} |
| }}, |
| {"purge 4", kUnder, kRGBA64Bytes, |
| /* purgeable */{ |
| kProxy32Instantiated, |
| kProxy32Instantiated, |
| kProxy32Instantiated, |
| kProxy32Instantiated |
| }, |
| /* unpurgeable */{}, |
| { |
| {kProxy64, 0, 2} |
| }}, |
| {"dont purge what we've reserved", kOver, kRGBA64Bytes, |
| /* purgeable */{kProxy64Instantiated}, |
| /* unpurgeable */{}, |
| { |
| {kProxy64, 0, 2}, |
| {kProxy64, 1, 3} |
| }}, |
| {"unpurgeable", kOver, kRGBA64Bytes, |
| /* purgeable */{}, |
| /* unpurgeable */{kProxy64Instantiated}, |
| { |
| {kProxy64, 0, 2} |
| }}, |
| {"lazy", kUnder, kRGBA64Bytes, |
| /* purgeable */{}, |
| /* unpurgeable */{}, |
| { |
| {kProxy64Lazy, 0, 2} |
| }}, |
| {"lazy, over", kOver, kRGBA64Bytes - 1, |
| /* purgeable */{}, |
| /* unpurgeable */{}, |
| { |
| {kProxy64Lazy, 0, 2} |
| }}, |
| {"fully-lazy", kUnder, kRGBA64Bytes, |
| /* purgeable */{}, |
| /* unpurgeable */{}, |
| { |
| {kProxy64FullyLazy, 0, 2} |
| }}, |
| {"fully-lazy, over", kOver, kRGBA64Bytes - 1, |
| /* purgeable */{}, |
| /* unpurgeable */{}, |
| { |
| {kProxy64FullyLazy, 0, 2} |
| }}, |
| }; |
| SkString match(""); |
| for (size_t i = 0; i < SK_ARRAY_COUNT(tests); i++) { |
| TestCase& test = tests[i]; |
| if (match.isEmpty() || match == SkString(test.fName)) { |
| // Create proxies |
| for (Interval& interval : test.fIntervals) { |
| interval.fProxy = make_proxy(dContext, interval.fParams); |
| } |
| reporter->push(SkString(test.fName)); |
| memory_budget_test(reporter, dContext, test); |
| reporter->pop(); |
| } |
| } |
| } |
| |