| /* |
| * Copyright 2018 Google Inc. |
| * |
| * Use of this source code is governed by a BSD-style license that can be |
| * found in the LICENSE file. |
| */ |
| |
| #include "tools/DDLPromiseImageHelper.h" |
| |
| #include "include/core/SkDeferredDisplayListRecorder.h" |
| #include "include/core/SkPicture.h" |
| #include "include/core/SkSerialProcs.h" |
| #include "include/core/SkYUVAIndex.h" |
| #include "include/core/SkYUVASizeInfo.h" |
| #include "include/gpu/GrDirectContext.h" |
| #include "src/codec/SkCodecImageGenerator.h" |
| #include "src/core/SkCachedData.h" |
| #include "src/core/SkMipmap.h" |
| #include "src/core/SkTaskGroup.h" |
| #include "src/gpu/GrContextPriv.h" |
| #include "src/image/SkImage_Base.h" |
| #include "src/image/SkImage_GpuYUVA.h" |
| |
| DDLPromiseImageHelper::PromiseImageInfo::PromiseImageInfo(int index, |
| uint32_t originalUniqueID, |
| const SkImageInfo& ii) |
| : fIndex(index) |
| , fOriginalUniqueID(originalUniqueID) |
| , fImageInfo(ii) { |
| } |
| |
| DDLPromiseImageHelper::PromiseImageInfo::PromiseImageInfo(PromiseImageInfo&& other) |
| : fIndex(other.fIndex) |
| , fOriginalUniqueID(other.fOriginalUniqueID) |
| , fImageInfo(other.fImageInfo) |
| , fBaseLevel(other.fBaseLevel) |
| , fMipLevels(std::move(other.fMipLevels)) |
| , fYUVColorSpace(other.fYUVColorSpace) { |
| memcpy(fYUVAIndices, other.fYUVAIndices, sizeof(fYUVAIndices)); |
| for (int i = 0; i < SkYUVASizeInfo::kMaxCount; ++i) { |
| fYUVPlanes[i] = std::move(other.fYUVPlanes[i]); |
| fCallbackContexts[i] = std::move(other.fCallbackContexts[i]); |
| } |
| } |
| |
| DDLPromiseImageHelper::PromiseImageInfo::~PromiseImageInfo() {} |
| |
| const std::unique_ptr<SkPixmap[]> DDLPromiseImageHelper::PromiseImageInfo::normalMipLevels() const { |
| SkASSERT(!this->isYUV()); |
| std::unique_ptr<SkPixmap[]> pixmaps(new SkPixmap[this->numMipLevels()]); |
| pixmaps[0] = fBaseLevel.pixmap(); |
| if (fMipLevels) { |
| for (int i = 0; i < fMipLevels->countLevels(); ++i) { |
| SkMipmap::Level mipLevel; |
| fMipLevels->getLevel(i, &mipLevel); |
| pixmaps[i+1] = mipLevel.fPixmap; |
| } |
| } |
| return pixmaps; |
| } |
| |
| int DDLPromiseImageHelper::PromiseImageInfo::numMipLevels() const { |
| SkASSERT(!this->isYUV()); |
| return fMipLevels ? fMipLevels->countLevels()+1 : 1; |
| } |
| |
| void DDLPromiseImageHelper::PromiseImageInfo::setMipLevels(const SkBitmap& baseLevel, |
| std::unique_ptr<SkMipmap> mipLevels) { |
| fBaseLevel = baseLevel; |
| fMipLevels = std::move(mipLevels); |
| } |
| |
| void DDLPromiseImageHelper::PromiseImageInfo::setYUVPlanes( |
| const SkYUVASizeInfo& yuvaSizeInfo, |
| const SkYUVAIndex yuvaIndices[SkYUVAIndex::kIndexCount], |
| SkYUVColorSpace cs, |
| std::unique_ptr<char[]> planes[SkYUVAIndex::kIndexCount]) { |
| memcpy(fYUVAIndices, yuvaIndices, sizeof(fYUVAIndices)); |
| fYUVColorSpace = cs; |
| SkColorType colorTypes[] = {kUnknown_SkColorType, |
| kUnknown_SkColorType, |
| kUnknown_SkColorType, |
| kUnknown_SkColorType}; |
| for (int yuvIndex = 0; yuvIndex < SkYUVAIndex::kIndexCount; ++yuvIndex) { |
| int texIdx = yuvaIndices[yuvIndex].fIndex; |
| if (texIdx < 0) { |
| SkASSERT(SkYUVAIndex::kA_Index == yuvIndex); |
| continue; |
| } |
| // In current testing only ever expect A8, RG_88 |
| if (colorTypes[texIdx] == kUnknown_SkColorType) { |
| colorTypes[texIdx] = kAlpha_8_SkColorType; |
| } else { |
| SkASSERT(colorTypes[texIdx] == kAlpha_8_SkColorType); |
| colorTypes[texIdx] = kR8G8_unorm_SkColorType; |
| } |
| } |
| |
| for (int i = 0; i < SkYUVASizeInfo::kMaxCount; ++i) { |
| if (yuvaSizeInfo.fSizes[i].isEmpty()) { |
| SkASSERT(!yuvaSizeInfo.fWidthBytes[i] && colorTypes[i] == kUnknown_SkColorType); |
| continue; |
| } |
| auto info = SkImageInfo::Make(yuvaSizeInfo.fSizes[i], colorTypes[i], kPremul_SkAlphaType); |
| auto release = [](void* addr, void*) { std::unique_ptr<char[]>(static_cast<char*>(addr)); }; |
| fYUVPlanes[i].installPixels(info, planes[i].get(), yuvaSizeInfo.fWidthBytes[i], release, |
| planes[i].release()); |
| } |
| } |
| |
| /////////////////////////////////////////////////////////////////////////////////////////////////// |
| PromiseImageCallbackContext::~PromiseImageCallbackContext() { |
| SkASSERT(fDoneCnt == fNumImages); |
| SkASSERT(!fUnreleasedFulfills); |
| SkASSERT(fTotalReleases == fTotalFulfills); |
| SkASSERT(!fTotalFulfills || fDoneCnt); |
| |
| if (fPromiseImageTexture) { |
| fContext->deleteBackendTexture(fPromiseImageTexture->backendTexture()); |
| } |
| } |
| |
| void PromiseImageCallbackContext::setBackendTexture(const GrBackendTexture& backendTexture) { |
| SkASSERT(!fPromiseImageTexture); |
| SkASSERT(fBackendFormat == backendTexture.getBackendFormat()); |
| fPromiseImageTexture = SkPromiseImageTexture::Make(backendTexture); |
| } |
| |
| void PromiseImageCallbackContext::destroyBackendTexture() { |
| SkASSERT(!fPromiseImageTexture || fPromiseImageTexture->unique()); |
| |
| if (fPromiseImageTexture) { |
| fContext->deleteBackendTexture(fPromiseImageTexture->backendTexture()); |
| } |
| fPromiseImageTexture = nullptr; |
| } |
| |
| /////////////////////////////////////////////////////////////////////////////////////////////////// |
| |
| sk_sp<SkData> DDLPromiseImageHelper::deflateSKP(const SkPicture* inputPicture) { |
| SkSerialProcs procs; |
| |
| procs.fImageCtx = this; |
| procs.fImageProc = [](SkImage* image, void* ctx) -> sk_sp<SkData> { |
| auto helper = static_cast<DDLPromiseImageHelper*>(ctx); |
| |
| int id = helper->findOrDefineImage(image); |
| |
| // Even if 'id' is invalid (i.e., -1) write it to the SKP |
| return SkData::MakeWithCopy(&id, sizeof(id)); |
| }; |
| |
| return inputPicture->serialize(&procs); |
| } |
| |
| static GrBackendTexture create_yuva_texture(GrDirectContext* direct, const SkPixmap& pm, |
| const SkYUVAIndex yuvaIndices[4], int texIndex) { |
| SkASSERT(texIndex >= 0 && texIndex <= 3); |
| |
| #ifdef SK_DEBUG |
| int channelCount = 0; |
| for (int i = 0; i < SkYUVAIndex::kIndexCount; ++i) { |
| if (yuvaIndices[i].fIndex == texIndex) { |
| ++channelCount; |
| } |
| } |
| if (2 == channelCount) { |
| SkASSERT(kR8G8_unorm_SkColorType == pm.colorType()); |
| } |
| #endif |
| bool finishedBECreate = false; |
| auto markFinished = [](void* context) { |
| *(bool*)context = true; |
| }; |
| auto beTex = direct->createBackendTexture(&pm, 1, GrRenderable::kNo, GrProtected::kNo, |
| markFinished, &finishedBECreate); |
| if (beTex.isValid()) { |
| direct->submit(); |
| while (!finishedBECreate) { |
| direct->checkAsyncWorkCompletion(); |
| } |
| } |
| return beTex; |
| } |
| |
| /* |
| * Create backend textures and upload data to them for all the textures required to satisfy |
| * a single promise image. |
| * For YUV textures this will result in up to 4 actual textures. |
| */ |
| void DDLPromiseImageHelper::CreateBETexturesForPromiseImage(GrDirectContext* direct, |
| PromiseImageInfo* info) { |
| if (info->isYUV()) { |
| int numPixmaps; |
| SkAssertResult(SkYUVAIndex::AreValidIndices(info->yuvaIndices(), &numPixmaps)); |
| for (int j = 0; j < numPixmaps; ++j) { |
| const SkPixmap& yuvPixmap = info->yuvPixmap(j); |
| |
| PromiseImageCallbackContext* callbackContext = info->callbackContext(j); |
| SkASSERT(callbackContext); |
| |
| // DDL TODO: what should we do with mipmapped YUV images |
| callbackContext->setBackendTexture(create_yuva_texture(direct, yuvPixmap, |
| info->yuvaIndices(), j)); |
| SkASSERT(callbackContext->promiseImageTexture()); |
| } |
| } else { |
| PromiseImageCallbackContext* callbackContext = info->callbackContext(0); |
| if (!callbackContext) { |
| // This texture would've been too large to fit on the GPU |
| return; |
| } |
| |
| std::unique_ptr<SkPixmap[]> mipLevels = info->normalMipLevels(); |
| |
| bool finishedBECreate = false; |
| auto markFinished = [](void* context) { |
| *(bool*)context = true; |
| }; |
| auto backendTex = direct->createBackendTexture(mipLevels.get(), info->numMipLevels(), |
| GrRenderable::kNo, GrProtected::kNo, |
| markFinished, &finishedBECreate); |
| SkASSERT(backendTex.isValid()); |
| direct->submit(); |
| while (!finishedBECreate) { |
| direct->checkAsyncWorkCompletion(); |
| } |
| |
| callbackContext->setBackendTexture(backendTex); |
| } |
| } |
| |
| void DDLPromiseImageHelper::DeleteBETexturesForPromiseImage(GrDirectContext* direct, |
| PromiseImageInfo* info) { |
| if (info->isYUV()) { |
| int numPixmaps; |
| SkAssertResult(SkYUVAIndex::AreValidIndices(info->yuvaIndices(), &numPixmaps)); |
| for (int j = 0; j < numPixmaps; ++j) { |
| PromiseImageCallbackContext* callbackContext = info->callbackContext(j); |
| SkASSERT(callbackContext); |
| |
| callbackContext->destroyBackendTexture(); |
| SkASSERT(!callbackContext->promiseImageTexture()); |
| } |
| } else { |
| PromiseImageCallbackContext* callbackContext = info->callbackContext(0); |
| if (!callbackContext) { |
| // This texture would've been too large to fit on the GPU |
| return; |
| } |
| |
| callbackContext->destroyBackendTexture(); |
| SkASSERT(!callbackContext->promiseImageTexture()); |
| } |
| } |
| |
| void DDLPromiseImageHelper::createCallbackContexts(GrDirectContext* direct) { |
| const GrCaps* caps = direct->priv().caps(); |
| const int maxDimension = caps->maxTextureSize(); |
| |
| for (int i = 0; i < fImageInfo.count(); ++i) { |
| PromiseImageInfo& info = fImageInfo[i]; |
| |
| if (info.isYUV()) { |
| int numPixmaps; |
| SkAssertResult(SkYUVAIndex::AreValidIndices(info.yuvaIndices(), &numPixmaps)); |
| |
| for (int j = 0; j < numPixmaps; ++j) { |
| const SkPixmap& yuvPixmap = info.yuvPixmap(j); |
| |
| GrBackendFormat backendFormat = direct->defaultBackendFormat(yuvPixmap.colorType(), |
| GrRenderable::kNo); |
| |
| sk_sp<PromiseImageCallbackContext> callbackContext( |
| new PromiseImageCallbackContext(direct, backendFormat)); |
| |
| info.setCallbackContext(j, std::move(callbackContext)); |
| } |
| } else { |
| const SkBitmap& baseLevel = info.baseLevel(); |
| |
| // TODO: explicitly mark the PromiseImageInfo as too big and check in uploadAllToGPU |
| if (maxDimension < std::max(baseLevel.width(), baseLevel.height())) { |
| // This won't fit on the GPU. Fallback to a raster-backed image per tile. |
| continue; |
| } |
| |
| GrBackendFormat backendFormat = direct->defaultBackendFormat(baseLevel.colorType(), |
| GrRenderable::kNo); |
| if (!caps->isFormatTexturable(backendFormat)) { |
| continue; |
| } |
| |
| sk_sp<PromiseImageCallbackContext> callbackContext( |
| new PromiseImageCallbackContext(direct, backendFormat)); |
| |
| info.setCallbackContext(0, std::move(callbackContext)); |
| } |
| } |
| } |
| |
| void DDLPromiseImageHelper::uploadAllToGPU(SkTaskGroup* taskGroup, GrDirectContext* direct) { |
| if (taskGroup) { |
| for (int i = 0; i < fImageInfo.count(); ++i) { |
| PromiseImageInfo* info = &fImageInfo[i]; |
| |
| taskGroup->add([direct, info]() { CreateBETexturesForPromiseImage(direct, info); }); |
| } |
| } else { |
| for (int i = 0; i < fImageInfo.count(); ++i) { |
| CreateBETexturesForPromiseImage(direct, &fImageInfo[i]); |
| } |
| } |
| } |
| |
| void DDLPromiseImageHelper::deleteAllFromGPU(SkTaskGroup* taskGroup, GrDirectContext* direct) { |
| if (taskGroup) { |
| for (int i = 0; i < fImageInfo.count(); ++i) { |
| PromiseImageInfo* info = &fImageInfo[i]; |
| |
| taskGroup->add([direct, info]() { DeleteBETexturesForPromiseImage(direct, info); }); |
| } |
| } else { |
| for (int i = 0; i < fImageInfo.count(); ++i) { |
| DeleteBETexturesForPromiseImage(direct, &fImageInfo[i]); |
| } |
| } |
| } |
| |
| sk_sp<SkPicture> DDLPromiseImageHelper::reinflateSKP( |
| SkDeferredDisplayListRecorder* recorder, |
| SkData* compressedPictureData, |
| SkTArray<sk_sp<SkImage>>* promiseImages) const { |
| PerRecorderContext perRecorderContext { recorder, this, promiseImages }; |
| |
| SkDeserialProcs procs; |
| procs.fImageCtx = (void*) &perRecorderContext; |
| procs.fImageProc = CreatePromiseImages; |
| |
| return SkPicture::MakeFromData(compressedPictureData, &procs); |
| } |
| |
| // This generates promise images to replace the indices in the compressed picture. This |
| // reconstitution is performed separately in each thread so we end up with multiple |
| // promise images referring to the same GrBackendTexture. |
| sk_sp<SkImage> DDLPromiseImageHelper::CreatePromiseImages(const void* rawData, |
| size_t length, void* ctxIn) { |
| PerRecorderContext* perRecorderContext = static_cast<PerRecorderContext*>(ctxIn); |
| const DDLPromiseImageHelper* helper = perRecorderContext->fHelper; |
| SkDeferredDisplayListRecorder* recorder = perRecorderContext->fRecorder; |
| |
| SkASSERT(length == sizeof(int)); |
| |
| const int* indexPtr = static_cast<const int*>(rawData); |
| if (!helper->isValidID(*indexPtr)) { |
| return nullptr; |
| } |
| |
| const DDLPromiseImageHelper::PromiseImageInfo& curImage = helper->getInfo(*indexPtr); |
| |
| // If there is no callback context that means 'createCallbackContexts' determined the |
| // texture wouldn't fit on the GPU. Create a separate bitmap-backed image for each thread. |
| if (!curImage.isYUV() && !curImage.callbackContext(0)) { |
| SkASSERT(curImage.baseLevel().isImmutable()); |
| return SkImage::MakeFromBitmap(curImage.baseLevel()); |
| } |
| |
| SkASSERT(curImage.index() == *indexPtr); |
| |
| sk_sp<SkImage> image; |
| if (curImage.isYUV()) { |
| GrBackendFormat backendFormats[SkYUVASizeInfo::kMaxCount]; |
| void* contexts[SkYUVASizeInfo::kMaxCount] = { nullptr, nullptr, nullptr, nullptr }; |
| SkISize sizes[SkYUVASizeInfo::kMaxCount]; |
| // TODO: store this value somewhere? |
| int textureCount; |
| SkAssertResult(SkYUVAIndex::AreValidIndices(curImage.yuvaIndices(), &textureCount)); |
| for (int i = 0; i < textureCount; ++i) { |
| backendFormats[i] = curImage.backendFormat(i); |
| SkASSERT(backendFormats[i].isValid()); |
| contexts[i] = curImage.refCallbackContext(i).release(); |
| sizes[i].set(curImage.yuvPixmap(i).width(), curImage.yuvPixmap(i).height()); |
| } |
| for (int i = textureCount; i < SkYUVASizeInfo::kMaxCount; ++i) { |
| sizes[i] = SkISize::MakeEmpty(); |
| } |
| |
| image = recorder->makeYUVAPromiseTexture( |
| curImage.yuvColorSpace(), |
| backendFormats, |
| sizes, |
| curImage.yuvaIndices(), |
| curImage.overallWidth(), |
| curImage.overallHeight(), |
| GrSurfaceOrigin::kTopLeft_GrSurfaceOrigin, |
| curImage.refOverallColorSpace(), |
| PromiseImageCallbackContext::PromiseImageFulfillProc, |
| PromiseImageCallbackContext::PromiseImageReleaseProc, |
| PromiseImageCallbackContext::PromiseImageDoneProc, |
| contexts, |
| SkDeferredDisplayListRecorder::PromiseImageApiVersion::kNew); |
| for (int i = 0; i < textureCount; ++i) { |
| curImage.callbackContext(i)->wasAddedToImage(); |
| } |
| |
| #ifdef SK_DEBUG |
| { |
| // By the peekProxy contract this image should not have a single backing proxy so |
| // should return null. The call should also not trigger the conversion to RGBA. |
| SkImage_GpuYUVA* yuva = reinterpret_cast<SkImage_GpuYUVA*>(image.get()); |
| SkASSERT(!yuva->peekProxy()); |
| SkASSERT(!yuva->peekProxy()); // the first call didn't force a conversion to RGBA |
| } |
| #endif |
| } else { |
| GrBackendFormat backendFormat = curImage.backendFormat(0); |
| SkASSERT(backendFormat.isValid()); |
| |
| // Each DDL recorder gets its own ref on the promise callback context for the |
| // promise images it creates. |
| image = recorder->makePromiseTexture( |
| backendFormat, |
| curImage.overallWidth(), |
| curImage.overallHeight(), |
| curImage.mipMapped(0), |
| GrSurfaceOrigin::kTopLeft_GrSurfaceOrigin, |
| curImage.overallColorType(), |
| curImage.overallAlphaType(), |
| curImage.refOverallColorSpace(), |
| PromiseImageCallbackContext::PromiseImageFulfillProc, |
| PromiseImageCallbackContext::PromiseImageReleaseProc, |
| PromiseImageCallbackContext::PromiseImageDoneProc, |
| (void*)curImage.refCallbackContext(0).release(), |
| SkDeferredDisplayListRecorder::PromiseImageApiVersion::kNew); |
| curImage.callbackContext(0)->wasAddedToImage(); |
| } |
| perRecorderContext->fPromiseImages->push_back(image); |
| SkASSERT(image); |
| return image; |
| } |
| |
| int DDLPromiseImageHelper::findImage(SkImage* image) const { |
| for (int i = 0; i < fImageInfo.count(); ++i) { |
| if (fImageInfo[i].originalUniqueID() == image->uniqueID()) { // trying to dedup here |
| SkASSERT(fImageInfo[i].index() == i); |
| SkASSERT(this->isValidID(i) && this->isValidID(fImageInfo[i].index())); |
| return i; |
| } |
| } |
| return -1; |
| } |
| |
| int DDLPromiseImageHelper::addImage(SkImage* image) { |
| SkImage_Base* ib = as_IB(image); |
| |
| SkImageInfo overallII = SkImageInfo::Make(image->width(), image->height(), |
| image->colorType() == kBGRA_8888_SkColorType |
| ? kRGBA_8888_SkColorType |
| : image->colorType(), |
| image->alphaType(), |
| image->refColorSpace()); |
| |
| PromiseImageInfo& newImageInfo = fImageInfo.emplace_back(fImageInfo.count(), |
| image->uniqueID(), |
| overallII); |
| |
| SkYUVASizeInfo yuvaSizeInfo; |
| SkYUVAIndex yuvaIndices[SkYUVAIndex::kIndexCount]; |
| SkYUVColorSpace yuvColorSpace; |
| auto codec = SkCodecImageGenerator::MakeFromEncodedCodec(ib->refEncodedData()); |
| std::unique_ptr<char[]> planes[SkYUVASizeInfo::kMaxCount]; |
| if (codec && codec->queryYUVA8(&yuvaSizeInfo, yuvaIndices, &yuvColorSpace)) { |
| void* data[4]; |
| for (int i = 0; i < 4; ++i) { |
| size_t size = yuvaSizeInfo.fSizes[i].height() * yuvaSizeInfo.fWidthBytes[i]; |
| if (size) { |
| planes[i].reset(new char[size]); |
| } |
| data[i] = planes[i].get(); |
| } |
| SkAssertResult(codec->getYUVA8Planes(yuvaSizeInfo, yuvaIndices, data)); |
| newImageInfo.setYUVPlanes(yuvaSizeInfo, yuvaIndices, yuvColorSpace, planes); |
| } else { |
| sk_sp<SkImage> rasterImage = image->makeRasterImage(); // force decoding of lazy images |
| if (!rasterImage) { |
| return -1; |
| } |
| |
| SkBitmap tmp; |
| tmp.allocPixels(overallII); |
| |
| if (!rasterImage->readPixels(tmp.pixmap(), 0, 0)) { |
| return -1; |
| } |
| |
| tmp.setImmutable(); |
| |
| // Given how the DDL testing harness works (i.e., only modifying the SkImages w/in an |
| // SKP) we don't know if a given SkImage will require mipmapping. To work around this |
| // we just try to create all the backend textures as mipmapped but, failing that, fall |
| // back to un-mipped. |
| std::unique_ptr<SkMipmap> mipmaps(SkMipmap::Build(tmp.pixmap(), nullptr)); |
| |
| newImageInfo.setMipLevels(tmp, std::move(mipmaps)); |
| } |
| // In either case newImageInfo's PromiseImageCallbackContext is filled in by uploadAllToGPU |
| |
| return fImageInfo.count()-1; |
| } |
| |
| int DDLPromiseImageHelper::findOrDefineImage(SkImage* image) { |
| int preExistingID = this->findImage(image); |
| if (preExistingID >= 0) { |
| SkASSERT(this->isValidID(preExistingID)); |
| return preExistingID; |
| } |
| |
| int newID = this->addImage(image); |
| return newID; |
| } |