Remove support for manual memory management in GrVkMemory.

These code paths are no longer used now that the GrBackendTexture code
uses the normal Ganesh code paths.

Change-Id: Idecf6876d96c7be587919f969d4736374d82c199
Reviewed-on: https://skia-review.googlesource.com/c/skia/+/289623
Reviewed-by: Jim Van Verth <jvanverth@google.com>
Commit-Queue: Greg Daniel <egdaniel@google.com>
diff --git a/src/gpu/vk/GrVkMemory.cpp b/src/gpu/vk/GrVkMemory.cpp
index ff07d3b..7d82ae7 100644
--- a/src/gpu/vk/GrVkMemory.cpp
+++ b/src/gpu/vk/GrVkMemory.cpp
@@ -76,12 +76,9 @@
 
 void GrVkMemory::FreeBufferMemory(const GrVkGpu* gpu, GrVkBuffer::Type type,
                                   const GrVkAlloc& alloc) {
-    if (alloc.fBackendMemory) {
-        GrVkMemoryAllocator* allocator = gpu->memoryAllocator();
-        allocator->freeMemory(alloc.fBackendMemory);
-    } else {
-        GR_VK_CALL(gpu->vkInterface(), FreeMemory(gpu->device(), alloc.fMemory, nullptr));
-    }
+    SkASSERT(alloc.fBackendMemory);
+    GrVkMemoryAllocator* allocator = gpu->memoryAllocator();
+    allocator->freeMemory(alloc.fBackendMemory);
 }
 
 const VkDeviceSize kMaxSmallImageSize = 256 * 1024;
@@ -128,38 +125,22 @@
 
 void GrVkMemory::FreeImageMemory(const GrVkGpu* gpu, bool linearTiling,
                                  const GrVkAlloc& alloc) {
-    if (alloc.fBackendMemory) {
-        GrVkMemoryAllocator* allocator = gpu->memoryAllocator();
-        allocator->freeMemory(alloc.fBackendMemory);
-    } else {
-        GR_VK_CALL(gpu->vkInterface(), FreeMemory(gpu->device(), alloc.fMemory, nullptr));
-    }
+    SkASSERT(alloc.fBackendMemory);
+    GrVkMemoryAllocator* allocator = gpu->memoryAllocator();
+    allocator->freeMemory(alloc.fBackendMemory);
 }
 
 void* GrVkMemory::MapAlloc(GrVkGpu* gpu, const GrVkAlloc& alloc) {
     SkASSERT(GrVkAlloc::kMappable_Flag & alloc.fFlags);
-    if (alloc.fBackendMemory) {
-        GrVkMemoryAllocator* allocator = gpu->memoryAllocator();
-        return allocator->mapMemory(alloc.fBackendMemory);
-    }
-
-    void* mapPtr;
-    VkResult err;
-    GR_VK_CALL_RESULT(gpu, err, MapMemory(gpu->device(), alloc.fMemory, alloc.fOffset, alloc.fSize,
-                                          0, &mapPtr));
-    if (err) {
-        mapPtr = nullptr;
-    }
-    return mapPtr;
+    SkASSERT(alloc.fBackendMemory);
+    GrVkMemoryAllocator* allocator = gpu->memoryAllocator();
+    return allocator->mapMemory(alloc.fBackendMemory);
 }
 
 void GrVkMemory::UnmapAlloc(const GrVkGpu* gpu, const GrVkAlloc& alloc) {
-    if (alloc.fBackendMemory) {
-        GrVkMemoryAllocator* allocator = gpu->memoryAllocator();
-        allocator->unmapMemory(alloc.fBackendMemory);
-    } else {
-        GR_VK_CALL(gpu->vkInterface(), UnmapMemory(gpu->device(), alloc.fMemory));
-    }
+    SkASSERT(alloc.fBackendMemory);
+    GrVkMemoryAllocator* allocator = gpu->memoryAllocator();
+    allocator->unmapMemory(alloc.fBackendMemory);
 }
 
 void GrVkMemory::GetNonCoherentMappedMemoryRange(const GrVkAlloc& alloc, VkDeviceSize offset,
@@ -190,17 +171,9 @@
     if (alloc.fFlags & GrVkAlloc::kNoncoherent_Flag) {
         SkASSERT(offset == 0);
         SkASSERT(size <= alloc.fSize);
-        if (alloc.fBackendMemory) {
-            GrVkMemoryAllocator* allocator = gpu->memoryAllocator();
-            allocator->flushMappedMemory(alloc.fBackendMemory, offset, size);
-        } else {
-            VkDeviceSize alignment = gpu->physicalDeviceProperties().limits.nonCoherentAtomSize;
-            VkMappedMemoryRange mappedMemoryRange;
-            GrVkMemory::GetNonCoherentMappedMemoryRange(alloc, offset, size, alignment,
-                                                        &mappedMemoryRange);
-            GR_VK_CALL(gpu->vkInterface(), FlushMappedMemoryRanges(gpu->device(), 1,
-                                                                   &mappedMemoryRange));
-        }
+        SkASSERT(alloc.fBackendMemory);
+        GrVkMemoryAllocator* allocator = gpu->memoryAllocator();
+        allocator->flushMappedMemory(alloc.fBackendMemory, offset, size);
     }
 }
 
@@ -209,17 +182,9 @@
     if (alloc.fFlags & GrVkAlloc::kNoncoherent_Flag) {
         SkASSERT(offset == 0);
         SkASSERT(size <= alloc.fSize);
-        if (alloc.fBackendMemory) {
-            GrVkMemoryAllocator* allocator = gpu->memoryAllocator();
-            allocator->invalidateMappedMemory(alloc.fBackendMemory, offset, size);
-        } else {
-            VkDeviceSize alignment = gpu->physicalDeviceProperties().limits.nonCoherentAtomSize;
-            VkMappedMemoryRange mappedMemoryRange;
-            GrVkMemory::GetNonCoherentMappedMemoryRange(alloc, offset, size, alignment,
-                                                        &mappedMemoryRange);
-            GR_VK_CALL(gpu->vkInterface(), InvalidateMappedMemoryRanges(gpu->device(), 1,
-                                                                        &mappedMemoryRange));
-        }
+        SkASSERT(alloc.fBackendMemory);
+        GrVkMemoryAllocator* allocator = gpu->memoryAllocator();
+        allocator->invalidateMappedMemory(alloc.fBackendMemory, offset, size);
     }
 }