blob: 18be792861fa538c88ffe98c51bd08691dee23ed [file] [log] [blame]
Greg Danielb76a72a2017-07-13 15:07:54 -04001/*
2 * Copyright 2017 Google Inc.
3 *
4 * Use of this source code is governed by a BSD-style license that can be
5 * found in the LICENSE file.
6 */
7
8#include "MtlTestContext.h"
9
10#include "GrContext.h"
11#include "GrContextOptions.h"
12
13#import <Metal/Metal.h>
14
15#ifdef SK_METAL
16
17namespace {
18/**
19 * Implements sk_gpu_test::FenceSync for Metal.
20 */
21
22// TODO
23#if 0
24class MtlFenceSync : public sk_gpu_test::FenceSync {
25public:
26 MtlFenceSync(sk_sp<const GrVkInterface> vk, VkDevice device, VkQueue queue,
27 uint32_t queueFamilyIndex)
28 : fVk(std::move(vk))
29 , fDevice(device)
30 , fQueue(queue) {
31 SkDEBUGCODE(fUnfinishedSyncs = 0;)
32 VkCommandPoolCreateInfo createInfo;
33 createInfo.sType = VK_STRUCTURE_TYPE_COMMAND_POOL_CREATE_INFO;
34 createInfo.pNext = nullptr;
35 createInfo.flags = 0;
36 createInfo.queueFamilyIndex = queueFamilyIndex;
37 GR_VK_CALL_ERRCHECK(fVk, CreateCommandPool(fDevice, &createInfo, nullptr, &fCommandPool));
38
39 VkCommandBufferAllocateInfo allocateInfo;
40 allocateInfo.sType = VK_STRUCTURE_TYPE_COMMAND_BUFFER_ALLOCATE_INFO;
41 allocateInfo.pNext = nullptr;
42 allocateInfo.commandBufferCount = 1;
43 allocateInfo.commandPool = fCommandPool;
44 allocateInfo.level = VK_COMMAND_BUFFER_LEVEL_PRIMARY;
45 GR_VK_CALL_ERRCHECK(fVk, AllocateCommandBuffers(fDevice, &allocateInfo, &fCommandBuffer));
46
47 VkCommandBufferBeginInfo beginInfo;
48 beginInfo.sType = VK_STRUCTURE_TYPE_COMMAND_BUFFER_BEGIN_INFO;
49 beginInfo.pNext = nullptr;
50 beginInfo.flags = VK_COMMAND_BUFFER_USAGE_SIMULTANEOUS_USE_BIT;
51 beginInfo.pInheritanceInfo = nullptr;
52 GR_VK_CALL_ERRCHECK(fVk, BeginCommandBuffer(fCommandBuffer, &beginInfo));
53 GR_VK_CALL_ERRCHECK(fVk, EndCommandBuffer(fCommandBuffer));
54 }
55
56 ~VkFenceSync() override {
57 SkASSERT(!fUnfinishedSyncs);
58 // If the above assertion is true then the command buffer should not be in flight.
59 GR_VK_CALL(fVk, FreeCommandBuffers(fDevice, fCommandPool, 1, &fCommandBuffer));
60 GR_VK_CALL(fVk, DestroyCommandPool(fDevice, fCommandPool, nullptr));
61 }
62
63 sk_gpu_test::PlatformFence SK_WARN_UNUSED_RESULT insertFence() const override {
64 VkFence fence;
65 VkFenceCreateInfo info;
66 info.sType = VK_STRUCTURE_TYPE_FENCE_CREATE_INFO;
67 info.pNext = nullptr;
68 info.flags = 0;
69 GR_VK_CALL_ERRCHECK(fVk, CreateFence(fDevice, &info, nullptr, &fence));
70 VkSubmitInfo submitInfo;
71 submitInfo.sType = VK_STRUCTURE_TYPE_SUBMIT_INFO;
72 submitInfo.pNext = nullptr;
73 submitInfo.waitSemaphoreCount = 0;
74 submitInfo.pWaitSemaphores = nullptr;
75 submitInfo.pWaitDstStageMask = nullptr;
76 submitInfo.commandBufferCount = 1;
77 submitInfo.pCommandBuffers = &fCommandBuffer;
78 submitInfo.signalSemaphoreCount = 0;
79 submitInfo.pSignalSemaphores = nullptr;
80 GR_VK_CALL_ERRCHECK(fVk, QueueSubmit(fQueue, 1, &submitInfo, fence));
81 SkDEBUGCODE(++fUnfinishedSyncs;)
82 return (sk_gpu_test::PlatformFence)fence;
83 }
84
85 bool waitFence(sk_gpu_test::PlatformFence opaqueFence) const override {
86 VkFence fence = (VkFence)opaqueFence;
87 static constexpr uint64_t kForever = ~((uint64_t)0);
88 auto result = GR_VK_CALL(fVk, WaitForFences(fDevice, 1, &fence, true, kForever));
89 return result != VK_TIMEOUT;
90 }
91
92 void deleteFence(sk_gpu_test::PlatformFence opaqueFence) const override {
93 VkFence fence = (VkFence)opaqueFence;
94 GR_VK_CALL(fVk, DestroyFence(fDevice, fence, nullptr));
95 SkDEBUGCODE(--fUnfinishedSyncs;)
96 }
97
98private:
99 sk_sp<const GrVkInterface> fVk;
100 VkDevice fDevice;
101 VkQueue fQueue;
102 VkCommandPool fCommandPool;
103 VkCommandBuffer fCommandBuffer;
104 SkDEBUGCODE(mutable int fUnfinishedSyncs;)
105 typedef sk_gpu_test::FenceSync INHERITED;
106};
107
108GR_STATIC_ASSERT(sizeof(VkFence) <= sizeof(sk_gpu_test::PlatformFence));
109#endif
110
111class MtlTestContext : public sk_gpu_test::TestContext {
112public:
113 static MtlTestContext* Create(TestContext* sharedContext) {
114 SkASSERT(!sharedContext);
115 id<MTLDevice> device = MTLCreateSystemDefaultDevice();
116 id<MTLCommandQueue> queue = [device newCommandQueue];
117
118 return new MtlTestContext(device, queue);
119 }
120
121 ~MtlTestContext() override { this->teardown(); }
122
Greg Danielbdf12ad2018-10-12 09:31:11 -0400123 GrBackendApi backend() override { return GrBackendApi::kMetal; }
Greg Danielb76a72a2017-07-13 15:07:54 -0400124
Greg Danielb76a72a2017-07-13 15:07:54 -0400125 void testAbandon() override {}
126
127 // There is really nothing to here since we don't own any unqueued command buffers here.
128 void submit() override {}
129
130 void finish() override {}
131
132 sk_sp<GrContext> makeGrContext(const GrContextOptions& options) override {
133 return GrContext::MakeMetal((__bridge_retained void*)fDevice,
134 (__bridge_retained void*)fQueue,
135 options);
136 }
137
138private:
139 MtlTestContext(id<MTLDevice> device, id<MTLCommandQueue> queue)
140 : fDevice(device), fQueue(queue) {
141 fFenceSync.reset(nullptr);
142 }
143
144 void onPlatformMakeCurrent() const override {}
Brian Salomon55ad7742017-11-17 09:25:23 -0500145 std::function<void()> onPlatformGetAutoContextRestore() const override { return nullptr; }
Greg Danielb76a72a2017-07-13 15:07:54 -0400146 void onPlatformSwapBuffers() const override {}
147
148 id<MTLDevice> fDevice;
149 id<MTLCommandQueue> fQueue;
150
151 typedef sk_gpu_test::TestContext INHERITED;
152};
153
154} // anonymous namespace
155
156namespace sk_gpu_test {
157
158TestContext* CreatePlatformMtlTestContext(TestContext* sharedContext) {
159 return MtlTestContext::Create(sharedContext);
160}
161} // namespace sk_gpu_test
162
163
164#endif