| Greg Daniel | 164a9f0 | 2016-02-22 09:56:40 -0500 | [diff] [blame] | 1 | /* |
| 2 | * Copyright 2015 Google Inc. |
| 3 | * |
| 4 | * Use of this source code is governed by a BSD-style license that can be |
| 5 | * found in the LICENSE file. |
| 6 | */ |
| 7 | |
| 8 | #include "GrVkUtil.h" |
| 9 | |
| egdaniel | 88987d8 | 2016-09-19 10:17:34 -0700 | [diff] [blame] | 10 | #include "vk/GrVkGpu.h" |
| egdaniel | 88987d8 | 2016-09-19 10:17:34 -0700 | [diff] [blame] | 11 | #include "SkSLCompiler.h" |
| egdaniel | 88987d8 | 2016-09-19 10:17:34 -0700 | [diff] [blame] | 12 | |
| Greg Daniel | 164a9f0 | 2016-02-22 09:56:40 -0500 | [diff] [blame] | 13 | bool GrPixelConfigToVkFormat(GrPixelConfig config, VkFormat* format) { |
| 14 | VkFormat dontCare; |
| 15 | if (!format) { |
| 16 | format = &dontCare; |
| 17 | } |
| 18 | |
| 19 | switch (config) { |
| Brian Salomon | bf7b620 | 2016-11-11 16:08:03 -0500 | [diff] [blame] | 20 | case kUnknown_GrPixelConfig: |
| 21 | return false; |
| Greg Daniel | 164a9f0 | 2016-02-22 09:56:40 -0500 | [diff] [blame] | 22 | case kRGBA_8888_GrPixelConfig: |
| 23 | *format = VK_FORMAT_R8G8B8A8_UNORM; |
| Brian Salomon | bf7b620 | 2016-11-11 16:08:03 -0500 | [diff] [blame] | 24 | return true; |
| Greg Daniel | 164a9f0 | 2016-02-22 09:56:40 -0500 | [diff] [blame] | 25 | case kBGRA_8888_GrPixelConfig: |
| 26 | *format = VK_FORMAT_B8G8R8A8_UNORM; |
| Brian Salomon | bf7b620 | 2016-11-11 16:08:03 -0500 | [diff] [blame] | 27 | return true; |
| Greg Daniel | 164a9f0 | 2016-02-22 09:56:40 -0500 | [diff] [blame] | 28 | case kSRGBA_8888_GrPixelConfig: |
| 29 | *format = VK_FORMAT_R8G8B8A8_SRGB; |
| Brian Salomon | bf7b620 | 2016-11-11 16:08:03 -0500 | [diff] [blame] | 30 | return true; |
| jvanverth | 9f37246 | 2016-04-06 06:08:59 -0700 | [diff] [blame] | 31 | case kSBGRA_8888_GrPixelConfig: |
| 32 | *format = VK_FORMAT_B8G8R8A8_SRGB; |
| Brian Salomon | bf7b620 | 2016-11-11 16:08:03 -0500 | [diff] [blame] | 33 | return true; |
| 34 | case kRGBA_8888_sint_GrPixelConfig: |
| 35 | *format = VK_FORMAT_R8G8B8A8_SINT; |
| 36 | return true; |
| Greg Daniel | 164a9f0 | 2016-02-22 09:56:40 -0500 | [diff] [blame] | 37 | case kRGB_565_GrPixelConfig: |
| 38 | *format = VK_FORMAT_R5G6B5_UNORM_PACK16; |
| Brian Salomon | bf7b620 | 2016-11-11 16:08:03 -0500 | [diff] [blame] | 39 | return true; |
| Greg Daniel | 164a9f0 | 2016-02-22 09:56:40 -0500 | [diff] [blame] | 40 | case kRGBA_4444_GrPixelConfig: |
| egdaniel | 3fe0327 | 2016-08-15 10:59:17 -0700 | [diff] [blame] | 41 | // R4G4B4A4 is not required to be supported so we actually |
| 42 | // store the data is if it was B4G4R4A4 and swizzle in shaders |
| 43 | *format = VK_FORMAT_B4G4R4A4_UNORM_PACK16; |
| Brian Salomon | bf7b620 | 2016-11-11 16:08:03 -0500 | [diff] [blame] | 44 | return true; |
| Greg Daniel | ef59d87 | 2017-11-17 16:47:21 -0500 | [diff] [blame] | 45 | case kAlpha_8_GrPixelConfig: // fall through |
| 46 | case kAlpha_8_as_Red_GrPixelConfig: |
| Greg Daniel | 164a9f0 | 2016-02-22 09:56:40 -0500 | [diff] [blame] | 47 | *format = VK_FORMAT_R8_UNORM; |
| Brian Salomon | bf7b620 | 2016-11-11 16:08:03 -0500 | [diff] [blame] | 48 | return true; |
| Greg Daniel | ef59d87 | 2017-11-17 16:47:21 -0500 | [diff] [blame] | 49 | case kAlpha_8_as_Alpha_GrPixelConfig: |
| 50 | return false; |
| Brian Osman | 986563b | 2017-01-10 14:20:02 -0500 | [diff] [blame] | 51 | case kGray_8_GrPixelConfig: |
| Greg Daniel | 7af060a | 2017-12-05 16:27:11 -0500 | [diff] [blame^] | 52 | case kGray_8_as_Red_GrPixelConfig: |
| Brian Osman | 986563b | 2017-01-10 14:20:02 -0500 | [diff] [blame] | 53 | *format = VK_FORMAT_R8_UNORM; |
| 54 | return true; |
| Greg Daniel | 7af060a | 2017-12-05 16:27:11 -0500 | [diff] [blame^] | 55 | case kGray_8_as_Lum_GrPixelConfig: |
| 56 | return false; |
| Greg Daniel | 164a9f0 | 2016-02-22 09:56:40 -0500 | [diff] [blame] | 57 | case kRGBA_float_GrPixelConfig: |
| 58 | *format = VK_FORMAT_R32G32B32A32_SFLOAT; |
| Brian Salomon | bf7b620 | 2016-11-11 16:08:03 -0500 | [diff] [blame] | 59 | return true; |
| csmartdalton | 6aa0e11 | 2017-02-08 16:14:11 -0500 | [diff] [blame] | 60 | case kRG_float_GrPixelConfig: |
| 61 | *format = VK_FORMAT_R32G32_SFLOAT; |
| 62 | return true; |
| Greg Daniel | 164a9f0 | 2016-02-22 09:56:40 -0500 | [diff] [blame] | 63 | case kRGBA_half_GrPixelConfig: |
| 64 | *format = VK_FORMAT_R16G16B16A16_SFLOAT; |
| Brian Salomon | bf7b620 | 2016-11-11 16:08:03 -0500 | [diff] [blame] | 65 | return true; |
| Greg Daniel | ef59d87 | 2017-11-17 16:47:21 -0500 | [diff] [blame] | 66 | case kAlpha_half_GrPixelConfig: // fall through |
| 67 | case kAlpha_half_as_Red_GrPixelConfig: |
| Greg Daniel | 164a9f0 | 2016-02-22 09:56:40 -0500 | [diff] [blame] | 68 | *format = VK_FORMAT_R16_SFLOAT; |
| Brian Salomon | bf7b620 | 2016-11-11 16:08:03 -0500 | [diff] [blame] | 69 | return true; |
| Greg Daniel | 164a9f0 | 2016-02-22 09:56:40 -0500 | [diff] [blame] | 70 | } |
| Ben Wagner | b4aab9a | 2017-08-16 10:53:04 -0400 | [diff] [blame] | 71 | SK_ABORT("Unexpected config"); |
| Brian Salomon | bf7b620 | 2016-11-11 16:08:03 -0500 | [diff] [blame] | 72 | return false; |
| Greg Daniel | 164a9f0 | 2016-02-22 09:56:40 -0500 | [diff] [blame] | 73 | } |
| 74 | |
| Greg Daniel | 9440345 | 2017-04-18 15:52:36 -0400 | [diff] [blame] | 75 | GrPixelConfig GrVkFormatToPixelConfig(VkFormat format) { |
| jvanverth | 9f37246 | 2016-04-06 06:08:59 -0700 | [diff] [blame] | 76 | switch (format) { |
| 77 | case VK_FORMAT_R8G8B8A8_UNORM: |
| Greg Daniel | 9440345 | 2017-04-18 15:52:36 -0400 | [diff] [blame] | 78 | return kRGBA_8888_GrPixelConfig; |
| jvanverth | 9f37246 | 2016-04-06 06:08:59 -0700 | [diff] [blame] | 79 | case VK_FORMAT_B8G8R8A8_UNORM: |
| Greg Daniel | 9440345 | 2017-04-18 15:52:36 -0400 | [diff] [blame] | 80 | return kBGRA_8888_GrPixelConfig; |
| jvanverth | 9f37246 | 2016-04-06 06:08:59 -0700 | [diff] [blame] | 81 | case VK_FORMAT_R8G8B8A8_SRGB: |
| Greg Daniel | 9440345 | 2017-04-18 15:52:36 -0400 | [diff] [blame] | 82 | return kSRGBA_8888_GrPixelConfig; |
| jvanverth | 9f37246 | 2016-04-06 06:08:59 -0700 | [diff] [blame] | 83 | case VK_FORMAT_B8G8R8A8_SRGB: |
| Greg Daniel | 9440345 | 2017-04-18 15:52:36 -0400 | [diff] [blame] | 84 | return kSBGRA_8888_GrPixelConfig; |
| Brian Salomon | bf7b620 | 2016-11-11 16:08:03 -0500 | [diff] [blame] | 85 | case VK_FORMAT_R8G8B8A8_SINT: |
| Greg Daniel | 9440345 | 2017-04-18 15:52:36 -0400 | [diff] [blame] | 86 | return kRGBA_8888_sint_GrPixelConfig; |
| jvanverth | 9f37246 | 2016-04-06 06:08:59 -0700 | [diff] [blame] | 87 | case VK_FORMAT_R5G6B5_UNORM_PACK16: |
| Greg Daniel | 9440345 | 2017-04-18 15:52:36 -0400 | [diff] [blame] | 88 | return kRGB_565_GrPixelConfig; |
| jvanverth | 9f37246 | 2016-04-06 06:08:59 -0700 | [diff] [blame] | 89 | break; |
| egdaniel | 3fe0327 | 2016-08-15 10:59:17 -0700 | [diff] [blame] | 90 | case VK_FORMAT_B4G4R4A4_UNORM_PACK16: |
| 91 | // R4G4B4A4 is not required to be supported so we actually |
| 92 | // store RGBA_4444 data as B4G4R4A4. |
| Greg Daniel | 9440345 | 2017-04-18 15:52:36 -0400 | [diff] [blame] | 93 | return kRGBA_4444_GrPixelConfig; |
| jvanverth | 9f37246 | 2016-04-06 06:08:59 -0700 | [diff] [blame] | 94 | case VK_FORMAT_R8_UNORM: |
| Greg Daniel | 9440345 | 2017-04-18 15:52:36 -0400 | [diff] [blame] | 95 | return kAlpha_8_GrPixelConfig; |
| jvanverth | 9f37246 | 2016-04-06 06:08:59 -0700 | [diff] [blame] | 96 | case VK_FORMAT_R32G32B32A32_SFLOAT: |
| Greg Daniel | 9440345 | 2017-04-18 15:52:36 -0400 | [diff] [blame] | 97 | return kRGBA_float_GrPixelConfig; |
| csmartdalton | 6aa0e11 | 2017-02-08 16:14:11 -0500 | [diff] [blame] | 98 | case VK_FORMAT_R32G32_SFLOAT: |
| Greg Daniel | 9440345 | 2017-04-18 15:52:36 -0400 | [diff] [blame] | 99 | return kRG_float_GrPixelConfig; |
| jvanverth | 9f37246 | 2016-04-06 06:08:59 -0700 | [diff] [blame] | 100 | case VK_FORMAT_R16G16B16A16_SFLOAT: |
| Greg Daniel | 9440345 | 2017-04-18 15:52:36 -0400 | [diff] [blame] | 101 | return kRGBA_half_GrPixelConfig; |
| jvanverth | 9f37246 | 2016-04-06 06:08:59 -0700 | [diff] [blame] | 102 | case VK_FORMAT_R16_SFLOAT: |
| Greg Daniel | 9440345 | 2017-04-18 15:52:36 -0400 | [diff] [blame] | 103 | return kAlpha_half_GrPixelConfig; |
| jvanverth | 9f37246 | 2016-04-06 06:08:59 -0700 | [diff] [blame] | 104 | default: |
| Greg Daniel | 9440345 | 2017-04-18 15:52:36 -0400 | [diff] [blame] | 105 | return kUnknown_GrPixelConfig; |
| jvanverth | 9f37246 | 2016-04-06 06:08:59 -0700 | [diff] [blame] | 106 | } |
| jvanverth | 9f37246 | 2016-04-06 06:08:59 -0700 | [diff] [blame] | 107 | } |
| 108 | |
| brianosman | f05ab1b | 2016-05-12 11:01:10 -0700 | [diff] [blame] | 109 | bool GrVkFormatIsSRGB(VkFormat format, VkFormat* linearFormat) { |
| 110 | VkFormat linearFmt = format; |
| 111 | switch (format) { |
| 112 | case VK_FORMAT_R8_SRGB: |
| 113 | linearFmt = VK_FORMAT_R8_UNORM; |
| 114 | break; |
| 115 | case VK_FORMAT_R8G8_SRGB: |
| 116 | linearFmt = VK_FORMAT_R8G8_UNORM; |
| 117 | break; |
| 118 | case VK_FORMAT_R8G8B8_SRGB: |
| 119 | linearFmt = VK_FORMAT_R8G8B8_UNORM; |
| 120 | break; |
| 121 | case VK_FORMAT_B8G8R8_SRGB: |
| 122 | linearFmt = VK_FORMAT_B8G8R8_UNORM; |
| 123 | break; |
| 124 | case VK_FORMAT_R8G8B8A8_SRGB: |
| 125 | linearFmt = VK_FORMAT_R8G8B8A8_UNORM; |
| 126 | break; |
| 127 | case VK_FORMAT_B8G8R8A8_SRGB: |
| 128 | linearFmt = VK_FORMAT_B8G8R8A8_UNORM; |
| 129 | break; |
| 130 | case VK_FORMAT_A8B8G8R8_SRGB_PACK32: |
| 131 | linearFmt = VK_FORMAT_A8B8G8R8_UNORM_PACK32; |
| 132 | break; |
| 133 | case VK_FORMAT_BC1_RGB_SRGB_BLOCK: |
| 134 | linearFmt = VK_FORMAT_BC1_RGB_UNORM_BLOCK; |
| 135 | break; |
| 136 | case VK_FORMAT_BC1_RGBA_SRGB_BLOCK: |
| 137 | linearFmt = VK_FORMAT_BC1_RGBA_UNORM_BLOCK; |
| 138 | break; |
| 139 | case VK_FORMAT_BC2_SRGB_BLOCK: |
| 140 | linearFmt = VK_FORMAT_BC2_UNORM_BLOCK; |
| 141 | break; |
| 142 | case VK_FORMAT_BC3_SRGB_BLOCK: |
| 143 | linearFmt = VK_FORMAT_BC3_UNORM_BLOCK; |
| 144 | break; |
| 145 | case VK_FORMAT_BC7_SRGB_BLOCK: |
| 146 | linearFmt = VK_FORMAT_BC7_UNORM_BLOCK; |
| 147 | break; |
| 148 | case VK_FORMAT_ETC2_R8G8B8_SRGB_BLOCK: |
| 149 | linearFmt = VK_FORMAT_ETC2_R8G8B8_UNORM_BLOCK; |
| 150 | break; |
| 151 | case VK_FORMAT_ETC2_R8G8B8A1_SRGB_BLOCK: |
| 152 | linearFmt = VK_FORMAT_ETC2_R8G8B8A1_UNORM_BLOCK; |
| 153 | break; |
| 154 | case VK_FORMAT_ETC2_R8G8B8A8_SRGB_BLOCK: |
| 155 | linearFmt = VK_FORMAT_ETC2_R8G8B8A8_UNORM_BLOCK; |
| 156 | break; |
| 157 | case VK_FORMAT_ASTC_4x4_SRGB_BLOCK: |
| 158 | linearFmt = VK_FORMAT_ASTC_4x4_UNORM_BLOCK; |
| 159 | break; |
| 160 | case VK_FORMAT_ASTC_5x4_SRGB_BLOCK: |
| 161 | linearFmt = VK_FORMAT_ASTC_5x4_UNORM_BLOCK; |
| 162 | break; |
| 163 | case VK_FORMAT_ASTC_5x5_SRGB_BLOCK: |
| 164 | linearFmt = VK_FORMAT_ASTC_5x5_UNORM_BLOCK; |
| 165 | break; |
| 166 | case VK_FORMAT_ASTC_6x5_SRGB_BLOCK: |
| 167 | linearFmt = VK_FORMAT_ASTC_6x5_UNORM_BLOCK; |
| 168 | break; |
| 169 | case VK_FORMAT_ASTC_6x6_SRGB_BLOCK: |
| 170 | linearFmt = VK_FORMAT_ASTC_6x6_UNORM_BLOCK; |
| 171 | break; |
| 172 | case VK_FORMAT_ASTC_8x5_SRGB_BLOCK: |
| 173 | linearFmt = VK_FORMAT_ASTC_8x5_UNORM_BLOCK; |
| 174 | break; |
| 175 | case VK_FORMAT_ASTC_8x6_SRGB_BLOCK: |
| 176 | linearFmt = VK_FORMAT_ASTC_8x6_UNORM_BLOCK; |
| 177 | break; |
| 178 | case VK_FORMAT_ASTC_8x8_SRGB_BLOCK: |
| 179 | linearFmt = VK_FORMAT_ASTC_8x8_UNORM_BLOCK; |
| 180 | break; |
| 181 | case VK_FORMAT_ASTC_10x5_SRGB_BLOCK: |
| 182 | linearFmt = VK_FORMAT_ASTC_10x5_UNORM_BLOCK; |
| 183 | break; |
| 184 | case VK_FORMAT_ASTC_10x6_SRGB_BLOCK: |
| 185 | linearFmt = VK_FORMAT_ASTC_10x6_UNORM_BLOCK; |
| 186 | break; |
| 187 | case VK_FORMAT_ASTC_10x8_SRGB_BLOCK: |
| 188 | linearFmt = VK_FORMAT_ASTC_10x8_UNORM_BLOCK; |
| 189 | break; |
| 190 | case VK_FORMAT_ASTC_10x10_SRGB_BLOCK: |
| 191 | linearFmt = VK_FORMAT_ASTC_10x10_UNORM_BLOCK; |
| 192 | break; |
| 193 | case VK_FORMAT_ASTC_12x10_SRGB_BLOCK: |
| 194 | linearFmt = VK_FORMAT_ASTC_12x10_UNORM_BLOCK; |
| 195 | break; |
| 196 | case VK_FORMAT_ASTC_12x12_SRGB_BLOCK: |
| 197 | linearFmt = VK_FORMAT_ASTC_12x12_UNORM_BLOCK; |
| 198 | break; |
| 199 | default: |
| 200 | break; |
| 201 | } |
| 202 | if (linearFormat) { |
| 203 | *linearFormat = linearFmt; |
| 204 | } |
| 205 | return (linearFmt != format); |
| 206 | } |
| 207 | |
| Greg Daniel | 164a9f0 | 2016-02-22 09:56:40 -0500 | [diff] [blame] | 208 | bool GrSampleCountToVkSampleCount(uint32_t samples, VkSampleCountFlagBits* vkSamples) { |
| 209 | switch (samples) { |
| 210 | case 0: // fall through |
| 211 | case 1: |
| 212 | *vkSamples = VK_SAMPLE_COUNT_1_BIT; |
| 213 | return true; |
| 214 | case 2: |
| 215 | *vkSamples = VK_SAMPLE_COUNT_2_BIT; |
| 216 | return true; |
| 217 | case 4: |
| egdaniel | bf63e61 | 2016-08-17 06:26:16 -0700 | [diff] [blame] | 218 | *vkSamples = VK_SAMPLE_COUNT_4_BIT; |
| Greg Daniel | 164a9f0 | 2016-02-22 09:56:40 -0500 | [diff] [blame] | 219 | return true; |
| 220 | case 8: |
| egdaniel | bf63e61 | 2016-08-17 06:26:16 -0700 | [diff] [blame] | 221 | *vkSamples = VK_SAMPLE_COUNT_8_BIT; |
| Greg Daniel | 164a9f0 | 2016-02-22 09:56:40 -0500 | [diff] [blame] | 222 | return true; |
| 223 | case 16: |
| egdaniel | bf63e61 | 2016-08-17 06:26:16 -0700 | [diff] [blame] | 224 | *vkSamples = VK_SAMPLE_COUNT_16_BIT; |
| Greg Daniel | 164a9f0 | 2016-02-22 09:56:40 -0500 | [diff] [blame] | 225 | return true; |
| 226 | case 32: |
| egdaniel | bf63e61 | 2016-08-17 06:26:16 -0700 | [diff] [blame] | 227 | *vkSamples = VK_SAMPLE_COUNT_32_BIT; |
| Greg Daniel | 164a9f0 | 2016-02-22 09:56:40 -0500 | [diff] [blame] | 228 | return true; |
| 229 | case 64: |
| egdaniel | bf63e61 | 2016-08-17 06:26:16 -0700 | [diff] [blame] | 230 | *vkSamples = VK_SAMPLE_COUNT_64_BIT; |
| Greg Daniel | 164a9f0 | 2016-02-22 09:56:40 -0500 | [diff] [blame] | 231 | return true; |
| 232 | default: |
| 233 | return false; |
| 234 | } |
| 235 | } |
| egdaniel | 88987d8 | 2016-09-19 10:17:34 -0700 | [diff] [blame] | 236 | |
| egdaniel | 88987d8 | 2016-09-19 10:17:34 -0700 | [diff] [blame] | 237 | SkSL::Program::Kind vk_shader_stage_to_skiasl_kind(VkShaderStageFlagBits stage) { |
| 238 | if (VK_SHADER_STAGE_VERTEX_BIT == stage) { |
| 239 | return SkSL::Program::kVertex_Kind; |
| 240 | } |
| Chris Dalton | 33607c6 | 2017-07-07 11:00:48 -0600 | [diff] [blame] | 241 | if (VK_SHADER_STAGE_GEOMETRY_BIT == stage) { |
| 242 | return SkSL::Program::kGeometry_Kind; |
| 243 | } |
| egdaniel | 88987d8 | 2016-09-19 10:17:34 -0700 | [diff] [blame] | 244 | SkASSERT(VK_SHADER_STAGE_FRAGMENT_BIT == stage); |
| 245 | return SkSL::Program::kFragment_Kind; |
| 246 | } |
| Ethan Nicholas | 941e7e2 | 2016-12-12 15:33:30 -0500 | [diff] [blame] | 247 | |
| 248 | VkShaderStageFlagBits skiasl_kind_to_vk_shader_stage(SkSL::Program::Kind kind) { |
| 249 | if (SkSL::Program::kVertex_Kind == kind) { |
| 250 | return VK_SHADER_STAGE_VERTEX_BIT; |
| egdaniel | 88987d8 | 2016-09-19 10:17:34 -0700 | [diff] [blame] | 251 | } |
| Chris Dalton | 33607c6 | 2017-07-07 11:00:48 -0600 | [diff] [blame] | 252 | if (SkSL::Program::kGeometry_Kind == kind) { |
| 253 | return VK_SHADER_STAGE_GEOMETRY_BIT; |
| 254 | } |
| Ethan Nicholas | 941e7e2 | 2016-12-12 15:33:30 -0500 | [diff] [blame] | 255 | SkASSERT(SkSL::Program::kFragment_Kind == kind); |
| 256 | return VK_SHADER_STAGE_FRAGMENT_BIT; |
| egdaniel | 88987d8 | 2016-09-19 10:17:34 -0700 | [diff] [blame] | 257 | } |
| egdaniel | 88987d8 | 2016-09-19 10:17:34 -0700 | [diff] [blame] | 258 | |
| 259 | bool GrCompileVkShaderModule(const GrVkGpu* gpu, |
| 260 | const char* shaderString, |
| 261 | VkShaderStageFlagBits stage, |
| 262 | VkShaderModule* shaderModule, |
| Ethan Nicholas | 941e7e2 | 2016-12-12 15:33:30 -0500 | [diff] [blame] | 263 | VkPipelineShaderStageCreateInfo* stageInfo, |
| 264 | const SkSL::Program::Settings& settings, |
| 265 | SkSL::Program::Inputs* outInputs) { |
| 266 | std::unique_ptr<SkSL::Program> program = gpu->shaderCompiler()->convertProgram( |
| 267 | vk_shader_stage_to_skiasl_kind(stage), |
| Brian Osman | 93ba0a4 | 2017-08-14 14:48:10 -0400 | [diff] [blame] | 268 | SkSL::String(shaderString), |
| Ethan Nicholas | 941e7e2 | 2016-12-12 15:33:30 -0500 | [diff] [blame] | 269 | settings); |
| 270 | if (!program) { |
| 271 | SkDebugf("SkSL error:\n%s\n", gpu->shaderCompiler()->errorText().c_str()); |
| 272 | SkASSERT(false); |
| 273 | } |
| 274 | *outInputs = program->fInputs; |
| Ethan Nicholas | 0df1b04 | 2017-03-31 13:56:23 -0400 | [diff] [blame] | 275 | SkSL::String code; |
| Ethan Nicholas | 941e7e2 | 2016-12-12 15:33:30 -0500 | [diff] [blame] | 276 | if (!gpu->shaderCompiler()->toSPIRV(*program, &code)) { |
| 277 | SkDebugf("%s\n", gpu->shaderCompiler()->errorText().c_str()); |
| 278 | return false; |
| 279 | } |
| 280 | |
| egdaniel | 88987d8 | 2016-09-19 10:17:34 -0700 | [diff] [blame] | 281 | VkShaderModuleCreateInfo moduleCreateInfo; |
| 282 | memset(&moduleCreateInfo, 0, sizeof(VkShaderModuleCreateInfo)); |
| 283 | moduleCreateInfo.sType = VK_STRUCTURE_TYPE_SHADER_MODULE_CREATE_INFO; |
| 284 | moduleCreateInfo.pNext = nullptr; |
| 285 | moduleCreateInfo.flags = 0; |
| Ethan Nicholas | 941e7e2 | 2016-12-12 15:33:30 -0500 | [diff] [blame] | 286 | moduleCreateInfo.codeSize = code.size(); |
| 287 | moduleCreateInfo.pCode = (const uint32_t*)code.c_str(); |
| egdaniel | 88987d8 | 2016-09-19 10:17:34 -0700 | [diff] [blame] | 288 | |
| 289 | VkResult err = GR_VK_CALL(gpu->vkInterface(), CreateShaderModule(gpu->device(), |
| 290 | &moduleCreateInfo, |
| 291 | nullptr, |
| 292 | shaderModule)); |
| egdaniel | 88987d8 | 2016-09-19 10:17:34 -0700 | [diff] [blame] | 293 | if (err) { |
| 294 | return false; |
| 295 | } |
| 296 | |
| 297 | memset(stageInfo, 0, sizeof(VkPipelineShaderStageCreateInfo)); |
| 298 | stageInfo->sType = VK_STRUCTURE_TYPE_PIPELINE_SHADER_STAGE_CREATE_INFO; |
| 299 | stageInfo->pNext = nullptr; |
| 300 | stageInfo->flags = 0; |
| Ethan Nicholas | 941e7e2 | 2016-12-12 15:33:30 -0500 | [diff] [blame] | 301 | stageInfo->stage = skiasl_kind_to_vk_shader_stage(program->fKind); |
| egdaniel | 88987d8 | 2016-09-19 10:17:34 -0700 | [diff] [blame] | 302 | stageInfo->module = *shaderModule; |
| 303 | stageInfo->pName = "main"; |
| 304 | stageInfo->pSpecializationInfo = nullptr; |
| 305 | |
| 306 | return true; |
| 307 | } |