Bug 13632 Header 84 Assorted questions and comments on vulkan.h
diff --git a/demos/cube.c b/demos/cube.c
index f046c02..5522c1f 100644
--- a/demos/cube.c
+++ b/demos/cube.c
@@ -28,7 +28,7 @@
VkImageLayout imageLayout;
uint32_t num_mem;
- VkGpuMemory *mem;
+ VkDeviceMemory *mem;
VkImageView view;
int32_t tex_width, tex_height;
};
@@ -203,12 +203,12 @@
bool use_staging_buffer;
VkInstance inst;
- VkPhysicalGpu gpu;
+ VkPhysicalDevice gpu;
VkDevice device;
VkQueue queue;
uint32_t graphics_queue_node_index;
- VkPhysicalGpuProperties *gpu_props;
- VkPhysicalGpuQueueProperties *queue_props;
+ VkPhysicalDeviceProperties *gpu_props;
+ VkPhysicalDeviceQueueProperties *queue_props;
VkFramebuffer framebuffer;
int width, height;
@@ -216,7 +216,7 @@
struct {
VkImage image;
- VkGpuMemory mem;
+ VkDeviceMemory mem;
VkCmdBuffer cmd;
VkColorAttachmentView view;
@@ -228,7 +228,7 @@
VkImage image;
uint32_t num_mem;
- VkGpuMemory *mem;
+ VkDeviceMemory *mem;
VkDepthStencilView view;
} depth;
@@ -237,7 +237,7 @@
struct {
VkBuffer buf;
uint32_t num_mem;
- VkGpuMemory *mem;
+ VkDeviceMemory *mem;
VkBufferView view;
VkBufferViewAttachInfo attach;
} uniform_data;
@@ -294,14 +294,14 @@
static void demo_add_mem_refs(
struct demo *demo,
- int num_refs, VkGpuMemory *mem)
+ int num_refs, VkDeviceMemory *mem)
{
vkQueueAddMemReferences(demo->queue, num_refs, mem);
}
static void demo_remove_mem_refs(
struct demo *demo,
- int num_refs, VkGpuMemory *mem)
+ int num_refs, VkDeviceMemory *mem)
{
vkQueueRemoveMemReferences(demo->queue, num_refs, mem);
}
@@ -328,7 +328,7 @@
VkCmdBufferBeginInfo cmd_buf_info = {
.sType = VK_STRUCTURE_TYPE_CMD_BUFFER_BEGIN_INFO,
.pNext = NULL,
- .flags = VK_CMD_BUFFER_OPTIMIZE_GPU_SMALL_BATCH_BIT |
+ .flags = VK_CMD_BUFFER_OPTIMIZE_SMALL_BATCH_BIT |
VK_CMD_BUFFER_OPTIMIZE_ONE_TIME_SUBMIT_BIT,
};
err = vkBeginCommandBuffer(demo->cmd, &cmd_buf_info);
@@ -359,16 +359,7 @@
VkPipeEvent set_events[] = { VK_PIPE_EVENT_TOP_OF_PIPE };
- VkPipelineBarrier pipeline_barrier;
- pipeline_barrier.sType = VK_STRUCTURE_TYPE_PIPELINE_BARRIER;
- pipeline_barrier.pNext = NULL;
- pipeline_barrier.eventCount = 1;
- pipeline_barrier.pEvents = set_events;
- pipeline_barrier.waitEvent = VK_WAIT_EVENT_TOP_OF_PIPE;
- pipeline_barrier.memBarrierCount = 1;
- pipeline_barrier.ppMemBarriers = (const void **)&pmemory_barrier;
-
- vkCmdPipelineBarrier(demo->cmd, &pipeline_barrier);
+ vkCmdPipelineBarrier(demo->cmd, VK_WAIT_EVENT_TOP_OF_PIPE, 1, set_events, 1, (const void **)&pmemory_barrier);
}
static void demo_draw_build_cmd(struct demo *demo, VkCmdBuffer cmd_buf)
@@ -390,7 +381,7 @@
VkCmdBufferBeginInfo cmd_buf_info = {
.sType = VK_STRUCTURE_TYPE_CMD_BUFFER_BEGIN_INFO,
.pNext = NULL,
- .flags = VK_CMD_BUFFER_OPTIMIZE_GPU_SMALL_BATCH_BIT |
+ .flags = VK_CMD_BUFFER_OPTIMIZE_SMALL_BATCH_BIT |
VK_CMD_BUFFER_OPTIMIZE_ONE_TIME_SUBMIT_BIT,
};
VkResult err;
@@ -422,7 +413,7 @@
rp_info.pColorLoadOps = &load_op;
rp_info.pColorStoreOps = &store_op;
rp_info.pColorLoadClearValues = &clear_color;
- rp_info.depthStencilFormat = VK_FMT_D16_UNORM;
+ rp_info.depthStencilFormat = VK_FORMAT_D16_UNORM;
rp_info.depthStencilLayout = depth_stencil.layout;
rp_info.depthLoadOp = VK_ATTACHMENT_LOAD_OP_DONT_CARE;
rp_info.depthLoadClearValue = clear_depth;
@@ -441,11 +432,11 @@
vkCmdBindDescriptorSets(cmd_buf, VK_PIPELINE_BIND_POINT_GRAPHICS,
demo->desc_layout_chain, 0, 1, &demo->desc_set, NULL);
- vkCmdBindDynamicStateObject(cmd_buf, VK_STATE_BIND_VIEWPORT, demo->viewport);
- vkCmdBindDynamicStateObject(cmd_buf, VK_STATE_BIND_RASTER, demo->raster);
- vkCmdBindDynamicStateObject(cmd_buf, VK_STATE_BIND_COLOR_BLEND,
+ vkCmdBindDynamicStateObject(cmd_buf, VK_STATE_BIND_POINT_VIEWPORT, demo->viewport);
+ vkCmdBindDynamicStateObject(cmd_buf, VK_STATE_BIND_POINT_RASTER, demo->raster);
+ vkCmdBindDynamicStateObject(cmd_buf, VK_STATE_BIND_POINT_COLOR_BLEND,
demo->color_blend);
- vkCmdBindDynamicStateObject(cmd_buf, VK_STATE_BIND_DEPTH_STENCIL,
+ vkCmdBindDynamicStateObject(cmd_buf, VK_STATE_BIND_POINT_DEPTH_STENCIL,
demo->depth_stencil);
vkCmdBeginRenderPass(cmd_buf, &rp_begin);
@@ -576,17 +567,17 @@
static void demo_prepare_depth(struct demo *demo)
{
- const VkFormat depth_format = VK_FMT_D16_UNORM;
+ const VkFormat depth_format = VK_FORMAT_D16_UNORM;
const VkImageCreateInfo image = {
.sType = VK_STRUCTURE_TYPE_IMAGE_CREATE_INFO,
.pNext = NULL,
- .imageType = VK_IMAGE_2D,
+ .imageType = VK_IMAGE_TYPE_2D,
.format = depth_format,
.extent = { demo->width, demo->height, 1 },
.mipLevels = 1,
.arraySize = 1,
.samples = 1,
- .tiling = VK_OPTIMAL_TILING,
+ .tiling = VK_IMAGE_TILING_OPTIMAL,
.usage = VK_IMAGE_USAGE_DEPTH_STENCIL_BIT,
.flags = 0,
};
@@ -594,7 +585,7 @@
.sType = VK_STRUCTURE_TYPE_MEMORY_ALLOC_INFO,
.pNext = NULL,
.allocationSize = 0,
- .memProps = VK_MEMORY_PROPERTY_GPU_ONLY,
+ .memProps = VK_MEMORY_PROPERTY_DEVICE_ONLY,
.memPriority = VK_MEMORY_PRIORITY_NORMAL,
};
VkDepthStencilViewCreateInfo view = {
@@ -619,13 +610,13 @@
&demo->depth.image);
assert(!err);
- err = vkGetObjectInfo(demo->depth.image, VK_INFO_TYPE_MEMORY_ALLOCATION_COUNT, &num_alloc_size, &num_allocations);
+ err = vkGetObjectInfo(demo->depth.image, VK_OBJECT_INFO_TYPE_MEMORY_ALLOCATION_COUNT, &num_alloc_size, &num_allocations);
assert(!err && num_alloc_size == sizeof(num_allocations));
mem_reqs = malloc(num_allocations * sizeof(VkMemoryRequirements));
- demo->depth.mem = malloc(num_allocations * sizeof(VkGpuMemory));
+ demo->depth.mem = malloc(num_allocations * sizeof(VkDeviceMemory));
demo->depth.num_mem = num_allocations;
err = vkGetObjectInfo(demo->depth.image,
- VK_INFO_TYPE_MEMORY_REQUIREMENTS,
+ VK_OBJECT_INFO_TYPE_MEMORY_REQUIREMENTS,
&mem_reqs_size, mem_reqs);
assert(!err && mem_reqs_size == num_allocations * sizeof(VkMemoryRequirements));
for (uint32_t i = 0; i < num_allocations; i ++) {
@@ -809,7 +800,7 @@
VkImageTiling tiling,
VkFlags mem_props)
{
- const VkFormat tex_format = VK_FMT_B8G8R8A8_UNORM;
+ const VkFormat tex_format = VK_FORMAT_B8G8R8A8_UNORM;
int32_t tex_width;
int32_t tex_height;
VkResult err;
@@ -823,7 +814,7 @@
const VkImageCreateInfo image_create_info = {
.sType = VK_STRUCTURE_TYPE_IMAGE_CREATE_INFO,
.pNext = NULL,
- .imageType = VK_IMAGE_2D,
+ .imageType = VK_IMAGE_TYPE_2D,
.format = tex_format,
.extent = { tex_width, tex_height, 1 },
.mipLevels = 1,
@@ -851,16 +842,16 @@
assert(!err);
err = vkGetObjectInfo(tex_obj->image,
- VK_INFO_TYPE_MEMORY_ALLOCATION_COUNT,
+ VK_OBJECT_INFO_TYPE_MEMORY_ALLOCATION_COUNT,
&num_alloc_size, &num_allocations);
assert(!err && num_alloc_size == sizeof(num_allocations));
mem_reqs = malloc(num_allocations * sizeof(VkMemoryRequirements));
- tex_obj->mem = malloc(num_allocations * sizeof(VkGpuMemory));
+ tex_obj->mem = malloc(num_allocations * sizeof(VkDeviceMemory));
err = vkGetObjectInfo(tex_obj->image,
- VK_INFO_TYPE_MEMORY_REQUIREMENTS,
+ VK_OBJECT_INFO_TYPE_MEMORY_REQUIREMENTS,
&mem_reqs_size, mem_reqs);
assert(!err && mem_reqs_size == num_allocations * sizeof(VkMemoryRequirements));
- mem_alloc.memProps = VK_MEMORY_PROPERTY_CPU_VISIBLE_BIT;
+ mem_alloc.memProps = VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT;
for (uint32_t j = 0; j < num_allocations; j ++) {
mem_alloc.allocationSize = mem_reqs[j].size;
@@ -878,7 +869,7 @@
tex_obj->num_mem = num_allocations;
- if (mem_props & VK_MEMORY_PROPERTY_CPU_VISIBLE_BIT) {
+ if (mem_props & VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT) {
const VkImageSubresource subres = {
.aspect = VK_IMAGE_ASPECT_COLOR,
.mipLevel = 0,
@@ -889,7 +880,7 @@
void *data;
err = vkGetImageSubresourceInfo(tex_obj->image, &subres,
- VK_INFO_TYPE_SUBRESOURCE_LAYOUT,
+ VK_SUBRESOURCE_INFO_TYPE_LAYOUT,
&layout_size, &layout);
assert(!err && layout_size == sizeof(layout));
/* Linear texture must be within a single memory object */
@@ -927,33 +918,33 @@
static void demo_prepare_textures(struct demo *demo)
{
- const VkFormat tex_format = VK_FMT_R8G8B8A8_UNORM;
+ const VkFormat tex_format = VK_FORMAT_R8G8B8A8_UNORM;
VkFormatProperties props;
size_t size = sizeof(props);
VkResult err;
uint32_t i;
err = vkGetFormatInfo(demo->device, tex_format,
- VK_INFO_TYPE_FORMAT_PROPERTIES,
+ VK_FORMAT_INFO_TYPE_PROPERTIES,
&size, &props);
assert(!err);
for (i = 0; i < DEMO_TEXTURE_COUNT; i++) {
- if (props.linearTilingFeatures & VK_FORMAT_SAMPLED_IMAGE_BIT && !demo->use_staging_buffer) {
+ if (props.linearTilingFeatures & VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT && !demo->use_staging_buffer) {
/* Device can texture using linear textures */
demo_prepare_texture_image(demo, tex_files[i], &demo->textures[i],
- VK_LINEAR_TILING, VK_MEMORY_PROPERTY_CPU_VISIBLE_BIT);
- } else if (props.optimalTilingFeatures & VK_FORMAT_SAMPLED_IMAGE_BIT) {
+ VK_IMAGE_TILING_LINEAR, VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT);
+ } else if (props.optimalTilingFeatures & VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT) {
/* Must use staging buffer to copy linear texture to optimized */
struct texture_object staging_texture;
memset(&staging_texture, 0, sizeof(staging_texture));
demo_prepare_texture_image(demo, tex_files[i], &staging_texture,
- VK_LINEAR_TILING, VK_MEMORY_PROPERTY_CPU_VISIBLE_BIT);
+ VK_IMAGE_TILING_LINEAR, VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT);
demo_prepare_texture_image(demo, tex_files[i], &demo->textures[i],
- VK_OPTIMAL_TILING, VK_MEMORY_PROPERTY_GPU_ONLY);
+ VK_IMAGE_TILING_OPTIMAL, VK_MEMORY_PROPERTY_DEVICE_ONLY);
demo_set_image_layout(demo, staging_texture.image,
staging_texture.imageLayout,
@@ -987,7 +978,7 @@
demo_destroy_texture_image(demo, &staging_texture);
demo_remove_mem_refs(demo, staging_texture.num_mem, staging_texture.mem);
} else {
- /* Can't support VK_FMT_B8G8R8A8_UNORM !? */
+ /* Can't support VK_FORMAT_B8G8R8A8_UNORM !? */
assert(!"No support for tB8G8R8A8_UNORM as texture image format");
}
@@ -996,23 +987,23 @@
.pNext = NULL,
.magFilter = VK_TEX_FILTER_NEAREST,
.minFilter = VK_TEX_FILTER_NEAREST,
- .mipMode = VK_TEX_MIPMAP_BASE,
+ .mipMode = VK_TEX_MIPMAP_MODE_BASE,
.addressU = VK_TEX_ADDRESS_CLAMP,
.addressV = VK_TEX_ADDRESS_CLAMP,
.addressW = VK_TEX_ADDRESS_CLAMP,
.mipLodBias = 0.0f,
.maxAnisotropy = 1,
- .compareFunc = VK_COMPARE_NEVER,
+ .compareOp = VK_COMPARE_OP_NEVER,
.minLod = 0.0f,
.maxLod = 0.0f,
- .borderColorType = VK_BORDER_COLOR_OPAQUE_WHITE,
+ .borderColor = VK_BORDER_COLOR_OPAQUE_WHITE,
};
VkImageViewCreateInfo view = {
.sType = VK_STRUCTURE_TYPE_IMAGE_VIEW_CREATE_INFO,
.pNext = NULL,
.image = VK_NULL_HANDLE,
- .viewType = VK_IMAGE_VIEW_2D,
+ .viewType = VK_IMAGE_VIEW_TYPE_2D,
.format = tex_format,
.channels = { VK_CHANNEL_SWIZZLE_R,
VK_CHANNEL_SWIZZLE_G,
@@ -1043,7 +1034,7 @@
.sType = VK_STRUCTURE_TYPE_MEMORY_ALLOC_INFO,
.pNext = NULL,
.allocationSize = 0,
- .memProps = VK_MEMORY_PROPERTY_CPU_VISIBLE_BIT,
+ .memProps = VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT,
.memPriority = VK_MEMORY_PRIORITY_NORMAL,
};
VkMemoryRequirements *mem_reqs;
@@ -1080,14 +1071,14 @@
assert(!err);
err = vkGetObjectInfo(demo->uniform_data.buf,
- VK_INFO_TYPE_MEMORY_ALLOCATION_COUNT,
+ VK_OBJECT_INFO_TYPE_MEMORY_ALLOCATION_COUNT,
&num_alloc_size, &num_allocations);
assert(!err && num_alloc_size == sizeof(num_allocations));
mem_reqs = malloc(num_allocations * sizeof(VkMemoryRequirements));
- demo->uniform_data.mem = malloc(num_allocations * sizeof(VkGpuMemory));
+ demo->uniform_data.mem = malloc(num_allocations * sizeof(VkDeviceMemory));
demo->uniform_data.num_mem = num_allocations;
err = vkGetObjectInfo(demo->uniform_data.buf,
- VK_INFO_TYPE_MEMORY_REQUIREMENTS,
+ VK_OBJECT_INFO_TYPE_MEMORY_REQUIREMENTS,
&mem_reqs_size, mem_reqs);
assert(!err && mem_reqs_size == num_allocations * sizeof(*mem_reqs));
for (uint32_t i = 0; i < num_allocations; i ++) {
@@ -1113,7 +1104,7 @@
memset(&view_info, 0, sizeof(view_info));
view_info.sType = VK_STRUCTURE_TYPE_BUFFER_VIEW_CREATE_INFO;
view_info.buffer = demo->uniform_data.buf;
- view_info.viewType = VK_BUFFER_VIEW_RAW;
+ view_info.viewType = VK_BUFFER_VIEW_TYPE_RAW;
view_info.offset = 0;
view_info.range = sizeof(data);
@@ -1130,13 +1121,13 @@
[0] = {
.descriptorType = VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER,
.count = 1,
- .stageFlags = VK_SHADER_STAGE_FLAGS_VERTEX_BIT,
+ .stageFlags = VK_SHADER_STAGE_VERTEX_BIT,
.pImmutableSamplers = NULL,
},
[1] = {
.descriptorType = VK_DESCRIPTOR_TYPE_COMBINED_IMAGE_SAMPLER,
.count = DEMO_TEXTURE_COUNT,
- .stageFlags = VK_SHADER_STAGE_FLAGS_FRAGMENT_BIT,
+ .stageFlags = VK_SHADER_STAGE_FRAGMENT_BIT,
.pImmutableSamplers = NULL,
},
};
@@ -1158,7 +1149,7 @@
}
static VkShader demo_prepare_shader(struct demo *demo,
- VkPipelineShaderStage stage,
+ VkShaderStage stage,
const void *code,
size_t size)
{
@@ -1186,7 +1177,7 @@
createInfo.pCode = malloc(createInfo.codeSize);
createInfo.flags = 0;
- /* try version 0 first: VkPipelineShaderStage followed by GLSL */
+ /* try version 0 first: VkShaderStage followed by GLSL */
((uint32_t *) createInfo.pCode)[0] = ICD_SPV_MAGIC;
((uint32_t *) createInfo.pCode)[1] = 0;
((uint32_t *) createInfo.pCode)[2] = stage;
@@ -1306,12 +1297,12 @@
memset(&ia, 0, sizeof(ia));
ia.sType = VK_STRUCTURE_TYPE_PIPELINE_IA_STATE_CREATE_INFO;
- ia.topology = VK_TOPOLOGY_TRIANGLE_LIST;
+ ia.topology = VK_PRIMITIVE_TOPOLOGY_TRIANGLE_LIST;
memset(&rs, 0, sizeof(rs));
rs.sType = VK_STRUCTURE_TYPE_PIPELINE_RS_STATE_CREATE_INFO;
- rs.fillMode = VK_FILL_SOLID;
- rs.cullMode = VK_CULL_BACK;
+ rs.fillMode = VK_FILL_MODE_SOLID;
+ rs.cullMode = VK_CULL_MODE_BACK;
rs.frontFace = VK_FRONT_FACE_CCW;
memset(&cb, 0, sizeof(cb));
@@ -1326,7 +1317,7 @@
memset(&vp, 0, sizeof(vp));
vp.sType = VK_STRUCTURE_TYPE_PIPELINE_VP_STATE_CREATE_INFO;
- vp.numViewports = 1;
+ vp.viewportCount = 1;
vp.clipOrigin = VK_COORDINATE_ORIGIN_LOWER_LEFT;
memset(&ds, 0, sizeof(ds));
@@ -1334,11 +1325,11 @@
ds.format = demo->depth.format;
ds.depthTestEnable = VK_TRUE;
ds.depthWriteEnable = VK_TRUE;
- ds.depthFunc = VK_COMPARE_LESS_EQUAL;
+ ds.depthCompareOp = VK_COMPARE_OP_LESS_EQUAL;
ds.depthBoundsEnable = VK_FALSE;
ds.back.stencilFailOp = VK_STENCIL_OP_KEEP;
ds.back.stencilPassOp = VK_STENCIL_OP_KEEP;
- ds.back.stencilFunc = VK_COMPARE_ALWAYS;
+ ds.back.stencilCompareOp = VK_COMPARE_OP_ALWAYS;
ds.stencilTestEnable = VK_FALSE;
ds.front = ds.back;
@@ -1727,24 +1718,24 @@
err = vkCreateDevice(demo->gpu, &device, &demo->device);
assert(!err);
- err = vkGetGpuInfo(demo->gpu, VK_INFO_TYPE_PHYSICAL_GPU_PROPERTIES,
+ err = vkGetPhysicalDeviceInfo(demo->gpu, VK_PHYSICAL_DEVICE_INFO_TYPE_PROPERTIES,
&data_size, NULL);
assert(!err);
- demo->gpu_props = (VkPhysicalGpuProperties *) malloc(data_size);
- err = vkGetGpuInfo(demo->gpu, VK_INFO_TYPE_PHYSICAL_GPU_PROPERTIES,
+ demo->gpu_props = (VkPhysicalDeviceProperties *) malloc(data_size);
+ err = vkGetPhysicalDeviceInfo(demo->gpu, VK_PHYSICAL_DEVICE_INFO_TYPE_PROPERTIES,
&data_size, demo->gpu_props);
assert(!err);
- err = vkGetGpuInfo(demo->gpu, VK_INFO_TYPE_PHYSICAL_GPU_QUEUE_PROPERTIES,
+ err = vkGetPhysicalDeviceInfo(demo->gpu, VK_PHYSICAL_DEVICE_INFO_TYPE_QUEUE_PROPERTIES,
&data_size, NULL);
assert(!err);
- demo->queue_props = (VkPhysicalGpuQueueProperties *) malloc(data_size);
- err = vkGetGpuInfo(demo->gpu, VK_INFO_TYPE_PHYSICAL_GPU_QUEUE_PROPERTIES,
+ demo->queue_props = (VkPhysicalDeviceQueueProperties *) malloc(data_size);
+ err = vkGetPhysicalDeviceInfo(demo->gpu, VK_PHYSICAL_DEVICE_INFO_TYPE_QUEUE_PROPERTIES,
&data_size, demo->queue_props);
assert(!err);
- queue_count = (uint32_t)(data_size / sizeof(VkPhysicalGpuQueueProperties));
+ queue_count = (uint32_t)(data_size / sizeof(VkPhysicalDeviceQueueProperties));
assert(queue_count >= 1);
// Graphics queue and MemMgr queue can be separate.
@@ -1803,7 +1794,7 @@
demo->width = 500;
demo->height = 500;
- demo->format = VK_FMT_B8G8R8A8_UNORM;
+ demo->format = VK_FORMAT_B8G8R8A8_UNORM;
demo->spin_angle = 0.01f;
demo->spin_increment = 0.01f;
diff --git a/demos/tri.c b/demos/tri.c
index a594cf9..bea3a7d 100644
--- a/demos/tri.c
+++ b/demos/tri.c
@@ -28,7 +28,7 @@
VkImageLayout imageLayout;
uint32_t num_mem;
- VkGpuMemory *mem;
+ VkDeviceMemory *mem;
VkImageView view;
int32_t tex_width, tex_height;
};
@@ -38,11 +38,11 @@
xcb_screen_t *screen;
VkInstance inst;
- VkPhysicalGpu gpu;
+ VkPhysicalDevice gpu;
VkDevice device;
VkQueue queue;
- VkPhysicalGpuProperties *gpu_props;
- VkPhysicalGpuQueueProperties *queue_props;
+ VkPhysicalDeviceProperties *gpu_props;
+ VkPhysicalDeviceQueueProperties *queue_props;
uint32_t graphics_queue_node_index;
int width, height;
@@ -51,7 +51,7 @@
struct {
VkImage image;
uint32_t num_mem;
- VkGpuMemory mem;
+ VkDeviceMemory mem;
VkColorAttachmentView view;
VkFence fence;
@@ -62,7 +62,7 @@
VkImage image;
uint32_t num_mem;
- VkGpuMemory *mem;
+ VkDeviceMemory *mem;
VkDepthStencilView view;
} depth;
@@ -71,7 +71,7 @@
struct {
VkBuffer buf;
uint32_t num_mem;
- VkGpuMemory *mem;
+ VkDeviceMemory *mem;
VkPipelineVertexInputCreateInfo vi;
VkVertexInputBindingDescription vi_bindings[1];
@@ -123,14 +123,14 @@
static void demo_add_mem_refs(
struct demo *demo,
- int num_refs, VkGpuMemory *mem)
+ int num_refs, VkDeviceMemory *mem)
{
vkQueueAddMemReferences(demo->queue, num_refs, mem);
}
static void demo_remove_mem_refs(
struct demo *demo,
- int num_refs, VkGpuMemory *mem)
+ int num_refs, VkDeviceMemory *mem)
{
vkQueueRemoveMemReferences(demo->queue, num_refs, mem);
}
@@ -157,7 +157,7 @@
VkCmdBufferBeginInfo cmd_buf_info = {
.sType = VK_STRUCTURE_TYPE_CMD_BUFFER_BEGIN_INFO,
.pNext = NULL,
- .flags = VK_CMD_BUFFER_OPTIMIZE_GPU_SMALL_BATCH_BIT |
+ .flags = VK_CMD_BUFFER_OPTIMIZE_SMALL_BATCH_BIT |
VK_CMD_BUFFER_OPTIMIZE_ONE_TIME_SUBMIT_BIT,
};
err = vkBeginCommandBuffer(demo->cmd, &cmd_buf_info);
@@ -188,16 +188,7 @@
VkPipeEvent set_events[] = { VK_PIPE_EVENT_TOP_OF_PIPE };
- VkPipelineBarrier pipeline_barrier;
- pipeline_barrier.sType = VK_STRUCTURE_TYPE_PIPELINE_BARRIER;
- pipeline_barrier.pNext = NULL;
- pipeline_barrier.eventCount = 1;
- pipeline_barrier.pEvents = set_events;
- pipeline_barrier.waitEvent = VK_WAIT_EVENT_TOP_OF_PIPE;
- pipeline_barrier.memBarrierCount = 1;
- pipeline_barrier.ppMemBarriers = (const void **)&pmemory_barrier;
-
- vkCmdPipelineBarrier(demo->cmd, &pipeline_barrier);
+ vkCmdPipelineBarrier(demo->cmd, VK_WAIT_EVENT_TOP_OF_PIPE, 1, set_events, 1, (const void **)&pmemory_barrier);
}
static void demo_draw_build_cmd(struct demo *demo)
@@ -219,7 +210,7 @@
VkCmdBufferBeginInfo cmd_buf_info = {
.sType = VK_STRUCTURE_TYPE_CMD_BUFFER_BEGIN_INFO,
.pNext = NULL,
- .flags = VK_CMD_BUFFER_OPTIMIZE_GPU_SMALL_BATCH_BIT |
+ .flags = VK_CMD_BUFFER_OPTIMIZE_SMALL_BATCH_BIT |
VK_CMD_BUFFER_OPTIMIZE_ONE_TIME_SUBMIT_BIT,
};
VkResult err;
@@ -251,7 +242,7 @@
rp_info.pColorLoadOps = &load_op;
rp_info.pColorStoreOps = &store_op;
rp_info.pColorLoadClearValues = &clear_color;
- rp_info.depthStencilFormat = VK_FMT_D16_UNORM;
+ rp_info.depthStencilFormat = VK_FORMAT_D16_UNORM;
rp_info.depthStencilLayout = depth_stencil.layout;
rp_info.depthLoadOp = VK_ATTACHMENT_LOAD_OP_DONT_CARE;
rp_info.depthLoadClearValue = clear_depth;
@@ -270,14 +261,14 @@
vkCmdBindDescriptorSets(demo->cmd, VK_PIPELINE_BIND_POINT_GRAPHICS,
demo->desc_layout_chain, 0, 1, & demo->desc_set, NULL);
- vkCmdBindDynamicStateObject(demo->cmd, VK_STATE_BIND_VIEWPORT, demo->viewport);
- vkCmdBindDynamicStateObject(demo->cmd, VK_STATE_BIND_RASTER, demo->raster);
- vkCmdBindDynamicStateObject(demo->cmd, VK_STATE_BIND_COLOR_BLEND,
+ vkCmdBindDynamicStateObject(demo->cmd, VK_STATE_BIND_POINT_VIEWPORT, demo->viewport);
+ vkCmdBindDynamicStateObject(demo->cmd, VK_STATE_BIND_POINT_RASTER, demo->raster);
+ vkCmdBindDynamicStateObject(demo->cmd, VK_STATE_BIND_POINT_COLOR_BLEND,
demo->color_blend);
- vkCmdBindDynamicStateObject(demo->cmd, VK_STATE_BIND_DEPTH_STENCIL,
+ vkCmdBindDynamicStateObject(demo->cmd, VK_STATE_BIND_POINT_DEPTH_STENCIL,
demo->depth_stencil);
- VkGpuSize offsets[1] = {0};
+ VkDeviceSize offsets[1] = {0};
vkCmdBindVertexBuffers(demo->cmd, VERTEX_BUFFER_BIND_ID, 1, &demo->vertices.buf, offsets);
vkCmdBeginRenderPass(demo->cmd, &rp_begin);
@@ -384,17 +375,17 @@
static void demo_prepare_depth(struct demo *demo)
{
- const VkFormat depth_format = VK_FMT_D16_UNORM;
+ const VkFormat depth_format = VK_FORMAT_D16_UNORM;
const VkImageCreateInfo image = {
.sType = VK_STRUCTURE_TYPE_IMAGE_CREATE_INFO,
.pNext = NULL,
- .imageType = VK_IMAGE_2D,
+ .imageType = VK_IMAGE_TYPE_2D,
.format = depth_format,
.extent = { demo->width, demo->height, 1 },
.mipLevels = 1,
.arraySize = 1,
.samples = 1,
- .tiling = VK_OPTIMAL_TILING,
+ .tiling = VK_IMAGE_TILING_OPTIMAL,
.usage = VK_IMAGE_USAGE_DEPTH_STENCIL_BIT,
.flags = 0,
};
@@ -402,7 +393,7 @@
.sType = VK_STRUCTURE_TYPE_MEMORY_ALLOC_INFO,
.pNext = NULL,
.allocationSize = 0,
- .memProps = VK_MEMORY_PROPERTY_GPU_ONLY,
+ .memProps = VK_MEMORY_PROPERTY_DEVICE_ONLY,
.memPriority = VK_MEMORY_PRIORITY_NORMAL,
};
VkDepthStencilViewCreateInfo view = {
@@ -428,13 +419,13 @@
&demo->depth.image);
assert(!err);
- err = vkGetObjectInfo(demo->depth.image, VK_INFO_TYPE_MEMORY_ALLOCATION_COUNT, &num_alloc_size, &num_allocations);
+ err = vkGetObjectInfo(demo->depth.image, VK_OBJECT_INFO_TYPE_MEMORY_ALLOCATION_COUNT, &num_alloc_size, &num_allocations);
assert(!err && num_alloc_size == sizeof(num_allocations));
mem_reqs = malloc(num_allocations * sizeof(VkMemoryRequirements));
- demo->depth.mem = malloc(num_allocations * sizeof(VkGpuMemory));
+ demo->depth.mem = malloc(num_allocations * sizeof(VkDeviceMemory));
demo->depth.num_mem = num_allocations;
err = vkGetObjectInfo(demo->depth.image,
- VK_INFO_TYPE_MEMORY_REQUIREMENTS,
+ VK_OBJECT_INFO_TYPE_MEMORY_REQUIREMENTS,
&mem_reqs_size, mem_reqs);
assert(!err && mem_reqs_size == num_allocations * sizeof(VkMemoryRequirements));
for (uint32_t i = 0; i < num_allocations; i ++) {
@@ -470,7 +461,7 @@
VkImageTiling tiling,
VkFlags mem_props)
{
- const VkFormat tex_format = VK_FMT_B8G8R8A8_UNORM;
+ const VkFormat tex_format = VK_FORMAT_B8G8R8A8_UNORM;
const int32_t tex_width = 2;
const int32_t tex_height = 2;
VkResult err;
@@ -481,7 +472,7 @@
const VkImageCreateInfo image_create_info = {
.sType = VK_STRUCTURE_TYPE_IMAGE_CREATE_INFO,
.pNext = NULL,
- .imageType = VK_IMAGE_2D,
+ .imageType = VK_IMAGE_TYPE_2D,
.format = tex_format,
.extent = { tex_width, tex_height, 1 },
.mipLevels = 1,
@@ -509,16 +500,16 @@
assert(!err);
err = vkGetObjectInfo(tex_obj->image,
- VK_INFO_TYPE_MEMORY_ALLOCATION_COUNT,
+ VK_OBJECT_INFO_TYPE_MEMORY_ALLOCATION_COUNT,
&num_alloc_size, &num_allocations);
assert(!err && num_alloc_size == sizeof(num_allocations));
mem_reqs = malloc(num_allocations * sizeof(VkMemoryRequirements));
- tex_obj->mem = malloc(num_allocations * sizeof(VkGpuMemory));
+ tex_obj->mem = malloc(num_allocations * sizeof(VkDeviceMemory));
err = vkGetObjectInfo(tex_obj->image,
- VK_INFO_TYPE_MEMORY_REQUIREMENTS,
+ VK_OBJECT_INFO_TYPE_MEMORY_REQUIREMENTS,
&mem_reqs_size, mem_reqs);
assert(!err && mem_reqs_size == num_allocations * sizeof(VkMemoryRequirements));
- mem_alloc.memProps = VK_MEMORY_PROPERTY_CPU_VISIBLE_BIT;
+ mem_alloc.memProps = VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT;
for (uint32_t j = 0; j < num_allocations; j ++) {
mem_alloc.allocationSize = mem_reqs[j].size;
@@ -536,7 +527,7 @@
tex_obj->num_mem = num_allocations;
- if (mem_props & VK_MEMORY_PROPERTY_CPU_VISIBLE_BIT) {
+ if (mem_props & VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT) {
const VkImageSubresource subres = {
.aspect = VK_IMAGE_ASPECT_COLOR,
.mipLevel = 0,
@@ -548,7 +539,7 @@
int32_t x, y;
err = vkGetImageSubresourceInfo(tex_obj->image, &subres,
- VK_INFO_TYPE_SUBRESOURCE_LAYOUT,
+ VK_SUBRESOURCE_INFO_TYPE_LAYOUT,
&layout_size, &layout);
assert(!err && layout_size == sizeof(layout));
/* Linear texture must be within a single memory object */
@@ -588,7 +579,7 @@
static void demo_prepare_textures(struct demo *demo)
{
- const VkFormat tex_format = VK_FMT_B8G8R8A8_UNORM;
+ const VkFormat tex_format = VK_FORMAT_B8G8R8A8_UNORM;
VkFormatProperties props;
size_t size = sizeof(props);
const uint32_t tex_colors[DEMO_TEXTURE_COUNT][2] = {
@@ -598,25 +589,25 @@
uint32_t i;
err = vkGetFormatInfo(demo->device, tex_format,
- VK_INFO_TYPE_FORMAT_PROPERTIES,
+ VK_FORMAT_INFO_TYPE_PROPERTIES,
&size, &props);
assert(!err);
for (i = 0; i < DEMO_TEXTURE_COUNT; i++) {
- if ((props.linearTilingFeatures & VK_FORMAT_SAMPLED_IMAGE_BIT) && !demo->use_staging_buffer) {
+ if ((props.linearTilingFeatures & VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT) && !demo->use_staging_buffer) {
/* Device can texture using linear textures */
demo_prepare_texture_image(demo, tex_colors[i], &demo->textures[i],
- VK_LINEAR_TILING, VK_MEMORY_PROPERTY_CPU_VISIBLE_BIT);
- } else if (props.optimalTilingFeatures & VK_FORMAT_SAMPLED_IMAGE_BIT){
+ VK_IMAGE_TILING_LINEAR, VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT);
+ } else if (props.optimalTilingFeatures & VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT){
/* Must use staging buffer to copy linear texture to optimized */
struct texture_object staging_texture;
memset(&staging_texture, 0, sizeof(staging_texture));
demo_prepare_texture_image(demo, tex_colors[i], &staging_texture,
- VK_LINEAR_TILING, VK_MEMORY_PROPERTY_CPU_VISIBLE_BIT);
+ VK_IMAGE_TILING_LINEAR, VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT);
demo_prepare_texture_image(demo, tex_colors[i], &demo->textures[i],
- VK_OPTIMAL_TILING, VK_MEMORY_PROPERTY_GPU_ONLY);
+ VK_IMAGE_TILING_OPTIMAL, VK_MEMORY_PROPERTY_DEVICE_ONLY);
demo_set_image_layout(demo, staging_texture.image,
staging_texture.imageLayout,
@@ -650,7 +641,7 @@
demo_destroy_texture_image(demo, &staging_texture);
demo_remove_mem_refs(demo, staging_texture.num_mem, staging_texture.mem);
} else {
- /* Can't support VK_FMT_B8G8R8A8_UNORM !? */
+ /* Can't support VK_FORMAT_B8G8R8A8_UNORM !? */
assert(!"No support for B8G8R8A8_UNORM as texture image format");
}
@@ -659,22 +650,22 @@
.pNext = NULL,
.magFilter = VK_TEX_FILTER_NEAREST,
.minFilter = VK_TEX_FILTER_NEAREST,
- .mipMode = VK_TEX_MIPMAP_BASE,
+ .mipMode = VK_TEX_MIPMAP_MODE_BASE,
.addressU = VK_TEX_ADDRESS_WRAP,
.addressV = VK_TEX_ADDRESS_WRAP,
.addressW = VK_TEX_ADDRESS_WRAP,
.mipLodBias = 0.0f,
.maxAnisotropy = 1,
- .compareFunc = VK_COMPARE_NEVER,
+ .compareOp = VK_COMPARE_OP_NEVER,
.minLod = 0.0f,
.maxLod = 0.0f,
- .borderColorType = VK_BORDER_COLOR_OPAQUE_WHITE,
+ .borderColor = VK_BORDER_COLOR_OPAQUE_WHITE,
};
VkImageViewCreateInfo view = {
.sType = VK_STRUCTURE_TYPE_IMAGE_VIEW_CREATE_INFO,
.pNext = NULL,
.image = VK_NULL_HANDLE,
- .viewType = VK_IMAGE_VIEW_2D,
+ .viewType = VK_IMAGE_VIEW_TYPE_2D,
.format = tex_format,
.channels = { VK_CHANNEL_SWIZZLE_R,
VK_CHANNEL_SWIZZLE_G,
@@ -716,7 +707,7 @@
.sType = VK_STRUCTURE_TYPE_MEMORY_ALLOC_INFO,
.pNext = NULL,
.allocationSize = 0,
- .memProps = VK_MEMORY_PROPERTY_CPU_VISIBLE_BIT,
+ .memProps = VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT,
.memPriority = VK_MEMORY_PRIORITY_NORMAL,
};
VkMemoryRequirements *mem_reqs;
@@ -732,14 +723,14 @@
assert(!err);
err = vkGetObjectInfo(demo->vertices.buf,
- VK_INFO_TYPE_MEMORY_ALLOCATION_COUNT,
+ VK_OBJECT_INFO_TYPE_MEMORY_ALLOCATION_COUNT,
&num_alloc_size, &num_allocations);
assert(!err && num_alloc_size == sizeof(num_allocations));
mem_reqs = malloc(num_allocations * sizeof(VkMemoryRequirements));
- demo->vertices.mem = malloc(num_allocations * sizeof(VkGpuMemory));
+ demo->vertices.mem = malloc(num_allocations * sizeof(VkDeviceMemory));
demo->vertices.num_mem = num_allocations;
err = vkGetObjectInfo(demo->vertices.buf,
- VK_INFO_TYPE_MEMORY_REQUIREMENTS,
+ VK_OBJECT_INFO_TYPE_MEMORY_REQUIREMENTS,
&mem_reqs_size, mem_reqs);
assert(!err && mem_reqs_size == sizeof(*mem_reqs));
for (uint32_t i = 0; i < num_allocations; i ++) {
@@ -775,12 +766,12 @@
demo->vertices.vi_attrs[0].binding = VERTEX_BUFFER_BIND_ID;
demo->vertices.vi_attrs[0].location = 0;
- demo->vertices.vi_attrs[0].format = VK_FMT_R32G32B32_SFLOAT;
+ demo->vertices.vi_attrs[0].format = VK_FORMAT_R32G32B32_SFLOAT;
demo->vertices.vi_attrs[0].offsetInBytes = 0;
demo->vertices.vi_attrs[1].binding = VERTEX_BUFFER_BIND_ID;
demo->vertices.vi_attrs[1].location = 1;
- demo->vertices.vi_attrs[1].format = VK_FMT_R32G32_SFLOAT;
+ demo->vertices.vi_attrs[1].format = VK_FORMAT_R32G32_SFLOAT;
demo->vertices.vi_attrs[1].offsetInBytes = sizeof(float) * 3;
}
@@ -789,7 +780,7 @@
const VkDescriptorSetLayoutBinding layout_binding = {
.descriptorType = VK_DESCRIPTOR_TYPE_COMBINED_IMAGE_SAMPLER,
.count = DEMO_TEXTURE_COUNT,
- .stageFlags = VK_SHADER_STAGE_FLAGS_FRAGMENT_BIT,
+ .stageFlags = VK_SHADER_STAGE_FRAGMENT_BIT,
.pImmutableSamplers = NULL,
};
const VkDescriptorSetLayoutCreateInfo descriptor_layout = {
@@ -810,7 +801,7 @@
}
static VkShader demo_prepare_shader(struct demo *demo,
- VkPipelineShaderStage stage,
+ VkShaderStage stage,
const void *code,
size_t size)
{
@@ -827,7 +818,7 @@
createInfo.pCode = malloc(createInfo.codeSize);
createInfo.flags = 0;
- /* try version 0 first: VkPipelineShaderStage followed by GLSL */
+ /* try version 0 first: VkShaderStage followed by GLSL */
((uint32_t *) createInfo.pCode)[0] = ICD_SPV_MAGIC;
((uint32_t *) createInfo.pCode)[1] = 0;
((uint32_t *) createInfo.pCode)[2] = stage;
@@ -900,12 +891,12 @@
memset(&ia, 0, sizeof(ia));
ia.sType = VK_STRUCTURE_TYPE_PIPELINE_IA_STATE_CREATE_INFO;
- ia.topology = VK_TOPOLOGY_TRIANGLE_LIST;
+ ia.topology = VK_PRIMITIVE_TOPOLOGY_TRIANGLE_LIST;
memset(&rs, 0, sizeof(rs));
rs.sType = VK_STRUCTURE_TYPE_PIPELINE_RS_STATE_CREATE_INFO;
- rs.fillMode = VK_FILL_SOLID;
- rs.cullMode = VK_CULL_NONE;
+ rs.fillMode = VK_FILL_MODE_SOLID;
+ rs.cullMode = VK_CULL_MODE_NONE;
rs.frontFace = VK_FRONT_FACE_CCW;
memset(&cb, 0, sizeof(cb));
@@ -921,7 +912,7 @@
memset(&vp, 0, sizeof(vp));
vp.sType = VK_STRUCTURE_TYPE_PIPELINE_VP_STATE_CREATE_INFO;
- vp.numViewports = 1;
+ vp.viewportCount = 1;
vp.clipOrigin = VK_COORDINATE_ORIGIN_UPPER_LEFT;
memset(&ds, 0, sizeof(ds));
@@ -929,11 +920,11 @@
ds.format = demo->depth.format;
ds.depthTestEnable = VK_TRUE;
ds.depthWriteEnable = VK_TRUE;
- ds.depthFunc = VK_COMPARE_LESS_EQUAL;
+ ds.depthCompareOp = VK_COMPARE_OP_LESS_EQUAL;
ds.depthBoundsEnable = VK_FALSE;
ds.back.stencilFailOp = VK_STENCIL_OP_KEEP;
ds.back.stencilPassOp = VK_STENCIL_OP_KEEP;
- ds.back.stencilFunc = VK_COMPARE_ALWAYS;
+ ds.back.stencilCompareOp = VK_COMPARE_OP_ALWAYS;
ds.stencilTestEnable = VK_FALSE;
ds.front = ds.back;
@@ -1266,24 +1257,24 @@
err = vkCreateDevice(demo->gpu, &device, &demo->device);
assert(!err);
- err = vkGetGpuInfo(demo->gpu, VK_INFO_TYPE_PHYSICAL_GPU_PROPERTIES,
+ err = vkGetPhysicalDeviceInfo(demo->gpu, VK_PHYSICAL_DEVICE_INFO_TYPE_PROPERTIES,
&data_size, NULL);
assert(!err);
- demo->gpu_props = (VkPhysicalGpuProperties *) malloc(data_size);
- err = vkGetGpuInfo(demo->gpu, VK_INFO_TYPE_PHYSICAL_GPU_PROPERTIES,
+ demo->gpu_props = (VkPhysicalDeviceProperties *) malloc(data_size);
+ err = vkGetPhysicalDeviceInfo(demo->gpu, VK_PHYSICAL_DEVICE_INFO_TYPE_PROPERTIES,
&data_size, demo->gpu_props);
assert(!err);
- err = vkGetGpuInfo(demo->gpu, VK_INFO_TYPE_PHYSICAL_GPU_QUEUE_PROPERTIES,
+ err = vkGetPhysicalDeviceInfo(demo->gpu, VK_PHYSICAL_DEVICE_INFO_TYPE_QUEUE_PROPERTIES,
&data_size, NULL);
assert(!err);
- demo->queue_props = (VkPhysicalGpuQueueProperties *) malloc(data_size);
- err = vkGetGpuInfo(demo->gpu, VK_INFO_TYPE_PHYSICAL_GPU_QUEUE_PROPERTIES,
+ demo->queue_props = (VkPhysicalDeviceQueueProperties *) malloc(data_size);
+ err = vkGetPhysicalDeviceInfo(demo->gpu, VK_PHYSICAL_DEVICE_INFO_TYPE_QUEUE_PROPERTIES,
&data_size, demo->queue_props);
assert(!err);
- queue_count = (uint32_t) (data_size / sizeof(VkPhysicalGpuQueueProperties));
+ queue_count = (uint32_t) (data_size / sizeof(VkPhysicalDeviceQueueProperties));
assert(queue_count >= 1);
for (i = 0; i < queue_count; i++) {
@@ -1336,7 +1327,7 @@
demo->width = 300;
demo->height = 300;
- demo->format = VK_FMT_B8G8R8A8_UNORM;
+ demo->format = VK_FORMAT_B8G8R8A8_UNORM;
}
static void demo_cleanup(struct demo *demo)
diff --git a/demos/vulkaninfo.c b/demos/vulkaninfo.c
index 4189e73..2fc76eb 100644
--- a/demos/vulkaninfo.c
+++ b/demos/vulkaninfo.c
@@ -48,21 +48,21 @@
VkDevice obj;
- VkFormatProperties format_props[VK_NUM_FMT];
+ VkFormatProperties format_props[VK_NUM_FORMAT];
};
struct app_gpu {
uint32_t id;
- VkPhysicalGpu obj;
+ VkPhysicalDevice obj;
- VkPhysicalGpuProperties props;
- VkPhysicalGpuPerformance perf;
+ VkPhysicalDeviceProperties props;
+ VkPhysicalDevicePerformance perf;
uint32_t queue_count;
- VkPhysicalGpuQueueProperties *queue_props;
+ VkPhysicalDeviceQueueProperties *queue_props;
VkDeviceQueueCreateInfo *queue_reqs;
- VkPhysicalGpuMemoryProperties memory_props;
+ VkPhysicalDeviceMemoryProperties memory_props;
uint32_t extension_count;
char **extensions;
@@ -83,8 +83,8 @@
STR(VK_ERROR_UNKNOWN);
STR(VK_ERROR_UNAVAILABLE);
STR(VK_ERROR_INITIALIZATION_FAILED);
- STR(VK_ERROR_OUT_OF_MEMORY);
- STR(VK_ERROR_OUT_OF_GPU_MEMORY);
+ STR(VK_ERROR_OUT_OF_HOST_MEMORY);
+ STR(VK_ERROR_OUT_OF_DEVICE_MEMORY);
STR(VK_ERROR_DEVICE_ALREADY_CREATED);
STR(VK_ERROR_DEVICE_LOST);
STR(VK_ERROR_INVALID_POINTER);
@@ -119,23 +119,23 @@
}
}
-static const char *vk_gpu_type_string(VkPhysicalGpuType type)
+static const char *vk_physical_device_type_string(VkPhysicalDeviceType type)
{
switch (type) {
-#define STR(r) case VK_GPU_TYPE_ ##r: return #r
+#define STR(r) case VK_PHYSICAL_DEVICE_TYPE_ ##r: return #r
STR(OTHER);
- STR(INTEGRATED);
- STR(DISCRETE);
- STR(VIRTUAL);
+ STR(INTEGRATED_GPU);
+ STR(DISCRETE_GPU);
+ STR(VIRTUAL_GPU);
#undef STR
- default: return "UNKNOWN_GPU";
+ default: return "UNKNOWN_DEVICE";
}
}
static const char *vk_format_string(VkFormat fmt)
{
switch (fmt) {
-#define STR(r) case VK_FMT_ ##r: return #r
+#define STR(r) case VK_FORMAT_ ##r: return #r
STR(UNDEFINED);
STR(R4G4_UNORM);
STR(R4G4_USCALED);
@@ -312,13 +312,13 @@
{
VkFormat f;
- for (f = 0; f < VK_NUM_FMT; f++) {
+ for (f = 0; f < VK_NUM_FORMAT; f++) {
const VkFormat fmt = f;
VkResult err;
size_t size = sizeof(dev->format_props[f]);
err = vkGetFormatInfo(dev->obj, fmt,
- VK_INFO_TYPE_FORMAT_PROPERTIES,
+ VK_FORMAT_INFO_TYPE_PROPERTIES,
&size, &dev->format_props[f]);
if (err) {
memset(&dev->format_props[f], 0,
@@ -384,7 +384,7 @@
gpu->extensions =
malloc(sizeof(gpu->extensions[0]) * gpu->extension_count);
if (!gpu->extensions)
- ERR_EXIT(VK_ERROR_OUT_OF_MEMORY);
+ ERR_EXIT(VK_ERROR_OUT_OF_HOST_MEMORY);
gpu->extension_count = 0;
for (i = 0; i < ARRAY_SIZE(known_extensions); i++) {
@@ -396,7 +396,7 @@
}
}
-static void app_gpu_init(struct app_gpu *gpu, uint32_t id, VkPhysicalGpu obj)
+static void app_gpu_init(struct app_gpu *gpu, uint32_t id, VkPhysicalDevice obj)
{
size_t size;
VkResult err;
@@ -407,22 +407,22 @@
gpu->id = id;
gpu->obj = obj;
size = sizeof(gpu->props);
- err = vkGetGpuInfo(gpu->obj,
- VK_INFO_TYPE_PHYSICAL_GPU_PROPERTIES,
+ err = vkGetPhysicalDeviceInfo(gpu->obj,
+ VK_PHYSICAL_DEVICE_INFO_TYPE_PROPERTIES,
&size, &gpu->props);
if (err || size != sizeof(gpu->props))
ERR_EXIT(err);
size = sizeof(gpu->perf);
- err = vkGetGpuInfo(gpu->obj,
- VK_INFO_TYPE_PHYSICAL_GPU_PERFORMANCE,
+ err = vkGetPhysicalDeviceInfo(gpu->obj,
+ VK_PHYSICAL_DEVICE_INFO_TYPE_PERFORMANCE,
&size, &gpu->perf);
if (err || size != sizeof(gpu->perf))
ERR_EXIT(err);
/* get queue count */
- err = vkGetGpuInfo(gpu->obj,
- VK_INFO_TYPE_PHYSICAL_GPU_QUEUE_PROPERTIES,
+ err = vkGetPhysicalDeviceInfo(gpu->obj,
+ VK_PHYSICAL_DEVICE_INFO_TYPE_QUEUE_PROPERTIES,
&size, NULL);
if (err || size % sizeof(gpu->queue_props[0]))
ERR_EXIT(err);
@@ -432,9 +432,9 @@
malloc(sizeof(gpu->queue_props[0]) * gpu->queue_count);
size = sizeof(gpu->queue_props[0]) * gpu->queue_count;
if (!gpu->queue_props)
- ERR_EXIT(VK_ERROR_OUT_OF_MEMORY);
- err = vkGetGpuInfo(gpu->obj,
- VK_INFO_TYPE_PHYSICAL_GPU_QUEUE_PROPERTIES,
+ ERR_EXIT(VK_ERROR_OUT_OF_HOST_MEMORY);
+ err = vkGetPhysicalDeviceInfo(gpu->obj,
+ VK_PHYSICAL_DEVICE_INFO_TYPE_QUEUE_PROPERTIES,
&size, gpu->queue_props);
if (err || size != sizeof(gpu->queue_props[0]) * gpu->queue_count)
ERR_EXIT(err);
@@ -443,15 +443,15 @@
size = sizeof(*gpu->queue_reqs) * gpu->queue_count;
gpu->queue_reqs = malloc(sizeof(*gpu->queue_reqs) * gpu->queue_count);
if (!gpu->queue_reqs)
- ERR_EXIT(VK_ERROR_OUT_OF_MEMORY);
+ ERR_EXIT(VK_ERROR_OUT_OF_HOST_MEMORY);
for (i = 0; i < gpu->queue_count; i++) {
gpu->queue_reqs[i].queueNodeIndex = i;
gpu->queue_reqs[i].queueCount = gpu->queue_props[i].queueCount;
}
size = sizeof(gpu->memory_props);
- err = vkGetGpuInfo(gpu->obj,
- VK_INFO_TYPE_PHYSICAL_GPU_MEMORY_PROPERTIES,
+ err = vkGetPhysicalDeviceInfo(gpu->obj,
+ VK_PHYSICAL_DEVICE_INFO_TYPE_MEMORY_PROPERTIES,
&size, &gpu->memory_props);
if (err || size != sizeof(gpu->memory_props))
ERR_EXIT(err);
@@ -492,21 +492,21 @@
continue;
printf("\t%s tiling image =%s%s%s\n", tilings[i].name,
- (tilings[i].flags & VK_FORMAT_SAMPLED_IMAGE_BIT) ? " sampled" : "",
- (tilings[i].flags & VK_FORMAT_STORAGE_IMAGE_BIT) ? " storage" : "",
- (tilings[i].flags & VK_FORMAT_STORAGE_IMAGE_ATOMIC_BIT) ? " atomic" : "");
+ (tilings[i].flags & VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT) ? " sampled" : "",
+ (tilings[i].flags & VK_FORMAT_FEATURE_STORAGE_IMAGE_BIT) ? " storage" : "",
+ (tilings[i].flags & VK_FORMAT_FEATURE_STORAGE_IMAGE_ATOMIC_BIT) ? " atomic" : "");
printf("\t%s tiling texel =%s%s%s\n", tilings[i].name,
- (tilings[i].flags & VK_FORMAT_UNIFORM_TEXEL_BUFFER_BIT) ? " TBO" : "",
- (tilings[i].flags & VK_FORMAT_STORAGE_TEXEL_BUFFER_BIT) ? " IBO" : "",
- (tilings[i].flags & VK_FORMAT_STORAGE_TEXEL_BUFFER_ATOMIC_BIT) ? " atomic" : "");
+ (tilings[i].flags & VK_FORMAT_FEATURE_UNIFORM_TEXEL_BUFFER_BIT) ? " TBO" : "",
+ (tilings[i].flags & VK_FORMAT_FEATURE_STORAGE_TEXEL_BUFFER_BIT) ? " IBO" : "",
+ (tilings[i].flags & VK_FORMAT_FEATURE_STORAGE_TEXEL_BUFFER_ATOMIC_BIT) ? " atomic" : "");
printf("\t%s tiling attachment =%s%s%s\n", tilings[i].name,
- (tilings[i].flags & VK_FORMAT_COLOR_ATTACHMENT_BIT) ? " color" : "",
- (tilings[i].flags & VK_FORMAT_COLOR_ATTACHMENT_BLEND_BIT) ? " blend" : "",
- (tilings[i].flags & VK_FORMAT_DEPTH_STENCIL_ATTACHMENT_BIT) ? " depth/stencil" : "");
+ (tilings[i].flags & VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT) ? " color" : "",
+ (tilings[i].flags & VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BLEND_BIT) ? " blend" : "",
+ (tilings[i].flags & VK_FORMAT_FEATURE_DEPTH_STENCIL_ATTACHMENT_BIT) ? " depth/stencil" : "");
printf("\t%s tiling vertex = %u\n", tilings[i].name,
- (bool) (tilings[i].flags & VK_FORMAT_VERTEX_BUFFER_BIT));
+ (bool) (tilings[i].flags & VK_FORMAT_FEATURE_VERTEX_BUFFER_BIT));
printf("\t%s tiling conversion = %u\n", tilings[i].name,
- (bool) (tilings[i].flags & VK_FORMAT_CONVERSION_BIT));
+ (bool) (tilings[i].flags & VK_FORMAT_FEATURE_CONVERSION_BIT));
}
}
@@ -516,24 +516,24 @@
{
VkFormat fmt;
- for (fmt = 0; fmt < VK_NUM_FMT; fmt++) {
+ for (fmt = 0; fmt < VK_NUM_FORMAT; fmt++) {
app_dev_dump_format_props(dev, fmt);
}
}
static void app_gpu_dump_multi_compat(const struct app_gpu *gpu, const struct app_gpu *other,
- const VkGpuCompatibilityInfo *info)
+ const VkPhysicalDeviceCompatibilityInfo *info)
{
- printf("VkGpuCompatibilityInfo[GPU%d]\n", other->id);
+ printf("VkPhysicalDeviceCompatibilityInfo[GPU%d]\n", other->id);
-#define TEST(info, b) printf(#b " = %u\n", (bool) (info->compatibilityFlags & VK_GPU_COMPAT_ ##b## _BIT))
- TEST(info, ASIC_FEATURES);
+#define TEST(info, b) printf(#b " = %u\n", (bool) (info->compatibilityFlags & VK_PHYSICAL_DEVICE_COMPATIBILITY_ ##b## _BIT))
+ TEST(info, FEATURES);
TEST(info, IQ_MATCH);
TEST(info, PEER_TRANSFER);
TEST(info, SHARED_MEMORY);
TEST(info, SHARED_SYNC);
- TEST(info, SHARED_GPU0_DISPLAY);
- TEST(info, SHARED_GPU1_DISPLAY);
+ TEST(info, SHARED_DEVICE0_DISPLAY);
+ TEST(info, SHARED_DEVICE1_DISPLAY);
#undef TEST
}
@@ -544,12 +544,12 @@
for (i = 0; i < gpu_count; i++) {
for (j = 0; j < gpu_count; j++) {
- VkGpuCompatibilityInfo info;
+ VkPhysicalDeviceCompatibilityInfo info;
if (i == j)
continue;
- err = vkGetMultiGpuCompatibility(gpus[i].obj,
+ err = vkGetMultiDeviceCompatibility(gpus[i].obj,
gpus[j].obj, &info);
if (err)
ERR_EXIT(err);
@@ -561,15 +561,15 @@
static void app_gpu_dump_props(const struct app_gpu *gpu)
{
- const VkPhysicalGpuProperties *props = &gpu->props;
+ const VkPhysicalDeviceProperties *props = &gpu->props;
- printf("VkPhysicalGpuProperties\n");
+ printf("VkPhysicalDeviceProperties\n");
printf("\tapiVersion = %u\n", props->apiVersion);
printf("\tdriverVersion = %u\n", props->driverVersion);
printf("\tvendorId = 0x%04x\n", props->vendorId);
printf("\tdeviceId = 0x%04x\n", props->deviceId);
- printf("\tgpuType = %s\n", vk_gpu_type_string(props->gpuType));
- printf("\tgpuName = %s\n", props->gpuName);
+ printf("\tdeviceType = %s\n", vk_physical_device_type_string(props->deviceType));
+ printf("\tdeviceName = %s\n", props->deviceName);
printf("\tmaxInlineMemoryUpdateSize = %zu\n", props->maxInlineMemoryUpdateSize);
printf("\tmaxBoundDescriptorSets = %u\n", props->maxBoundDescriptorSets);
printf("\tmaxThreadGroupSize = %u\n", props->maxThreadGroupSize);
@@ -579,10 +579,10 @@
static void app_gpu_dump_perf(const struct app_gpu *gpu)
{
- const VkPhysicalGpuPerformance *perf = &gpu->perf;
+ const VkPhysicalDevicePerformance *perf = &gpu->perf;
- printf("VkPhysicalGpuPerformance\n");
- printf("\tmaxGpuClock = %f\n", perf->maxGpuClock);
+ printf("VkPhysicalDevicePerformance\n");
+ printf("\tmaxGpuClock = %f\n", perf->maxDeviceClock);
printf("\taluPerClock = %f\n", perf->aluPerClock);
printf("\ttexPerClock = %f\n", perf->texPerClock);
printf("\tprimsPerClock = %f\n", perf->primsPerClock);
@@ -605,9 +605,9 @@
static void app_gpu_dump_queue_props(const struct app_gpu *gpu, uint32_t id)
{
- const VkPhysicalGpuQueueProperties *props = &gpu->queue_props[id];
+ const VkPhysicalDeviceQueueProperties *props = &gpu->queue_props[id];
- printf("VkPhysicalGpuQueueProperties[%d]\n", id);
+ printf("VkPhysicalDeviceQueueProperties[%d]\n", id);
printf("\tqueueFlags = %c%c%c%c\n",
(props->queueFlags & VK_QUEUE_GRAPHICS_BIT) ? 'G' : '.',
(props->queueFlags & VK_QUEUE_COMPUTE_BIT) ? 'C' : '.',
@@ -621,9 +621,9 @@
static void app_gpu_dump_memory_props(const struct app_gpu *gpu)
{
- const VkPhysicalGpuMemoryProperties *props = &gpu->memory_props;
+ const VkPhysicalDeviceMemoryProperties *props = &gpu->memory_props;
- printf("VkPhysicalGpuMemoryProperties\n");
+ printf("VkPhysicalDeviceMemoryProperties\n");
printf("\tsupportsMigration = %u\n", props->supportsMigration);
printf("\tsupportsPinning = %u\n", props->supportsPinning);
}
@@ -668,7 +668,7 @@
.ppEnabledExtensionNames = NULL,
};
struct app_gpu gpus[MAX_GPUS];
- VkPhysicalGpu objs[MAX_GPUS];
+ VkPhysicalDevice objs[MAX_GPUS];
VkInstance inst;
uint32_t gpu_count, i;
VkResult err;
diff --git a/docs/vk_ds.dot b/docs/vk_ds.dot
index 514614c..d587629 100644
--- a/docs/vk_ds.dot
+++ b/docs/vk_ds.dot
@@ -26,13 +26,13 @@
{
label="vkCreateSampler - multiple calls return unique VkSampler handles"
"_VK_SAMPLER_CREATE_INFO_0" [
-label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkSamplerCreateInfo</TD></TR><TR><TD PORT="f1">VkStructureType</TD><TD PORT="f2">sType=<BR/>VK_STRUCTURE_TYPE_SAMPLER_CREATE_INFO</TD></TR><TR><TD PORT="f3">const void*</TD><TD PORT="f4">pNext=NULL</TD></TR><TR><TD PORT="f5">VkTexFilter</TD><TD PORT="f6">magFilter</TD></TR><TR><TD PORT="f7">VkTexFilter</TD><TD PORT="f8">minFilter</TD></TR><TR><TD PORT="f9">VkTexMipmapMode</TD><TD PORT="f10">mipMode</TD></TR><TR><TD PORT="f11">VkTexAddress</TD><TD PORT="f12">addressU</TD></TR><TR><TD PORT="f13">VkTexAddress</TD><TD PORT="f14">addressV</TD></TR><TR><TD PORT="f15">VkTexAddress</TD><TD PORT="f16">addressW</TD></TR><TR><TD PORT="f17">float</TD><TD PORT="f18">mipLodBias</TD></TR><TR><TD PORT="f19">uint32_t</TD><TD PORT="f20">maxAnisotropy</TD></TR><TR><TD PORT="f21">VkCompareFunc</TD><TD PORT="f22">compareFunc</TD></TR><TR><TD PORT="f23">float</TD><TD PORT="f24">minLod</TD></TR><TR><TD PORT="f25">float</TD><TD PORT="f26">maxLod</TD></TR><TR><TD PORT="f27">VkBorderColorType</TD><TD PORT="f28">borderColorType</TD></TR></TABLE>>
+label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkSamplerCreateInfo</TD></TR><TR><TD PORT="f1">VkStructureType</TD><TD PORT="f2">sType=<BR/>VK_STRUCTURE_TYPE_SAMPLER_CREATE_INFO</TD></TR><TR><TD PORT="f3">const void*</TD><TD PORT="f4">pNext=NULL</TD></TR><TR><TD PORT="f5">VkTexFilter</TD><TD PORT="f6">magFilter</TD></TR><TR><TD PORT="f7">VkTexFilter</TD><TD PORT="f8">minFilter</TD></TR><TR><TD PORT="f9">VkTexMipmapMode</TD><TD PORT="f10">mipMode</TD></TR><TR><TD PORT="f11">VkTexAddress</TD><TD PORT="f12">addressU</TD></TR><TR><TD PORT="f13">VkTexAddress</TD><TD PORT="f14">addressV</TD></TR><TR><TD PORT="f15">VkTexAddress</TD><TD PORT="f16">addressW</TD></TR><TR><TD PORT="f17">float</TD><TD PORT="f18">mipLodBias</TD></TR><TR><TD PORT="f19">uint32_t</TD><TD PORT="f20">maxAnisotropy</TD></TR><TR><TD PORT="f21">VkCompareOp</TD><TD PORT="f22">compareFunc</TD></TR><TR><TD PORT="f23">float</TD><TD PORT="f24">minLod</TD></TR><TR><TD PORT="f25">float</TD><TD PORT="f26">maxLod</TD></TR><TR><TD PORT="f27">VkBorderColor</TD><TD PORT="f28">borderColorType</TD></TR></TABLE>>
];
"SAMPLER_ELLIPSES" [
label = "..."
];
"_VK_SAMPLER_CREATE_INFO_19" [
-label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkSamplerCreateInfo</TD></TR><TR><TD PORT="f1">VkStructureType</TD><TD PORT="f2">sType=<BR/>VK_STRUCTURE_TYPE_SAMPLER_CREATE_INFO</TD></TR><TR><TD PORT="f3">const void*</TD><TD PORT="f4">pNext=NULL</TD></TR><TR><TD PORT="f5">VkTexFilter</TD><TD PORT="f6">magFilter</TD></TR><TR><TD PORT="f7">VkTexFilter</TD><TD PORT="f8">minFilter</TD></TR><TR><TD PORT="f9">VkTexMipmapMode</TD><TD PORT="f10">mipMode</TD></TR><TR><TD PORT="f11">VkTexAddress</TD><TD PORT="f12">addressU</TD></TR><TR><TD PORT="f13">VkTexAddress</TD><TD PORT="f14">addressV</TD></TR><TR><TD PORT="f15">VkTexAddress</TD><TD PORT="f16">addressW</TD></TR><TR><TD PORT="f17">float</TD><TD PORT="f18">mipLodBias</TD></TR><TR><TD PORT="f19">uint32_t</TD><TD PORT="f20">maxAnisotropy</TD></TR><TR><TD PORT="f21">VkCompareFunc</TD><TD PORT="f22">compareFunc</TD></TR><TR><TD PORT="f23">float</TD><TD PORT="f24">minLod</TD></TR><TR><TD PORT="f25">float</TD><TD PORT="f26">maxLod</TD></TR><TR><TD PORT="f27">VkBorderColorType</TD><TD PORT="f28">borderColorType</TD></TR></TABLE>>
+label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkSamplerCreateInfo</TD></TR><TR><TD PORT="f1">VkStructureType</TD><TD PORT="f2">sType=<BR/>VK_STRUCTURE_TYPE_SAMPLER_CREATE_INFO</TD></TR><TR><TD PORT="f3">const void*</TD><TD PORT="f4">pNext=NULL</TD></TR><TR><TD PORT="f5">VkTexFilter</TD><TD PORT="f6">magFilter</TD></TR><TR><TD PORT="f7">VkTexFilter</TD><TD PORT="f8">minFilter</TD></TR><TR><TD PORT="f9">VkTexMipmapMode</TD><TD PORT="f10">mipMode</TD></TR><TR><TD PORT="f11">VkTexAddress</TD><TD PORT="f12">addressU</TD></TR><TR><TD PORT="f13">VkTexAddress</TD><TD PORT="f14">addressV</TD></TR><TR><TD PORT="f15">VkTexAddress</TD><TD PORT="f16">addressW</TD></TR><TR><TD PORT="f17">float</TD><TD PORT="f18">mipLodBias</TD></TR><TR><TD PORT="f19">uint32_t</TD><TD PORT="f20">maxAnisotropy</TD></TR><TR><TD PORT="f21">VkCompareOp</TD><TD PORT="f22">compareFunc</TD></TR><TR><TD PORT="f23">float</TD><TD PORT="f24">minLod</TD></TR><TR><TD PORT="f25">float</TD><TD PORT="f26">maxLod</TD></TR><TR><TD PORT="f27">VkBorderColor</TD><TD PORT="f28">borderColorType</TD></TR></TABLE>>
];
}
subgraph clusterSamplerAttach
@@ -52,13 +52,13 @@
{
label="vkAttachMemoryViewDescriptors - pMemViews array of VK_MEMORY_VIEW_ATTACH_INFO structs"
"_VK_MEMORY_VIEW_ATTACH_INFO_3" [
-label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VK_MEMORY_VIEW_ATTACH_INFO</TD></TR><TR><TD PORT="f1">VkStructureType</TD><TD PORT="f2">sType=<BR/>VK_STRUCTURE_TYPE_MEMORY_VIEW_ATTACH_INFO</TD></TR><TR><TD PORT="f3">void*</TD><TD PORT="f4">pNext=NULL</TD></TR><TR><TD PORT="f5">VkGpuMemory</TD><TD PORT="f6">mem</TD></TR><TR><TD PORT="f7">VkGpuSize</TD><TD PORT="f8">offset</TD></TR><TR><TD PORT="f9">VkGpuSize</TD><TD PORT="f10">range</TD></TR><TR><TD PORT="f11">VkGpuSize</TD><TD PORT="f12">stride</TD></TR><TR><TD PORT="f13">VkFormat</TD><TD PORT="f14">format</TD></TR><TR><TD PORT="f15">VK_MEMORY_STATE</TD><TD PORT="f16">state</TD></TR></TABLE>>
+label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VK_MEMORY_VIEW_ATTACH_INFO</TD></TR><TR><TD PORT="f1">VkStructureType</TD><TD PORT="f2">sType=<BR/>VK_STRUCTURE_TYPE_MEMORY_VIEW_ATTACH_INFO</TD></TR><TR><TD PORT="f3">void*</TD><TD PORT="f4">pNext=NULL</TD></TR><TR><TD PORT="f5">VkDeviceMemory</TD><TD PORT="f6">mem</TD></TR><TR><TD PORT="f7">VkDeviceSize</TD><TD PORT="f8">offset</TD></TR><TR><TD PORT="f9">VkDeviceSize</TD><TD PORT="f10">range</TD></TR><TR><TD PORT="f11">VkDeviceSize</TD><TD PORT="f12">stride</TD></TR><TR><TD PORT="f13">VkFormat</TD><TD PORT="f14">format</TD></TR><TR><TD PORT="f15">VK_MEMORY_STATE</TD><TD PORT="f16">state</TD></TR></TABLE>>
];
"MEM_VIEW_ELLIPSES" [
label = "..."
];
"_VK_MEMORY_VIEW_ATTACH_INFO_0" [
-label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VK_MEMORY_VIEW_ATTACH_INFO</TD></TR><TR><TD PORT="f1">VkStructureType</TD><TD PORT="f2">sType=<BR/>VK_STRUCTURE_TYPE_MEMORY_VIEW_ATTACH_INFO</TD></TR><TR><TD PORT="f3">void*</TD><TD PORT="f4">pNext=NULL</TD></TR><TR><TD PORT="f5">VkGpuMemory</TD><TD PORT="f6">mem</TD></TR><TR><TD PORT="f7">VkGpuSize</TD><TD PORT="f8">offset</TD></TR><TR><TD PORT="f9">VkGpuSize</TD><TD PORT="f10">range</TD></TR><TR><TD PORT="f11">VkGpuSize</TD><TD PORT="f12">stride</TD></TR><TR><TD PORT="f13">VkFormat</TD><TD PORT="f14">format</TD></TR><TR><TD PORT="f15">VK_MEMORY_STATE</TD><TD PORT="f16">state</TD></TR></TABLE>>
+label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VK_MEMORY_VIEW_ATTACH_INFO</TD></TR><TR><TD PORT="f1">VkStructureType</TD><TD PORT="f2">sType=<BR/>VK_STRUCTURE_TYPE_MEMORY_VIEW_ATTACH_INFO</TD></TR><TR><TD PORT="f3">void*</TD><TD PORT="f4">pNext=NULL</TD></TR><TR><TD PORT="f5">VkDeviceMemory</TD><TD PORT="f6">mem</TD></TR><TR><TD PORT="f7">VkDeviceSize</TD><TD PORT="f8">offset</TD></TR><TR><TD PORT="f9">VkDeviceSize</TD><TD PORT="f10">range</TD></TR><TR><TD PORT="f11">VkDeviceSize</TD><TD PORT="f12">stride</TD></TR><TR><TD PORT="f13">VkFormat</TD><TD PORT="f14">format</TD></TR><TR><TD PORT="f15">VK_MEMORY_STATE</TD><TD PORT="f16">state</TD></TR></TABLE>>
];
}
subgraph clusterImageView
diff --git a/docs/vk_full_pipeline_ds.dot b/docs/vk_full_pipeline_ds.dot
index 894e235..15f0c26 100644
--- a/docs/vk_full_pipeline_ds.dot
+++ b/docs/vk_full_pipeline_ds.dot
@@ -12,7 +12,7 @@
label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkGraphicsPipelineCreateInfo</TD></TR><TR><TD>VkStructureType</TD><TD>sType=<BR/>VK_STRUCTURE_TYPE_GRAPHICS_PIPELINE_CREATE_INFO</TD></TR><TR><TD>const void*</TD><TD PORT="f2">pNext</TD></TR><TR><TD>VkFlags</TD><TD>flags</TD></TR></TABLE>>
];
"VkPipelineIaStateCreateInfo_" [
-label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkPipelineIaStateCreateInfo</TD></TR><TR><TD>VkStructureType</TD><TD>sType=<BR/>VK_STRUCTURE_TYPE_PIPELINE_IA_STATE_CREATE_INFO</TD></TR><TR><TD>const void*</TD><TD PORT="f2">pNext</TD></TR><TR><TD>VkPrimitiveTopology</TD><TD>topology</TD></TR><TR><TD>bool32_t</TD><TD>disableVertexReuse</TD></TR><TR><TD>VkProvokingVertexConvention</TD><TD>provokingVertex</TD></TR><TR><TD>bool32_t</TD><TD>primitiveRestartEnable</TD></TR><TR><TD>uint32_t</TD><TD>primitiveRestartIndex</TD></TR></TABLE>>
+label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkPipelineIaStateCreateInfo</TD></TR><TR><TD>VkStructureType</TD><TD>sType=<BR/>VK_STRUCTURE_TYPE_PIPELINE_IA_STATE_CREATE_INFO</TD></TR><TR><TD>const void*</TD><TD PORT="f2">pNext</TD></TR><TR><TD>VkPrimitiveTopology</TD><TD>topology</TD></TR><TR><TD>bool32_t</TD><TD>disableVertexReuse</TD></TR><TR><TD>VkProvokingVertex</TD><TD>provokingVertex</TD></TR><TR><TD>bool32_t</TD><TD>primitiveRestartEnable</TD></TR><TR><TD>uint32_t</TD><TD>primitiveRestartIndex</TD></TR></TABLE>>
];
"VkPipelineTessStateCreateInfo_" [
label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkPipelineTessStateCreateInfo</TD></TR><TR><TD>VkStructureType</TD><TD>sType=<BR/>VK_STRUCTURE_TYPE_PIPELINE_TESS_STATE_CREATE_INFO</TD></TR><TR><TD>const void*</TD><TD PORT="f2">pNext</TD></TR><TR><TD>uint32_t</TD><TD>patchControlPoints</TD></TR> <TR><TD>float</TD><TD>optimalTessFactor</TD></TR><TR><TD>float</TD><TD>fixedTessFactor</TD></TR></TABLE>>
@@ -42,19 +42,19 @@
label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkPipelineShaderStageCreateInfo</TD></TR><TR><TD>VkStructureType</TD><TD>sType=<BR/>VK_STRUCTURE_TYPE_PIPELINE_SHADER_STAGE_CREATE_INFO</TD></TR> <TR><TD>const void*</TD><TD PORT="f2">pNext=NULL</TD></TR> <TR><TD>VkPipelineShader</TD><TD PORT="f4">shader</TD></TR> </TABLE>>
];
"VS_VK_PIPELINE_SHADER" [
-label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkPipelineShader</TD></TR><TR><TD PORT="f1">VkPipelineShaderStage</TD><TD PORT="f2">stage=VS</TD></TR><TR><TD PORT="f3">VkShader</TD><TD PORT="f4">shader</TD></TR><TR><TD PORT="f5">VK_DESCRIPTOR_SET_MAPPING</TD><TD PORT="f6">descriptorSetMapping[2]</TD></TR><TR><TD PORT="f7">uint32_t</TD><TD PORT="f8">linkConstBufferCount</TD></TR><TR><TD PORT="f9">const VkLinkConstBuffer*</TD><TD PORT="f10">pLinkConstBufferInfo</TD></TR><TR><TD PORT="f11">VK_DYNAMIC_MEMORY_VIEW_SLOT_INFO</TD><TD PORT="f12">dynamicMemoryViewMapping</TD></TR></TABLE>>
+label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkPipelineShader</TD></TR><TR><TD PORT="f1">VkShaderStage</TD><TD PORT="f2">stage=VS</TD></TR><TR><TD PORT="f3">VkShader</TD><TD PORT="f4">shader</TD></TR><TR><TD PORT="f5">VK_DESCRIPTOR_SET_MAPPING</TD><TD PORT="f6">descriptorSetMapping[2]</TD></TR><TR><TD PORT="f7">uint32_t</TD><TD PORT="f8">linkConstBufferCount</TD></TR><TR><TD PORT="f9">const VkLinkConstBuffer*</TD><TD PORT="f10">pLinkConstBufferInfo</TD></TR><TR><TD PORT="f11">VK_DYNAMIC_MEMORY_VIEW_SLOT_INFO</TD><TD PORT="f12">dynamicMemoryViewMapping</TD></TR></TABLE>>
];
"TC_VK_PIPELINE_SHADER" [
-label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkPipelineShader</TD></TR><TR><TD PORT="f1">VkPipelineShaderStage</TD><TD PORT="f2">stage=TC</TD></TR><TR><TD PORT="f3">VkShader</TD><TD PORT="f4">shader</TD></TR><TR><TD PORT="f5">VK_DESCRIPTOR_SET_MAPPING</TD><TD PORT="f6">descriptorSetMapping[2]</TD></TR><TR><TD PORT="f7">uint32_t</TD><TD PORT="f8">linkConstBufferCount</TD></TR><TR><TD PORT="f9">const VkLinkConstBuffer*</TD><TD PORT="f10">pLinkConstBufferInfo</TD></TR><TR><TD PORT="f11">VK_DYNAMIC_MEMORY_VIEW_SLOT_INFO</TD><TD PORT="f12">dynamicMemoryViewMapping</TD></TR></TABLE>>
+label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkPipelineShader</TD></TR><TR><TD PORT="f1">VkShaderStage</TD><TD PORT="f2">stage=TC</TD></TR><TR><TD PORT="f3">VkShader</TD><TD PORT="f4">shader</TD></TR><TR><TD PORT="f5">VK_DESCRIPTOR_SET_MAPPING</TD><TD PORT="f6">descriptorSetMapping[2]</TD></TR><TR><TD PORT="f7">uint32_t</TD><TD PORT="f8">linkConstBufferCount</TD></TR><TR><TD PORT="f9">const VkLinkConstBuffer*</TD><TD PORT="f10">pLinkConstBufferInfo</TD></TR><TR><TD PORT="f11">VK_DYNAMIC_MEMORY_VIEW_SLOT_INFO</TD><TD PORT="f12">dynamicMemoryViewMapping</TD></TR></TABLE>>
];
"TE_VK_PIPELINE_SHADER" [
-label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkPipelineShader</TD></TR><TR><TD PORT="f1">VkPipelineShaderStage</TD><TD PORT="f2">stage=TE</TD></TR><TR><TD PORT="f3">VkShader</TD><TD PORT="f4">shader</TD></TR><TR><TD PORT="f5">VK_DESCRIPTOR_SET_MAPPING</TD><TD PORT="f6">descriptorSetMapping[2]</TD></TR><TR><TD PORT="f7">uint32_t</TD><TD PORT="f8">linkConstBufferCount</TD></TR><TR><TD PORT="f9">const VkLinkConstBuffer*</TD><TD PORT="f10">pLinkConstBufferInfo</TD></TR><TR><TD PORT="f11">VK_DYNAMIC_MEMORY_VIEW_SLOT_INFO</TD><TD PORT="f12">dynamicMemoryViewMapping</TD></TR></TABLE>>
+label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkPipelineShader</TD></TR><TR><TD PORT="f1">VkShaderStage</TD><TD PORT="f2">stage=TE</TD></TR><TR><TD PORT="f3">VkShader</TD><TD PORT="f4">shader</TD></TR><TR><TD PORT="f5">VK_DESCRIPTOR_SET_MAPPING</TD><TD PORT="f6">descriptorSetMapping[2]</TD></TR><TR><TD PORT="f7">uint32_t</TD><TD PORT="f8">linkConstBufferCount</TD></TR><TR><TD PORT="f9">const VkLinkConstBuffer*</TD><TD PORT="f10">pLinkConstBufferInfo</TD></TR><TR><TD PORT="f11">VK_DYNAMIC_MEMORY_VIEW_SLOT_INFO</TD><TD PORT="f12">dynamicMemoryViewMapping</TD></TR></TABLE>>
];
"GS_VK_PIPELINE_SHADER" [
-label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkPipelineShader</TD></TR><TR><TD PORT="f1">VkPipelineShaderStage</TD><TD PORT="f2">stage=GS</TD></TR><TR><TD PORT="f3">VkShader</TD><TD PORT="f4">shader</TD></TR><TR><TD PORT="f5">VK_DESCRIPTOR_SET_MAPPING</TD><TD PORT="f6">descriptorSetMapping[2]</TD></TR><TR><TD PORT="f7">uint32_t</TD><TD PORT="f8">linkConstBufferCount</TD></TR><TR><TD PORT="f9">const VkLinkConstBuffer*</TD><TD PORT="f10">pLinkConstBufferInfo</TD></TR><TR><TD PORT="f11">VK_DYNAMIC_MEMORY_VIEW_SLOT_INFO</TD><TD PORT="f12">dynamicMemoryViewMapping</TD></TR></TABLE>>
+label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkPipelineShader</TD></TR><TR><TD PORT="f1">VkShaderStage</TD><TD PORT="f2">stage=GS</TD></TR><TR><TD PORT="f3">VkShader</TD><TD PORT="f4">shader</TD></TR><TR><TD PORT="f5">VK_DESCRIPTOR_SET_MAPPING</TD><TD PORT="f6">descriptorSetMapping[2]</TD></TR><TR><TD PORT="f7">uint32_t</TD><TD PORT="f8">linkConstBufferCount</TD></TR><TR><TD PORT="f9">const VkLinkConstBuffer*</TD><TD PORT="f10">pLinkConstBufferInfo</TD></TR><TR><TD PORT="f11">VK_DYNAMIC_MEMORY_VIEW_SLOT_INFO</TD><TD PORT="f12">dynamicMemoryViewMapping</TD></TR></TABLE>>
];
"FS_VK_PIPELINE_SHADER" [
-label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkPipelineShader</TD></TR><TR><TD PORT="f1">VkPipelineShaderStage</TD><TD PORT="f2">stage=FS</TD></TR><TR><TD PORT="f3">VkShader</TD><TD PORT="f4">shader</TD></TR><TR><TD PORT="f5">VK_DESCRIPTOR_SET_MAPPING</TD><TD PORT="f6">descriptorSetMapping[2]</TD></TR><TR><TD PORT="f7">uint32_t</TD><TD PORT="f8">linkConstBufferCount</TD></TR><TR><TD PORT="f9">const VkLinkConstBuffer*</TD><TD PORT="f10">pLinkConstBufferInfo</TD></TR><TR><TD PORT="f11">VK_DYNAMIC_MEMORY_VIEW_SLOT_INFO</TD><TD PORT="f12">dynamicMemoryViewMapping</TD></TR></TABLE>>
+label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkPipelineShader</TD></TR><TR><TD PORT="f1">VkShaderStage</TD><TD PORT="f2">stage=FS</TD></TR><TR><TD PORT="f3">VkShader</TD><TD PORT="f4">shader</TD></TR><TR><TD PORT="f5">VK_DESCRIPTOR_SET_MAPPING</TD><TD PORT="f6">descriptorSetMapping[2]</TD></TR><TR><TD PORT="f7">uint32_t</TD><TD PORT="f8">linkConstBufferCount</TD></TR><TR><TD PORT="f9">const VkLinkConstBuffer*</TD><TD PORT="f10">pLinkConstBufferInfo</TD></TR><TR><TD PORT="f11">VK_DYNAMIC_MEMORY_VIEW_SLOT_INFO</TD><TD PORT="f12">dynamicMemoryViewMapping</TD></TR></TABLE>>
];
"VkGraphicsPipelineCreateInfo_":f2 -> "VkPipelineIaStateCreateInfo_" [
id = 100
@@ -119,13 +119,13 @@
{
label="vkCreateSampler - multiple calls return unique VkSampler handles"
"_VK_SAMPLER_CREATE_INFO_0" [
-label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkSamplerCreateInfo</TD></TR><TR><TD PORT="f1">VkStructureType</TD><TD PORT="f2">sType=<BR/>VK_STRUCTURE_TYPE_SAMPLER_CREATE_INFO</TD></TR><TR><TD PORT="f3">const void*</TD><TD PORT="f4">pNext=NULL</TD></TR><TR><TD PORT="f5">VkTexFilter</TD><TD PORT="f6">magFilter</TD></TR><TR><TD PORT="f7">VkTexFilter</TD><TD PORT="f8">minFilter</TD></TR><TR><TD PORT="f9">VkTexMipmapMode</TD><TD PORT="f10">mipMode</TD></TR><TR><TD PORT="f11">VkTexAddress</TD><TD PORT="f12">addressU</TD></TR><TR><TD PORT="f13">VkTexAddress</TD><TD PORT="f14">addressV</TD></TR><TR><TD PORT="f15">VkTexAddress</TD><TD PORT="f16">addressW</TD></TR><TR><TD PORT="f17">float</TD><TD PORT="f18">mipLodBias</TD></TR><TR><TD PORT="f19">uint32_t</TD><TD PORT="f20">maxAnisotropy</TD></TR><TR><TD PORT="f21">VkCompareFunc</TD><TD PORT="f22">compareFunc</TD></TR><TR><TD PORT="f23">float</TD><TD PORT="f24">minLod</TD></TR><TR><TD PORT="f25">float</TD><TD PORT="f26">maxLod</TD></TR><TR><TD PORT="f27">VkBorderColorType</TD><TD PORT="f28">borderColorType</TD></TR></TABLE>>
+label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkSamplerCreateInfo</TD></TR><TR><TD PORT="f1">VkStructureType</TD><TD PORT="f2">sType=<BR/>VK_STRUCTURE_TYPE_SAMPLER_CREATE_INFO</TD></TR><TR><TD PORT="f3">const void*</TD><TD PORT="f4">pNext=NULL</TD></TR><TR><TD PORT="f5">VkTexFilter</TD><TD PORT="f6">magFilter</TD></TR><TR><TD PORT="f7">VkTexFilter</TD><TD PORT="f8">minFilter</TD></TR><TR><TD PORT="f9">VkTexMipmapMode</TD><TD PORT="f10">mipMode</TD></TR><TR><TD PORT="f11">VkTexAddress</TD><TD PORT="f12">addressU</TD></TR><TR><TD PORT="f13">VkTexAddress</TD><TD PORT="f14">addressV</TD></TR><TR><TD PORT="f15">VkTexAddress</TD><TD PORT="f16">addressW</TD></TR><TR><TD PORT="f17">float</TD><TD PORT="f18">mipLodBias</TD></TR><TR><TD PORT="f19">uint32_t</TD><TD PORT="f20">maxAnisotropy</TD></TR><TR><TD PORT="f21">VkCompareOp</TD><TD PORT="f22">compareFunc</TD></TR><TR><TD PORT="f23">float</TD><TD PORT="f24">minLod</TD></TR><TR><TD PORT="f25">float</TD><TD PORT="f26">maxLod</TD></TR><TR><TD PORT="f27">VkBorderColor</TD><TD PORT="f28">borderColorType</TD></TR></TABLE>>
];
"SAMPLER_ELLIPSES" [
label = "..."
];
"_VK_SAMPLER_CREATE_INFO_19" [
-label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkSamplerCreateInfo</TD></TR><TR><TD PORT="f1">VkStructureType</TD><TD PORT="f2">sType=<BR/>VK_STRUCTURE_TYPE_SAMPLER_CREATE_INFO</TD></TR><TR><TD PORT="f3">const void*</TD><TD PORT="f4">pNext=NULL</TD></TR><TR><TD PORT="f5">VkTexFilter</TD><TD PORT="f6">magFilter</TD></TR><TR><TD PORT="f7">VkTexFilter</TD><TD PORT="f8">minFilter</TD></TR><TR><TD PORT="f9">VkTexMipmapMode</TD><TD PORT="f10">mipMode</TD></TR><TR><TD PORT="f11">VkTexAddress</TD><TD PORT="f12">addressU</TD></TR><TR><TD PORT="f13">VkTexAddress</TD><TD PORT="f14">addressV</TD></TR><TR><TD PORT="f15">VkTexAddress</TD><TD PORT="f16">addressW</TD></TR><TR><TD PORT="f17">float</TD><TD PORT="f18">mipLodBias</TD></TR><TR><TD PORT="f19">uint32_t</TD><TD PORT="f20">maxAnisotropy</TD></TR><TR><TD PORT="f21">VkCompareFunc</TD><TD PORT="f22">compareFunc</TD></TR><TR><TD PORT="f23">float</TD><TD PORT="f24">minLod</TD></TR><TR><TD PORT="f25">float</TD><TD PORT="f26">maxLod</TD></TR><TR><TD PORT="f27">VkBorderColorType</TD><TD PORT="f28">borderColorType</TD></TR></TABLE>>
+label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkSamplerCreateInfo</TD></TR><TR><TD PORT="f1">VkStructureType</TD><TD PORT="f2">sType=<BR/>VK_STRUCTURE_TYPE_SAMPLER_CREATE_INFO</TD></TR><TR><TD PORT="f3">const void*</TD><TD PORT="f4">pNext=NULL</TD></TR><TR><TD PORT="f5">VkTexFilter</TD><TD PORT="f6">magFilter</TD></TR><TR><TD PORT="f7">VkTexFilter</TD><TD PORT="f8">minFilter</TD></TR><TR><TD PORT="f9">VkTexMipmapMode</TD><TD PORT="f10">mipMode</TD></TR><TR><TD PORT="f11">VkTexAddress</TD><TD PORT="f12">addressU</TD></TR><TR><TD PORT="f13">VkTexAddress</TD><TD PORT="f14">addressV</TD></TR><TR><TD PORT="f15">VkTexAddress</TD><TD PORT="f16">addressW</TD></TR><TR><TD PORT="f17">float</TD><TD PORT="f18">mipLodBias</TD></TR><TR><TD PORT="f19">uint32_t</TD><TD PORT="f20">maxAnisotropy</TD></TR><TR><TD PORT="f21">VkCompareOp</TD><TD PORT="f22">compareFunc</TD></TR><TR><TD PORT="f23">float</TD><TD PORT="f24">minLod</TD></TR><TR><TD PORT="f25">float</TD><TD PORT="f26">maxLod</TD></TR><TR><TD PORT="f27">VkBorderColor</TD><TD PORT="f28">borderColorType</TD></TR></TABLE>>
];
}
subgraph clusterSamplerAttach
@@ -145,13 +145,13 @@
{
label="vkAttachMemoryViewDescriptors - pMemViews array of VK_MEMORY_VIEW_ATTACH_INFO structs"
"_VK_MEMORY_VIEW_ATTACH_INFO_3" [
-label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VK_MEMORY_VIEW_ATTACH_INFO</TD></TR><TR><TD PORT="f1">VkStructureType</TD><TD PORT="f2">sType=<BR/>VK_STRUCTURE_TYPE_MEMORY_VIEW_ATTACH_INFO</TD></TR><TR><TD PORT="f3">void*</TD><TD PORT="f4">pNext=NULL</TD></TR><TR><TD PORT="f5">VkGpuMemory</TD><TD PORT="f6">mem</TD></TR><TR><TD PORT="f7">VkGpuSize</TD><TD PORT="f8">offset</TD></TR><TR><TD PORT="f9">VkGpuSize</TD><TD PORT="f10">range</TD></TR><TR><TD PORT="f11">VkGpuSize</TD><TD PORT="f12">stride</TD></TR><TR><TD PORT="f13">VkFormat</TD><TD PORT="f14">format</TD></TR><TR><TD PORT="f15">VK_MEMORY_STATE</TD><TD PORT="f16">state</TD></TR></TABLE>>
+label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VK_MEMORY_VIEW_ATTACH_INFO</TD></TR><TR><TD PORT="f1">VkStructureType</TD><TD PORT="f2">sType=<BR/>VK_STRUCTURE_TYPE_MEMORY_VIEW_ATTACH_INFO</TD></TR><TR><TD PORT="f3">void*</TD><TD PORT="f4">pNext=NULL</TD></TR><TR><TD PORT="f5">VkDeviceMemory</TD><TD PORT="f6">mem</TD></TR><TR><TD PORT="f7">VkDeviceSize</TD><TD PORT="f8">offset</TD></TR><TR><TD PORT="f9">VkDeviceSize</TD><TD PORT="f10">range</TD></TR><TR><TD PORT="f11">VkDeviceSize</TD><TD PORT="f12">stride</TD></TR><TR><TD PORT="f13">VkFormat</TD><TD PORT="f14">format</TD></TR><TR><TD PORT="f15">VK_MEMORY_STATE</TD><TD PORT="f16">state</TD></TR></TABLE>>
];
"MEM_VIEW_ELLIPSES" [
label = "..."
];
"_VK_MEMORY_VIEW_ATTACH_INFO_0" [
-label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VK_MEMORY_VIEW_ATTACH_INFO</TD></TR><TR><TD PORT="f1">VkStructureType</TD><TD PORT="f2">sType=<BR/>VK_STRUCTURE_TYPE_MEMORY_VIEW_ATTACH_INFO</TD></TR><TR><TD PORT="f3">void*</TD><TD PORT="f4">pNext=NULL</TD></TR><TR><TD PORT="f5">VkGpuMemory</TD><TD PORT="f6">mem</TD></TR><TR><TD PORT="f7">VkGpuSize</TD><TD PORT="f8">offset</TD></TR><TR><TD PORT="f9">VkGpuSize</TD><TD PORT="f10">range</TD></TR><TR><TD PORT="f11">VkGpuSize</TD><TD PORT="f12">stride</TD></TR><TR><TD PORT="f13">VkFormat</TD><TD PORT="f14">format</TD></TR><TR><TD PORT="f15">VK_MEMORY_STATE</TD><TD PORT="f16">state</TD></TR></TABLE>>
+label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VK_MEMORY_VIEW_ATTACH_INFO</TD></TR><TR><TD PORT="f1">VkStructureType</TD><TD PORT="f2">sType=<BR/>VK_STRUCTURE_TYPE_MEMORY_VIEW_ATTACH_INFO</TD></TR><TR><TD PORT="f3">void*</TD><TD PORT="f4">pNext=NULL</TD></TR><TR><TD PORT="f5">VkDeviceMemory</TD><TD PORT="f6">mem</TD></TR><TR><TD PORT="f7">VkDeviceSize</TD><TD PORT="f8">offset</TD></TR><TR><TD PORT="f9">VkDeviceSize</TD><TD PORT="f10">range</TD></TR><TR><TD PORT="f11">VkDeviceSize</TD><TD PORT="f12">stride</TD></TR><TR><TD PORT="f13">VkFormat</TD><TD PORT="f14">format</TD></TR><TR><TD PORT="f15">VK_MEMORY_STATE</TD><TD PORT="f16">state</TD></TR></TABLE>>
];
}
subgraph clusterImageView
diff --git a/docs/vk_graphics_pipeline.dot b/docs/vk_graphics_pipeline.dot
index 6aea1c4..c659cff 100644
--- a/docs/vk_graphics_pipeline.dot
+++ b/docs/vk_graphics_pipeline.dot
@@ -12,7 +12,7 @@
label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkGraphicsPipelineCreateInfo</TD></TR><TR><TD>VkStructureType</TD><TD>sType=<BR/>VK_STRUCTURE_TYPE_GRAPHICS_PIPELINE_CREATE_INFO</TD></TR><TR><TD>const void*</TD><TD PORT="f2">pNext</TD></TR><TR><TD>VkFlags</TD><TD>flags</TD></TR></TABLE>>
];
"VkPipelineIaStateCreateInfo_" [
-label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkPipelineIaStateCreateInfo</TD></TR><TR><TD>VkStructureType</TD><TD>sType=<BR/>VK_STRUCTURE_TYPE_PIPELINE_IA_STATE_CREATE_INFO</TD></TR><TR><TD>const void*</TD><TD PORT="f2">pNext</TD></TR><TR><TD>VkPrimitiveTopology</TD><TD>topology</TD></TR><TR><TD>bool32_t</TD><TD>disableVertexReuse</TD></TR><TR><TD>VkProvokingVertexConvention</TD><TD>provokingVertex</TD></TR><TR><TD>bool32_t</TD><TD>primitiveRestartEnable</TD></TR><TR><TD>uint32_t</TD><TD>primitiveRestartIndex</TD></TR></TABLE>>
+label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkPipelineIaStateCreateInfo</TD></TR><TR><TD>VkStructureType</TD><TD>sType=<BR/>VK_STRUCTURE_TYPE_PIPELINE_IA_STATE_CREATE_INFO</TD></TR><TR><TD>const void*</TD><TD PORT="f2">pNext</TD></TR><TR><TD>VkPrimitiveTopology</TD><TD>topology</TD></TR><TR><TD>bool32_t</TD><TD>disableVertexReuse</TD></TR><TR><TD>VkProvokingVertex</TD><TD>provokingVertex</TD></TR><TR><TD>bool32_t</TD><TD>primitiveRestartEnable</TD></TR><TR><TD>uint32_t</TD><TD>primitiveRestartIndex</TD></TR></TABLE>>
];
"VkPipelineTessStateCreateInfo_" [
label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkPipelineTessStateCreateInfo</TD></TR><TR><TD>VkStructureType</TD><TD>sType=<BR/>VK_STRUCTURE_TYPE_PIPELINE_TESS_STATE_CREATE_INFO</TD></TR><TR><TD>const void*</TD><TD PORT="f2">pNext</TD></TR><TR><TD>uint32_t</TD><TD>patchControlPoints</TD></TR> <TR><TD>float</TD><TD>optimalTessFactor</TD></TR><TR><TD>float</TD><TD>fixedTessFactor</TD></TR></TABLE>>
@@ -42,19 +42,19 @@
label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkPipelineShaderStageCreateInfo</TD></TR><TR><TD>VkStructureType</TD><TD>sType=<BR/>VK_STRUCTURE_TYPE_PIPELINE_SHADER_STAGE_CREATE_INFO</TD></TR> <TR><TD>const void*</TD><TD PORT="f2">pNext=NULL</TD></TR> <TR><TD>VkPipelineShader</TD><TD PORT="f4">shader</TD></TR> </TABLE>>
];
"VS_VK_PIPELINE_SHADER" [
-label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkPipelineShader</TD></TR><TR><TD PORT="f1">VkPipelineShaderStage</TD><TD PORT="f2">stage=VS</TD></TR><TR><TD PORT="f3">VkShader</TD><TD PORT="f4">shader</TD></TR><TR><TD PORT="f5">VK_DESCRIPTOR_SET_MAPPING</TD><TD PORT="f6">descriptorSetMapping[2]</TD></TR><TR><TD PORT="f7">uint32_t</TD><TD PORT="f8">linkConstBufferCount</TD></TR><TR><TD PORT="f9">const VkLinkConstBuffer*</TD><TD PORT="f10">pLinkConstBufferInfo</TD></TR><TR><TD PORT="f11">VK_DYNAMIC_MEMORY_VIEW_SLOT_INFO</TD><TD PORT="f12">dynamicMemoryViewMapping</TD></TR></TABLE>>
+label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkPipelineShader</TD></TR><TR><TD PORT="f1">VkShaderStage</TD><TD PORT="f2">stage=VS</TD></TR><TR><TD PORT="f3">VkShader</TD><TD PORT="f4">shader</TD></TR><TR><TD PORT="f5">VK_DESCRIPTOR_SET_MAPPING</TD><TD PORT="f6">descriptorSetMapping[2]</TD></TR><TR><TD PORT="f7">uint32_t</TD><TD PORT="f8">linkConstBufferCount</TD></TR><TR><TD PORT="f9">const VkLinkConstBuffer*</TD><TD PORT="f10">pLinkConstBufferInfo</TD></TR><TR><TD PORT="f11">VK_DYNAMIC_MEMORY_VIEW_SLOT_INFO</TD><TD PORT="f12">dynamicMemoryViewMapping</TD></TR></TABLE>>
];
"TC_VK_PIPELINE_SHADER" [
-label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkPipelineShader</TD></TR><TR><TD PORT="f1">VkPipelineShaderStage</TD><TD PORT="f2">stage=TC</TD></TR><TR><TD PORT="f3">VkShader</TD><TD PORT="f4">shader</TD></TR><TR><TD PORT="f5">VK_DESCRIPTOR_SET_MAPPING</TD><TD PORT="f6">descriptorSetMapping[2]</TD></TR><TR><TD PORT="f7">uint32_t</TD><TD PORT="f8">linkConstBufferCount</TD></TR><TR><TD PORT="f9">const VkLinkConstBuffer*</TD><TD PORT="f10">pLinkConstBufferInfo</TD></TR><TR><TD PORT="f11">VK_DYNAMIC_MEMORY_VIEW_SLOT_INFO</TD><TD PORT="f12">dynamicMemoryViewMapping</TD></TR></TABLE>>
+label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkPipelineShader</TD></TR><TR><TD PORT="f1">VkShaderStage</TD><TD PORT="f2">stage=TC</TD></TR><TR><TD PORT="f3">VkShader</TD><TD PORT="f4">shader</TD></TR><TR><TD PORT="f5">VK_DESCRIPTOR_SET_MAPPING</TD><TD PORT="f6">descriptorSetMapping[2]</TD></TR><TR><TD PORT="f7">uint32_t</TD><TD PORT="f8">linkConstBufferCount</TD></TR><TR><TD PORT="f9">const VkLinkConstBuffer*</TD><TD PORT="f10">pLinkConstBufferInfo</TD></TR><TR><TD PORT="f11">VK_DYNAMIC_MEMORY_VIEW_SLOT_INFO</TD><TD PORT="f12">dynamicMemoryViewMapping</TD></TR></TABLE>>
];
"TE_VK_PIPELINE_SHADER" [
-label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkPipelineShader</TD></TR><TR><TD PORT="f1">VkPipelineShaderStage</TD><TD PORT="f2">stage=TE</TD></TR><TR><TD PORT="f3">VkShader</TD><TD PORT="f4">shader</TD></TR><TR><TD PORT="f5">VK_DESCRIPTOR_SET_MAPPING</TD><TD PORT="f6">descriptorSetMapping[2]</TD></TR><TR><TD PORT="f7">uint32_t</TD><TD PORT="f8">linkConstBufferCount</TD></TR><TR><TD PORT="f9">const VkLinkConstBuffer*</TD><TD PORT="f10">pLinkConstBufferInfo</TD></TR><TR><TD PORT="f11">VK_DYNAMIC_MEMORY_VIEW_SLOT_INFO</TD><TD PORT="f12">dynamicMemoryViewMapping</TD></TR></TABLE>>
+label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkPipelineShader</TD></TR><TR><TD PORT="f1">VkShaderStage</TD><TD PORT="f2">stage=TE</TD></TR><TR><TD PORT="f3">VkShader</TD><TD PORT="f4">shader</TD></TR><TR><TD PORT="f5">VK_DESCRIPTOR_SET_MAPPING</TD><TD PORT="f6">descriptorSetMapping[2]</TD></TR><TR><TD PORT="f7">uint32_t</TD><TD PORT="f8">linkConstBufferCount</TD></TR><TR><TD PORT="f9">const VkLinkConstBuffer*</TD><TD PORT="f10">pLinkConstBufferInfo</TD></TR><TR><TD PORT="f11">VK_DYNAMIC_MEMORY_VIEW_SLOT_INFO</TD><TD PORT="f12">dynamicMemoryViewMapping</TD></TR></TABLE>>
];
"GS_VK_PIPELINE_SHADER" [
-label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkPipelineShader</TD></TR><TR><TD PORT="f1">VkPipelineShaderStage</TD><TD PORT="f2">stage=GS</TD></TR><TR><TD PORT="f3">VkShader</TD><TD PORT="f4">shader</TD></TR><TR><TD PORT="f5">VK_DESCRIPTOR_SET_MAPPING</TD><TD PORT="f6">descriptorSetMapping[2]</TD></TR><TR><TD PORT="f7">uint32_t</TD><TD PORT="f8">linkConstBufferCount</TD></TR><TR><TD PORT="f9">const VkLinkConstBuffer*</TD><TD PORT="f10">pLinkConstBufferInfo</TD></TR><TR><TD PORT="f11">VK_DYNAMIC_MEMORY_VIEW_SLOT_INFO</TD><TD PORT="f12">dynamicMemoryViewMapping</TD></TR></TABLE>>
+label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkPipelineShader</TD></TR><TR><TD PORT="f1">VkShaderStage</TD><TD PORT="f2">stage=GS</TD></TR><TR><TD PORT="f3">VkShader</TD><TD PORT="f4">shader</TD></TR><TR><TD PORT="f5">VK_DESCRIPTOR_SET_MAPPING</TD><TD PORT="f6">descriptorSetMapping[2]</TD></TR><TR><TD PORT="f7">uint32_t</TD><TD PORT="f8">linkConstBufferCount</TD></TR><TR><TD PORT="f9">const VkLinkConstBuffer*</TD><TD PORT="f10">pLinkConstBufferInfo</TD></TR><TR><TD PORT="f11">VK_DYNAMIC_MEMORY_VIEW_SLOT_INFO</TD><TD PORT="f12">dynamicMemoryViewMapping</TD></TR></TABLE>>
];
"FS_VK_PIPELINE_SHADER" [
-label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkPipelineShader</TD></TR><TR><TD PORT="f1">VkPipelineShaderStage</TD><TD PORT="f2">stage=FS</TD></TR><TR><TD PORT="f3">VkShader</TD><TD PORT="f4">shader</TD></TR><TR><TD PORT="f5">VK_DESCRIPTOR_SET_MAPPING</TD><TD PORT="f6">descriptorSetMapping[2]</TD></TR><TR><TD PORT="f7">uint32_t</TD><TD PORT="f8">linkConstBufferCount</TD></TR><TR><TD PORT="f9">const VkLinkConstBuffer*</TD><TD PORT="f10">pLinkConstBufferInfo</TD></TR><TR><TD PORT="f11">VK_DYNAMIC_MEMORY_VIEW_SLOT_INFO</TD><TD PORT="f12">dynamicMemoryViewMapping</TD></TR></TABLE>>
+label = <<TABLE BORDER="0" CELLBORDER="1" CELLSPACING="0"> <TR><TD COLSPAN="2" PORT="f0">VkPipelineShader</TD></TR><TR><TD PORT="f1">VkShaderStage</TD><TD PORT="f2">stage=FS</TD></TR><TR><TD PORT="f3">VkShader</TD><TD PORT="f4">shader</TD></TR><TR><TD PORT="f5">VK_DESCRIPTOR_SET_MAPPING</TD><TD PORT="f6">descriptorSetMapping[2]</TD></TR><TR><TD PORT="f7">uint32_t</TD><TD PORT="f8">linkConstBufferCount</TD></TR><TR><TD PORT="f9">const VkLinkConstBuffer*</TD><TD PORT="f10">pLinkConstBufferInfo</TD></TR><TR><TD PORT="f11">VK_DYNAMIC_MEMORY_VIEW_SLOT_INFO</TD><TD PORT="f12">dynamicMemoryViewMapping</TD></TR></TABLE>>
];
"VkGraphicsPipelineCreateInfo_":f2 -> "VkPipelineIaStateCreateInfo_" [
id = 0
diff --git a/icd/README.md b/icd/README.md
index 4077046..616f978 100644
--- a/icd/README.md
+++ b/icd/README.md
@@ -8,7 +8,7 @@
common/ provides helper and utility functions, as well as all VK entry points
except vkInitAndEnumerateGpus. Hardware drivers are required to provide that
function, and to embed a "VkLayerDispatchTable *" as the first member of
-VkPhysicalGpu and all VkBaseObject.
+VkPhysicalDevice and all VkBaseObject.
Thread safety
diff --git a/icd/common/icd-enumerate-drm.c b/icd/common/icd-enumerate-drm.c
index 65e3065..06e76b1 100644
--- a/icd/common/icd-enumerate-drm.c
+++ b/icd/common/icd-enumerate-drm.c
@@ -125,7 +125,7 @@
return devices;
} else {
dev = icd_instance_alloc(instance, sizeof(*dev), 0,
- VK_SYSTEM_ALLOC_INTERNAL_TEMP);
+ VK_SYSTEM_ALLOC_TYPE_INTERNAL_TEMP);
if (!dev)
return devices;
diff --git a/icd/common/icd-format.c b/icd/common/icd-format.c
index cffdb7f..e2c2ada 100644
--- a/icd/common/icd-format.c
+++ b/icd/common/icd-format.c
@@ -32,177 +32,177 @@
static const struct icd_format_info {
size_t size;
uint32_t channel_count;
-} icd_format_table[VK_NUM_FMT] = {
- [VK_FMT_UNDEFINED] = { 0, 0 },
- [VK_FMT_R4G4_UNORM] = { 1, 2 },
- [VK_FMT_R4G4_USCALED] = { 1, 2 },
- [VK_FMT_R4G4B4A4_UNORM] = { 2, 4 },
- [VK_FMT_R4G4B4A4_USCALED] = { 2, 4 },
- [VK_FMT_R5G6B5_UNORM] = { 2, 3 },
- [VK_FMT_R5G6B5_USCALED] = { 2, 3 },
- [VK_FMT_R5G5B5A1_UNORM] = { 2, 4 },
- [VK_FMT_R5G5B5A1_USCALED] = { 2, 4 },
- [VK_FMT_R8_UNORM] = { 1, 1 },
- [VK_FMT_R8_SNORM] = { 1, 1 },
- [VK_FMT_R8_USCALED] = { 1, 1 },
- [VK_FMT_R8_SSCALED] = { 1, 1 },
- [VK_FMT_R8_UINT] = { 1, 1 },
- [VK_FMT_R8_SINT] = { 1, 1 },
- [VK_FMT_R8_SRGB] = { 1, 1 },
- [VK_FMT_R8G8_UNORM] = { 2, 2 },
- [VK_FMT_R8G8_SNORM] = { 2, 2 },
- [VK_FMT_R8G8_USCALED] = { 2, 2 },
- [VK_FMT_R8G8_SSCALED] = { 2, 2 },
- [VK_FMT_R8G8_UINT] = { 2, 2 },
- [VK_FMT_R8G8_SINT] = { 2, 2 },
- [VK_FMT_R8G8_SRGB] = { 2, 2 },
- [VK_FMT_R8G8B8_UNORM] = { 3, 3 },
- [VK_FMT_R8G8B8_SNORM] = { 3, 3 },
- [VK_FMT_R8G8B8_USCALED] = { 3, 3 },
- [VK_FMT_R8G8B8_SSCALED] = { 3, 3 },
- [VK_FMT_R8G8B8_UINT] = { 3, 3 },
- [VK_FMT_R8G8B8_SINT] = { 3, 3 },
- [VK_FMT_R8G8B8_SRGB] = { 3, 3 },
- [VK_FMT_R8G8B8A8_UNORM] = { 4, 4 },
- [VK_FMT_R8G8B8A8_SNORM] = { 4, 4 },
- [VK_FMT_R8G8B8A8_USCALED] = { 4, 4 },
- [VK_FMT_R8G8B8A8_SSCALED] = { 4, 4 },
- [VK_FMT_R8G8B8A8_UINT] = { 4, 4 },
- [VK_FMT_R8G8B8A8_SINT] = { 4, 4 },
- [VK_FMT_R8G8B8A8_SRGB] = { 4, 4 },
- [VK_FMT_R10G10B10A2_UNORM] = { 4, 4 },
- [VK_FMT_R10G10B10A2_SNORM] = { 4, 4 },
- [VK_FMT_R10G10B10A2_USCALED] = { 4, 4 },
- [VK_FMT_R10G10B10A2_SSCALED] = { 4, 4 },
- [VK_FMT_R10G10B10A2_UINT] = { 4, 4 },
- [VK_FMT_R10G10B10A2_SINT] = { 4, 4 },
- [VK_FMT_R16_UNORM] = { 2, 1 },
- [VK_FMT_R16_SNORM] = { 2, 1 },
- [VK_FMT_R16_USCALED] = { 2, 1 },
- [VK_FMT_R16_SSCALED] = { 2, 1 },
- [VK_FMT_R16_UINT] = { 2, 1 },
- [VK_FMT_R16_SINT] = { 2, 1 },
- [VK_FMT_R16_SFLOAT] = { 2, 1 },
- [VK_FMT_R16G16_UNORM] = { 4, 2 },
- [VK_FMT_R16G16_SNORM] = { 4, 2 },
- [VK_FMT_R16G16_USCALED] = { 4, 2 },
- [VK_FMT_R16G16_SSCALED] = { 4, 2 },
- [VK_FMT_R16G16_UINT] = { 4, 2 },
- [VK_FMT_R16G16_SINT] = { 4, 2 },
- [VK_FMT_R16G16_SFLOAT] = { 4, 2 },
- [VK_FMT_R16G16B16_UNORM] = { 6, 3 },
- [VK_FMT_R16G16B16_SNORM] = { 6, 3 },
- [VK_FMT_R16G16B16_USCALED] = { 6, 3 },
- [VK_FMT_R16G16B16_SSCALED] = { 6, 3 },
- [VK_FMT_R16G16B16_UINT] = { 6, 3 },
- [VK_FMT_R16G16B16_SINT] = { 6, 3 },
- [VK_FMT_R16G16B16_SFLOAT] = { 6, 3 },
- [VK_FMT_R16G16B16A16_UNORM] = { 8, 4 },
- [VK_FMT_R16G16B16A16_SNORM] = { 8, 4 },
- [VK_FMT_R16G16B16A16_USCALED] = { 8, 4 },
- [VK_FMT_R16G16B16A16_SSCALED] = { 8, 4 },
- [VK_FMT_R16G16B16A16_UINT] = { 8, 4 },
- [VK_FMT_R16G16B16A16_SINT] = { 8, 4 },
- [VK_FMT_R16G16B16A16_SFLOAT] = { 8, 4 },
- [VK_FMT_R32_UINT] = { 4, 1 },
- [VK_FMT_R32_SINT] = { 4, 1 },
- [VK_FMT_R32_SFLOAT] = { 4, 1 },
- [VK_FMT_R32G32_UINT] = { 8, 2 },
- [VK_FMT_R32G32_SINT] = { 8, 2 },
- [VK_FMT_R32G32_SFLOAT] = { 8, 2 },
- [VK_FMT_R32G32B32_UINT] = { 12, 3 },
- [VK_FMT_R32G32B32_SINT] = { 12, 3 },
- [VK_FMT_R32G32B32_SFLOAT] = { 12, 3 },
- [VK_FMT_R32G32B32A32_UINT] = { 16, 4 },
- [VK_FMT_R32G32B32A32_SINT] = { 16, 4 },
- [VK_FMT_R32G32B32A32_SFLOAT] = { 16, 4 },
- [VK_FMT_R64_SFLOAT] = { 8, 1 },
- [VK_FMT_R64G64_SFLOAT] = { 16, 2 },
- [VK_FMT_R64G64B64_SFLOAT] = { 24, 3 },
- [VK_FMT_R64G64B64A64_SFLOAT] = { 32, 4 },
- [VK_FMT_R11G11B10_UFLOAT] = { 4, 3 },
- [VK_FMT_R9G9B9E5_UFLOAT] = { 4, 3 },
- [VK_FMT_D16_UNORM] = { 2, 1 },
- [VK_FMT_D24_UNORM] = { 3, 1 },
- [VK_FMT_D32_SFLOAT] = { 4, 1 },
- [VK_FMT_S8_UINT] = { 1, 1 },
- [VK_FMT_D16_UNORM_S8_UINT] = { 3, 2 },
- [VK_FMT_D24_UNORM_S8_UINT] = { 4, 2 },
- [VK_FMT_D32_SFLOAT_S8_UINT] = { 4, 2 },
- [VK_FMT_BC1_RGB_UNORM] = { 8, 4 },
- [VK_FMT_BC1_RGB_SRGB] = { 8, 4 },
- [VK_FMT_BC1_RGBA_UNORM] = { 8, 4 },
- [VK_FMT_BC1_RGBA_SRGB] = { 8, 4 },
- [VK_FMT_BC2_UNORM] = { 16, 4 },
- [VK_FMT_BC2_SRGB] = { 16, 4 },
- [VK_FMT_BC3_UNORM] = { 16, 4 },
- [VK_FMT_BC3_SRGB] = { 16, 4 },
- [VK_FMT_BC4_UNORM] = { 8, 4 },
- [VK_FMT_BC4_SNORM] = { 8, 4 },
- [VK_FMT_BC5_UNORM] = { 16, 4 },
- [VK_FMT_BC5_SNORM] = { 16, 4 },
- [VK_FMT_BC6H_UFLOAT] = { 16, 4 },
- [VK_FMT_BC6H_SFLOAT] = { 16, 4 },
- [VK_FMT_BC7_UNORM] = { 16, 4 },
- [VK_FMT_BC7_SRGB] = { 16, 4 },
+} icd_format_table[VK_NUM_FORMAT] = {
+ [VK_FORMAT_UNDEFINED] = { 0, 0 },
+ [VK_FORMAT_R4G4_UNORM] = { 1, 2 },
+ [VK_FORMAT_R4G4_USCALED] = { 1, 2 },
+ [VK_FORMAT_R4G4B4A4_UNORM] = { 2, 4 },
+ [VK_FORMAT_R4G4B4A4_USCALED] = { 2, 4 },
+ [VK_FORMAT_R5G6B5_UNORM] = { 2, 3 },
+ [VK_FORMAT_R5G6B5_USCALED] = { 2, 3 },
+ [VK_FORMAT_R5G5B5A1_UNORM] = { 2, 4 },
+ [VK_FORMAT_R5G5B5A1_USCALED] = { 2, 4 },
+ [VK_FORMAT_R8_UNORM] = { 1, 1 },
+ [VK_FORMAT_R8_SNORM] = { 1, 1 },
+ [VK_FORMAT_R8_USCALED] = { 1, 1 },
+ [VK_FORMAT_R8_SSCALED] = { 1, 1 },
+ [VK_FORMAT_R8_UINT] = { 1, 1 },
+ [VK_FORMAT_R8_SINT] = { 1, 1 },
+ [VK_FORMAT_R8_SRGB] = { 1, 1 },
+ [VK_FORMAT_R8G8_UNORM] = { 2, 2 },
+ [VK_FORMAT_R8G8_SNORM] = { 2, 2 },
+ [VK_FORMAT_R8G8_USCALED] = { 2, 2 },
+ [VK_FORMAT_R8G8_SSCALED] = { 2, 2 },
+ [VK_FORMAT_R8G8_UINT] = { 2, 2 },
+ [VK_FORMAT_R8G8_SINT] = { 2, 2 },
+ [VK_FORMAT_R8G8_SRGB] = { 2, 2 },
+ [VK_FORMAT_R8G8B8_UNORM] = { 3, 3 },
+ [VK_FORMAT_R8G8B8_SNORM] = { 3, 3 },
+ [VK_FORMAT_R8G8B8_USCALED] = { 3, 3 },
+ [VK_FORMAT_R8G8B8_SSCALED] = { 3, 3 },
+ [VK_FORMAT_R8G8B8_UINT] = { 3, 3 },
+ [VK_FORMAT_R8G8B8_SINT] = { 3, 3 },
+ [VK_FORMAT_R8G8B8_SRGB] = { 3, 3 },
+ [VK_FORMAT_R8G8B8A8_UNORM] = { 4, 4 },
+ [VK_FORMAT_R8G8B8A8_SNORM] = { 4, 4 },
+ [VK_FORMAT_R8G8B8A8_USCALED] = { 4, 4 },
+ [VK_FORMAT_R8G8B8A8_SSCALED] = { 4, 4 },
+ [VK_FORMAT_R8G8B8A8_UINT] = { 4, 4 },
+ [VK_FORMAT_R8G8B8A8_SINT] = { 4, 4 },
+ [VK_FORMAT_R8G8B8A8_SRGB] = { 4, 4 },
+ [VK_FORMAT_R10G10B10A2_UNORM] = { 4, 4 },
+ [VK_FORMAT_R10G10B10A2_SNORM] = { 4, 4 },
+ [VK_FORMAT_R10G10B10A2_USCALED] = { 4, 4 },
+ [VK_FORMAT_R10G10B10A2_SSCALED] = { 4, 4 },
+ [VK_FORMAT_R10G10B10A2_UINT] = { 4, 4 },
+ [VK_FORMAT_R10G10B10A2_SINT] = { 4, 4 },
+ [VK_FORMAT_R16_UNORM] = { 2, 1 },
+ [VK_FORMAT_R16_SNORM] = { 2, 1 },
+ [VK_FORMAT_R16_USCALED] = { 2, 1 },
+ [VK_FORMAT_R16_SSCALED] = { 2, 1 },
+ [VK_FORMAT_R16_UINT] = { 2, 1 },
+ [VK_FORMAT_R16_SINT] = { 2, 1 },
+ [VK_FORMAT_R16_SFLOAT] = { 2, 1 },
+ [VK_FORMAT_R16G16_UNORM] = { 4, 2 },
+ [VK_FORMAT_R16G16_SNORM] = { 4, 2 },
+ [VK_FORMAT_R16G16_USCALED] = { 4, 2 },
+ [VK_FORMAT_R16G16_SSCALED] = { 4, 2 },
+ [VK_FORMAT_R16G16_UINT] = { 4, 2 },
+ [VK_FORMAT_R16G16_SINT] = { 4, 2 },
+ [VK_FORMAT_R16G16_SFLOAT] = { 4, 2 },
+ [VK_FORMAT_R16G16B16_UNORM] = { 6, 3 },
+ [VK_FORMAT_R16G16B16_SNORM] = { 6, 3 },
+ [VK_FORMAT_R16G16B16_USCALED] = { 6, 3 },
+ [VK_FORMAT_R16G16B16_SSCALED] = { 6, 3 },
+ [VK_FORMAT_R16G16B16_UINT] = { 6, 3 },
+ [VK_FORMAT_R16G16B16_SINT] = { 6, 3 },
+ [VK_FORMAT_R16G16B16_SFLOAT] = { 6, 3 },
+ [VK_FORMAT_R16G16B16A16_UNORM] = { 8, 4 },
+ [VK_FORMAT_R16G16B16A16_SNORM] = { 8, 4 },
+ [VK_FORMAT_R16G16B16A16_USCALED] = { 8, 4 },
+ [VK_FORMAT_R16G16B16A16_SSCALED] = { 8, 4 },
+ [VK_FORMAT_R16G16B16A16_UINT] = { 8, 4 },
+ [VK_FORMAT_R16G16B16A16_SINT] = { 8, 4 },
+ [VK_FORMAT_R16G16B16A16_SFLOAT] = { 8, 4 },
+ [VK_FORMAT_R32_UINT] = { 4, 1 },
+ [VK_FORMAT_R32_SINT] = { 4, 1 },
+ [VK_FORMAT_R32_SFLOAT] = { 4, 1 },
+ [VK_FORMAT_R32G32_UINT] = { 8, 2 },
+ [VK_FORMAT_R32G32_SINT] = { 8, 2 },
+ [VK_FORMAT_R32G32_SFLOAT] = { 8, 2 },
+ [VK_FORMAT_R32G32B32_UINT] = { 12, 3 },
+ [VK_FORMAT_R32G32B32_SINT] = { 12, 3 },
+ [VK_FORMAT_R32G32B32_SFLOAT] = { 12, 3 },
+ [VK_FORMAT_R32G32B32A32_UINT] = { 16, 4 },
+ [VK_FORMAT_R32G32B32A32_SINT] = { 16, 4 },
+ [VK_FORMAT_R32G32B32A32_SFLOAT] = { 16, 4 },
+ [VK_FORMAT_R64_SFLOAT] = { 8, 1 },
+ [VK_FORMAT_R64G64_SFLOAT] = { 16, 2 },
+ [VK_FORMAT_R64G64B64_SFLOAT] = { 24, 3 },
+ [VK_FORMAT_R64G64B64A64_SFLOAT] = { 32, 4 },
+ [VK_FORMAT_R11G11B10_UFLOAT] = { 4, 3 },
+ [VK_FORMAT_R9G9B9E5_UFLOAT] = { 4, 3 },
+ [VK_FORMAT_D16_UNORM] = { 2, 1 },
+ [VK_FORMAT_D24_UNORM] = { 3, 1 },
+ [VK_FORMAT_D32_SFLOAT] = { 4, 1 },
+ [VK_FORMAT_S8_UINT] = { 1, 1 },
+ [VK_FORMAT_D16_UNORM_S8_UINT] = { 3, 2 },
+ [VK_FORMAT_D24_UNORM_S8_UINT] = { 4, 2 },
+ [VK_FORMAT_D32_SFLOAT_S8_UINT] = { 4, 2 },
+ [VK_FORMAT_BC1_RGB_UNORM] = { 8, 4 },
+ [VK_FORMAT_BC1_RGB_SRGB] = { 8, 4 },
+ [VK_FORMAT_BC1_RGBA_UNORM] = { 8, 4 },
+ [VK_FORMAT_BC1_RGBA_SRGB] = { 8, 4 },
+ [VK_FORMAT_BC2_UNORM] = { 16, 4 },
+ [VK_FORMAT_BC2_SRGB] = { 16, 4 },
+ [VK_FORMAT_BC3_UNORM] = { 16, 4 },
+ [VK_FORMAT_BC3_SRGB] = { 16, 4 },
+ [VK_FORMAT_BC4_UNORM] = { 8, 4 },
+ [VK_FORMAT_BC4_SNORM] = { 8, 4 },
+ [VK_FORMAT_BC5_UNORM] = { 16, 4 },
+ [VK_FORMAT_BC5_SNORM] = { 16, 4 },
+ [VK_FORMAT_BC6H_UFLOAT] = { 16, 4 },
+ [VK_FORMAT_BC6H_SFLOAT] = { 16, 4 },
+ [VK_FORMAT_BC7_UNORM] = { 16, 4 },
+ [VK_FORMAT_BC7_SRGB] = { 16, 4 },
/* TODO: Initialize remaining compressed formats. */
- [VK_FMT_ETC2_R8G8B8_UNORM] = { 0, 0 },
- [VK_FMT_ETC2_R8G8B8A1_UNORM] = { 0, 0 },
- [VK_FMT_ETC2_R8G8B8A8_UNORM] = { 0, 0 },
- [VK_FMT_EAC_R11_UNORM] = { 0, 0 },
- [VK_FMT_EAC_R11_SNORM] = { 0, 0 },
- [VK_FMT_EAC_R11G11_UNORM] = { 0, 0 },
- [VK_FMT_EAC_R11G11_SNORM] = { 0, 0 },
- [VK_FMT_ASTC_4x4_UNORM] = { 0, 0 },
- [VK_FMT_ASTC_4x4_SRGB] = { 0, 0 },
- [VK_FMT_ASTC_5x4_UNORM] = { 0, 0 },
- [VK_FMT_ASTC_5x4_SRGB] = { 0, 0 },
- [VK_FMT_ASTC_5x5_UNORM] = { 0, 0 },
- [VK_FMT_ASTC_5x5_SRGB] = { 0, 0 },
- [VK_FMT_ASTC_6x5_UNORM] = { 0, 0 },
- [VK_FMT_ASTC_6x5_SRGB] = { 0, 0 },
- [VK_FMT_ASTC_6x6_UNORM] = { 0, 0 },
- [VK_FMT_ASTC_6x6_SRGB] = { 0, 0 },
- [VK_FMT_ASTC_8x5_UNORM] = { 0, 0 },
- [VK_FMT_ASTC_8x5_SRGB] = { 0, 0 },
- [VK_FMT_ASTC_8x6_UNORM] = { 0, 0 },
- [VK_FMT_ASTC_8x6_SRGB] = { 0, 0 },
- [VK_FMT_ASTC_8x8_UNORM] = { 0, 0 },
- [VK_FMT_ASTC_8x8_SRGB] = { 0, 0 },
- [VK_FMT_ASTC_10x5_UNORM] = { 0, 0 },
- [VK_FMT_ASTC_10x5_SRGB] = { 0, 0 },
- [VK_FMT_ASTC_10x6_UNORM] = { 0, 0 },
- [VK_FMT_ASTC_10x6_SRGB] = { 0, 0 },
- [VK_FMT_ASTC_10x8_UNORM] = { 0, 0 },
- [VK_FMT_ASTC_10x8_SRGB] = { 0, 0 },
- [VK_FMT_ASTC_10x10_UNORM] = { 0, 0 },
- [VK_FMT_ASTC_10x10_SRGB] = { 0, 0 },
- [VK_FMT_ASTC_12x10_UNORM] = { 0, 0 },
- [VK_FMT_ASTC_12x10_SRGB] = { 0, 0 },
- [VK_FMT_ASTC_12x12_UNORM] = { 0, 0 },
- [VK_FMT_ASTC_12x12_SRGB] = { 0, 0 },
- [VK_FMT_B5G6R5_UNORM] = { 2, 3 },
- [VK_FMT_B5G6R5_USCALED] = { 2, 3 },
- [VK_FMT_B8G8R8_UNORM] = { 3, 3 },
- [VK_FMT_B8G8R8_SNORM] = { 3, 3 },
- [VK_FMT_B8G8R8_USCALED] = { 3, 3 },
- [VK_FMT_B8G8R8_SSCALED] = { 3, 3 },
- [VK_FMT_B8G8R8_UINT] = { 3, 3 },
- [VK_FMT_B8G8R8_SINT] = { 3, 3 },
- [VK_FMT_B8G8R8_SRGB] = { 3, 3 },
- [VK_FMT_B8G8R8A8_UNORM] = { 4, 4 },
- [VK_FMT_B8G8R8A8_SNORM] = { 4, 4 },
- [VK_FMT_B8G8R8A8_USCALED] = { 4, 4 },
- [VK_FMT_B8G8R8A8_SSCALED] = { 4, 4 },
- [VK_FMT_B8G8R8A8_UINT] = { 4, 4 },
- [VK_FMT_B8G8R8A8_SINT] = { 4, 4 },
- [VK_FMT_B8G8R8A8_SRGB] = { 4, 4 },
- [VK_FMT_B10G10R10A2_UNORM] = { 4, 4 },
- [VK_FMT_B10G10R10A2_SNORM] = { 4, 4 },
- [VK_FMT_B10G10R10A2_USCALED] = { 4, 4 },
- [VK_FMT_B10G10R10A2_SSCALED] = { 4, 4 },
- [VK_FMT_B10G10R10A2_UINT] = { 4, 4 },
- [VK_FMT_B10G10R10A2_SINT] = { 4, 4 },
+ [VK_FORMAT_ETC2_R8G8B8_UNORM] = { 0, 0 },
+ [VK_FORMAT_ETC2_R8G8B8A1_UNORM] = { 0, 0 },
+ [VK_FORMAT_ETC2_R8G8B8A8_UNORM] = { 0, 0 },
+ [VK_FORMAT_EAC_R11_UNORM] = { 0, 0 },
+ [VK_FORMAT_EAC_R11_SNORM] = { 0, 0 },
+ [VK_FORMAT_EAC_R11G11_UNORM] = { 0, 0 },
+ [VK_FORMAT_EAC_R11G11_SNORM] = { 0, 0 },
+ [VK_FORMAT_ASTC_4x4_UNORM] = { 0, 0 },
+ [VK_FORMAT_ASTC_4x4_SRGB] = { 0, 0 },
+ [VK_FORMAT_ASTC_5x4_UNORM] = { 0, 0 },
+ [VK_FORMAT_ASTC_5x4_SRGB] = { 0, 0 },
+ [VK_FORMAT_ASTC_5x5_UNORM] = { 0, 0 },
+ [VK_FORMAT_ASTC_5x5_SRGB] = { 0, 0 },
+ [VK_FORMAT_ASTC_6x5_UNORM] = { 0, 0 },
+ [VK_FORMAT_ASTC_6x5_SRGB] = { 0, 0 },
+ [VK_FORMAT_ASTC_6x6_UNORM] = { 0, 0 },
+ [VK_FORMAT_ASTC_6x6_SRGB] = { 0, 0 },
+ [VK_FORMAT_ASTC_8x5_UNORM] = { 0, 0 },
+ [VK_FORMAT_ASTC_8x5_SRGB] = { 0, 0 },
+ [VK_FORMAT_ASTC_8x6_UNORM] = { 0, 0 },
+ [VK_FORMAT_ASTC_8x6_SRGB] = { 0, 0 },
+ [VK_FORMAT_ASTC_8x8_UNORM] = { 0, 0 },
+ [VK_FORMAT_ASTC_8x8_SRGB] = { 0, 0 },
+ [VK_FORMAT_ASTC_10x5_UNORM] = { 0, 0 },
+ [VK_FORMAT_ASTC_10x5_SRGB] = { 0, 0 },
+ [VK_FORMAT_ASTC_10x6_UNORM] = { 0, 0 },
+ [VK_FORMAT_ASTC_10x6_SRGB] = { 0, 0 },
+ [VK_FORMAT_ASTC_10x8_UNORM] = { 0, 0 },
+ [VK_FORMAT_ASTC_10x8_SRGB] = { 0, 0 },
+ [VK_FORMAT_ASTC_10x10_UNORM] = { 0, 0 },
+ [VK_FORMAT_ASTC_10x10_SRGB] = { 0, 0 },
+ [VK_FORMAT_ASTC_12x10_UNORM] = { 0, 0 },
+ [VK_FORMAT_ASTC_12x10_SRGB] = { 0, 0 },
+ [VK_FORMAT_ASTC_12x12_UNORM] = { 0, 0 },
+ [VK_FORMAT_ASTC_12x12_SRGB] = { 0, 0 },
+ [VK_FORMAT_B5G6R5_UNORM] = { 2, 3 },
+ [VK_FORMAT_B5G6R5_USCALED] = { 2, 3 },
+ [VK_FORMAT_B8G8R8_UNORM] = { 3, 3 },
+ [VK_FORMAT_B8G8R8_SNORM] = { 3, 3 },
+ [VK_FORMAT_B8G8R8_USCALED] = { 3, 3 },
+ [VK_FORMAT_B8G8R8_SSCALED] = { 3, 3 },
+ [VK_FORMAT_B8G8R8_UINT] = { 3, 3 },
+ [VK_FORMAT_B8G8R8_SINT] = { 3, 3 },
+ [VK_FORMAT_B8G8R8_SRGB] = { 3, 3 },
+ [VK_FORMAT_B8G8R8A8_UNORM] = { 4, 4 },
+ [VK_FORMAT_B8G8R8A8_SNORM] = { 4, 4 },
+ [VK_FORMAT_B8G8R8A8_USCALED] = { 4, 4 },
+ [VK_FORMAT_B8G8R8A8_SSCALED] = { 4, 4 },
+ [VK_FORMAT_B8G8R8A8_UINT] = { 4, 4 },
+ [VK_FORMAT_B8G8R8A8_SINT] = { 4, 4 },
+ [VK_FORMAT_B8G8R8A8_SRGB] = { 4, 4 },
+ [VK_FORMAT_B10G10R10A2_UNORM] = { 4, 4 },
+ [VK_FORMAT_B10G10R10A2_SNORM] = { 4, 4 },
+ [VK_FORMAT_B10G10R10A2_USCALED] = { 4, 4 },
+ [VK_FORMAT_B10G10R10A2_SSCALED] = { 4, 4 },
+ [VK_FORMAT_B10G10R10A2_UINT] = { 4, 4 },
+ [VK_FORMAT_B10G10R10A2_SINT] = { 4, 4 },
};
bool icd_format_is_ds(VkFormat format)
@@ -210,13 +210,13 @@
bool is_ds = false;
switch (format) {
- case VK_FMT_D16_UNORM:
- case VK_FMT_D24_UNORM:
- case VK_FMT_D32_SFLOAT:
- case VK_FMT_S8_UINT:
- case VK_FMT_D16_UNORM_S8_UINT:
- case VK_FMT_D24_UNORM_S8_UINT:
- case VK_FMT_D32_SFLOAT_S8_UINT:
+ case VK_FORMAT_D16_UNORM:
+ case VK_FORMAT_D24_UNORM:
+ case VK_FORMAT_D32_SFLOAT:
+ case VK_FORMAT_S8_UINT:
+ case VK_FORMAT_D16_UNORM_S8_UINT:
+ case VK_FORMAT_D24_UNORM_S8_UINT:
+ case VK_FORMAT_D32_SFLOAT_S8_UINT:
is_ds = true;
break;
default:
@@ -231,64 +231,64 @@
bool is_norm = false;
switch (format) {
- case VK_FMT_R4G4_UNORM:
- case VK_FMT_R4G4B4A4_UNORM:
- case VK_FMT_R5G6B5_UNORM:
- case VK_FMT_R5G5B5A1_UNORM:
- case VK_FMT_R8_UNORM:
- case VK_FMT_R8_SNORM:
- case VK_FMT_R8G8_UNORM:
- case VK_FMT_R8G8_SNORM:
- case VK_FMT_R8G8B8_UNORM:
- case VK_FMT_R8G8B8_SNORM:
- case VK_FMT_R8G8B8A8_UNORM:
- case VK_FMT_R8G8B8A8_SNORM:
- case VK_FMT_R10G10B10A2_UNORM:
- case VK_FMT_R10G10B10A2_SNORM:
- case VK_FMT_R16_UNORM:
- case VK_FMT_R16_SNORM:
- case VK_FMT_R16G16_UNORM:
- case VK_FMT_R16G16_SNORM:
- case VK_FMT_R16G16B16_UNORM:
- case VK_FMT_R16G16B16_SNORM:
- case VK_FMT_R16G16B16A16_UNORM:
- case VK_FMT_R16G16B16A16_SNORM:
- case VK_FMT_BC1_RGB_UNORM:
- case VK_FMT_BC2_UNORM:
- case VK_FMT_BC3_UNORM:
- case VK_FMT_BC4_UNORM:
- case VK_FMT_BC4_SNORM:
- case VK_FMT_BC5_UNORM:
- case VK_FMT_BC5_SNORM:
- case VK_FMT_BC7_UNORM:
- case VK_FMT_ETC2_R8G8B8_UNORM:
- case VK_FMT_ETC2_R8G8B8A1_UNORM:
- case VK_FMT_ETC2_R8G8B8A8_UNORM:
- case VK_FMT_EAC_R11_UNORM:
- case VK_FMT_EAC_R11_SNORM:
- case VK_FMT_EAC_R11G11_UNORM:
- case VK_FMT_EAC_R11G11_SNORM:
- case VK_FMT_ASTC_4x4_UNORM:
- case VK_FMT_ASTC_5x4_UNORM:
- case VK_FMT_ASTC_5x5_UNORM:
- case VK_FMT_ASTC_6x5_UNORM:
- case VK_FMT_ASTC_6x6_UNORM:
- case VK_FMT_ASTC_8x5_UNORM:
- case VK_FMT_ASTC_8x6_UNORM:
- case VK_FMT_ASTC_8x8_UNORM:
- case VK_FMT_ASTC_10x5_UNORM:
- case VK_FMT_ASTC_10x6_UNORM:
- case VK_FMT_ASTC_10x8_UNORM:
- case VK_FMT_ASTC_10x10_UNORM:
- case VK_FMT_ASTC_12x10_UNORM:
- case VK_FMT_ASTC_12x12_UNORM:
- case VK_FMT_B5G6R5_UNORM:
- case VK_FMT_B8G8R8_UNORM:
- case VK_FMT_B8G8R8_SNORM:
- case VK_FMT_B8G8R8A8_UNORM:
- case VK_FMT_B8G8R8A8_SNORM:
- case VK_FMT_B10G10R10A2_UNORM:
- case VK_FMT_B10G10R10A2_SNORM:
+ case VK_FORMAT_R4G4_UNORM:
+ case VK_FORMAT_R4G4B4A4_UNORM:
+ case VK_FORMAT_R5G6B5_UNORM:
+ case VK_FORMAT_R5G5B5A1_UNORM:
+ case VK_FORMAT_R8_UNORM:
+ case VK_FORMAT_R8_SNORM:
+ case VK_FORMAT_R8G8_UNORM:
+ case VK_FORMAT_R8G8_SNORM:
+ case VK_FORMAT_R8G8B8_UNORM:
+ case VK_FORMAT_R8G8B8_SNORM:
+ case VK_FORMAT_R8G8B8A8_UNORM:
+ case VK_FORMAT_R8G8B8A8_SNORM:
+ case VK_FORMAT_R10G10B10A2_UNORM:
+ case VK_FORMAT_R10G10B10A2_SNORM:
+ case VK_FORMAT_R16_UNORM:
+ case VK_FORMAT_R16_SNORM:
+ case VK_FORMAT_R16G16_UNORM:
+ case VK_FORMAT_R16G16_SNORM:
+ case VK_FORMAT_R16G16B16_UNORM:
+ case VK_FORMAT_R16G16B16_SNORM:
+ case VK_FORMAT_R16G16B16A16_UNORM:
+ case VK_FORMAT_R16G16B16A16_SNORM:
+ case VK_FORMAT_BC1_RGB_UNORM:
+ case VK_FORMAT_BC2_UNORM:
+ case VK_FORMAT_BC3_UNORM:
+ case VK_FORMAT_BC4_UNORM:
+ case VK_FORMAT_BC4_SNORM:
+ case VK_FORMAT_BC5_UNORM:
+ case VK_FORMAT_BC5_SNORM:
+ case VK_FORMAT_BC7_UNORM:
+ case VK_FORMAT_ETC2_R8G8B8_UNORM:
+ case VK_FORMAT_ETC2_R8G8B8A1_UNORM:
+ case VK_FORMAT_ETC2_R8G8B8A8_UNORM:
+ case VK_FORMAT_EAC_R11_UNORM:
+ case VK_FORMAT_EAC_R11_SNORM:
+ case VK_FORMAT_EAC_R11G11_UNORM:
+ case VK_FORMAT_EAC_R11G11_SNORM:
+ case VK_FORMAT_ASTC_4x4_UNORM:
+ case VK_FORMAT_ASTC_5x4_UNORM:
+ case VK_FORMAT_ASTC_5x5_UNORM:
+ case VK_FORMAT_ASTC_6x5_UNORM:
+ case VK_FORMAT_ASTC_6x6_UNORM:
+ case VK_FORMAT_ASTC_8x5_UNORM:
+ case VK_FORMAT_ASTC_8x6_UNORM:
+ case VK_FORMAT_ASTC_8x8_UNORM:
+ case VK_FORMAT_ASTC_10x5_UNORM:
+ case VK_FORMAT_ASTC_10x6_UNORM:
+ case VK_FORMAT_ASTC_10x8_UNORM:
+ case VK_FORMAT_ASTC_10x10_UNORM:
+ case VK_FORMAT_ASTC_12x10_UNORM:
+ case VK_FORMAT_ASTC_12x12_UNORM:
+ case VK_FORMAT_B5G6R5_UNORM:
+ case VK_FORMAT_B8G8R8_UNORM:
+ case VK_FORMAT_B8G8R8_SNORM:
+ case VK_FORMAT_B8G8R8A8_UNORM:
+ case VK_FORMAT_B8G8R8A8_SNORM:
+ case VK_FORMAT_B10G10R10A2_UNORM:
+ case VK_FORMAT_B10G10R10A2_SNORM:
is_norm = true;
break;
default:
@@ -303,38 +303,38 @@
bool is_int = false;
switch (format) {
- case VK_FMT_R8_UINT:
- case VK_FMT_R8_SINT:
- case VK_FMT_R8G8_UINT:
- case VK_FMT_R8G8_SINT:
- case VK_FMT_R8G8B8_UINT:
- case VK_FMT_R8G8B8_SINT:
- case VK_FMT_R8G8B8A8_UINT:
- case VK_FMT_R8G8B8A8_SINT:
- case VK_FMT_R10G10B10A2_UINT:
- case VK_FMT_R10G10B10A2_SINT:
- case VK_FMT_R16_UINT:
- case VK_FMT_R16_SINT:
- case VK_FMT_R16G16_UINT:
- case VK_FMT_R16G16_SINT:
- case VK_FMT_R16G16B16_UINT:
- case VK_FMT_R16G16B16_SINT:
- case VK_FMT_R16G16B16A16_UINT:
- case VK_FMT_R16G16B16A16_SINT:
- case VK_FMT_R32_UINT:
- case VK_FMT_R32_SINT:
- case VK_FMT_R32G32_UINT:
- case VK_FMT_R32G32_SINT:
- case VK_FMT_R32G32B32_UINT:
- case VK_FMT_R32G32B32_SINT:
- case VK_FMT_R32G32B32A32_UINT:
- case VK_FMT_R32G32B32A32_SINT:
- case VK_FMT_B8G8R8_UINT:
- case VK_FMT_B8G8R8_SINT:
- case VK_FMT_B8G8R8A8_UINT:
- case VK_FMT_B8G8R8A8_SINT:
- case VK_FMT_B10G10R10A2_UINT:
- case VK_FMT_B10G10R10A2_SINT:
+ case VK_FORMAT_R8_UINT:
+ case VK_FORMAT_R8_SINT:
+ case VK_FORMAT_R8G8_UINT:
+ case VK_FORMAT_R8G8_SINT:
+ case VK_FORMAT_R8G8B8_UINT:
+ case VK_FORMAT_R8G8B8_SINT:
+ case VK_FORMAT_R8G8B8A8_UINT:
+ case VK_FORMAT_R8G8B8A8_SINT:
+ case VK_FORMAT_R10G10B10A2_UINT:
+ case VK_FORMAT_R10G10B10A2_SINT:
+ case VK_FORMAT_R16_UINT:
+ case VK_FORMAT_R16_SINT:
+ case VK_FORMAT_R16G16_UINT:
+ case VK_FORMAT_R16G16_SINT:
+ case VK_FORMAT_R16G16B16_UINT:
+ case VK_FORMAT_R16G16B16_SINT:
+ case VK_FORMAT_R16G16B16A16_UINT:
+ case VK_FORMAT_R16G16B16A16_SINT:
+ case VK_FORMAT_R32_UINT:
+ case VK_FORMAT_R32_SINT:
+ case VK_FORMAT_R32G32_UINT:
+ case VK_FORMAT_R32G32_SINT:
+ case VK_FORMAT_R32G32B32_UINT:
+ case VK_FORMAT_R32G32B32_SINT:
+ case VK_FORMAT_R32G32B32A32_UINT:
+ case VK_FORMAT_R32G32B32A32_SINT:
+ case VK_FORMAT_B8G8R8_UINT:
+ case VK_FORMAT_B8G8R8_SINT:
+ case VK_FORMAT_B8G8R8A8_UINT:
+ case VK_FORMAT_B8G8R8A8_SINT:
+ case VK_FORMAT_B10G10R10A2_UINT:
+ case VK_FORMAT_B10G10R10A2_SINT:
is_int = true;
break;
default:
@@ -349,22 +349,22 @@
bool is_float = false;
switch (format) {
- case VK_FMT_R16_SFLOAT:
- case VK_FMT_R16G16_SFLOAT:
- case VK_FMT_R16G16B16_SFLOAT:
- case VK_FMT_R16G16B16A16_SFLOAT:
- case VK_FMT_R32_SFLOAT:
- case VK_FMT_R32G32_SFLOAT:
- case VK_FMT_R32G32B32_SFLOAT:
- case VK_FMT_R32G32B32A32_SFLOAT:
- case VK_FMT_R64_SFLOAT:
- case VK_FMT_R64G64_SFLOAT:
- case VK_FMT_R64G64B64_SFLOAT:
- case VK_FMT_R64G64B64A64_SFLOAT:
- case VK_FMT_R11G11B10_UFLOAT:
- case VK_FMT_R9G9B9E5_UFLOAT:
- case VK_FMT_BC6H_UFLOAT:
- case VK_FMT_BC6H_SFLOAT:
+ case VK_FORMAT_R16_SFLOAT:
+ case VK_FORMAT_R16G16_SFLOAT:
+ case VK_FORMAT_R16G16B16_SFLOAT:
+ case VK_FORMAT_R16G16B16A16_SFLOAT:
+ case VK_FORMAT_R32_SFLOAT:
+ case VK_FORMAT_R32G32_SFLOAT:
+ case VK_FORMAT_R32G32B32_SFLOAT:
+ case VK_FORMAT_R32G32B32A32_SFLOAT:
+ case VK_FORMAT_R64_SFLOAT:
+ case VK_FORMAT_R64G64_SFLOAT:
+ case VK_FORMAT_R64G64B64_SFLOAT:
+ case VK_FORMAT_R64G64B64A64_SFLOAT:
+ case VK_FORMAT_R11G11B10_UFLOAT:
+ case VK_FORMAT_R9G9B9E5_UFLOAT:
+ case VK_FORMAT_BC6H_UFLOAT:
+ case VK_FORMAT_BC6H_SFLOAT:
is_float = true;
break;
default:
@@ -379,30 +379,30 @@
bool is_srgb = false;
switch (format) {
- case VK_FMT_R8_SRGB:
- case VK_FMT_R8G8_SRGB:
- case VK_FMT_R8G8B8_SRGB:
- case VK_FMT_R8G8B8A8_SRGB:
- case VK_FMT_BC1_RGB_SRGB:
- case VK_FMT_BC2_SRGB:
- case VK_FMT_BC3_SRGB:
- case VK_FMT_BC7_SRGB:
- case VK_FMT_ASTC_4x4_SRGB:
- case VK_FMT_ASTC_5x4_SRGB:
- case VK_FMT_ASTC_5x5_SRGB:
- case VK_FMT_ASTC_6x5_SRGB:
- case VK_FMT_ASTC_6x6_SRGB:
- case VK_FMT_ASTC_8x5_SRGB:
- case VK_FMT_ASTC_8x6_SRGB:
- case VK_FMT_ASTC_8x8_SRGB:
- case VK_FMT_ASTC_10x5_SRGB:
- case VK_FMT_ASTC_10x6_SRGB:
- case VK_FMT_ASTC_10x8_SRGB:
- case VK_FMT_ASTC_10x10_SRGB:
- case VK_FMT_ASTC_12x10_SRGB:
- case VK_FMT_ASTC_12x12_SRGB:
- case VK_FMT_B8G8R8_SRGB:
- case VK_FMT_B8G8R8A8_SRGB:
+ case VK_FORMAT_R8_SRGB:
+ case VK_FORMAT_R8G8_SRGB:
+ case VK_FORMAT_R8G8B8_SRGB:
+ case VK_FORMAT_R8G8B8A8_SRGB:
+ case VK_FORMAT_BC1_RGB_SRGB:
+ case VK_FORMAT_BC2_SRGB:
+ case VK_FORMAT_BC3_SRGB:
+ case VK_FORMAT_BC7_SRGB:
+ case VK_FORMAT_ASTC_4x4_SRGB:
+ case VK_FORMAT_ASTC_5x4_SRGB:
+ case VK_FORMAT_ASTC_5x5_SRGB:
+ case VK_FORMAT_ASTC_6x5_SRGB:
+ case VK_FORMAT_ASTC_6x6_SRGB:
+ case VK_FORMAT_ASTC_8x5_SRGB:
+ case VK_FORMAT_ASTC_8x6_SRGB:
+ case VK_FORMAT_ASTC_8x8_SRGB:
+ case VK_FORMAT_ASTC_10x5_SRGB:
+ case VK_FORMAT_ASTC_10x6_SRGB:
+ case VK_FORMAT_ASTC_10x8_SRGB:
+ case VK_FORMAT_ASTC_10x10_SRGB:
+ case VK_FORMAT_ASTC_12x10_SRGB:
+ case VK_FORMAT_ASTC_12x12_SRGB:
+ case VK_FORMAT_B8G8R8_SRGB:
+ case VK_FORMAT_B8G8R8A8_SRGB:
is_srgb = true;
break;
default:
@@ -415,55 +415,55 @@
bool icd_format_is_compressed(VkFormat format)
{
switch (format) {
- case VK_FMT_BC1_RGB_UNORM:
- case VK_FMT_BC1_RGB_SRGB:
- case VK_FMT_BC2_UNORM:
- case VK_FMT_BC2_SRGB:
- case VK_FMT_BC3_UNORM:
- case VK_FMT_BC3_SRGB:
- case VK_FMT_BC4_UNORM:
- case VK_FMT_BC4_SNORM:
- case VK_FMT_BC5_UNORM:
- case VK_FMT_BC5_SNORM:
- case VK_FMT_BC6H_UFLOAT:
- case VK_FMT_BC6H_SFLOAT:
- case VK_FMT_BC7_UNORM:
- case VK_FMT_BC7_SRGB:
- case VK_FMT_ETC2_R8G8B8_UNORM:
- case VK_FMT_ETC2_R8G8B8A1_UNORM:
- case VK_FMT_ETC2_R8G8B8A8_UNORM:
- case VK_FMT_EAC_R11_UNORM:
- case VK_FMT_EAC_R11_SNORM:
- case VK_FMT_EAC_R11G11_UNORM:
- case VK_FMT_EAC_R11G11_SNORM:
- case VK_FMT_ASTC_4x4_UNORM:
- case VK_FMT_ASTC_4x4_SRGB:
- case VK_FMT_ASTC_5x4_UNORM:
- case VK_FMT_ASTC_5x4_SRGB:
- case VK_FMT_ASTC_5x5_UNORM:
- case VK_FMT_ASTC_5x5_SRGB:
- case VK_FMT_ASTC_6x5_UNORM:
- case VK_FMT_ASTC_6x5_SRGB:
- case VK_FMT_ASTC_6x6_UNORM:
- case VK_FMT_ASTC_6x6_SRGB:
- case VK_FMT_ASTC_8x5_UNORM:
- case VK_FMT_ASTC_8x5_SRGB:
- case VK_FMT_ASTC_8x6_UNORM:
- case VK_FMT_ASTC_8x6_SRGB:
- case VK_FMT_ASTC_8x8_UNORM:
- case VK_FMT_ASTC_8x8_SRGB:
- case VK_FMT_ASTC_10x5_UNORM:
- case VK_FMT_ASTC_10x5_SRGB:
- case VK_FMT_ASTC_10x6_UNORM:
- case VK_FMT_ASTC_10x6_SRGB:
- case VK_FMT_ASTC_10x8_UNORM:
- case VK_FMT_ASTC_10x8_SRGB:
- case VK_FMT_ASTC_10x10_UNORM:
- case VK_FMT_ASTC_10x10_SRGB:
- case VK_FMT_ASTC_12x10_UNORM:
- case VK_FMT_ASTC_12x10_SRGB:
- case VK_FMT_ASTC_12x12_UNORM:
- case VK_FMT_ASTC_12x12_SRGB:
+ case VK_FORMAT_BC1_RGB_UNORM:
+ case VK_FORMAT_BC1_RGB_SRGB:
+ case VK_FORMAT_BC2_UNORM:
+ case VK_FORMAT_BC2_SRGB:
+ case VK_FORMAT_BC3_UNORM:
+ case VK_FORMAT_BC3_SRGB:
+ case VK_FORMAT_BC4_UNORM:
+ case VK_FORMAT_BC4_SNORM:
+ case VK_FORMAT_BC5_UNORM:
+ case VK_FORMAT_BC5_SNORM:
+ case VK_FORMAT_BC6H_UFLOAT:
+ case VK_FORMAT_BC6H_SFLOAT:
+ case VK_FORMAT_BC7_UNORM:
+ case VK_FORMAT_BC7_SRGB:
+ case VK_FORMAT_ETC2_R8G8B8_UNORM:
+ case VK_FORMAT_ETC2_R8G8B8A1_UNORM:
+ case VK_FORMAT_ETC2_R8G8B8A8_UNORM:
+ case VK_FORMAT_EAC_R11_UNORM:
+ case VK_FORMAT_EAC_R11_SNORM:
+ case VK_FORMAT_EAC_R11G11_UNORM:
+ case VK_FORMAT_EAC_R11G11_SNORM:
+ case VK_FORMAT_ASTC_4x4_UNORM:
+ case VK_FORMAT_ASTC_4x4_SRGB:
+ case VK_FORMAT_ASTC_5x4_UNORM:
+ case VK_FORMAT_ASTC_5x4_SRGB:
+ case VK_FORMAT_ASTC_5x5_UNORM:
+ case VK_FORMAT_ASTC_5x5_SRGB:
+ case VK_FORMAT_ASTC_6x5_UNORM:
+ case VK_FORMAT_ASTC_6x5_SRGB:
+ case VK_FORMAT_ASTC_6x6_UNORM:
+ case VK_FORMAT_ASTC_6x6_SRGB:
+ case VK_FORMAT_ASTC_8x5_UNORM:
+ case VK_FORMAT_ASTC_8x5_SRGB:
+ case VK_FORMAT_ASTC_8x6_UNORM:
+ case VK_FORMAT_ASTC_8x6_SRGB:
+ case VK_FORMAT_ASTC_8x8_UNORM:
+ case VK_FORMAT_ASTC_8x8_SRGB:
+ case VK_FORMAT_ASTC_10x5_UNORM:
+ case VK_FORMAT_ASTC_10x5_SRGB:
+ case VK_FORMAT_ASTC_10x6_UNORM:
+ case VK_FORMAT_ASTC_10x6_SRGB:
+ case VK_FORMAT_ASTC_10x8_UNORM:
+ case VK_FORMAT_ASTC_10x8_SRGB:
+ case VK_FORMAT_ASTC_10x10_UNORM:
+ case VK_FORMAT_ASTC_10x10_SRGB:
+ case VK_FORMAT_ASTC_12x10_UNORM:
+ case VK_FORMAT_ASTC_12x10_SRGB:
+ case VK_FORMAT_ASTC_12x12_UNORM:
+ case VK_FORMAT_ASTC_12x12_SRGB:
return true;
default:
return false;
@@ -490,228 +490,228 @@
{
/* assume little-endian */
switch (format) {
- case VK_FMT_UNDEFINED:
+ case VK_FORMAT_UNDEFINED:
break;
- case VK_FMT_R4G4_UNORM:
- case VK_FMT_R4G4_USCALED:
+ case VK_FORMAT_R4G4_UNORM:
+ case VK_FORMAT_R4G4_USCALED:
((uint8_t *) value)[0] = (color[0] & 0xf) << 0 |
(color[1] & 0xf) << 4;
break;
- case VK_FMT_R4G4B4A4_UNORM:
- case VK_FMT_R4G4B4A4_USCALED:
+ case VK_FORMAT_R4G4B4A4_UNORM:
+ case VK_FORMAT_R4G4B4A4_USCALED:
((uint16_t *) value)[0] = (color[0] & 0xf) << 0 |
(color[1] & 0xf) << 4 |
(color[2] & 0xf) << 8 |
(color[3] & 0xf) << 12;
break;
- case VK_FMT_R5G6B5_UNORM:
- case VK_FMT_R5G6B5_USCALED:
+ case VK_FORMAT_R5G6B5_UNORM:
+ case VK_FORMAT_R5G6B5_USCALED:
((uint16_t *) value)[0] = (color[0] & 0x1f) << 0 |
(color[1] & 0x3f) << 5 |
(color[2] & 0x1f) << 11;
break;
- case VK_FMT_B5G6R5_UNORM:
+ case VK_FORMAT_B5G6R5_UNORM:
((uint16_t *) value)[0] = (color[2] & 0x1f) << 0 |
(color[1] & 0x3f) << 5 |
(color[0] & 0x1f) << 11;
break;
- case VK_FMT_R5G5B5A1_UNORM:
- case VK_FMT_R5G5B5A1_USCALED:
+ case VK_FORMAT_R5G5B5A1_UNORM:
+ case VK_FORMAT_R5G5B5A1_USCALED:
((uint16_t *) value)[0] = (color[0] & 0x1f) << 0 |
(color[1] & 0x1f) << 5 |
(color[2] & 0x1f) << 10 |
(color[3] & 0x1) << 15;
break;
- case VK_FMT_R8_UNORM:
- case VK_FMT_R8_SNORM:
- case VK_FMT_R8_USCALED:
- case VK_FMT_R8_SSCALED:
- case VK_FMT_R8_UINT:
- case VK_FMT_R8_SINT:
- case VK_FMT_R8_SRGB:
+ case VK_FORMAT_R8_UNORM:
+ case VK_FORMAT_R8_SNORM:
+ case VK_FORMAT_R8_USCALED:
+ case VK_FORMAT_R8_SSCALED:
+ case VK_FORMAT_R8_UINT:
+ case VK_FORMAT_R8_SINT:
+ case VK_FORMAT_R8_SRGB:
((uint8_t *) value)[0] = (uint8_t) color[0];
break;
- case VK_FMT_R8G8_UNORM:
- case VK_FMT_R8G8_SNORM:
- case VK_FMT_R8G8_USCALED:
- case VK_FMT_R8G8_SSCALED:
- case VK_FMT_R8G8_UINT:
- case VK_FMT_R8G8_SINT:
- case VK_FMT_R8G8_SRGB:
+ case VK_FORMAT_R8G8_UNORM:
+ case VK_FORMAT_R8G8_SNORM:
+ case VK_FORMAT_R8G8_USCALED:
+ case VK_FORMAT_R8G8_SSCALED:
+ case VK_FORMAT_R8G8_UINT:
+ case VK_FORMAT_R8G8_SINT:
+ case VK_FORMAT_R8G8_SRGB:
((uint8_t *) value)[0] = (uint8_t) color[0];
((uint8_t *) value)[1] = (uint8_t) color[1];
break;
- case VK_FMT_R8G8B8A8_UNORM:
- case VK_FMT_R8G8B8A8_SNORM:
- case VK_FMT_R8G8B8A8_USCALED:
- case VK_FMT_R8G8B8A8_SSCALED:
- case VK_FMT_R8G8B8A8_UINT:
- case VK_FMT_R8G8B8A8_SINT:
- case VK_FMT_R8G8B8A8_SRGB:
+ case VK_FORMAT_R8G8B8A8_UNORM:
+ case VK_FORMAT_R8G8B8A8_SNORM:
+ case VK_FORMAT_R8G8B8A8_USCALED:
+ case VK_FORMAT_R8G8B8A8_SSCALED:
+ case VK_FORMAT_R8G8B8A8_UINT:
+ case VK_FORMAT_R8G8B8A8_SINT:
+ case VK_FORMAT_R8G8B8A8_SRGB:
((uint8_t *) value)[0] = (uint8_t) color[0];
((uint8_t *) value)[1] = (uint8_t) color[1];
((uint8_t *) value)[2] = (uint8_t) color[2];
((uint8_t *) value)[3] = (uint8_t) color[3];
break;
- case VK_FMT_B8G8R8A8_UNORM:
- case VK_FMT_B8G8R8A8_SRGB:
+ case VK_FORMAT_B8G8R8A8_UNORM:
+ case VK_FORMAT_B8G8R8A8_SRGB:
((uint8_t *) value)[0] = (uint8_t) color[2];
((uint8_t *) value)[1] = (uint8_t) color[1];
((uint8_t *) value)[2] = (uint8_t) color[0];
((uint8_t *) value)[3] = (uint8_t) color[3];
break;
- case VK_FMT_R11G11B10_UFLOAT:
+ case VK_FORMAT_R11G11B10_UFLOAT:
((uint32_t *) value)[0] = (color[0] & 0x7ff) << 0 |
(color[1] & 0x7ff) << 11 |
(color[2] & 0x3ff) << 22;
break;
- case VK_FMT_R10G10B10A2_UNORM:
- case VK_FMT_R10G10B10A2_SNORM:
- case VK_FMT_R10G10B10A2_USCALED:
- case VK_FMT_R10G10B10A2_SSCALED:
- case VK_FMT_R10G10B10A2_UINT:
- case VK_FMT_R10G10B10A2_SINT:
+ case VK_FORMAT_R10G10B10A2_UNORM:
+ case VK_FORMAT_R10G10B10A2_SNORM:
+ case VK_FORMAT_R10G10B10A2_USCALED:
+ case VK_FORMAT_R10G10B10A2_SSCALED:
+ case VK_FORMAT_R10G10B10A2_UINT:
+ case VK_FORMAT_R10G10B10A2_SINT:
((uint32_t *) value)[0] = (color[0] & 0x3ff) << 0 |
(color[1] & 0x3ff) << 10 |
(color[2] & 0x3ff) << 20 |
(color[3] & 0x3) << 30;
break;
- case VK_FMT_R16_UNORM:
- case VK_FMT_R16_SNORM:
- case VK_FMT_R16_USCALED:
- case VK_FMT_R16_SSCALED:
- case VK_FMT_R16_UINT:
- case VK_FMT_R16_SINT:
- case VK_FMT_R16_SFLOAT:
+ case VK_FORMAT_R16_UNORM:
+ case VK_FORMAT_R16_SNORM:
+ case VK_FORMAT_R16_USCALED:
+ case VK_FORMAT_R16_SSCALED:
+ case VK_FORMAT_R16_UINT:
+ case VK_FORMAT_R16_SINT:
+ case VK_FORMAT_R16_SFLOAT:
((uint16_t *) value)[0] = (uint16_t) color[0];
break;
- case VK_FMT_R16G16_UNORM:
- case VK_FMT_R16G16_SNORM:
- case VK_FMT_R16G16_USCALED:
- case VK_FMT_R16G16_SSCALED:
- case VK_FMT_R16G16_UINT:
- case VK_FMT_R16G16_SINT:
- case VK_FMT_R16G16_SFLOAT:
+ case VK_FORMAT_R16G16_UNORM:
+ case VK_FORMAT_R16G16_SNORM:
+ case VK_FORMAT_R16G16_USCALED:
+ case VK_FORMAT_R16G16_SSCALED:
+ case VK_FORMAT_R16G16_UINT:
+ case VK_FORMAT_R16G16_SINT:
+ case VK_FORMAT_R16G16_SFLOAT:
((uint16_t *) value)[0] = (uint16_t) color[0];
((uint16_t *) value)[1] = (uint16_t) color[1];
break;
- case VK_FMT_R16G16B16A16_UNORM:
- case VK_FMT_R16G16B16A16_SNORM:
- case VK_FMT_R16G16B16A16_USCALED:
- case VK_FMT_R16G16B16A16_SSCALED:
- case VK_FMT_R16G16B16A16_UINT:
- case VK_FMT_R16G16B16A16_SINT:
- case VK_FMT_R16G16B16A16_SFLOAT:
+ case VK_FORMAT_R16G16B16A16_UNORM:
+ case VK_FORMAT_R16G16B16A16_SNORM:
+ case VK_FORMAT_R16G16B16A16_USCALED:
+ case VK_FORMAT_R16G16B16A16_SSCALED:
+ case VK_FORMAT_R16G16B16A16_UINT:
+ case VK_FORMAT_R16G16B16A16_SINT:
+ case VK_FORMAT_R16G16B16A16_SFLOAT:
((uint16_t *) value)[0] = (uint16_t) color[0];
((uint16_t *) value)[1] = (uint16_t) color[1];
((uint16_t *) value)[2] = (uint16_t) color[2];
((uint16_t *) value)[3] = (uint16_t) color[3];
break;
- case VK_FMT_R32_UINT:
- case VK_FMT_R32_SINT:
- case VK_FMT_R32_SFLOAT:
+ case VK_FORMAT_R32_UINT:
+ case VK_FORMAT_R32_SINT:
+ case VK_FORMAT_R32_SFLOAT:
((uint32_t *) value)[0] = color[0];
break;
- case VK_FMT_R32G32_UINT:
- case VK_FMT_R32G32_SINT:
- case VK_FMT_R32G32_SFLOAT:
+ case VK_FORMAT_R32G32_UINT:
+ case VK_FORMAT_R32G32_SINT:
+ case VK_FORMAT_R32G32_SFLOAT:
((uint32_t *) value)[0] = color[0];
((uint32_t *) value)[1] = color[1];
break;
- case VK_FMT_R32G32B32_UINT:
- case VK_FMT_R32G32B32_SINT:
- case VK_FMT_R32G32B32_SFLOAT:
+ case VK_FORMAT_R32G32B32_UINT:
+ case VK_FORMAT_R32G32B32_SINT:
+ case VK_FORMAT_R32G32B32_SFLOAT:
((uint32_t *) value)[0] = color[0];
((uint32_t *) value)[1] = color[1];
((uint32_t *) value)[2] = color[2];
break;
- case VK_FMT_R32G32B32A32_UINT:
- case VK_FMT_R32G32B32A32_SINT:
- case VK_FMT_R32G32B32A32_SFLOAT:
+ case VK_FORMAT_R32G32B32A32_UINT:
+ case VK_FORMAT_R32G32B32A32_SINT:
+ case VK_FORMAT_R32G32B32A32_SFLOAT:
((uint32_t *) value)[0] = color[0];
((uint32_t *) value)[1] = color[1];
((uint32_t *) value)[2] = color[2];
((uint32_t *) value)[3] = color[3];
break;
- case VK_FMT_D16_UNORM_S8_UINT:
+ case VK_FORMAT_D16_UNORM_S8_UINT:
((uint16_t *) value)[0] = (uint16_t) color[0];
((char *) value)[2] = (uint8_t) color[1];
break;
- case VK_FMT_D32_SFLOAT_S8_UINT:
+ case VK_FORMAT_D32_SFLOAT_S8_UINT:
((uint32_t *) value)[0] = (uint32_t) color[0];
((char *) value)[4] = (uint8_t) color[1];
break;
- case VK_FMT_R9G9B9E5_UFLOAT:
+ case VK_FORMAT_R9G9B9E5_UFLOAT:
((uint32_t *) value)[0] = (color[0] & 0x1ff) << 0 |
(color[1] & 0x1ff) << 9 |
(color[2] & 0x1ff) << 18 |
(color[3] & 0x1f) << 27;
break;
- case VK_FMT_BC1_RGB_UNORM:
- case VK_FMT_BC1_RGB_SRGB:
- case VK_FMT_BC4_UNORM:
- case VK_FMT_BC4_SNORM:
+ case VK_FORMAT_BC1_RGB_UNORM:
+ case VK_FORMAT_BC1_RGB_SRGB:
+ case VK_FORMAT_BC4_UNORM:
+ case VK_FORMAT_BC4_SNORM:
memcpy(value, color, 8);
break;
- case VK_FMT_BC2_UNORM:
- case VK_FMT_BC2_SRGB:
- case VK_FMT_BC3_UNORM:
- case VK_FMT_BC3_SRGB:
- case VK_FMT_BC5_UNORM:
- case VK_FMT_BC5_SNORM:
- case VK_FMT_BC6H_UFLOAT:
- case VK_FMT_BC6H_SFLOAT:
- case VK_FMT_BC7_UNORM:
- case VK_FMT_BC7_SRGB:
+ case VK_FORMAT_BC2_UNORM:
+ case VK_FORMAT_BC2_SRGB:
+ case VK_FORMAT_BC3_UNORM:
+ case VK_FORMAT_BC3_SRGB:
+ case VK_FORMAT_BC5_UNORM:
+ case VK_FORMAT_BC5_SNORM:
+ case VK_FORMAT_BC6H_UFLOAT:
+ case VK_FORMAT_BC6H_SFLOAT:
+ case VK_FORMAT_BC7_UNORM:
+ case VK_FORMAT_BC7_SRGB:
memcpy(value, color, 16);
break;
- case VK_FMT_R8G8B8_UNORM:
- case VK_FMT_R8G8B8_SNORM:
- case VK_FMT_R8G8B8_USCALED:
- case VK_FMT_R8G8B8_SSCALED:
- case VK_FMT_R8G8B8_UINT:
- case VK_FMT_R8G8B8_SINT:
- case VK_FMT_R8G8B8_SRGB:
+ case VK_FORMAT_R8G8B8_UNORM:
+ case VK_FORMAT_R8G8B8_SNORM:
+ case VK_FORMAT_R8G8B8_USCALED:
+ case VK_FORMAT_R8G8B8_SSCALED:
+ case VK_FORMAT_R8G8B8_UINT:
+ case VK_FORMAT_R8G8B8_SINT:
+ case VK_FORMAT_R8G8B8_SRGB:
((uint8_t *) value)[0] = (uint8_t) color[0];
((uint8_t *) value)[1] = (uint8_t) color[1];
((uint8_t *) value)[2] = (uint8_t) color[2];
break;
- case VK_FMT_R16G16B16_UNORM:
- case VK_FMT_R16G16B16_SNORM:
- case VK_FMT_R16G16B16_USCALED:
- case VK_FMT_R16G16B16_SSCALED:
- case VK_FMT_R16G16B16_UINT:
- case VK_FMT_R16G16B16_SINT:
- case VK_FMT_R16G16B16_SFLOAT:
+ case VK_FORMAT_R16G16B16_UNORM:
+ case VK_FORMAT_R16G16B16_SNORM:
+ case VK_FORMAT_R16G16B16_USCALED:
+ case VK_FORMAT_R16G16B16_SSCALED:
+ case VK_FORMAT_R16G16B16_UINT:
+ case VK_FORMAT_R16G16B16_SINT:
+ case VK_FORMAT_R16G16B16_SFLOAT:
((uint16_t *) value)[0] = (uint16_t) color[0];
((uint16_t *) value)[1] = (uint16_t) color[1];
((uint16_t *) value)[2] = (uint16_t) color[2];
break;
- case VK_FMT_B10G10R10A2_UNORM:
- case VK_FMT_B10G10R10A2_SNORM:
- case VK_FMT_B10G10R10A2_USCALED:
- case VK_FMT_B10G10R10A2_SSCALED:
- case VK_FMT_B10G10R10A2_UINT:
- case VK_FMT_B10G10R10A2_SINT:
+ case VK_FORMAT_B10G10R10A2_UNORM:
+ case VK_FORMAT_B10G10R10A2_SNORM:
+ case VK_FORMAT_B10G10R10A2_USCALED:
+ case VK_FORMAT_B10G10R10A2_SSCALED:
+ case VK_FORMAT_B10G10R10A2_UINT:
+ case VK_FORMAT_B10G10R10A2_SINT:
((uint32_t *) value)[0] = (color[2] & 0x3ff) << 0 |
(color[1] & 0x3ff) << 10 |
(color[0] & 0x3ff) << 20 |
(color[3] & 0x3) << 30;
break;
- case VK_FMT_R64_SFLOAT:
+ case VK_FORMAT_R64_SFLOAT:
/* higher 32 bits always 0 */
((uint64_t *) value)[0] = color[0];
break;
- case VK_FMT_R64G64_SFLOAT:
+ case VK_FORMAT_R64G64_SFLOAT:
((uint64_t *) value)[0] = color[0];
((uint64_t *) value)[1] = color[1];
break;
- case VK_FMT_R64G64B64_SFLOAT:
+ case VK_FORMAT_R64G64B64_SFLOAT:
((uint64_t *) value)[0] = color[0];
((uint64_t *) value)[1] = color[1];
((uint64_t *) value)[2] = color[2];
break;
- case VK_FMT_R64G64B64A64_SFLOAT:
+ case VK_FORMAT_R64G64B64A64_SFLOAT:
((uint64_t *) value)[0] = color[0];
((uint64_t *) value)[1] = color[1];
((uint64_t *) value)[2] = color[2];
diff --git a/icd/common/icd-format.h b/icd/common/icd-format.h
index fee887f..e3649c0 100644
--- a/icd/common/icd-format.h
+++ b/icd/common/icd-format.h
@@ -33,7 +33,7 @@
static inline bool icd_format_is_undef(VkFormat format)
{
- return (format == VK_FMT_UNDEFINED);
+ return (format == VK_FORMAT_UNDEFINED);
}
bool icd_format_is_ds(VkFormat format);
diff --git a/icd/common/icd-instance.c b/icd/common/icd-instance.c
index 6d6c2cf..5a6e9f7 100644
--- a/icd/common/icd-instance.c
+++ b/icd/common/icd-instance.c
@@ -76,7 +76,7 @@
alloc_cb = &default_alloc_cb;
instance = alloc_cb->pfnAlloc(alloc_cb->pUserData, sizeof(*instance), 0,
- VK_SYSTEM_ALLOC_API_OBJECT);
+ VK_SYSTEM_ALLOC_TYPE_API_OBJECT);
if (!instance)
return NULL;
@@ -85,7 +85,7 @@
name = (app_info->pAppName) ? app_info->pAppName : "unnamed";
len = strlen(name);
instance->name = alloc_cb->pfnAlloc(alloc_cb->pUserData, len + 1, 0,
- VK_SYSTEM_ALLOC_INTERNAL);
+ VK_SYSTEM_ALLOC_TYPE_INTERNAL);
if (!instance->name) {
alloc_cb->pfnFree(alloc_cb->pUserData, instance);
return NULL;
@@ -150,9 +150,9 @@
if (!logger) {
logger = icd_instance_alloc(instance, sizeof(*logger), 0,
- VK_SYSTEM_ALLOC_DEBUG);
+ VK_SYSTEM_ALLOC_TYPE_DEBUG);
if (!logger)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
logger->func = func;
logger->next = instance->loggers;
diff --git a/icd/intel/buf.c b/icd/intel/buf.c
index adcb2a1..f42571c 100644
--- a/icd/intel/buf.c
+++ b/icd/intel/buf.c
@@ -43,7 +43,7 @@
VkResult ret = VK_SUCCESS;
switch (type) {
- case VK_INFO_TYPE_MEMORY_REQUIREMENTS:
+ case VK_OBJECT_INFO_TYPE_MEMORY_REQUIREMENTS:
{
VkMemoryRequirements *mem_req = data;
@@ -87,7 +87,7 @@
buf = (struct intel_buf *) intel_base_create(&dev->base.handle,
sizeof(*buf), dev->base.dbg, VK_DBG_OBJECT_BUFFER, info, 0);
if (!buf)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
buf->size = info->size;
buf->usage = info->usage;
diff --git a/icd/intel/buf.h b/icd/intel/buf.h
index 2ef8293..e6162ea 100644
--- a/icd/intel/buf.h
+++ b/icd/intel/buf.h
@@ -34,7 +34,7 @@
struct intel_buf {
struct intel_obj obj;
- VkGpuSize size;
+ VkDeviceSize size;
VkFlags usage;
};
diff --git a/icd/intel/cmd.c b/icd/intel/cmd.c
index f4f6520..992edb4 100644
--- a/icd/intel/cmd.c
+++ b/icd/intel/cmd.c
@@ -107,7 +107,7 @@
/* reuse the old bo */
cmd_writer_discard(cmd, which);
} else {
- return VK_ERROR_OUT_OF_GPU_MEMORY;
+ return VK_ERROR_OUT_OF_DEVICE_MEMORY;
}
writer->used = 0;
@@ -152,7 +152,7 @@
new_bo = alloc_writer_bo(cmd->dev->winsys, which, new_size);
if (!new_bo) {
cmd_writer_discard(cmd, which);
- cmd_fail(cmd, VK_ERROR_OUT_OF_GPU_MEMORY);
+ cmd_fail(cmd, VK_ERROR_OUT_OF_DEVICE_MEMORY);
return;
}
@@ -192,10 +192,10 @@
struct intel_cmd_item *items;
items = intel_alloc(cmd, sizeof(writer->items[0]) * new_alloc,
- 0, VK_SYSTEM_ALLOC_DEBUG);
+ 0, VK_SYSTEM_ALLOC_TYPE_DEBUG);
if (!items) {
writer->item_used = 0;
- cmd_fail(cmd, VK_ERROR_OUT_OF_MEMORY);
+ cmd_fail(cmd, VK_ERROR_OUT_OF_HOST_MEMORY);
return;
}
@@ -275,7 +275,7 @@
cmd = (struct intel_cmd *) intel_base_create(&dev->base.handle,
sizeof(*cmd), dev->base.dbg, VK_DBG_OBJECT_CMD_BUFFER, info, 0);
if (!cmd)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
cmd->obj.destroy = cmd_destroy;
@@ -290,10 +290,10 @@
*/
cmd->reloc_count = dev->gpu->batch_buffer_reloc_count;
cmd->relocs = intel_alloc(cmd, sizeof(cmd->relocs[0]) * cmd->reloc_count,
- 4096, VK_SYSTEM_ALLOC_INTERNAL);
+ 4096, VK_SYSTEM_ALLOC_TYPE_INTERNAL);
if (!cmd->relocs) {
intel_cmd_destroy(cmd);
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
}
*cmd_ret = cmd;
@@ -345,7 +345,7 @@
const uint32_t size = cmd->dev->gpu->max_batch_buffer_size / 2;
uint32_t divider = 1;
- if (flags & VK_CMD_BUFFER_OPTIMIZE_GPU_SMALL_BATCH_BIT)
+ if (flags & VK_CMD_BUFFER_OPTIMIZE_SMALL_BATCH_BIT)
divider *= 4;
cmd->writers[INTEL_CMD_WRITER_BATCH].size = size / divider;
@@ -495,7 +495,7 @@
uint32_t startCounter,
uint32_t counterCount,
VkBuffer srcBuffer,
- VkGpuSize srcOffset)
+ VkDeviceSize srcOffset)
{
}
@@ -505,7 +505,7 @@
uint32_t startCounter,
uint32_t counterCount,
VkBuffer destBuffer,
- VkGpuSize destOffset)
+ VkDeviceSize destOffset)
{
}
diff --git a/icd/intel/cmd.h b/icd/intel/cmd.h
index cc37f7f..d03a46c 100644
--- a/icd/intel/cmd.h
+++ b/icd/intel/cmd.h
@@ -179,12 +179,12 @@
struct {
const struct intel_buf *buf[INTEL_MAX_VERTEX_BINDING_COUNT];
- VkGpuSize offset[INTEL_MAX_VERTEX_BINDING_COUNT];
+ VkDeviceSize offset[INTEL_MAX_VERTEX_BINDING_COUNT];
} vertex;
struct {
const struct intel_buf *buf;
- VkGpuSize offset;
+ VkDeviceSize offset;
VkIndexType type;
} index;
@@ -253,7 +253,7 @@
void intel_cmd_decode(struct intel_cmd *cmd, bool decode_inst_writer);
static inline struct intel_bo *intel_cmd_get_batch(const struct intel_cmd *cmd,
- VkGpuSize *used)
+ VkDeviceSize *used)
{
const struct intel_cmd_writer *writer =
&cmd->writers[INTEL_CMD_WRITER_BATCH];
diff --git a/icd/intel/cmd_barrier.c b/icd/intel/cmd_barrier.c
index 93f7d33..e87fcea 100644
--- a/icd/intel/cmd_barrier.c
+++ b/icd/intel/cmd_barrier.c
@@ -271,8 +271,12 @@
}
ICD_EXPORT void VKAPI vkCmdWaitEvents(
- VkCmdBuffer cmdBuffer,
- const VkEventWaitInfo* pWaitInfo)
+ VkCmdBuffer cmdBuffer,
+ VkWaitEvent waitEvent,
+ uint32_t eventCount,
+ const VkEvent* pEvents,
+ uint32_t memBarrierCount,
+ const void** ppMemBarriers)
{
struct intel_cmd *cmd = intel_cmd(cmdBuffer);
@@ -290,12 +294,16 @@
*/
cmd_memory_barriers(cmd,
GEN6_PIPE_CONTROL_CS_STALL,
- pWaitInfo->memBarrierCount, pWaitInfo->ppMemBarriers);
+ memBarrierCount, ppMemBarriers);
}
ICD_EXPORT void VKAPI vkCmdPipelineBarrier(
- VkCmdBuffer cmdBuffer,
- const VkPipelineBarrier* pBarrier)
+ VkCmdBuffer cmdBuffer,
+ VkWaitEvent waitEvent,
+ uint32_t pipeEventCount,
+ const VkPipeEvent* pPipeEvents,
+ uint32_t memBarrierCount,
+ const void** ppMemBarriers)
{
struct intel_cmd *cmd = intel_cmd(cmdBuffer);
uint32_t pipe_control_flags = 0;
@@ -308,10 +316,10 @@
/* Cache control is done with PIPE_CONTROL flags.
* With no GEN6_PIPE_CONTROL_CS_STALL flag set, it behaves as VK_PIPE_EVENT_TOP_OF_PIPE.
- * All other pEvents values will behave as VK_PIPE_EVENT_GPU_COMMANDS_COMPLETE.
+ * All other pEvents values will behave as VK_PIPE_EVENT_COMMANDS_COMPLETE.
*/
- for (i = 0; i < pBarrier->eventCount; i++) {
- switch(pBarrier->pEvents[i])
+ for (i = 0; i < pipeEventCount; i++) {
+ switch(pPipeEvents[i])
{
case VK_PIPE_EVENT_TOP_OF_PIPE:
break;
@@ -321,7 +329,7 @@
case VK_PIPE_EVENT_GRAPHICS_PIPELINE_COMPLETE:
case VK_PIPE_EVENT_COMPUTE_PIPELINE_COMPLETE:
case VK_PIPE_EVENT_TRANSFER_COMPLETE:
- case VK_PIPE_EVENT_GPU_COMMANDS_COMPLETE:
+ case VK_PIPE_EVENT_COMMANDS_COMPLETE:
pipe_control_flags |= GEN6_PIPE_CONTROL_CS_STALL;
break;
default:
@@ -336,5 +344,5 @@
*/
cmd_memory_barriers(cmd,
pipe_control_flags,
- pBarrier->memBarrierCount, pBarrier->ppMemBarriers);
+ memBarrierCount, ppMemBarriers);
}
diff --git a/icd/intel/cmd_meta.c b/icd/intel/cmd_meta.c
index f1a7a13..711b4de 100644
--- a/icd/intel/cmd_meta.c
+++ b/icd/intel/cmd_meta.c
@@ -33,17 +33,17 @@
static VkResult cmd_meta_create_buf_view(struct intel_cmd *cmd,
VkBuffer buf,
- VkGpuSize range,
+ VkDeviceSize range,
VkFormat format,
struct intel_buf_view **view)
{
VkBufferViewCreateInfo info;
- VkGpuSize stride;
+ VkDeviceSize stride;
memset(&info, 0, sizeof(info));
info.sType = VK_STRUCTURE_TYPE_BUFFER_VIEW_CREATE_INFO;
info.buffer = buf;
- info.viewType = VK_BUFFER_VIEW_FORMATTED;
+ info.viewType = VK_BUFFER_VIEW_TYPE_FORMATTED;
info.format = format;
info.range = range;
@@ -129,14 +129,14 @@
info.image = (VkImage) img;
switch (img->type) {
- case VK_IMAGE_1D:
- info.viewType = VK_IMAGE_VIEW_1D;
+ case VK_IMAGE_TYPE_1D:
+ info.viewType = VK_IMAGE_VIEW_TYPE_1D;
break;
- case VK_IMAGE_2D:
- info.viewType = VK_IMAGE_VIEW_2D;
+ case VK_IMAGE_TYPE_2D:
+ info.viewType = VK_IMAGE_VIEW_TYPE_2D;
break;
- case VK_IMAGE_3D:
- info.viewType = VK_IMAGE_VIEW_3D;
+ case VK_IMAGE_TYPE_3D:
+ info.viewType = VK_IMAGE_VIEW_TYPE_3D;
break;
default:
break;
@@ -275,7 +275,7 @@
static void cmd_meta_set_src_for_writer(struct intel_cmd *cmd,
enum intel_cmd_writer_type writer,
- VkGpuSize size,
+ VkDeviceSize size,
VkFormat format,
struct intel_cmd_meta *meta)
{
@@ -343,16 +343,16 @@
enum intel_dev_meta_shader shader_id;
switch (img->type) {
- case VK_IMAGE_1D:
+ case VK_IMAGE_TYPE_1D:
shader_id = (copy_array) ?
INTEL_DEV_META_FS_COPY_1D_ARRAY : INTEL_DEV_META_FS_COPY_1D;
break;
- case VK_IMAGE_2D:
+ case VK_IMAGE_TYPE_2D:
shader_id = (img->samples > 1) ? INTEL_DEV_META_FS_COPY_2D_MS :
(copy_array) ? INTEL_DEV_META_FS_COPY_2D_ARRAY :
INTEL_DEV_META_FS_COPY_2D;
break;
- case VK_IMAGE_3D:
+ case VK_IMAGE_TYPE_3D:
default:
shader_id = INTEL_DEV_META_FS_COPY_2D_ARRAY;
break;
@@ -362,9 +362,9 @@
}
static bool cmd_meta_mem_dword_aligned(const struct intel_cmd *cmd,
- VkGpuSize src_offset,
- VkGpuSize dst_offset,
- VkGpuSize size)
+ VkDeviceSize src_offset,
+ VkDeviceSize dst_offset,
+ VkDeviceSize size)
{
return !((src_offset | dst_offset | size) & 0x3);
}
@@ -374,23 +374,23 @@
{
switch (icd_format_get_size(format)) {
case 1:
- format = VK_FMT_R8_UINT;
+ format = VK_FORMAT_R8_UINT;
break;
case 2:
- format = VK_FMT_R16_UINT;
+ format = VK_FORMAT_R16_UINT;
break;
case 4:
- format = VK_FMT_R32_UINT;
+ format = VK_FORMAT_R32_UINT;
break;
case 8:
- format = VK_FMT_R32G32_UINT;
+ format = VK_FORMAT_R32G32_UINT;
break;
case 16:
- format = VK_FMT_R32G32B32A32_UINT;
+ format = VK_FORMAT_R32G32B32A32_UINT;
break;
default:
assert(!"unsupported image format for raw blit op");
- format = VK_FMT_UNDEFINED;
+ format = VK_FORMAT_UNDEFINED;
break;
}
@@ -417,7 +417,7 @@
meta.height = 1;
meta.samples = 1;
- format = VK_FMT_UNDEFINED;
+ format = VK_FORMAT_UNDEFINED;
for (i = 0; i < regionCount; i++) {
const VkBufferCopy *region = &pRegions[i];
@@ -438,7 +438,7 @@
* INTEL_DEV_META_VS_COPY_MEM is untyped but expects the stride to
* be 16
*/
- fmt = VK_FMT_R32G32B32A32_UINT;
+ fmt = VK_FORMAT_R32G32B32A32_UINT;
} else {
if (cmd_gen(cmd) == INTEL_GEN(6)) {
intel_dev_log(cmd->dev, VK_DBG_MSG_ERROR,
@@ -454,7 +454,7 @@
* INTEL_DEV_META_VS_COPY_MEM_UNALIGNED is untyped but expects the
* stride to be 4
*/
- fmt = VK_FMT_R8G8B8A8_UINT;
+ fmt = VK_FORMAT_R8G8B8A8_UINT;
}
if (format != fmt) {
@@ -647,33 +647,33 @@
/* buf_format is ignored by hw, but we derive stride from it */
switch (img_format) {
- case VK_FMT_R8_UINT:
+ case VK_FORMAT_R8_UINT:
meta.shader_id = INTEL_DEV_META_VS_COPY_R8_TO_MEM;
- buf_format = VK_FMT_R8G8B8A8_UINT;
+ buf_format = VK_FORMAT_R8G8B8A8_UINT;
break;
- case VK_FMT_R16_UINT:
+ case VK_FORMAT_R16_UINT:
meta.shader_id = INTEL_DEV_META_VS_COPY_R16_TO_MEM;
- buf_format = VK_FMT_R8G8B8A8_UINT;
+ buf_format = VK_FORMAT_R8G8B8A8_UINT;
break;
- case VK_FMT_R32_UINT:
+ case VK_FORMAT_R32_UINT:
meta.shader_id = INTEL_DEV_META_VS_COPY_R32_TO_MEM;
- buf_format = VK_FMT_R32G32B32A32_UINT;
+ buf_format = VK_FORMAT_R32G32B32A32_UINT;
break;
- case VK_FMT_R32G32_UINT:
+ case VK_FORMAT_R32G32_UINT:
meta.shader_id = INTEL_DEV_META_VS_COPY_R32G32_TO_MEM;
- buf_format = VK_FMT_R32G32B32A32_UINT;
+ buf_format = VK_FORMAT_R32G32B32A32_UINT;
break;
- case VK_FMT_R32G32B32A32_UINT:
+ case VK_FORMAT_R32G32B32A32_UINT:
meta.shader_id = INTEL_DEV_META_VS_COPY_R32G32B32A32_TO_MEM;
- buf_format = VK_FMT_R32G32B32A32_UINT;
+ buf_format = VK_FORMAT_R32G32B32A32_UINT;
break;
default:
- img_format = VK_FMT_UNDEFINED;
- buf_format = VK_FMT_UNDEFINED;
+ img_format = VK_FORMAT_UNDEFINED;
+ buf_format = VK_FORMAT_UNDEFINED;
break;
}
- if (img_format == VK_FMT_UNDEFINED ||
+ if (img_format == VK_FORMAT_UNDEFINED ||
(cmd_gen(cmd) == INTEL_GEN(6) &&
icd_format_get_size(img_format) < 4)) {
intel_dev_log(cmd->dev, VK_DBG_MSG_ERROR,
@@ -762,8 +762,8 @@
ICD_EXPORT void VKAPI vkCmdUpdateBuffer(
VkCmdBuffer cmdBuffer,
VkBuffer destBuffer,
- VkGpuSize destOffset,
- VkGpuSize dataSize,
+ VkDeviceSize destOffset,
+ VkDeviceSize dataSize,
const uint32_t* pData)
{
struct intel_cmd *cmd = intel_cmd(cmdBuffer);
@@ -798,7 +798,7 @@
/*
* INTEL_DEV_META_VS_COPY_MEM is untyped but expects the stride to be 16
*/
- format = VK_FMT_R32G32B32A32_UINT;
+ format = VK_FORMAT_R32G32B32A32_UINT;
cmd_meta_set_src_for_writer(cmd, INTEL_CMD_WRITER_STATE,
offset + dataSize, format, &meta);
@@ -810,8 +810,8 @@
ICD_EXPORT void VKAPI vkCmdFillBuffer(
VkCmdBuffer cmdBuffer,
VkBuffer destBuffer,
- VkGpuSize destOffset,
- VkGpuSize fillSize,
+ VkDeviceSize destOffset,
+ VkDeviceSize fillSize,
uint32_t data)
{
struct intel_cmd *cmd = intel_cmd(cmdBuffer);
@@ -840,7 +840,7 @@
/*
* INTEL_DEV_META_VS_FILL_MEM is untyped but expects the stride to be 16
*/
- format = VK_FMT_R32G32B32A32_UINT;
+ format = VK_FORMAT_R32G32B32A32_UINT;
cmd_meta_set_dst_for_buf(cmd, dst, format, &meta);
diff --git a/icd/intel/cmd_mi.c b/icd/intel/cmd_mi.c
index 73d53db..6add788 100644
--- a/icd/intel/cmd_mi.c
+++ b/icd/intel/cmd_mi.c
@@ -87,7 +87,7 @@
static void cmd_query_pipeline_statistics(struct intel_cmd *cmd,
struct intel_bo *bo,
- VkGpuSize offset)
+ VkDeviceSize offset)
{
const uint32_t regs[] = {
GEN6_REG_PS_INVOCATION_COUNT,
@@ -130,13 +130,13 @@
struct intel_cmd *cmd = intel_cmd(cmdBuffer);
struct intel_query *query = intel_query(queryPool);
struct intel_bo *bo = query->obj.mem->bo;
- const VkGpuSize offset = query->slot_stride * slot;
+ const VkDeviceSize offset = query->slot_stride * slot;
switch (query->type) {
- case VK_QUERY_OCCLUSION:
+ case VK_QUERY_TYPE_OCCLUSION:
cmd_batch_depth_count(cmd, bo, offset);
break;
- case VK_QUERY_PIPELINE_STATISTICS:
+ case VK_QUERY_TYPE_PIPELINE_STATISTICS:
cmd_query_pipeline_statistics(cmd, bo, offset);
break;
default:
@@ -153,13 +153,13 @@
struct intel_cmd *cmd = intel_cmd(cmdBuffer);
struct intel_query *query = intel_query(queryPool);
struct intel_bo *bo = query->obj.mem->bo;
- const VkGpuSize offset = query->slot_stride * slot;
+ const VkDeviceSize offset = query->slot_stride * slot;
switch (query->type) {
- case VK_QUERY_OCCLUSION:
+ case VK_QUERY_TYPE_OCCLUSION:
cmd_batch_depth_count(cmd, bo, offset + sizeof(uint64_t));
break;
- case VK_QUERY_PIPELINE_STATISTICS:
+ case VK_QUERY_TYPE_PIPELINE_STATISTICS:
cmd_query_pipeline_statistics(cmd, bo,
offset + sizeof(VkPipelineStatisticsData));
break;
@@ -185,7 +185,7 @@
/* Event setting is done with PIPE_CONTROL post-sync write immediate.
* With no other PIPE_CONTROL flags set, it behaves as VK_PIPE_EVENT_TOP_OF_PIPE.
- * All other pipeEvent values will behave as VK_PIPE_EVENT_GPU_COMMANDS_COMPLETE.
+ * All other pipeEvent values will behave as VK_PIPE_EVENT_COMMANDS_COMPLETE.
*/
switch(pipeEvent)
{
@@ -198,7 +198,7 @@
case VK_PIPE_EVENT_GRAPHICS_PIPELINE_COMPLETE:
case VK_PIPE_EVENT_COMPUTE_PIPELINE_COMPLETE:
case VK_PIPE_EVENT_TRANSFER_COMPLETE:
- case VK_PIPE_EVENT_GPU_COMMANDS_COMPLETE:
+ case VK_PIPE_EVENT_COMMANDS_COMPLETE:
pipe_control_flags = GEN6_PIPE_CONTROL_CS_STALL;
break;
default:
@@ -237,8 +237,8 @@
uint32_t startQuery,
uint32_t queryCount,
VkBuffer destBuffer,
- VkGpuSize destOffset,
- VkGpuSize destStride,
+ VkDeviceSize destOffset,
+ VkDeviceSize destStride,
VkFlags flags)
{
/* TODO: Fill in functionality here */
@@ -248,20 +248,20 @@
VkCmdBuffer cmdBuffer,
VkTimestampType timestampType,
VkBuffer destBuffer,
- VkGpuSize destOffset)
+ VkDeviceSize destOffset)
{
struct intel_cmd *cmd = intel_cmd(cmdBuffer);
struct intel_buf *buf = intel_buf(destBuffer);
switch (timestampType) {
- case VK_TIMESTAMP_TOP:
+ case VK_TIMESTAMP_TYPE_TOP:
/* XXX we are not supposed to use two commands... */
gen6_MI_STORE_REGISTER_MEM(cmd, buf->obj.mem->bo,
destOffset, GEN6_REG_TIMESTAMP);
gen6_MI_STORE_REGISTER_MEM(cmd, buf->obj.mem->bo,
destOffset + 4, GEN6_REG_TIMESTAMP + 4);
break;
- case VK_TIMESTAMP_BOTTOM:
+ case VK_TIMESTAMP_TYPE_BOTTOM:
cmd_batch_timestamp(cmd, buf->obj.mem->bo, destOffset);
break;
default:
diff --git a/icd/intel/cmd_pipeline.c b/icd/intel/cmd_pipeline.c
index 612595f..01dc658 100644
--- a/icd/intel/cmd_pipeline.c
+++ b/icd/intel/cmd_pipeline.c
@@ -218,13 +218,13 @@
return false;
switch (cmd->bind.index.type) {
- case VK_INDEX_8:
+ case VK_INDEX_TYPE_UINT8:
supported = (p->primitive_restart_index != 0xffu);
break;
- case VK_INDEX_16:
+ case VK_INDEX_TYPE_UINT16:
supported = (p->primitive_restart_index != 0xffffu);
break;
- case VK_INDEX_32:
+ case VK_INDEX_TYPE_UINT32:
supported = (p->primitive_restart_index != 0xffffffffu);
break;
default:
@@ -237,7 +237,7 @@
static void gen6_3DSTATE_INDEX_BUFFER(struct intel_cmd *cmd,
const struct intel_buf *buf,
- VkGpuSize offset,
+ VkDeviceSize offset,
VkIndexType type,
bool enable_cut_index)
{
@@ -257,15 +257,15 @@
dw0 |= GEN6_IB_DW0_CUT_INDEX_ENABLE;
switch (type) {
- case VK_INDEX_8:
+ case VK_INDEX_TYPE_UINT8:
dw0 |= GEN6_IB_DW0_FORMAT_BYTE;
offset_align = 1;
break;
- case VK_INDEX_16:
+ case VK_INDEX_TYPE_UINT16:
dw0 |= GEN6_IB_DW0_FORMAT_WORD;
offset_align = 2;
break;
- case VK_INDEX_32:
+ case VK_INDEX_TYPE_UINT32:
dw0 |= GEN6_IB_DW0_FORMAT_DWORD;
offset_align = 4;
break;
@@ -391,11 +391,11 @@
int format;
switch (pipeline->db_format) {
- case VK_FMT_D16_UNORM:
+ case VK_FORMAT_D16_UNORM:
format = GEN6_ZFORMAT_D16_UNORM;
break;
- case VK_FMT_D32_SFLOAT:
- case VK_FMT_D32_SFLOAT_S8_UINT:
+ case VK_FORMAT_D32_SFLOAT:
+ case VK_FORMAT_D32_SFLOAT_S8_UINT:
format = GEN6_ZFORMAT_D32_FLOAT;
break;
default:
@@ -1357,7 +1357,7 @@
void cmd_batch_depth_count(struct intel_cmd *cmd,
struct intel_bo *bo,
- VkGpuSize offset)
+ VkDeviceSize offset)
{
cmd_wa_gen6_pre_depth_stall_write(cmd);
@@ -1369,7 +1369,7 @@
void cmd_batch_timestamp(struct intel_cmd *cmd,
struct intel_bo *bo,
- VkGpuSize offset)
+ VkDeviceSize offset)
{
/* need any WA or stall? */
gen6_PIPE_CONTROL(cmd, GEN6_PIPE_CONTROL_WRITE_TIMESTAMP, bo, offset, 0);
@@ -1378,7 +1378,7 @@
void cmd_batch_immediate(struct intel_cmd *cmd,
uint32_t pipe_control_flags,
struct intel_bo *bo,
- VkGpuSize offset,
+ VkDeviceSize offset,
uint64_t val)
{
/* need any WA or stall? */
@@ -1638,7 +1638,7 @@
static uint32_t emit_binding_table(struct intel_cmd *cmd,
const struct intel_pipeline_rmap *rmap,
- const VkPipelineShaderStage stage)
+ const VkShaderStage stage)
{
const struct intel_desc_region *region = cmd->dev->desc_region;
const struct intel_cmd_dset_data *data = &cmd->bind.dset.graphics_data;
@@ -1796,7 +1796,7 @@
if (cmd->bind.vertex.buf[i]) {
const struct intel_buf *buf = cmd->bind.vertex.buf[i];
- const VkGpuSize offset = cmd->bind.vertex.offset[i];
+ const VkDeviceSize offset = cmd->bind.vertex.offset[i];
cmd_reserve_reloc(cmd, 2);
cmd_batch_reloc(cmd, pos + 1, buf->obj.mem->bo, offset, 0);
@@ -2010,7 +2010,7 @@
void *entries;
entries = intel_alloc(cmd, sizeof(cache->entries[0]) * count, 0,
- VK_SYSTEM_ALLOC_INTERNAL);
+ VK_SYSTEM_ALLOC_TYPE_INTERNAL);
if (entries) {
if (cache->entries) {
memcpy(entries, cache->entries,
@@ -3044,9 +3044,9 @@
data->set_offsets = intel_alloc(cmd,
sizeof(data->set_offsets[0]) * chain->layout_count,
- sizeof(data->set_offsets[0]), VK_SYSTEM_ALLOC_INTERNAL);
+ sizeof(data->set_offsets[0]), VK_SYSTEM_ALLOC_TYPE_INTERNAL);
if (!data->set_offsets) {
- cmd_fail(cmd, VK_ERROR_OUT_OF_MEMORY);
+ cmd_fail(cmd, VK_ERROR_OUT_OF_HOST_MEMORY);
data->set_offset_count = 0;
return false;
}
@@ -3060,9 +3060,9 @@
data->dynamic_offsets = intel_alloc(cmd,
sizeof(data->dynamic_offsets[0]) * chain->total_dynamic_desc_count,
- sizeof(data->dynamic_offsets[0]), VK_SYSTEM_ALLOC_INTERNAL);
+ sizeof(data->dynamic_offsets[0]), VK_SYSTEM_ALLOC_TYPE_INTERNAL);
if (!data->dynamic_offsets) {
- cmd_fail(cmd, VK_ERROR_OUT_OF_MEMORY);
+ cmd_fail(cmd, VK_ERROR_OUT_OF_HOST_MEMORY);
data->dynamic_offset_count = 0;
return false;
}
@@ -3097,7 +3097,7 @@
static void cmd_bind_vertex_data(struct intel_cmd *cmd,
const struct intel_buf *buf,
- VkGpuSize offset, uint32_t binding)
+ VkDeviceSize offset, uint32_t binding)
{
if (binding >= ARRAY_SIZE(cmd->bind.vertex.buf)) {
cmd_fail(cmd, VK_ERROR_UNKNOWN);
@@ -3110,7 +3110,7 @@
static void cmd_bind_index_data(struct intel_cmd *cmd,
const struct intel_buf *buf,
- VkGpuSize offset, VkIndexType type)
+ VkDeviceSize offset, VkIndexType type)
{
cmd->bind.index.buf = buf;
cmd->bind.index.offset = offset;
@@ -3349,19 +3349,19 @@
struct intel_cmd *cmd = intel_cmd(cmdBuffer);
switch (stateBindPoint) {
- case VK_STATE_BIND_VIEWPORT:
+ case VK_STATE_BIND_POINT_VIEWPORT:
cmd_bind_viewport_state(cmd,
intel_dynamic_vp((VkDynamicVpState) state));
break;
- case VK_STATE_BIND_RASTER:
+ case VK_STATE_BIND_POINT_RASTER:
cmd_bind_raster_state(cmd,
intel_dynamic_rs((VkDynamicRsState) state));
break;
- case VK_STATE_BIND_DEPTH_STENCIL:
+ case VK_STATE_BIND_POINT_DEPTH_STENCIL:
cmd_bind_ds_state(cmd,
intel_dynamic_ds((VkDynamicDsState) state));
break;
- case VK_STATE_BIND_COLOR_BLEND:
+ case VK_STATE_BIND_POINT_COLOR_BLEND:
cmd_bind_blend_state(cmd,
intel_dynamic_cb((VkDynamicCbState) state));
break;
@@ -3413,12 +3413,13 @@
}
}
+
ICD_EXPORT void VKAPI vkCmdBindVertexBuffers(
VkCmdBuffer cmdBuffer,
uint32_t startBinding,
uint32_t bindingCount,
const VkBuffer* pBuffers,
- const VkGpuSize* pOffsets)
+ const VkDeviceSize* pOffsets)
{
struct intel_cmd *cmd = intel_cmd(cmdBuffer);
@@ -3431,7 +3432,7 @@
ICD_EXPORT void VKAPI vkCmdBindIndexBuffer(
VkCmdBuffer cmdBuffer,
VkBuffer buffer,
- VkGpuSize offset,
+ VkDeviceSize offset,
VkIndexType indexType)
{
struct intel_cmd *cmd = intel_cmd(cmdBuffer);
@@ -3470,7 +3471,7 @@
ICD_EXPORT void VKAPI vkCmdDrawIndirect(
VkCmdBuffer cmdBuffer,
VkBuffer buffer,
- VkGpuSize offset,
+ VkDeviceSize offset,
uint32_t count,
uint32_t stride)
{
@@ -3482,7 +3483,7 @@
ICD_EXPORT void VKAPI vkCmdDrawIndexedIndirect(
VkCmdBuffer cmdBuffer,
VkBuffer buffer,
- VkGpuSize offset,
+ VkDeviceSize offset,
uint32_t count,
uint32_t stride)
{
@@ -3505,7 +3506,7 @@
ICD_EXPORT void VKAPI vkCmdDispatchIndirect(
VkCmdBuffer cmdBuffer,
VkBuffer buffer,
- VkGpuSize offset)
+ VkDeviceSize offset)
{
struct intel_cmd *cmd = intel_cmd(cmdBuffer);
diff --git a/icd/intel/cmd_priv.h b/icd/intel/cmd_priv.h
index 3834686..4f6b4b7 100644
--- a/icd/intel/cmd_priv.h
+++ b/icd/intel/cmd_priv.h
@@ -506,16 +506,16 @@
void cmd_batch_depth_count(struct intel_cmd *cmd,
struct intel_bo *bo,
- VkGpuSize offset);
+ VkDeviceSize offset);
void cmd_batch_timestamp(struct intel_cmd *cmd,
struct intel_bo *bo,
- VkGpuSize offset);
+ VkDeviceSize offset);
void cmd_batch_immediate(struct intel_cmd *cmd,
uint32_t pipe_control_flags,
struct intel_bo *bo,
- VkGpuSize offset,
+ VkDeviceSize offset,
uint64_t val);
void cmd_draw_meta(struct intel_cmd *cmd, const struct intel_cmd_meta *meta);
diff --git a/icd/intel/compiler/pipeline/pipeline_compiler_interface.cpp b/icd/intel/compiler/pipeline/pipeline_compiler_interface.cpp
index 3d0d584..6797460 100644
--- a/icd/intel/compiler/pipeline/pipeline_compiler_interface.cpp
+++ b/icd/intel/compiler/pipeline/pipeline_compiler_interface.cpp
@@ -296,7 +296,7 @@
uint32_t surface_count, i;
rmap = (struct intel_pipeline_rmap *)
- intel_alloc(gpu, sizeof(*rmap), 0, VK_SYSTEM_ALLOC_INTERNAL);
+ intel_alloc(gpu, sizeof(*rmap), 0, VK_SYSTEM_ALLOC_TYPE_INTERNAL);
if (!rmap)
return NULL;
@@ -313,7 +313,7 @@
rmap->slots = (struct intel_pipeline_rmap_slot *)
intel_alloc(gpu, sizeof(rmap->slots[0]) * rmap->slot_count,
- 0, VK_SYSTEM_ALLOC_INTERNAL);
+ 0, VK_SYSTEM_ALLOC_TYPE_INTERNAL);
if (!rmap->slots) {
intel_free(gpu, rmap);
return NULL;
@@ -427,9 +427,9 @@
{
pipe_shader->codeSize = get_vs_program_size(brw->shader_prog);
- pipe_shader->pCode = intel_alloc(gpu, pipe_shader->codeSize, 0, VK_SYSTEM_ALLOC_INTERNAL_SHADER);
+ pipe_shader->pCode = intel_alloc(gpu, pipe_shader->codeSize, 0, VK_SYSTEM_ALLOC_TYPE_INTERNAL_SHADER);
if (!pipe_shader->pCode) {
- status = VK_ERROR_OUT_OF_MEMORY;
+ status = VK_ERROR_OUT_OF_HOST_MEMORY;
break;
}
@@ -546,9 +546,9 @@
pipe_shader->codeSize = get_wm_program_size(brw->shader_prog);
- pipe_shader->pCode = intel_alloc(gpu, pipe_shader->codeSize, 0, VK_SYSTEM_ALLOC_INTERNAL_SHADER);
+ pipe_shader->pCode = intel_alloc(gpu, pipe_shader->codeSize, 0, VK_SYSTEM_ALLOC_TYPE_INTERNAL_SHADER);
if (!pipe_shader->pCode) {
- status = VK_ERROR_OUT_OF_MEMORY;
+ status = VK_ERROR_OUT_OF_HOST_MEMORY;
break;
}
@@ -698,7 +698,7 @@
pipe_shader->rmap = rmap_create(gpu, chain, &bt);
if (!pipe_shader->rmap) {
intel_pipeline_shader_cleanup(pipe_shader, gpu);
- status = VK_ERROR_OUT_OF_MEMORY;
+ status = VK_ERROR_OUT_OF_HOST_MEMORY;
}
}
diff --git a/icd/intel/compiler/pipeline/pipeline_compiler_interface_meta.cpp b/icd/intel/compiler/pipeline/pipeline_compiler_interface_meta.cpp
index ca0f462..b357785 100644
--- a/icd/intel/compiler/pipeline/pipeline_compiler_interface_meta.cpp
+++ b/icd/intel/compiler/pipeline/pipeline_compiler_interface_meta.cpp
@@ -619,7 +619,7 @@
prog = get_program(&prog_size, stderr);
- code = intel_alloc(gpu, prog_size, 0, VK_SYSTEM_ALLOC_INTERNAL);
+ code = intel_alloc(gpu, prog_size, 0, VK_SYSTEM_ALLOC_TYPE_INTERNAL);
if (!code)
return NULL;
diff --git a/icd/intel/desc.c b/icd/intel/desc.c
index 66a2ea7..a8d51fe 100644
--- a/icd/intel/desc.c
+++ b/icd/intel/desc.c
@@ -124,9 +124,9 @@
const uint32_t sampler_count = 16384;
struct intel_desc_region *region;
- region = intel_alloc(dev, sizeof(*region), 0, VK_SYSTEM_ALLOC_INTERNAL);
+ region = intel_alloc(dev, sizeof(*region), 0, VK_SYSTEM_ALLOC_TYPE_INTERNAL);
if (!region)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
memset(region, 0, sizeof(*region));
@@ -140,18 +140,18 @@
region->sampler_desc_size * sampler_count);
region->surfaces = intel_alloc(dev, region->size.surface,
- 64, VK_SYSTEM_ALLOC_INTERNAL);
+ 64, VK_SYSTEM_ALLOC_TYPE_INTERNAL);
if (!region->surfaces) {
intel_free(dev, region);
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
}
region->samplers = intel_alloc(dev, region->size.sampler,
- 64, VK_SYSTEM_ALLOC_INTERNAL);
+ 64, VK_SYSTEM_ALLOC_TYPE_INTERNAL);
if (!region->samplers) {
intel_free(dev, region->surfaces);
intel_free(dev, region);
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
}
*region_ret = region;
@@ -234,7 +234,7 @@
intel_desc_offset_add(end, ®ion->cur, &alloc);
if (!intel_desc_offset_within(end, ®ion->size))
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
/* increment the writer pointer */
region->cur = *end;
@@ -348,7 +348,7 @@
void intel_desc_region_read_surface(const struct intel_desc_region *region,
const struct intel_desc_offset *offset,
- VkPipelineShaderStage stage,
+ VkShaderStage stage,
const struct intel_mem **mem,
bool *read_only,
const uint32_t **cmd,
@@ -421,7 +421,7 @@
sizeof(*pool), dev->base.dbg, VK_DBG_OBJECT_DESCRIPTOR_POOL,
info, 0);
if (!pool)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
pool->dev = dev;
@@ -458,7 +458,7 @@
intel_desc_offset_add(end, &pool->cur, &layout->region_size);
if (!intel_desc_offset_within(end, &pool->region_end))
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
/* increment the writer pointer */
pool->cur = *end;
@@ -492,7 +492,7 @@
sizeof(*set), dev->base.dbg, VK_DBG_OBJECT_DESCRIPTOR_SET,
NULL, 0);
if (!set)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
set->region = dev->desc_region;
ret = intel_desc_pool_alloc(pool, layout,
@@ -711,9 +711,9 @@
/* allocate bindings */
layout->bindings = intel_alloc(layout, sizeof(layout->bindings[0]) *
- info->count, 0, VK_SYSTEM_ALLOC_INTERNAL);
+ info->count, 0, VK_SYSTEM_ALLOC_TYPE_INTERNAL);
if (!layout->bindings)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
memset(layout->bindings, 0, sizeof(layout->bindings[0]) * info->count);
layout->binding_count = info->count;
@@ -766,9 +766,9 @@
} else {
binding->immutable_samplers = intel_alloc(layout,
sizeof(binding->immutable_samplers[0]) * lb->count,
- 0, VK_SYSTEM_ALLOC_INTERNAL);
+ 0, VK_SYSTEM_ALLOC_TYPE_INTERNAL);
if (!binding->immutable_samplers)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
for (j = 0; j < lb->count; j++) {
binding->immutable_samplers[j] =
@@ -799,7 +799,7 @@
sizeof(*layout), dev->base.dbg,
VK_DBG_OBJECT_DESCRIPTOR_SET_LAYOUT, info, 0);
if (!layout)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
ret = desc_layout_init_bindings(layout, dev->desc_region, info);
if (ret != VK_SUCCESS) {
@@ -848,21 +848,21 @@
intel_base_create(&dev->base.handle, sizeof(*chain), dev->base.dbg,
VK_DBG_OBJECT_DESCRIPTOR_SET_LAYOUT_CHAIN, NULL, 0);
if (!chain)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
chain->layouts = intel_alloc(chain, sizeof(chain->layouts[0]) * count,
- 0, VK_SYSTEM_ALLOC_INTERNAL);
+ 0, VK_SYSTEM_ALLOC_TYPE_INTERNAL);
if (!chain) {
intel_desc_layout_chain_destroy(chain);
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
}
chain->dynamic_desc_indices = intel_alloc(chain,
sizeof(chain->dynamic_desc_indices[0]) * count,
- 0, VK_SYSTEM_ALLOC_INTERNAL);
+ 0, VK_SYSTEM_ALLOC_TYPE_INTERNAL);
if (!chain->dynamic_desc_indices) {
intel_desc_layout_chain_destroy(chain);
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
}
for (i = 0; i < count; i++) {
diff --git a/icd/intel/desc.h b/icd/intel/desc.h
index 300e5e1..6633726 100644
--- a/icd/intel/desc.h
+++ b/icd/intel/desc.h
@@ -255,7 +255,7 @@
void intel_desc_region_read_surface(const struct intel_desc_region *region,
const struct intel_desc_offset *offset,
- VkPipelineShaderStage stage,
+ VkShaderStage stage,
const struct intel_mem **mem,
bool *read_only,
const uint32_t **cmd,
diff --git a/icd/intel/dev.c b/icd/intel/dev.c
index 0f58657..d11d557 100644
--- a/icd/intel/dev.c
+++ b/icd/intel/dev.c
@@ -114,7 +114,7 @@
sizeof(*dev), info->flags,
VK_DBG_OBJECT_DEVICE, info, sizeof(struct intel_dev_dbg));
if (!dev)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
for (i = 0; i < info->extensionCount; i++) {
const enum intel_ext_type ext = intel_gpu_lookup_extension(gpu,
@@ -138,12 +138,12 @@
"command buffer scratch", 4096, false);
if (!dev->cmd_scratch_bo) {
intel_dev_destroy(dev);
- return VK_ERROR_OUT_OF_GPU_MEMORY;
+ return VK_ERROR_OUT_OF_DEVICE_MEMORY;
}
if (!dev_create_meta_shaders(dev)) {
intel_dev_destroy(dev);
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
}
ret = intel_desc_region_create(dev, &dev->desc_region);
@@ -233,9 +233,9 @@
f->triggered = false;
}
} else {
- f = intel_alloc(dev, sizeof(*f), 0, VK_SYSTEM_ALLOC_DEBUG);
+ f = intel_alloc(dev, sizeof(*f), 0, VK_SYSTEM_ALLOC_TYPE_DEBUG);
if (!f)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
f->msg_code = msg_code;
f->filter = filter;
@@ -320,7 +320,7 @@
}
ICD_EXPORT VkResult VKAPI vkCreateDevice(
- VkPhysicalGpu gpu_,
+ VkPhysicalDevice gpu_,
const VkDeviceCreateInfo* pCreateInfo,
VkDevice* pDevice)
{
diff --git a/icd/intel/event.c b/icd/intel/event.c
index 61520bf..5449f59 100644
--- a/icd/intel/event.c
+++ b/icd/intel/event.c
@@ -96,7 +96,7 @@
VkResult ret = VK_SUCCESS;
switch (type) {
- case VK_INFO_TYPE_MEMORY_REQUIREMENTS:
+ case VK_OBJECT_INFO_TYPE_MEMORY_REQUIREMENTS:
{
VkMemoryRequirements *mem_req = data;
@@ -126,7 +126,7 @@
event = (struct intel_event *) intel_base_create(&dev->base.handle,
sizeof(*event), dev->base.dbg, VK_DBG_OBJECT_EVENT, info, 0);
if (!event)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
event->obj.base.get_info = event_get_info;
event->obj.destroy = event_destroy;
diff --git a/icd/intel/fb.c b/icd/intel/fb.c
index 5a0ec80..665f043 100644
--- a/icd/intel/fb.c
+++ b/icd/intel/fb.c
@@ -49,7 +49,7 @@
fb = (struct intel_fb *) intel_base_create(&dev->base.handle,
sizeof(*fb), dev->base.dbg, VK_DBG_OBJECT_FRAMEBUFFER, info, 0);
if (!fb)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
width = info->width;
height = info->height;
@@ -148,7 +148,7 @@
rp = (struct intel_render_pass *) intel_base_create(&dev->base.handle,
sizeof(*rp), dev->base.dbg, VK_DBG_OBJECT_RENDER_PASS, info, 0);
if (!rp)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
rp->obj.destroy = render_pass_destroy;
diff --git a/icd/intel/fence.c b/icd/intel/fence.c
index cd94586..7b3cebd 100644
--- a/icd/intel/fence.c
+++ b/icd/intel/fence.c
@@ -47,7 +47,7 @@
fence = (struct intel_fence *) intel_base_create(&dev->base.handle,
sizeof(*fence), dev->base.dbg, VK_DBG_OBJECT_FENCE, info, 0);
if (!fence)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
if (dev->exts[INTEL_EXT_WSI_X11]) {
VkResult ret = intel_wsi_fence_init(fence);
diff --git a/icd/intel/format.c b/icd/intel/format.c
index 9f40cf3..258c84f 100644
--- a/icd/intel/format.c
+++ b/icd/intel/format.c
@@ -377,167 +377,167 @@
#undef CAP
};
-static const int intel_color_mapping[VK_NUM_FMT] = {
- [VK_FMT_R4G4_UNORM] = 0,
- [VK_FMT_R4G4_USCALED] = 0,
- [VK_FMT_R4G4B4A4_UNORM] = 0,
- [VK_FMT_R4G4B4A4_USCALED] = 0,
- [VK_FMT_R5G6B5_UNORM] = 0,
- [VK_FMT_R5G6B5_USCALED] = 0,
- [VK_FMT_R5G5B5A1_UNORM] = 0,
- [VK_FMT_R5G5B5A1_USCALED] = 0,
- [VK_FMT_R8_UNORM] = GEN6_FORMAT_R8_UNORM,
- [VK_FMT_R8_SNORM] = GEN6_FORMAT_R8_SNORM,
- [VK_FMT_R8_USCALED] = GEN6_FORMAT_R8_USCALED,
- [VK_FMT_R8_SSCALED] = GEN6_FORMAT_R8_SSCALED,
- [VK_FMT_R8_UINT] = GEN6_FORMAT_R8_UINT,
- [VK_FMT_R8_SINT] = GEN6_FORMAT_R8_SINT,
- [VK_FMT_R8_SRGB] = 0,
- [VK_FMT_R8G8_UNORM] = GEN6_FORMAT_R8G8_UNORM,
- [VK_FMT_R8G8_SNORM] = GEN6_FORMAT_R8G8_SNORM,
- [VK_FMT_R8G8_USCALED] = GEN6_FORMAT_R8G8_USCALED,
- [VK_FMT_R8G8_SSCALED] = GEN6_FORMAT_R8G8_SSCALED,
- [VK_FMT_R8G8_UINT] = GEN6_FORMAT_R8G8_UINT,
- [VK_FMT_R8G8_SINT] = GEN6_FORMAT_R8G8_SINT,
- [VK_FMT_R8G8_SRGB] = 0,
- [VK_FMT_R8G8B8_UNORM] = GEN6_FORMAT_R8G8B8_UNORM,
- [VK_FMT_R8G8B8_SNORM] = GEN6_FORMAT_R8G8B8_SNORM,
- [VK_FMT_R8G8B8_USCALED] = GEN6_FORMAT_R8G8B8_USCALED,
- [VK_FMT_R8G8B8_SSCALED] = GEN6_FORMAT_R8G8B8_SSCALED,
- [VK_FMT_R8G8B8_UINT] = GEN6_FORMAT_R8G8B8_UINT,
- [VK_FMT_R8G8B8_SINT] = GEN6_FORMAT_R8G8B8_SINT,
- [VK_FMT_R8G8B8_SRGB] = GEN6_FORMAT_R8G8B8_UNORM_SRGB,
- [VK_FMT_R8G8B8A8_UNORM] = GEN6_FORMAT_R8G8B8A8_UNORM,
- [VK_FMT_R8G8B8A8_SNORM] = GEN6_FORMAT_R8G8B8A8_SNORM,
- [VK_FMT_R8G8B8A8_USCALED] = GEN6_FORMAT_R8G8B8A8_USCALED,
- [VK_FMT_R8G8B8A8_SSCALED] = GEN6_FORMAT_R8G8B8A8_SSCALED,
- [VK_FMT_R8G8B8A8_UINT] = GEN6_FORMAT_R8G8B8A8_UINT,
- [VK_FMT_R8G8B8A8_SINT] = GEN6_FORMAT_R8G8B8A8_SINT,
- [VK_FMT_R8G8B8A8_SRGB] = GEN6_FORMAT_R8G8B8A8_UNORM_SRGB,
- [VK_FMT_R10G10B10A2_UNORM] = GEN6_FORMAT_R10G10B10A2_UNORM,
- [VK_FMT_R10G10B10A2_SNORM] = GEN6_FORMAT_R10G10B10A2_SNORM,
- [VK_FMT_R10G10B10A2_USCALED] = GEN6_FORMAT_R10G10B10A2_USCALED,
- [VK_FMT_R10G10B10A2_SSCALED] = GEN6_FORMAT_R10G10B10A2_SSCALED,
- [VK_FMT_R10G10B10A2_UINT] = GEN6_FORMAT_R10G10B10A2_UINT,
- [VK_FMT_R10G10B10A2_SINT] = GEN6_FORMAT_R10G10B10A2_SINT,
- [VK_FMT_R16_UNORM] = GEN6_FORMAT_R16_UNORM,
- [VK_FMT_R16_SNORM] = GEN6_FORMAT_R16_SNORM,
- [VK_FMT_R16_USCALED] = GEN6_FORMAT_R16_USCALED,
- [VK_FMT_R16_SSCALED] = GEN6_FORMAT_R16_SSCALED,
- [VK_FMT_R16_UINT] = GEN6_FORMAT_R16_UINT,
- [VK_FMT_R16_SINT] = GEN6_FORMAT_R16_SINT,
- [VK_FMT_R16_SFLOAT] = GEN6_FORMAT_R16_FLOAT,
- [VK_FMT_R16G16_UNORM] = GEN6_FORMAT_R16G16_UNORM,
- [VK_FMT_R16G16_SNORM] = GEN6_FORMAT_R16G16_SNORM,
- [VK_FMT_R16G16_USCALED] = GEN6_FORMAT_R16G16_USCALED,
- [VK_FMT_R16G16_SSCALED] = GEN6_FORMAT_R16G16_SSCALED,
- [VK_FMT_R16G16_UINT] = GEN6_FORMAT_R16G16_UINT,
- [VK_FMT_R16G16_SINT] = GEN6_FORMAT_R16G16_SINT,
- [VK_FMT_R16G16_SFLOAT] = GEN6_FORMAT_R16G16_FLOAT,
- [VK_FMT_R16G16B16_UNORM] = GEN6_FORMAT_R16G16B16_UNORM,
- [VK_FMT_R16G16B16_SNORM] = GEN6_FORMAT_R16G16B16_SNORM,
- [VK_FMT_R16G16B16_USCALED] = GEN6_FORMAT_R16G16B16_USCALED,
- [VK_FMT_R16G16B16_SSCALED] = GEN6_FORMAT_R16G16B16_SSCALED,
- [VK_FMT_R16G16B16_UINT] = GEN6_FORMAT_R16G16B16_UINT,
- [VK_FMT_R16G16B16_SINT] = GEN6_FORMAT_R16G16B16_SINT,
- [VK_FMT_R16G16B16_SFLOAT] = 0,
- [VK_FMT_R16G16B16A16_UNORM] = GEN6_FORMAT_R16G16B16A16_UNORM,
- [VK_FMT_R16G16B16A16_SNORM] = GEN6_FORMAT_R16G16B16A16_SNORM,
- [VK_FMT_R16G16B16A16_USCALED] = GEN6_FORMAT_R16G16B16A16_USCALED,
- [VK_FMT_R16G16B16A16_SSCALED] = GEN6_FORMAT_R16G16B16A16_SSCALED,
- [VK_FMT_R16G16B16A16_UINT] = GEN6_FORMAT_R16G16B16A16_UINT,
- [VK_FMT_R16G16B16A16_SINT] = GEN6_FORMAT_R16G16B16A16_SINT,
- [VK_FMT_R16G16B16A16_SFLOAT] = GEN6_FORMAT_R16G16B16A16_FLOAT,
- [VK_FMT_R32_UINT] = GEN6_FORMAT_R32_UINT,
- [VK_FMT_R32_SINT] = GEN6_FORMAT_R32_SINT,
- [VK_FMT_R32_SFLOAT] = GEN6_FORMAT_R32_FLOAT,
- [VK_FMT_R32G32_UINT] = GEN6_FORMAT_R32G32_UINT,
- [VK_FMT_R32G32_SINT] = GEN6_FORMAT_R32G32_SINT,
- [VK_FMT_R32G32_SFLOAT] = GEN6_FORMAT_R32G32_FLOAT,
- [VK_FMT_R32G32B32_UINT] = GEN6_FORMAT_R32G32B32_UINT,
- [VK_FMT_R32G32B32_SINT] = GEN6_FORMAT_R32G32B32_SINT,
- [VK_FMT_R32G32B32_SFLOAT] = GEN6_FORMAT_R32G32B32_FLOAT,
- [VK_FMT_R32G32B32A32_UINT] = GEN6_FORMAT_R32G32B32A32_UINT,
- [VK_FMT_R32G32B32A32_SINT] = GEN6_FORMAT_R32G32B32A32_SINT,
- [VK_FMT_R32G32B32A32_SFLOAT] = GEN6_FORMAT_R32G32B32A32_FLOAT,
- [VK_FMT_R64_SFLOAT] = GEN6_FORMAT_R64_FLOAT,
- [VK_FMT_R64G64_SFLOAT] = GEN6_FORMAT_R64G64_FLOAT,
- [VK_FMT_R64G64B64_SFLOAT] = GEN6_FORMAT_R64G64B64_FLOAT,
- [VK_FMT_R64G64B64A64_SFLOAT] = GEN6_FORMAT_R64G64B64A64_FLOAT,
- [VK_FMT_R11G11B10_UFLOAT] = GEN6_FORMAT_R11G11B10_FLOAT,
- [VK_FMT_R9G9B9E5_UFLOAT] = GEN6_FORMAT_R9G9B9E5_SHAREDEXP,
- [VK_FMT_BC1_RGB_UNORM] = GEN6_FORMAT_BC1_UNORM,
- [VK_FMT_BC1_RGB_SRGB] = GEN6_FORMAT_BC1_UNORM_SRGB,
- [VK_FMT_BC2_UNORM] = GEN6_FORMAT_BC2_UNORM,
- [VK_FMT_BC2_SRGB] = GEN6_FORMAT_BC2_UNORM_SRGB,
- [VK_FMT_BC3_UNORM] = GEN6_FORMAT_BC3_UNORM,
- [VK_FMT_BC3_SRGB] = GEN6_FORMAT_BC3_UNORM_SRGB,
- [VK_FMT_BC4_UNORM] = GEN6_FORMAT_BC4_UNORM,
- [VK_FMT_BC4_SNORM] = GEN6_FORMAT_BC4_SNORM,
- [VK_FMT_BC5_UNORM] = GEN6_FORMAT_BC5_UNORM,
- [VK_FMT_BC5_SNORM] = GEN6_FORMAT_BC5_SNORM,
- [VK_FMT_BC6H_UFLOAT] = GEN6_FORMAT_BC6H_UF16,
- [VK_FMT_BC6H_SFLOAT] = GEN6_FORMAT_BC6H_SF16,
- [VK_FMT_BC7_UNORM] = GEN6_FORMAT_BC7_UNORM,
- [VK_FMT_BC7_SRGB] = GEN6_FORMAT_BC7_UNORM_SRGB,
+static const int intel_color_mapping[VK_NUM_FORMAT] = {
+ [VK_FORMAT_R4G4_UNORM] = 0,
+ [VK_FORMAT_R4G4_USCALED] = 0,
+ [VK_FORMAT_R4G4B4A4_UNORM] = 0,
+ [VK_FORMAT_R4G4B4A4_USCALED] = 0,
+ [VK_FORMAT_R5G6B5_UNORM] = 0,
+ [VK_FORMAT_R5G6B5_USCALED] = 0,
+ [VK_FORMAT_R5G5B5A1_UNORM] = 0,
+ [VK_FORMAT_R5G5B5A1_USCALED] = 0,
+ [VK_FORMAT_R8_UNORM] = GEN6_FORMAT_R8_UNORM,
+ [VK_FORMAT_R8_SNORM] = GEN6_FORMAT_R8_SNORM,
+ [VK_FORMAT_R8_USCALED] = GEN6_FORMAT_R8_USCALED,
+ [VK_FORMAT_R8_SSCALED] = GEN6_FORMAT_R8_SSCALED,
+ [VK_FORMAT_R8_UINT] = GEN6_FORMAT_R8_UINT,
+ [VK_FORMAT_R8_SINT] = GEN6_FORMAT_R8_SINT,
+ [VK_FORMAT_R8_SRGB] = 0,
+ [VK_FORMAT_R8G8_UNORM] = GEN6_FORMAT_R8G8_UNORM,
+ [VK_FORMAT_R8G8_SNORM] = GEN6_FORMAT_R8G8_SNORM,
+ [VK_FORMAT_R8G8_USCALED] = GEN6_FORMAT_R8G8_USCALED,
+ [VK_FORMAT_R8G8_SSCALED] = GEN6_FORMAT_R8G8_SSCALED,
+ [VK_FORMAT_R8G8_UINT] = GEN6_FORMAT_R8G8_UINT,
+ [VK_FORMAT_R8G8_SINT] = GEN6_FORMAT_R8G8_SINT,
+ [VK_FORMAT_R8G8_SRGB] = 0,
+ [VK_FORMAT_R8G8B8_UNORM] = GEN6_FORMAT_R8G8B8_UNORM,
+ [VK_FORMAT_R8G8B8_SNORM] = GEN6_FORMAT_R8G8B8_SNORM,
+ [VK_FORMAT_R8G8B8_USCALED] = GEN6_FORMAT_R8G8B8_USCALED,
+ [VK_FORMAT_R8G8B8_SSCALED] = GEN6_FORMAT_R8G8B8_SSCALED,
+ [VK_FORMAT_R8G8B8_UINT] = GEN6_FORMAT_R8G8B8_UINT,
+ [VK_FORMAT_R8G8B8_SINT] = GEN6_FORMAT_R8G8B8_SINT,
+ [VK_FORMAT_R8G8B8_SRGB] = GEN6_FORMAT_R8G8B8_UNORM_SRGB,
+ [VK_FORMAT_R8G8B8A8_UNORM] = GEN6_FORMAT_R8G8B8A8_UNORM,
+ [VK_FORMAT_R8G8B8A8_SNORM] = GEN6_FORMAT_R8G8B8A8_SNORM,
+ [VK_FORMAT_R8G8B8A8_USCALED] = GEN6_FORMAT_R8G8B8A8_USCALED,
+ [VK_FORMAT_R8G8B8A8_SSCALED] = GEN6_FORMAT_R8G8B8A8_SSCALED,
+ [VK_FORMAT_R8G8B8A8_UINT] = GEN6_FORMAT_R8G8B8A8_UINT,
+ [VK_FORMAT_R8G8B8A8_SINT] = GEN6_FORMAT_R8G8B8A8_SINT,
+ [VK_FORMAT_R8G8B8A8_SRGB] = GEN6_FORMAT_R8G8B8A8_UNORM_SRGB,
+ [VK_FORMAT_R10G10B10A2_UNORM] = GEN6_FORMAT_R10G10B10A2_UNORM,
+ [VK_FORMAT_R10G10B10A2_SNORM] = GEN6_FORMAT_R10G10B10A2_SNORM,
+ [VK_FORMAT_R10G10B10A2_USCALED] = GEN6_FORMAT_R10G10B10A2_USCALED,
+ [VK_FORMAT_R10G10B10A2_SSCALED] = GEN6_FORMAT_R10G10B10A2_SSCALED,
+ [VK_FORMAT_R10G10B10A2_UINT] = GEN6_FORMAT_R10G10B10A2_UINT,
+ [VK_FORMAT_R10G10B10A2_SINT] = GEN6_FORMAT_R10G10B10A2_SINT,
+ [VK_FORMAT_R16_UNORM] = GEN6_FORMAT_R16_UNORM,
+ [VK_FORMAT_R16_SNORM] = GEN6_FORMAT_R16_SNORM,
+ [VK_FORMAT_R16_USCALED] = GEN6_FORMAT_R16_USCALED,
+ [VK_FORMAT_R16_SSCALED] = GEN6_FORMAT_R16_SSCALED,
+ [VK_FORMAT_R16_UINT] = GEN6_FORMAT_R16_UINT,
+ [VK_FORMAT_R16_SINT] = GEN6_FORMAT_R16_SINT,
+ [VK_FORMAT_R16_SFLOAT] = GEN6_FORMAT_R16_FLOAT,
+ [VK_FORMAT_R16G16_UNORM] = GEN6_FORMAT_R16G16_UNORM,
+ [VK_FORMAT_R16G16_SNORM] = GEN6_FORMAT_R16G16_SNORM,
+ [VK_FORMAT_R16G16_USCALED] = GEN6_FORMAT_R16G16_USCALED,
+ [VK_FORMAT_R16G16_SSCALED] = GEN6_FORMAT_R16G16_SSCALED,
+ [VK_FORMAT_R16G16_UINT] = GEN6_FORMAT_R16G16_UINT,
+ [VK_FORMAT_R16G16_SINT] = GEN6_FORMAT_R16G16_SINT,
+ [VK_FORMAT_R16G16_SFLOAT] = GEN6_FORMAT_R16G16_FLOAT,
+ [VK_FORMAT_R16G16B16_UNORM] = GEN6_FORMAT_R16G16B16_UNORM,
+ [VK_FORMAT_R16G16B16_SNORM] = GEN6_FORMAT_R16G16B16_SNORM,
+ [VK_FORMAT_R16G16B16_USCALED] = GEN6_FORMAT_R16G16B16_USCALED,
+ [VK_FORMAT_R16G16B16_SSCALED] = GEN6_FORMAT_R16G16B16_SSCALED,
+ [VK_FORMAT_R16G16B16_UINT] = GEN6_FORMAT_R16G16B16_UINT,
+ [VK_FORMAT_R16G16B16_SINT] = GEN6_FORMAT_R16G16B16_SINT,
+ [VK_FORMAT_R16G16B16_SFLOAT] = 0,
+ [VK_FORMAT_R16G16B16A16_UNORM] = GEN6_FORMAT_R16G16B16A16_UNORM,
+ [VK_FORMAT_R16G16B16A16_SNORM] = GEN6_FORMAT_R16G16B16A16_SNORM,
+ [VK_FORMAT_R16G16B16A16_USCALED] = GEN6_FORMAT_R16G16B16A16_USCALED,
+ [VK_FORMAT_R16G16B16A16_SSCALED] = GEN6_FORMAT_R16G16B16A16_SSCALED,
+ [VK_FORMAT_R16G16B16A16_UINT] = GEN6_FORMAT_R16G16B16A16_UINT,
+ [VK_FORMAT_R16G16B16A16_SINT] = GEN6_FORMAT_R16G16B16A16_SINT,
+ [VK_FORMAT_R16G16B16A16_SFLOAT] = GEN6_FORMAT_R16G16B16A16_FLOAT,
+ [VK_FORMAT_R32_UINT] = GEN6_FORMAT_R32_UINT,
+ [VK_FORMAT_R32_SINT] = GEN6_FORMAT_R32_SINT,
+ [VK_FORMAT_R32_SFLOAT] = GEN6_FORMAT_R32_FLOAT,
+ [VK_FORMAT_R32G32_UINT] = GEN6_FORMAT_R32G32_UINT,
+ [VK_FORMAT_R32G32_SINT] = GEN6_FORMAT_R32G32_SINT,
+ [VK_FORMAT_R32G32_SFLOAT] = GEN6_FORMAT_R32G32_FLOAT,
+ [VK_FORMAT_R32G32B32_UINT] = GEN6_FORMAT_R32G32B32_UINT,
+ [VK_FORMAT_R32G32B32_SINT] = GEN6_FORMAT_R32G32B32_SINT,
+ [VK_FORMAT_R32G32B32_SFLOAT] = GEN6_FORMAT_R32G32B32_FLOAT,
+ [VK_FORMAT_R32G32B32A32_UINT] = GEN6_FORMAT_R32G32B32A32_UINT,
+ [VK_FORMAT_R32G32B32A32_SINT] = GEN6_FORMAT_R32G32B32A32_SINT,
+ [VK_FORMAT_R32G32B32A32_SFLOAT] = GEN6_FORMAT_R32G32B32A32_FLOAT,
+ [VK_FORMAT_R64_SFLOAT] = GEN6_FORMAT_R64_FLOAT,
+ [VK_FORMAT_R64G64_SFLOAT] = GEN6_FORMAT_R64G64_FLOAT,
+ [VK_FORMAT_R64G64B64_SFLOAT] = GEN6_FORMAT_R64G64B64_FLOAT,
+ [VK_FORMAT_R64G64B64A64_SFLOAT] = GEN6_FORMAT_R64G64B64A64_FLOAT,
+ [VK_FORMAT_R11G11B10_UFLOAT] = GEN6_FORMAT_R11G11B10_FLOAT,
+ [VK_FORMAT_R9G9B9E5_UFLOAT] = GEN6_FORMAT_R9G9B9E5_SHAREDEXP,
+ [VK_FORMAT_BC1_RGB_UNORM] = GEN6_FORMAT_BC1_UNORM,
+ [VK_FORMAT_BC1_RGB_SRGB] = GEN6_FORMAT_BC1_UNORM_SRGB,
+ [VK_FORMAT_BC2_UNORM] = GEN6_FORMAT_BC2_UNORM,
+ [VK_FORMAT_BC2_SRGB] = GEN6_FORMAT_BC2_UNORM_SRGB,
+ [VK_FORMAT_BC3_UNORM] = GEN6_FORMAT_BC3_UNORM,
+ [VK_FORMAT_BC3_SRGB] = GEN6_FORMAT_BC3_UNORM_SRGB,
+ [VK_FORMAT_BC4_UNORM] = GEN6_FORMAT_BC4_UNORM,
+ [VK_FORMAT_BC4_SNORM] = GEN6_FORMAT_BC4_SNORM,
+ [VK_FORMAT_BC5_UNORM] = GEN6_FORMAT_BC5_UNORM,
+ [VK_FORMAT_BC5_SNORM] = GEN6_FORMAT_BC5_SNORM,
+ [VK_FORMAT_BC6H_UFLOAT] = GEN6_FORMAT_BC6H_UF16,
+ [VK_FORMAT_BC6H_SFLOAT] = GEN6_FORMAT_BC6H_SF16,
+ [VK_FORMAT_BC7_UNORM] = GEN6_FORMAT_BC7_UNORM,
+ [VK_FORMAT_BC7_SRGB] = GEN6_FORMAT_BC7_UNORM_SRGB,
/* TODO: Implement for remaining compressed formats. */
- [VK_FMT_ETC2_R8G8B8_UNORM] = 0,
- [VK_FMT_ETC2_R8G8B8A1_UNORM] = 0,
- [VK_FMT_ETC2_R8G8B8A8_UNORM] = 0,
- [VK_FMT_EAC_R11_UNORM] = 0,
- [VK_FMT_EAC_R11_SNORM] = 0,
- [VK_FMT_EAC_R11G11_UNORM] = 0,
- [VK_FMT_EAC_R11G11_SNORM] = 0,
- [VK_FMT_ASTC_4x4_UNORM] = 0,
- [VK_FMT_ASTC_4x4_SRGB] = 0,
- [VK_FMT_ASTC_5x4_UNORM] = 0,
- [VK_FMT_ASTC_5x4_SRGB] = 0,
- [VK_FMT_ASTC_5x5_UNORM] = 0,
- [VK_FMT_ASTC_5x5_SRGB] = 0,
- [VK_FMT_ASTC_6x5_UNORM] = 0,
- [VK_FMT_ASTC_6x5_SRGB] = 0,
- [VK_FMT_ASTC_6x6_UNORM] = 0,
- [VK_FMT_ASTC_6x6_SRGB] = 0,
- [VK_FMT_ASTC_8x5_UNORM] = 0,
- [VK_FMT_ASTC_8x5_SRGB] = 0,
- [VK_FMT_ASTC_8x6_UNORM] = 0,
- [VK_FMT_ASTC_8x6_SRGB] = 0,
- [VK_FMT_ASTC_8x8_UNORM] = 0,
- [VK_FMT_ASTC_8x8_SRGB] = 0,
- [VK_FMT_ASTC_10x5_UNORM] = 0,
- [VK_FMT_ASTC_10x5_SRGB] = 0,
- [VK_FMT_ASTC_10x6_UNORM] = 0,
- [VK_FMT_ASTC_10x6_SRGB] = 0,
- [VK_FMT_ASTC_10x8_UNORM] = 0,
- [VK_FMT_ASTC_10x8_SRGB] = 0,
- [VK_FMT_ASTC_10x10_UNORM] = 0,
- [VK_FMT_ASTC_10x10_SRGB] = 0,
- [VK_FMT_ASTC_12x10_UNORM] = 0,
- [VK_FMT_ASTC_12x10_SRGB] = 0,
- [VK_FMT_ASTC_12x12_UNORM] = 0,
- [VK_FMT_ASTC_12x12_SRGB] = 0,
- [VK_FMT_B5G6R5_UNORM] = GEN6_FORMAT_B5G6R5_UNORM,
- [VK_FMT_B5G6R5_USCALED] = 0,
- [VK_FMT_B8G8R8_UNORM] = 0,
- [VK_FMT_B8G8R8_SNORM] = 0,
- [VK_FMT_B8G8R8_USCALED] = 0,
- [VK_FMT_B8G8R8_SSCALED] = 0,
- [VK_FMT_B8G8R8_UINT] = 0,
- [VK_FMT_B8G8R8_SINT] = 0,
- [VK_FMT_B8G8R8_SRGB] = GEN6_FORMAT_B5G6R5_UNORM_SRGB,
- [VK_FMT_B8G8R8A8_UNORM] = GEN6_FORMAT_B8G8R8A8_UNORM,
- [VK_FMT_B8G8R8A8_SNORM] = 0,
- [VK_FMT_B8G8R8A8_USCALED] = 0,
- [VK_FMT_B8G8R8A8_SSCALED] = 0,
- [VK_FMT_B8G8R8A8_UINT] = 0,
- [VK_FMT_B8G8R8A8_SINT] = 0,
- [VK_FMT_B8G8R8A8_SRGB] = GEN6_FORMAT_B8G8R8A8_UNORM_SRGB,
- [VK_FMT_B10G10R10A2_UNORM] = GEN6_FORMAT_B10G10R10A2_UNORM,
- [VK_FMT_B10G10R10A2_SNORM] = GEN6_FORMAT_B10G10R10A2_SNORM,
- [VK_FMT_B10G10R10A2_USCALED] = GEN6_FORMAT_B10G10R10A2_USCALED,
- [VK_FMT_B10G10R10A2_SSCALED] = GEN6_FORMAT_B10G10R10A2_SSCALED,
- [VK_FMT_B10G10R10A2_UINT] = GEN6_FORMAT_B10G10R10A2_UINT,
- [VK_FMT_B10G10R10A2_SINT] = GEN6_FORMAT_B10G10R10A2_SINT
+ [VK_FORMAT_ETC2_R8G8B8_UNORM] = 0,
+ [VK_FORMAT_ETC2_R8G8B8A1_UNORM] = 0,
+ [VK_FORMAT_ETC2_R8G8B8A8_UNORM] = 0,
+ [VK_FORMAT_EAC_R11_UNORM] = 0,
+ [VK_FORMAT_EAC_R11_SNORM] = 0,
+ [VK_FORMAT_EAC_R11G11_UNORM] = 0,
+ [VK_FORMAT_EAC_R11G11_SNORM] = 0,
+ [VK_FORMAT_ASTC_4x4_UNORM] = 0,
+ [VK_FORMAT_ASTC_4x4_SRGB] = 0,
+ [VK_FORMAT_ASTC_5x4_UNORM] = 0,
+ [VK_FORMAT_ASTC_5x4_SRGB] = 0,
+ [VK_FORMAT_ASTC_5x5_UNORM] = 0,
+ [VK_FORMAT_ASTC_5x5_SRGB] = 0,
+ [VK_FORMAT_ASTC_6x5_UNORM] = 0,
+ [VK_FORMAT_ASTC_6x5_SRGB] = 0,
+ [VK_FORMAT_ASTC_6x6_UNORM] = 0,
+ [VK_FORMAT_ASTC_6x6_SRGB] = 0,
+ [VK_FORMAT_ASTC_8x5_UNORM] = 0,
+ [VK_FORMAT_ASTC_8x5_SRGB] = 0,
+ [VK_FORMAT_ASTC_8x6_UNORM] = 0,
+ [VK_FORMAT_ASTC_8x6_SRGB] = 0,
+ [VK_FORMAT_ASTC_8x8_UNORM] = 0,
+ [VK_FORMAT_ASTC_8x8_SRGB] = 0,
+ [VK_FORMAT_ASTC_10x5_UNORM] = 0,
+ [VK_FORMAT_ASTC_10x5_SRGB] = 0,
+ [VK_FORMAT_ASTC_10x6_UNORM] = 0,
+ [VK_FORMAT_ASTC_10x6_SRGB] = 0,
+ [VK_FORMAT_ASTC_10x8_UNORM] = 0,
+ [VK_FORMAT_ASTC_10x8_SRGB] = 0,
+ [VK_FORMAT_ASTC_10x10_UNORM] = 0,
+ [VK_FORMAT_ASTC_10x10_SRGB] = 0,
+ [VK_FORMAT_ASTC_12x10_UNORM] = 0,
+ [VK_FORMAT_ASTC_12x10_SRGB] = 0,
+ [VK_FORMAT_ASTC_12x12_UNORM] = 0,
+ [VK_FORMAT_ASTC_12x12_SRGB] = 0,
+ [VK_FORMAT_B5G6R5_UNORM] = GEN6_FORMAT_B5G6R5_UNORM,
+ [VK_FORMAT_B5G6R5_USCALED] = 0,
+ [VK_FORMAT_B8G8R8_UNORM] = 0,
+ [VK_FORMAT_B8G8R8_SNORM] = 0,
+ [VK_FORMAT_B8G8R8_USCALED] = 0,
+ [VK_FORMAT_B8G8R8_SSCALED] = 0,
+ [VK_FORMAT_B8G8R8_UINT] = 0,
+ [VK_FORMAT_B8G8R8_SINT] = 0,
+ [VK_FORMAT_B8G8R8_SRGB] = GEN6_FORMAT_B5G6R5_UNORM_SRGB,
+ [VK_FORMAT_B8G8R8A8_UNORM] = GEN6_FORMAT_B8G8R8A8_UNORM,
+ [VK_FORMAT_B8G8R8A8_SNORM] = 0,
+ [VK_FORMAT_B8G8R8A8_USCALED] = 0,
+ [VK_FORMAT_B8G8R8A8_SSCALED] = 0,
+ [VK_FORMAT_B8G8R8A8_UINT] = 0,
+ [VK_FORMAT_B8G8R8A8_SINT] = 0,
+ [VK_FORMAT_B8G8R8A8_SRGB] = GEN6_FORMAT_B8G8R8A8_UNORM_SRGB,
+ [VK_FORMAT_B10G10R10A2_UNORM] = GEN6_FORMAT_B10G10R10A2_UNORM,
+ [VK_FORMAT_B10G10R10A2_SNORM] = GEN6_FORMAT_B10G10R10A2_SNORM,
+ [VK_FORMAT_B10G10R10A2_USCALED] = GEN6_FORMAT_B10G10R10A2_USCALED,
+ [VK_FORMAT_B10G10R10A2_SSCALED] = GEN6_FORMAT_B10G10R10A2_SSCALED,
+ [VK_FORMAT_B10G10R10A2_UINT] = GEN6_FORMAT_B10G10R10A2_UINT,
+ [VK_FORMAT_B10G10R10A2_SINT] = GEN6_FORMAT_B10G10R10A2_SINT
};
int intel_format_translate_color(const struct intel_gpu *gpu,
@@ -552,7 +552,7 @@
/* TODO: Implement for remaining compressed formats. */
/* GEN6_FORMAT_R32G32B32A32_FLOAT happens to be 0 */
- if (format == VK_FMT_R32G32B32A32_SFLOAT)
+ if (format == VK_FORMAT_R32G32B32A32_SFLOAT)
assert(fmt == 0);
else if (!fmt)
fmt = -1;
@@ -577,7 +577,7 @@
vf = (fmt < ARRAY_SIZE(intel_vf_caps)) ? &intel_vf_caps[fmt] : NULL;
dp = (fmt < ARRAY_SIZE(intel_dp_caps)) ? &intel_dp_caps[fmt] : NULL;
- features = VK_FORMAT_STORAGE_IMAGE_BIT;
+ features = VK_FORMAT_FEATURE_STORAGE_IMAGE_BIT;
#define TEST(dev, func, cap) ((func) && (func)->cap && \
intel_gpu_gen((dev)->gpu) >= (func)->cap)
@@ -588,21 +588,21 @@
if (TEST(dev, sampler, sampling)) {
if (icd_format_is_int(format) ||
TEST(dev, sampler, filtering))
- features |= VK_FORMAT_SAMPLED_IMAGE_BIT;
+ features |= VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT;
}
if (TEST(dev, dp, typed_write))
- features |= VK_FORMAT_SAMPLED_IMAGE_BIT;
+ features |= VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT;
if (TEST(dev, dp, rt_write)) {
- features |= VK_FORMAT_COLOR_ATTACHMENT_BIT;
+ features |= VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT;
if (TEST(dev, dp, rt_write_blending))
- features |= VK_FORMAT_COLOR_ATTACHMENT_BLEND_BIT;
+ features |= VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BLEND_BIT;
- if (features & VK_FORMAT_SAMPLED_IMAGE_BIT) {
- features |= VK_FORMAT_SAMPLED_IMAGE_BIT |
- VK_FORMAT_CONVERSION_BIT;
+ if (features & VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT) {
+ features |= VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT |
+ VK_FORMAT_FEATURE_CONVERSION_BIT;
}
}
#undef TEST
@@ -618,19 +618,19 @@
assert(icd_format_is_ds(format));
switch (format) {
- case VK_FMT_S8_UINT:
- features = VK_FORMAT_DEPTH_STENCIL_ATTACHMENT_BIT;;
+ case VK_FORMAT_S8_UINT:
+ features = VK_FORMAT_FEATURE_DEPTH_STENCIL_ATTACHMENT_BIT;;
break;
- case VK_FMT_D16_UNORM:
- case VK_FMT_D24_UNORM:
- case VK_FMT_D32_SFLOAT:
- features = VK_FORMAT_DEPTH_STENCIL_ATTACHMENT_BIT;
+ case VK_FORMAT_D16_UNORM:
+ case VK_FORMAT_D24_UNORM:
+ case VK_FORMAT_D32_SFLOAT:
+ features = VK_FORMAT_FEATURE_DEPTH_STENCIL_ATTACHMENT_BIT;
break;
- case VK_FMT_D16_UNORM_S8_UINT:
- case VK_FMT_D24_UNORM_S8_UINT:
- case VK_FMT_D32_SFLOAT_S8_UINT:
- features = VK_FORMAT_DEPTH_STENCIL_ATTACHMENT_BIT |
- VK_FORMAT_DEPTH_STENCIL_ATTACHMENT_BIT;
+ case VK_FORMAT_D16_UNORM_S8_UINT:
+ case VK_FORMAT_D24_UNORM_S8_UINT:
+ case VK_FORMAT_D32_SFLOAT_S8_UINT:
+ features = VK_FORMAT_FEATURE_DEPTH_STENCIL_ATTACHMENT_BIT |
+ VK_FORMAT_FEATURE_DEPTH_STENCIL_ATTACHMENT_BIT;
break;
default:
features = 0;
@@ -643,8 +643,8 @@
static VkFlags intel_format_get_raw_features(const struct intel_dev *dev,
VkFormat format)
{
- return (format == VK_FMT_UNDEFINED) ?
- VK_FORMAT_STORAGE_IMAGE_BIT : 0;
+ return (format == VK_FORMAT_UNDEFINED) ?
+ VK_FORMAT_FEATURE_STORAGE_IMAGE_BIT : 0;
}
static void intel_format_get_props(const struct intel_dev *dev,
@@ -680,7 +680,7 @@
VkResult ret = VK_SUCCESS;
switch (infoType) {
- case VK_INFO_TYPE_FORMAT_PROPERTIES:
+ case VK_FORMAT_INFO_TYPE_PROPERTIES:
*pDataSize = sizeof(VkFormatProperties);
if (pData == NULL)
return ret;
diff --git a/icd/intel/format.h b/icd/intel/format.h
index 897de92..c464243 100644
--- a/icd/intel/format.h
+++ b/icd/intel/format.h
@@ -38,12 +38,12 @@
bool has_depth = false;
switch (format) {
- case VK_FMT_D16_UNORM:
- case VK_FMT_D24_UNORM:
- case VK_FMT_D32_SFLOAT:
- /* VK_FMT_D16_UNORM_S8_UINT is unsupported */
- case VK_FMT_D24_UNORM_S8_UINT:
- case VK_FMT_D32_SFLOAT_S8_UINT:
+ case VK_FORMAT_D16_UNORM:
+ case VK_FORMAT_D24_UNORM:
+ case VK_FORMAT_D32_SFLOAT:
+ /* VK_FORMAT_D16_UNORM_S8_UINT is unsupported */
+ case VK_FORMAT_D24_UNORM_S8_UINT:
+ case VK_FORMAT_D32_SFLOAT_S8_UINT:
has_depth = true;
break;
default:
diff --git a/icd/intel/gpu.c b/icd/intel/gpu.c
index 5c7eaac..a329e1a 100644
--- a/icd/intel/gpu.c
+++ b/icd/intel/gpu.c
@@ -164,9 +164,9 @@
return VK_ERROR_INITIALIZATION_FAILED;
}
- gpu = intel_alloc(instance, sizeof(*gpu), 0, VK_SYSTEM_ALLOC_API_OBJECT);
+ gpu = intel_alloc(instance, sizeof(*gpu), 0, VK_SYSTEM_ALLOC_TYPE_API_OBJECT);
if (!gpu)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
memset(gpu, 0, sizeof(*gpu));
/* there is no VK_DBG_OBJECT_GPU */
@@ -178,10 +178,10 @@
render_len = (render_node) ? strlen(render_node) : 0;
gpu->primary_node = intel_alloc(gpu, primary_len + 1 +
- ((render_len) ? (render_len + 1) : 0), 0, VK_SYSTEM_ALLOC_INTERNAL);
+ ((render_len) ? (render_len + 1) : 0), 0, VK_SYSTEM_ALLOC_TYPE_INTERNAL);
if (!gpu->primary_node) {
intel_free(instance, gpu);
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
}
memcpy(gpu->primary_node, primary_node, primary_len + 1);
@@ -221,7 +221,7 @@
}
void intel_gpu_get_props(const struct intel_gpu *gpu,
- VkPhysicalGpuProperties *props)
+ VkPhysicalDeviceProperties *props)
{
const char *name;
size_t name_len;
@@ -232,15 +232,15 @@
props->vendorId = 0x8086;
props->deviceId = gpu->devid;
- props->gpuType = VK_GPU_TYPE_INTEGRATED;
+ props->deviceType = VK_PHYSICAL_DEVICE_TYPE_INTEGRATED_GPU;
/* copy GPU name */
name = gpu_get_name(gpu);
name_len = strlen(name);
- if (name_len > sizeof(props->gpuName) - 1)
- name_len = sizeof(props->gpuName) - 1;
- memcpy(props->gpuName, name, name_len);
- props->gpuName[name_len] = '\0';
+ if (name_len > sizeof(props->deviceName) - 1)
+ name_len = sizeof(props->deviceName) - 1;
+ memcpy(props->deviceName, name, name_len);
+ props->deviceName[name_len] = '\0';
/* no size limit, but no bounded buffer could exceed 2GB */
@@ -255,10 +255,10 @@
}
void intel_gpu_get_perf(const struct intel_gpu *gpu,
- VkPhysicalGpuPerformance *perf)
+ VkPhysicalDevicePerformance *perf)
{
/* TODO */
- perf->maxGpuClock = 1.0f;
+ perf->maxDeviceClock = 1.0f;
perf->aluPerClock = 1.0f;
perf->texPerClock = 1.0f;
perf->primsPerClock = 1.0f;
@@ -267,7 +267,7 @@
void intel_gpu_get_queue_props(const struct intel_gpu *gpu,
enum intel_gpu_engine_type engine,
- VkPhysicalGpuQueueProperties *props)
+ VkPhysicalDeviceQueueProperties *props)
{
switch (engine) {
case INTEL_GPU_ENGINE_3D:
@@ -284,14 +284,14 @@
}
void intel_gpu_get_memory_props(const struct intel_gpu *gpu,
- VkPhysicalGpuMemoryProperties *props)
+ VkPhysicalDeviceMemoryProperties *props)
{
props->supportsMigration = false;
props->supportsPinning = true;
}
int intel_gpu_get_max_threads(const struct intel_gpu *gpu,
- VkPipelineShaderStage stage)
+ VkShaderStage stage)
{
switch (intel_gpu_gen(gpu)) {
case INTEL_GEN(7.5):
@@ -395,7 +395,7 @@
}
ICD_EXPORT VkResult VKAPI vkEnumerateLayers(
- VkPhysicalGpu gpu,
+ VkPhysicalDevice gpu,
size_t maxLayerCount,
size_t maxStringSize,
size_t* pOutLayerCount,
@@ -410,9 +410,9 @@
return VK_SUCCESS;
}
-ICD_EXPORT VkResult VKAPI vkGetGpuInfo(
- VkPhysicalGpu gpu_,
- VkPhysicalGpuInfoType infoType,
+ICD_EXPORT VkResult VKAPI vkGetPhysicalDeviceInfo(
+ VkPhysicalDevice gpu_,
+ VkPhysicalDeviceInfoType infoType,
size_t* pDataSize,
void* pData)
{
@@ -420,34 +420,34 @@
VkResult ret = VK_SUCCESS;
switch (infoType) {
- case VK_INFO_TYPE_PHYSICAL_GPU_PROPERTIES:
- *pDataSize = sizeof(VkPhysicalGpuProperties);
+ case VK_PHYSICAL_DEVICE_INFO_TYPE_PROPERTIES:
+ *pDataSize = sizeof(VkPhysicalDeviceProperties);
if (pData == NULL) {
return ret;
}
intel_gpu_get_props(gpu, pData);
break;
- case VK_INFO_TYPE_PHYSICAL_GPU_PERFORMANCE:
- *pDataSize = sizeof(VkPhysicalGpuPerformance);
+ case VK_PHYSICAL_DEVICE_INFO_TYPE_PERFORMANCE:
+ *pDataSize = sizeof(VkPhysicalDevicePerformance);
if (pData == NULL) {
return ret;
}
intel_gpu_get_perf(gpu, pData);
break;
- case VK_INFO_TYPE_PHYSICAL_GPU_QUEUE_PROPERTIES:
+ case VK_PHYSICAL_DEVICE_INFO_TYPE_QUEUE_PROPERTIES:
/*
* Vulkan Programmers guide, page 33:
* to determine the data size an application calls
- * vkGetGpuInfo() with a NULL data pointer. The
+ * vkGetPhysicalDeviceInfo() with a NULL data pointer. The
* expected data size for all queue property structures
* is returned in pDataSize
*/
- *pDataSize = sizeof(VkPhysicalGpuQueueProperties) *
+ *pDataSize = sizeof(VkPhysicalDeviceQueueProperties) *
INTEL_GPU_ENGINE_COUNT;
if (pData != NULL) {
- VkPhysicalGpuQueueProperties *dst = pData;
+ VkPhysicalDeviceQueueProperties *dst = pData;
int engine;
for (engine = 0; engine < INTEL_GPU_ENGINE_COUNT; engine++) {
@@ -457,8 +457,8 @@
}
break;
- case VK_INFO_TYPE_PHYSICAL_GPU_MEMORY_PROPERTIES:
- *pDataSize = sizeof(VkPhysicalGpuMemoryProperties);
+ case VK_PHYSICAL_DEVICE_INFO_TYPE_MEMORY_PROPERTIES:
+ *pDataSize = sizeof(VkPhysicalDeviceMemoryProperties);
if (pData == NULL) {
return ret;
}
@@ -513,7 +513,7 @@
}
ICD_EXPORT VkResult VKAPI vkGetPhysicalDeviceExtensionInfo(
- VkPhysicalGpu gpu,
+ VkPhysicalDevice gpu,
VkExtensionInfoType infoType,
uint32_t extensionIndex,
size_t* pDataSize,
@@ -554,21 +554,21 @@
return VK_SUCCESS;
}
-ICD_EXPORT VkResult VKAPI vkGetMultiGpuCompatibility(
- VkPhysicalGpu gpu0_,
- VkPhysicalGpu gpu1_,
- VkGpuCompatibilityInfo* pInfo)
+ICD_EXPORT VkResult VKAPI vkGetMultiDeviceCompatibility(
+ VkPhysicalDevice gpu0_,
+ VkPhysicalDevice gpu1_,
+ VkPhysicalDeviceCompatibilityInfo* pInfo)
{
const struct intel_gpu *gpu0 = intel_gpu(gpu0_);
const struct intel_gpu *gpu1 = intel_gpu(gpu1_);
- VkFlags compat = VK_GPU_COMPAT_IQ_MATCH_BIT |
- VK_GPU_COMPAT_PEER_TRANSFER_BIT |
- VK_GPU_COMPAT_SHARED_MEMORY_BIT |
- VK_GPU_COMPAT_SHARED_GPU0_DISPLAY_BIT |
- VK_GPU_COMPAT_SHARED_GPU1_DISPLAY_BIT;
+ VkFlags compat = VK_PHYSICAL_DEVICE_COMPATIBILITY_IQ_MATCH_BIT |
+ VK_PHYSICAL_DEVICE_COMPATIBILITY_PEER_TRANSFER_BIT |
+ VK_PHYSICAL_DEVICE_COMPATIBILITY_SHARED_MEMORY_BIT |
+ VK_PHYSICAL_DEVICE_COMPATIBILITY_SHARED_DEVICE0_DISPLAY_BIT |
+ VK_PHYSICAL_DEVICE_COMPATIBILITY_SHARED_DEVICE1_DISPLAY_BIT;
if (intel_gpu_gen(gpu0) == intel_gpu_gen(gpu1))
- compat |= VK_GPU_COMPAT_ASIC_FEATURES_BIT;
+ compat |= VK_PHYSICAL_DEVICE_COMPATIBILITY_FEATURES_BIT;
pInfo->compatibilityFlags = compat;
diff --git a/icd/intel/gpu.h b/icd/intel/gpu.h
index ba32f3b..30781da 100644
--- a/icd/intel/gpu.h
+++ b/icd/intel/gpu.h
@@ -74,7 +74,7 @@
int gen_opaque; /* always read this with intel_gpu_gen() */
int gt;
- VkGpuSize max_batch_buffer_size;
+ VkDeviceSize max_batch_buffer_size;
uint32_t batch_buffer_reloc_count;
/*
@@ -93,7 +93,7 @@
uint32_t display_count;
};
-static inline struct intel_gpu *intel_gpu(VkPhysicalGpu gpu)
+static inline struct intel_gpu *intel_gpu(VkPhysicalDevice gpu)
{
return (struct intel_gpu *) gpu;
}
@@ -113,17 +113,17 @@
void intel_gpu_destroy(struct intel_gpu *gpu);
void intel_gpu_get_props(const struct intel_gpu *gpu,
- VkPhysicalGpuProperties *props);
+ VkPhysicalDeviceProperties *props);
void intel_gpu_get_perf(const struct intel_gpu *gpu,
- VkPhysicalGpuPerformance *perf);
+ VkPhysicalDevicePerformance *perf);
void intel_gpu_get_queue_props(const struct intel_gpu *gpu,
enum intel_gpu_engine_type engine,
- VkPhysicalGpuQueueProperties *props);
+ VkPhysicalDeviceQueueProperties *props);
void intel_gpu_get_memory_props(const struct intel_gpu *gpu,
- VkPhysicalGpuMemoryProperties *props);
+ VkPhysicalDeviceMemoryProperties *props);
int intel_gpu_get_max_threads(const struct intel_gpu *gpu,
- VkPipelineShaderStage stage);
+ VkShaderStage stage);
int intel_gpu_get_primary_fd(struct intel_gpu *gpu);
diff --git a/icd/intel/img.c b/icd/intel/img.c
index 6796865..80e12b4 100644
--- a/icd/intel/img.c
+++ b/icd/intel/img.c
@@ -54,7 +54,7 @@
VkResult ret = VK_SUCCESS;
switch (type) {
- case VK_INFO_TYPE_MEMORY_REQUIREMENTS:
+ case VK_OBJECT_INFO_TYPE_MEMORY_REQUIREMENTS:
{
VkMemoryRequirements *mem_req = data;
@@ -85,7 +85,7 @@
img = (struct intel_img *) intel_base_create(&dev->base.handle,
sizeof(*img), dev->base.dbg, VK_DBG_OBJECT_IMAGE, info, 0);
if (!img)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
layout = &img->layout;
@@ -116,14 +116,14 @@
VkImageCreateInfo s8_info;
img->s8_layout = intel_alloc(img, sizeof(*img->s8_layout), 0,
- VK_SYSTEM_ALLOC_INTERNAL);
+ VK_SYSTEM_ALLOC_TYPE_INTERNAL);
if (!img->s8_layout) {
intel_img_destroy(img);
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
}
s8_info = *info;
- s8_info.format = VK_FMT_S8_UINT;
+ s8_info.format = VK_FORMAT_S8_UINT;
/* no stencil texturing */
s8_info.usage &= ~VK_IMAGE_USAGE_SAMPLED_BIT;
assert(icd_format_is_ds(info->format));
@@ -166,7 +166,7 @@
VkDevice device,
const VkPeerImageOpenInfo* pOpenInfo,
VkImage* pImage,
- VkGpuMemory* pMem)
+ VkDeviceMemory* pMem)
{
return VK_ERROR_UNAVAILABLE;
}
@@ -193,7 +193,7 @@
VkResult ret = VK_SUCCESS;
switch (infoType) {
- case VK_INFO_TYPE_SUBRESOURCE_LAYOUT:
+ case VK_SUBRESOURCE_INFO_TYPE_LAYOUT:
{
VkSubresourceLayout *layout = (VkSubresourceLayout *) pData;
unsigned x, y;
diff --git a/icd/intel/instance.c b/icd/intel/instance.c
index fd7265d..a515e2c 100644
--- a/icd/intel/instance.c
+++ b/icd/intel/instance.c
@@ -121,7 +121,7 @@
return NULL;
instance = icd_instance_alloc(icd, sizeof(*instance), 0,
- VK_SYSTEM_ALLOC_API_OBJECT);
+ VK_SYSTEM_ALLOC_TYPE_API_OBJECT);
if (!instance) {
icd_instance_destroy(icd);
return NULL;
@@ -150,7 +150,7 @@
instance = intel_instance_create(pCreateInfo);
if (!instance)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
*pInstance = (VkInstance) instance;
@@ -170,7 +170,7 @@
ICD_EXPORT VkResult VKAPI vkEnumeratePhysicalDevices(
VkInstance instance_,
uint32_t* pPhysicalDeviceCount,
- VkPhysicalGpu* pPhysicalDevices)
+ VkPhysicalDevice* pPhysicalDevices)
{
struct intel_instance *instance = intel_instance(instance_);
struct icd_drm_device *devices, *dev;
@@ -205,7 +205,7 @@
if (ret == VK_SUCCESS) {
intel_instance_add_gpu(instance, gpu);
- pPhysicalDevices[count++] = (VkPhysicalGpu) gpu;
+ pPhysicalDevices[count++] = (VkPhysicalDevice) gpu;
if (count >= *pPhysicalDeviceCount)
break;
}
diff --git a/icd/intel/intel.h b/icd/intel/intel.h
index de7bcbd..779e075 100644
--- a/icd/intel/intel.h
+++ b/icd/intel/intel.h
@@ -56,9 +56,9 @@
#define INTEL_MAX_VERTEX_ELEMENT_COUNT (INTEL_MAX_VERTEX_BINDING_COUNT + 1)
#define INTEL_MAX_RENDER_TARGETS 8
-#define INTEL_MEMORY_PROPERTY_ALL (VK_MEMORY_PROPERTY_GPU_ONLY | VK_MEMORY_PROPERTY_CPU_VISIBLE_BIT |\
- VK_MEMORY_PROPERTY_CPU_GPU_COHERENT_BIT | VK_MEMORY_PROPERTY_CPU_UNCACHED_BIT |\
- VK_MEMORY_PROPERTY_CPU_WRITE_COMBINED_BIT | VK_MEMORY_PROPERTY_PREFER_CPU_LOCAL |\
+#define INTEL_MEMORY_PROPERTY_ALL (VK_MEMORY_PROPERTY_DEVICE_ONLY | VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT |\
+ VK_MEMORY_PROPERTY_HOST_DEVICE_COHERENT_BIT | VK_MEMORY_PROPERTY_HOST_UNCACHED_BIT |\
+ VK_MEMORY_PROPERTY_HOST_WRITE_COMBINED_BIT | VK_MEMORY_PROPERTY_PREFER_HOST_LOCAL |\
VK_MEMORY_PROPERTY_SHAREABLE_BIT)
enum intel_debug_flags {
diff --git a/icd/intel/kmd/winsys_drm.c b/icd/intel/kmd/winsys_drm.c
index 349cc88..50b6f90 100644
--- a/icd/intel/kmd/winsys_drm.c
+++ b/icd/intel/kmd/winsys_drm.c
@@ -168,7 +168,7 @@
struct intel_winsys *winsys;
winsys = icd_instance_alloc(instance, sizeof(*winsys), 0,
- VK_SYSTEM_ALLOC_INTERNAL);
+ VK_SYSTEM_ALLOC_TYPE_INTERNAL);
if (!winsys)
return NULL;
diff --git a/icd/intel/layout.c b/icd/intel/layout.c
index dbc5e14..5826ec5 100644
--- a/icd/intel/layout.c
+++ b/icd/intel/layout.c
@@ -259,7 +259,7 @@
/* every LOD begins at tile boundaries */
if (info->mipLevels > 1) {
- assert(layout->format == VK_FMT_S8_UINT);
+ assert(layout->format == VK_FORMAT_S8_UINT);
cur_x = u_align(cur_x, 64);
cur_y = u_align(cur_y, 64);
}
@@ -403,11 +403,11 @@
} else if (info->usage & VK_IMAGE_USAGE_DEPTH_STENCIL_BIT) {
if (intel_gpu_gen(params->gpu) >= INTEL_GEN(7)) {
switch (layout->format) {
- case VK_FMT_D16_UNORM:
+ case VK_FORMAT_D16_UNORM:
layout->align_i = 8;
layout->align_j = 4;
break;
- case VK_FMT_S8_UINT:
+ case VK_FORMAT_S8_UINT:
layout->align_i = 8;
layout->align_j = 8;
break;
@@ -418,7 +418,7 @@
}
} else {
switch (layout->format) {
- case VK_FMT_S8_UINT:
+ case VK_FORMAT_S8_UINT:
layout->align_i = 4;
layout->align_j = 2;
break;
@@ -438,7 +438,7 @@
if (intel_gpu_gen(params->gpu) >= INTEL_GEN(7) &&
intel_gpu_gen(params->gpu) <= INTEL_GEN(7.5) && valign_4)
- assert(layout->format != VK_FMT_R32G32B32_SFLOAT);
+ assert(layout->format != VK_FORMAT_R32G32B32_SFLOAT);
layout->align_i = 4;
layout->align_j = (valign_4) ? 4 : 2;
@@ -477,7 +477,7 @@
if (params->scanout)
valid_tilings &= LAYOUT_TILING_X;
- if (info->tiling == VK_LINEAR_TILING)
+ if (info->tiling == VK_IMAGE_TILING_LINEAR)
valid_tilings &= LAYOUT_TILING_NONE;
/*
@@ -494,7 +494,7 @@
*/
if (info->usage & VK_IMAGE_USAGE_DEPTH_STENCIL_BIT) {
switch (format) {
- case VK_FMT_S8_UINT:
+ case VK_FORMAT_S8_UINT:
valid_tilings &= LAYOUT_TILING_W;
break;
default:
@@ -528,7 +528,7 @@
*/
if (intel_gpu_gen(params->gpu) >= INTEL_GEN(7) &&
intel_gpu_gen(params->gpu) <= INTEL_GEN(7.5) &&
- layout->format == VK_FMT_R32G32B32_SFLOAT)
+ layout->format == VK_FORMAT_R32G32B32_SFLOAT)
valid_tilings &= ~LAYOUT_TILING_Y;
valid_tilings &= ~LAYOUT_TILING_W;
@@ -609,7 +609,7 @@
* "note that the depth buffer and stencil buffer have an implied
* value of ARYSPC_FULL"
*/
- layout->walk = (info->imageType == VK_IMAGE_3D) ?
+ layout->walk = (info->imageType == VK_IMAGE_TYPE_3D) ?
INTEL_LAYOUT_WALK_3D : INTEL_LAYOUT_WALK_LAYER;
layout->interleaved_samples = true;
@@ -628,7 +628,7 @@
assert(info->mipLevels == 1);
layout->walk =
- (info->imageType == VK_IMAGE_3D) ? INTEL_LAYOUT_WALK_3D :
+ (info->imageType == VK_IMAGE_TYPE_3D) ? INTEL_LAYOUT_WALK_3D :
(info->mipLevels > 1) ? INTEL_LAYOUT_WALK_LAYER :
INTEL_LAYOUT_WALK_LOD;
@@ -652,8 +652,8 @@
* GEN6 does not support compact spacing otherwise.
*/
layout->walk =
- (params->info->imageType == VK_IMAGE_3D) ? INTEL_LAYOUT_WALK_3D :
- (layout->format == VK_FMT_S8_UINT) ? INTEL_LAYOUT_WALK_LOD :
+ (params->info->imageType == VK_IMAGE_TYPE_3D) ? INTEL_LAYOUT_WALK_3D :
+ (layout->format == VK_FORMAT_S8_UINT) ? INTEL_LAYOUT_WALK_LOD :
INTEL_LAYOUT_WALK_LAYER;
/* GEN6 supports only interleaved samples */
@@ -697,15 +697,15 @@
}
switch (format) {
- case VK_FMT_D24_UNORM_S8_UINT:
+ case VK_FORMAT_D24_UNORM_S8_UINT:
if (require_separate_stencil) {
- format = VK_FMT_D24_UNORM;
+ format = VK_FORMAT_D24_UNORM;
layout->separate_stencil = true;
}
break;
- case VK_FMT_D32_SFLOAT_S8_UINT:
+ case VK_FORMAT_D32_SFLOAT_S8_UINT:
if (require_separate_stencil) {
- format = VK_FMT_D32_SFLOAT;
+ format = VK_FORMAT_D32_SFLOAT;
layout->separate_stencil = true;
}
break;
@@ -732,7 +732,7 @@
if (intel_gpu_gen(params->gpu) < INTEL_GEN(7))
return false;
- if (info->imageType != VK_IMAGE_2D ||
+ if (info->imageType != VK_IMAGE_TYPE_2D ||
!(info->usage & VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT))
return false;
@@ -851,7 +851,7 @@
align_h = layout->align_j;
/* in case it is used as a cube */
- if (info->imageType == VK_IMAGE_2D)
+ if (info->imageType == VK_IMAGE_TYPE_2D)
pad_h += 2;
if (params->compressed && align_h < layout->align_j * 2)
diff --git a/icd/intel/mem.c b/icd/intel/mem.c
index e998514..93e7394 100644
--- a/icd/intel/mem.c
+++ b/icd/intel/mem.c
@@ -39,7 +39,7 @@
mem = (struct intel_mem *) intel_base_create(&dev->base.handle,
sizeof(*mem), dev->base.dbg, VK_DBG_OBJECT_GPU_MEMORY, info, 0);
if (!mem)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
mem->bo = intel_winsys_alloc_bo(dev->winsys,
"vk-gpu-memory", info->allocationSize, 0);
@@ -76,7 +76,7 @@
mem = (struct intel_mem *) intel_base_create(&dev->base.handle,
sizeof(*mem), dev->base.dbg, VK_DBG_OBJECT_GPU_MEMORY, NULL, 0);
if (!mem)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
mem->bo = intel_winsys_import_userptr(dev->winsys,
"vk-gpu-memory-userptr", (void *) userptr, size, 0);
@@ -102,7 +102,7 @@
ICD_EXPORT VkResult VKAPI vkAllocMemory(
VkDevice device,
const VkMemoryAllocInfo* pAllocInfo,
- VkGpuMemory* pMem)
+ VkDeviceMemory* pMem)
{
struct intel_dev *dev = intel_dev(device);
@@ -110,7 +110,7 @@
}
ICD_EXPORT VkResult VKAPI vkFreeMemory(
- VkGpuMemory mem_)
+ VkDeviceMemory mem_)
{
struct intel_mem *mem = intel_mem(mem_);
@@ -120,7 +120,7 @@
}
ICD_EXPORT VkResult VKAPI vkSetMemoryPriority(
- VkGpuMemory mem_,
+ VkDeviceMemory mem_,
VkMemoryPriority priority)
{
struct intel_mem *mem = intel_mem(mem_);
@@ -129,7 +129,7 @@
}
ICD_EXPORT VkResult VKAPI vkMapMemory(
- VkGpuMemory mem_,
+ VkDeviceMemory mem_,
VkFlags flags,
void** ppData)
{
@@ -142,7 +142,7 @@
}
ICD_EXPORT VkResult VKAPI vkUnmapMemory(
- VkGpuMemory mem_)
+ VkDeviceMemory mem_)
{
struct intel_mem *mem = intel_mem(mem_);
@@ -155,7 +155,7 @@
VkDevice device,
const void* pSysMem,
size_t memSize,
- VkGpuMemory* pMem)
+ VkDeviceMemory* pMem)
{
struct intel_dev *dev = intel_dev(device);
@@ -166,7 +166,7 @@
ICD_EXPORT VkResult VKAPI vkOpenSharedMemory(
VkDevice device,
const VkMemoryOpenInfo* pOpenInfo,
- VkGpuMemory* pMem)
+ VkDeviceMemory* pMem)
{
return VK_ERROR_UNAVAILABLE;
}
@@ -174,7 +174,7 @@
ICD_EXPORT VkResult VKAPI vkOpenPeerMemory(
VkDevice device,
const VkPeerMemoryOpenInfo* pOpenInfo,
- VkGpuMemory* pMem)
+ VkDeviceMemory* pMem)
{
return VK_ERROR_UNAVAILABLE;
}
diff --git a/icd/intel/mem.h b/icd/intel/mem.h
index 040e795..df2a532 100644
--- a/icd/intel/mem.h
+++ b/icd/intel/mem.h
@@ -36,7 +36,7 @@
struct intel_base base;
struct intel_bo *bo;
- VkGpuSize size;
+ VkDeviceSize size;
};
VkResult intel_mem_alloc(struct intel_dev *dev,
@@ -72,7 +72,7 @@
return intel_bo_is_busy(mem->bo);
}
-static inline struct intel_mem *intel_mem(VkGpuMemory mem)
+static inline struct intel_mem *intel_mem(VkDeviceMemory mem)
{
return (struct intel_mem *) mem;
}
diff --git a/icd/intel/obj.c b/icd/intel/obj.c
index 3e8c0a0..cd3d031 100644
--- a/icd/intel/obj.c
+++ b/icd/intel/obj.c
@@ -38,7 +38,7 @@
uint32_t *count;
switch (type) {
- case VK_INFO_TYPE_MEMORY_REQUIREMENTS:
+ case VK_OBJECT_INFO_TYPE_MEMORY_REQUIREMENTS:
{
s = sizeof(VkMemoryRequirements);
*size = s;
@@ -49,7 +49,7 @@
mem_req->memPropsAllowed = INTEL_MEMORY_PROPERTY_ALL;
break;
}
- case VK_INFO_TYPE_MEMORY_ALLOCATION_COUNT:
+ case VK_OBJECT_INFO_TYPE_MEMORY_ALLOCATION_COUNT:
*size = sizeof(uint32_t);
if (data == NULL)
return ret;
@@ -182,7 +182,7 @@
if (shallow_copy) {
dbg->create_info = intel_alloc(handle, shallow_copy, 0,
- VK_SYSTEM_ALLOC_DEBUG);
+ VK_SYSTEM_ALLOC_TYPE_DEBUG);
if (!dbg->create_info)
return false;
@@ -197,7 +197,7 @@
size = sizeof(*src);
dbg->create_info_size = size;
- dst = intel_alloc(handle, size, 0, VK_SYSTEM_ALLOC_DEBUG);
+ dst = intel_alloc(handle, size, 0, VK_SYSTEM_ALLOC_TYPE_DEBUG);
if (!dst)
return false;
memcpy(dst, src, sizeof(*src));
@@ -223,7 +223,7 @@
size += 1 + strlen(src->ppEnabledExtensionNames[i]);
}
- dst = intel_alloc(handle, size, 0, VK_SYSTEM_ALLOC_DEBUG);
+ dst = intel_alloc(handle, size, 0, VK_SYSTEM_ALLOC_TYPE_DEBUG);
if (!dst)
return false;
@@ -272,7 +272,7 @@
assert(dbg_size >= sizeof(*dbg));
- dbg = intel_alloc(handle, dbg_size, 0, VK_SYSTEM_ALLOC_DEBUG);
+ dbg = intel_alloc(handle, dbg_size, 0, VK_SYSTEM_ALLOC_TYPE_DEBUG);
if (!dbg)
return NULL;
@@ -317,7 +317,7 @@
assert(obj_size >= sizeof(*base));
- base = intel_alloc(handle, obj_size, 0, VK_SYSTEM_ALLOC_API_OBJECT);
+ base = intel_alloc(handle, obj_size, 0, VK_SYSTEM_ALLOC_TYPE_API_OBJECT);
if (!base)
return NULL;
@@ -370,8 +370,8 @@
VkQueue queue,
VkObject object,
uint32_t allocationIdx,
- VkGpuMemory mem_,
- VkGpuSize memOffset)
+ VkDeviceMemory mem_,
+ VkDeviceSize memOffset)
{
struct intel_obj *obj = intel_obj(object);
struct intel_mem *mem = intel_mem(mem_);
@@ -385,29 +385,29 @@
VkQueue queue,
VkObject object,
uint32_t allocationIdx,
- VkGpuSize rangeOffset,
- VkGpuSize rangeSize,
- VkGpuMemory mem,
- VkGpuSize memOffset)
+ VkDeviceSize rangeOffset,
+ VkDeviceSize rangeSize,
+ VkDeviceMemory mem,
+ VkDeviceSize memOffset)
{
return VK_ERROR_UNKNOWN;
}
ICD_EXPORT VkResult VKAPI vkQueueBindImageMemoryRange(
- VkQueue queue,
- VkImage image,
+ VkQueue queue,
+ VkImage image,
uint32_t allocationIdx,
- const VkImageMemoryBindInfo* pBindInfo,
- VkGpuMemory mem,
- VkGpuSize memOffset)
+ const VkImageMemoryBindInfo* pBindInfo,
+ VkDeviceMemory mem,
+ VkDeviceSize memOffset)
{
return VK_ERROR_UNKNOWN;
}
ICD_EXPORT VkResult VKAPI vkDbgSetObjectTag(
- VkBaseObject object,
- size_t tagSize,
- const void* pTag)
+ VkBaseObject object,
+ size_t tagSize,
+ const void* pTag)
{
struct intel_base *base = intel_base(object);
struct intel_base_dbg *dbg = base->dbg;
@@ -416,9 +416,9 @@
if (!dbg)
return VK_SUCCESS;
- tag = intel_alloc(base, tagSize, 0, VK_SYSTEM_ALLOC_DEBUG);
+ tag = intel_alloc(base, tagSize, 0, VK_SYSTEM_ALLOC_TYPE_DEBUG);
if (!tag)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
memcpy(tag, pTag, tagSize);
diff --git a/icd/intel/obj.h b/icd/intel/obj.h
index ccb3319..2dc4396 100644
--- a/icd/intel/obj.h
+++ b/icd/intel/obj.h
@@ -74,7 +74,7 @@
static inline void intel_obj_bind_mem(struct intel_obj *obj,
struct intel_mem *mem,
- VkGpuSize offset)
+ VkDeviceSize offset)
{
obj->mem = mem;
obj->offset = offset;
diff --git a/icd/intel/pipeline.c b/icd/intel/pipeline.c
index decb881..531afae 100644
--- a/icd/intel/pipeline.c
+++ b/icd/intel/pipeline.c
@@ -33,14 +33,14 @@
#include "shader.h"
#include "pipeline.h"
-static int translate_blend_func(VkBlendFunc func)
+static int translate_blend_func(VkBlendOp func)
{
switch (func) {
- case VK_BLEND_FUNC_ADD: return GEN6_BLENDFUNCTION_ADD;
- case VK_BLEND_FUNC_SUBTRACT: return GEN6_BLENDFUNCTION_SUBTRACT;
- case VK_BLEND_FUNC_REVERSE_SUBTRACT: return GEN6_BLENDFUNCTION_REVERSE_SUBTRACT;
- case VK_BLEND_FUNC_MIN: return GEN6_BLENDFUNCTION_MIN;
- case VK_BLEND_FUNC_MAX: return GEN6_BLENDFUNCTION_MAX;
+ case VK_BLEND_OP_ADD: return GEN6_BLENDFUNCTION_ADD;
+ case VK_BLEND_OP_SUBTRACT: return GEN6_BLENDFUNCTION_SUBTRACT;
+ case VK_BLEND_OP_REVERSE_SUBTRACT: return GEN6_BLENDFUNCTION_REVERSE_SUBTRACT;
+ case VK_BLEND_OP_MIN: return GEN6_BLENDFUNCTION_MIN;
+ case VK_BLEND_OP_MAX: return GEN6_BLENDFUNCTION_MAX;
default:
assert(!"unknown blend func");
return GEN6_BLENDFUNCTION_ADD;
@@ -75,17 +75,17 @@
};
}
-static int translate_compare_func(VkCompareFunc func)
+static int translate_compare_func(VkCompareOp func)
{
switch (func) {
- case VK_COMPARE_NEVER: return GEN6_COMPAREFUNCTION_NEVER;
- case VK_COMPARE_LESS: return GEN6_COMPAREFUNCTION_LESS;
- case VK_COMPARE_EQUAL: return GEN6_COMPAREFUNCTION_EQUAL;
- case VK_COMPARE_LESS_EQUAL: return GEN6_COMPAREFUNCTION_LEQUAL;
- case VK_COMPARE_GREATER: return GEN6_COMPAREFUNCTION_GREATER;
- case VK_COMPARE_NOT_EQUAL: return GEN6_COMPAREFUNCTION_NOTEQUAL;
- case VK_COMPARE_GREATER_EQUAL: return GEN6_COMPAREFUNCTION_GEQUAL;
- case VK_COMPARE_ALWAYS: return GEN6_COMPAREFUNCTION_ALWAYS;
+ case VK_COMPARE_OP_NEVER: return GEN6_COMPAREFUNCTION_NEVER;
+ case VK_COMPARE_OP_LESS: return GEN6_COMPAREFUNCTION_LESS;
+ case VK_COMPARE_OP_EQUAL: return GEN6_COMPAREFUNCTION_EQUAL;
+ case VK_COMPARE_OP_LESS_EQUAL: return GEN6_COMPAREFUNCTION_LEQUAL;
+ case VK_COMPARE_OP_GREATER: return GEN6_COMPAREFUNCTION_GREATER;
+ case VK_COMPARE_OP_NOT_EQUAL: return GEN6_COMPAREFUNCTION_NOTEQUAL;
+ case VK_COMPARE_OP_GREATER_EQUAL: return GEN6_COMPAREFUNCTION_GEQUAL;
+ case VK_COMPARE_OP_ALWAYS: return GEN6_COMPAREFUNCTION_ALWAYS;
default:
assert(!"unknown compare_func");
return GEN6_COMPAREFUNCTION_NEVER;
@@ -204,7 +204,7 @@
struct intel_pipeline_shader *sh;
VkResult ret;
- sh = intel_alloc(dev, sizeof(*sh), 0, VK_SYSTEM_ALLOC_INTERNAL);
+ sh = intel_alloc(dev, sizeof(*sh), 0, VK_SYSTEM_ALLOC_TYPE_INTERNAL);
if (!sh)
return NULL;
memset(sh, 0, sizeof(*sh));
@@ -316,37 +316,37 @@
pipeline->disable_vs_cache = info->ia.disableVertexReuse;
switch (info->ia.topology) {
- case VK_TOPOLOGY_POINT_LIST:
+ case VK_PRIMITIVE_TOPOLOGY_POINT_LIST:
pipeline->prim_type = GEN6_3DPRIM_POINTLIST;
break;
- case VK_TOPOLOGY_LINE_LIST:
+ case VK_PRIMITIVE_TOPOLOGY_LINE_LIST:
pipeline->prim_type = GEN6_3DPRIM_LINELIST;
break;
- case VK_TOPOLOGY_LINE_STRIP:
+ case VK_PRIMITIVE_TOPOLOGY_LINE_STRIP:
pipeline->prim_type = GEN6_3DPRIM_LINESTRIP;
break;
- case VK_TOPOLOGY_TRIANGLE_LIST:
+ case VK_PRIMITIVE_TOPOLOGY_TRIANGLE_LIST:
pipeline->prim_type = GEN6_3DPRIM_TRILIST;
break;
- case VK_TOPOLOGY_TRIANGLE_STRIP:
+ case VK_PRIMITIVE_TOPOLOGY_TRIANGLE_STRIP:
pipeline->prim_type = GEN6_3DPRIM_TRISTRIP;
break;
- case VK_TOPOLOGY_TRIANGLE_FAN:
+ case VK_PRIMITIVE_TOPOLOGY_TRIANGLE_FAN:
pipeline->prim_type = GEN6_3DPRIM_TRIFAN;
break;
- case VK_TOPOLOGY_LINE_LIST_ADJ:
+ case VK_PRIMITIVE_TOPOLOGY_LINE_LIST_ADJ:
pipeline->prim_type = GEN6_3DPRIM_LINELIST_ADJ;
break;
- case VK_TOPOLOGY_LINE_STRIP_ADJ:
+ case VK_PRIMITIVE_TOPOLOGY_LINE_STRIP_ADJ:
pipeline->prim_type = GEN6_3DPRIM_LINESTRIP_ADJ;
break;
- case VK_TOPOLOGY_TRIANGLE_LIST_ADJ:
+ case VK_PRIMITIVE_TOPOLOGY_TRIANGLE_LIST_ADJ:
pipeline->prim_type = GEN6_3DPRIM_TRILIST_ADJ;
break;
- case VK_TOPOLOGY_TRIANGLE_STRIP_ADJ:
+ case VK_PRIMITIVE_TOPOLOGY_TRIANGLE_STRIP_ADJ:
pipeline->prim_type = GEN6_3DPRIM_TRISTRIP_ADJ;
break;
- case VK_TOPOLOGY_PATCH:
+ case VK_PRIMITIVE_TOPOLOGY_PATCH:
if (!info->tess.patchControlPoints ||
info->tess.patchControlPoints > 32)
return VK_ERROR_BAD_PIPELINE_DATA;
@@ -388,15 +388,15 @@
}
switch (rs_state->fillMode) {
- case VK_FILL_POINTS:
+ case VK_FILL_MODE_POINTS:
pipeline->cmd_sf_fill |= GEN7_SF_DW1_FRONTFACE_POINT |
GEN7_SF_DW1_BACKFACE_POINT;
break;
- case VK_FILL_WIREFRAME:
+ case VK_FILL_MODE_WIREFRAME:
pipeline->cmd_sf_fill |= GEN7_SF_DW1_FRONTFACE_WIREFRAME |
GEN7_SF_DW1_BACKFACE_WIREFRAME;
break;
- case VK_FILL_SOLID:
+ case VK_FILL_MODE_SOLID:
default:
pipeline->cmd_sf_fill |= GEN7_SF_DW1_FRONTFACE_SOLID |
GEN7_SF_DW1_BACKFACE_SOLID;
@@ -414,20 +414,20 @@
}
switch (rs_state->cullMode) {
- case VK_CULL_NONE:
+ case VK_CULL_MODE_NONE:
default:
pipeline->cmd_sf_cull |= GEN7_SF_DW2_CULLMODE_NONE;
pipeline->cmd_clip_cull |= GEN7_CLIP_DW1_CULLMODE_NONE;
break;
- case VK_CULL_FRONT:
+ case VK_CULL_MODE_FRONT:
pipeline->cmd_sf_cull |= GEN7_SF_DW2_CULLMODE_FRONT;
pipeline->cmd_clip_cull |= GEN7_CLIP_DW1_CULLMODE_FRONT;
break;
- case VK_CULL_BACK:
+ case VK_CULL_MODE_BACK:
pipeline->cmd_sf_cull |= GEN7_SF_DW2_CULLMODE_BACK;
pipeline->cmd_clip_cull |= GEN7_CLIP_DW1_CULLMODE_BACK;
break;
- case VK_CULL_FRONT_AND_BACK:
+ case VK_CULL_MODE_FRONT_AND_BACK:
pipeline->cmd_sf_cull |= GEN7_SF_DW2_CULLMODE_BOTH;
pipeline->cmd_clip_cull |= GEN7_CLIP_DW1_CULLMODE_BOTH;
break;
@@ -478,7 +478,7 @@
VkResult ret = VK_SUCCESS;
switch (type) {
- case VK_INFO_TYPE_MEMORY_REQUIREMENTS:
+ case VK_OBJECT_INFO_TYPE_MEMORY_REQUIREMENTS:
{
VkMemoryRequirements *mem_req = data;
@@ -527,16 +527,16 @@
}
/*
- * VK_TOPOLOGY_PATCH primitive topology is only valid for tessellation pipelines.
+ * VK_PRIMITIVE_TOPOLOGY_PATCH primitive topology is only valid for tessellation pipelines.
* Mismatching primitive topology and tessellation fails graphics pipeline creation.
*/
if (pipeline->active_shaders & (SHADER_TESS_CONTROL_FLAG | SHADER_TESS_EVAL_FLAG) &&
- (pipeline->topology != VK_TOPOLOGY_PATCH)) {
+ (pipeline->topology != VK_PRIMITIVE_TOPOLOGY_PATCH)) {
// TODO: Log debug message: Invalid topology used with tessalation shader.
return VK_ERROR_BAD_PIPELINE_DATA;
}
- if ((pipeline->topology == VK_TOPOLOGY_PATCH) &&
+ if ((pipeline->topology == VK_PRIMITIVE_TOPOLOGY_PATCH) &&
(pipeline->active_shaders & ~(SHADER_TESS_CONTROL_FLAG | SHADER_TESS_EVAL_FLAG))) {
// TODO: Log debug message: Cannot use TOPOLOGY_PATCH on non-tessalation shader.
return VK_ERROR_BAD_PIPELINE_DATA;
@@ -949,7 +949,7 @@
body[2 + i] = hi << GEN8_SBE_SWIZ_HIGH__SHIFT | lo;
}
- if (info->ia.topology == VK_TOPOLOGY_POINT_LIST)
+ if (info->ia.topology == VK_PRIMITIVE_TOPOLOGY_POINT_LIST)
body[10] = fs->point_sprite_enables;
else
body[10] = 0;
@@ -1025,12 +1025,12 @@
if (info->db.stencilTestEnable) {
pipeline->cmd_depth_stencil = 1 << 31 |
- translate_compare_func(info->db.front.stencilFunc) << 28 |
+ translate_compare_func(info->db.front.stencilCompareOp) << 28 |
translate_stencil_op(info->db.front.stencilFailOp) << 25 |
translate_stencil_op(info->db.front.stencilDepthFailOp) << 22 |
translate_stencil_op(info->db.front.stencilPassOp) << 19 |
1 << 15 |
- translate_compare_func(info->db.back.stencilFunc) << 12 |
+ translate_compare_func(info->db.back.stencilCompareOp) << 12 |
translate_stencil_op(info->db.back.stencilFailOp) << 9 |
translate_stencil_op(info->db.back.stencilDepthFailOp) << 6 |
translate_stencil_op(info->db.back.stencilPassOp) << 3;
@@ -1053,7 +1053,7 @@
*/
if (info->db.depthTestEnable) {
pipeline->cmd_depth_test = GEN6_ZS_DW2_DEPTH_TEST_ENABLE |
- translate_compare_func(info->db.depthFunc) << 27;
+ translate_compare_func(info->db.depthCompareOp) << 27;
} else {
pipeline->cmd_depth_test = GEN6_COMPAREFUNCTION_ALWAYS << 27;
}
@@ -1105,14 +1105,14 @@
if (att->blendEnable) {
dw0 = 1 << 31 |
- translate_blend_func(att->blendFuncAlpha) << 26 |
+ translate_blend_func(att->blendOpAlpha) << 26 |
translate_blend(att->srcBlendAlpha) << 20 |
translate_blend(att->destBlendAlpha) << 15 |
- translate_blend_func(att->blendFuncColor) << 11 |
+ translate_blend_func(att->blendOpColor) << 11 |
translate_blend(att->srcBlendColor) << 5 |
translate_blend(att->destBlendColor);
- if (att->blendFuncAlpha != att->blendFuncColor ||
+ if (att->blendOpAlpha != att->blendOpColor ||
att->srcBlendAlpha != att->srcBlendColor ||
att->destBlendAlpha != att->destBlendColor)
dw0 |= 1 << 30;
@@ -1357,7 +1357,7 @@
sizeof(*pipeline), dev->base.dbg,
VK_DBG_OBJECT_GRAPHICS_PIPELINE, info_, 0);
if (!pipeline)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
pipeline->dev = dev;
pipeline->obj.base.get_info = pipeline_get_info;
diff --git a/icd/intel/pipeline.h b/icd/intel/pipeline.h
index 604a40a..bae2622 100644
--- a/icd/intel/pipeline.h
+++ b/icd/intel/pipeline.h
@@ -135,7 +135,7 @@
VkFlags barycentric_interps;
VkFlags point_sprite_enables;
- VkGpuSize per_thread_scratch_size;
+ VkDeviceSize per_thread_scratch_size;
enum intel_computed_depth_mode computed_depth_mode;
@@ -143,7 +143,7 @@
/* these are set up by the driver */
uint32_t max_threads;
- VkGpuSize scratch_offset;
+ VkDeviceSize scratch_offset;
};
/*
@@ -209,7 +209,7 @@
struct intel_pipeline_shader gs;
struct intel_pipeline_shader fs;
struct intel_pipeline_shader cs;
- VkGpuSize scratch_size;
+ VkDeviceSize scratch_size;
uint32_t wa_flags;
diff --git a/icd/intel/query.c b/icd/intel/query.c
index 8f2aa52..8b3cb91 100644
--- a/icd/intel/query.c
+++ b/icd/intel/query.c
@@ -43,7 +43,7 @@
VkResult ret = VK_SUCCESS;
switch (type) {
- case VK_INFO_TYPE_MEMORY_REQUIREMENTS:
+ case VK_OBJECT_INFO_TYPE_MEMORY_REQUIREMENTS:
{
VkMemoryRequirements *mem_req = data;
@@ -73,7 +73,7 @@
sizeof(*query), dev->base.dbg, VK_DBG_OBJECT_QUERY_POOL,
info, 0);
if (!query)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
query->type = info->queryType;
query->slot_count = info->slots;
@@ -84,10 +84,10 @@
* compare the differences to get the query results.
*/
switch (info->queryType) {
- case VK_QUERY_OCCLUSION:
+ case VK_QUERY_TYPE_OCCLUSION:
query->slot_stride = u_align(sizeof(uint64_t) * 2, 64);
break;
- case VK_QUERY_PIPELINE_STATISTICS:
+ case VK_QUERY_TYPE_PIPELINE_STATISTICS:
query->slot_stride =
u_align(sizeof(VkPipelineStatisticsData) * 2, 64);
break;
@@ -167,10 +167,10 @@
ptr += query->obj.offset + query->slot_stride * slot_start;
switch (query->type) {
- case VK_QUERY_OCCLUSION:
+ case VK_QUERY_TYPE_OCCLUSION:
query_process_occlusion(query, slot_count, ptr, results);
break;
- case VK_QUERY_PIPELINE_STATISTICS:
+ case VK_QUERY_TYPE_PIPELINE_STATISTICS:
query_process_pipeline_statistics(query, slot_count, ptr, results);
break;
default:
@@ -199,15 +199,16 @@
uint32_t startQuery,
uint32_t queryCount,
size_t* pDataSize,
- void* pData)
+ void* pData,
+ VkQueryResultFlags flags)
{
struct intel_query *query = intel_query(queryPool);
switch (query->type) {
- case VK_QUERY_OCCLUSION:
+ case VK_QUERY_TYPE_OCCLUSION:
*pDataSize = sizeof(uint64_t) * queryCount;
break;
- case VK_QUERY_PIPELINE_STATISTICS:
+ case VK_QUERY_TYPE_PIPELINE_STATISTICS:
*pDataSize = sizeof(VkPipelineStatisticsData) * queryCount;
break;
default:
diff --git a/icd/intel/queue.c b/icd/intel/queue.c
index cb42fef..a5eb625 100644
--- a/icd/intel/queue.c
+++ b/icd/intel/queue.c
@@ -47,7 +47,7 @@
static VkResult queue_submit_bo(struct intel_queue *queue,
struct intel_bo *bo,
- VkGpuSize used)
+ VkDeviceSize used)
{
struct intel_winsys *winsys = queue->dev->winsys;
int err;
@@ -61,7 +61,7 @@
}
static struct intel_bo *queue_create_bo(struct intel_queue *queue,
- VkGpuSize size,
+ VkDeviceSize size,
const void *cmd,
size_t cmd_len)
{
@@ -119,7 +119,7 @@
bo = queue_create_bo(queue, sizeof(pipeline_select_cmd),
pipeline_select_cmd, sizeof(pipeline_select_cmd));
if (!bo)
- return VK_ERROR_OUT_OF_GPU_MEMORY;
+ return VK_ERROR_OUT_OF_DEVICE_MEMORY;
switch (pipeline_select) {
case GEN6_PIPELINE_SELECT_DW0_SELECT_3D:
@@ -163,14 +163,14 @@
queue->atomic_bo = queue_create_bo(queue,
sizeof(uint32_t) * INTEL_QUEUE_ATOMIC_COUNTER_COUNT,
NULL, 0);
- return (queue->atomic_bo) ? VK_SUCCESS : VK_ERROR_OUT_OF_GPU_MEMORY;
+ return (queue->atomic_bo) ? VK_SUCCESS : VK_ERROR_OUT_OF_DEVICE_MEMORY;
}
bo = queue_create_bo(queue,
sizeof(uint32_t) * INTEL_QUEUE_ATOMIC_COUNTER_COUNT,
ctx_init_cmd, sizeof(ctx_init_cmd));
if (!bo)
- return VK_ERROR_OUT_OF_GPU_MEMORY;
+ return VK_ERROR_OUT_OF_DEVICE_MEMORY;
ret = queue_submit_bo(queue, bo, sizeof(ctx_init_cmd));
if (ret != VK_SUCCESS) {
@@ -203,7 +203,7 @@
{
uint32_t active[2], pending[2];
struct intel_bo *bo;
- VkGpuSize used;
+ VkDeviceSize used;
VkResult ret;
ret = queue_submit_cmd_prepare(queue, cmd);
@@ -241,7 +241,7 @@
struct intel_cmd *cmd)
{
struct intel_bo *bo;
- VkGpuSize used;
+ VkDeviceSize used;
VkResult ret;
ret = queue_submit_cmd_prepare(queue, cmd);
@@ -274,7 +274,7 @@
queue = (struct intel_queue *) intel_base_create(&dev->base.handle,
sizeof(*queue), dev->base.dbg, VK_DBG_OBJECT_QUEUE, NULL, 0);
if (!queue)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
queue->dev = dev;
queue->ring = ring;
@@ -321,7 +321,7 @@
ICD_EXPORT VkResult VKAPI vkQueueAddMemReferences(
VkQueue queue,
uint32_t count,
- const VkGpuMemory* pMems)
+ const VkDeviceMemory* pMems)
{
/*
* The winsys maintains the list of memory references. These are ignored
@@ -333,7 +333,7 @@
ICD_EXPORT VkResult VKAPI vkQueueRemoveMemReferences(
VkQueue queue,
uint32_t count,
- const VkGpuMemory* pMems)
+ const VkDeviceMemory* pMems)
{
/*
* The winsys maintains the list of memory references. These are ignored
diff --git a/icd/intel/sampler.c b/icd/intel/sampler.c
index f727a40..75dbd60 100644
--- a/icd/intel/sampler.c
+++ b/icd/intel/sampler.c
@@ -46,9 +46,9 @@
static int translate_tex_mipmap_mode(VkTexMipmapMode mode)
{
switch (mode) {
- case VK_TEX_MIPMAP_NEAREST: return GEN6_MIPFILTER_NEAREST;
- case VK_TEX_MIPMAP_LINEAR: return GEN6_MIPFILTER_LINEAR;
- case VK_TEX_MIPMAP_BASE: return GEN6_MIPFILTER_NONE;
+ case VK_TEX_MIPMAP_MODE_NEAREST: return GEN6_MIPFILTER_NEAREST;
+ case VK_TEX_MIPMAP_MODE_LINEAR: return GEN6_MIPFILTER_LINEAR;
+ case VK_TEX_MIPMAP_MODE_BASE: return GEN6_MIPFILTER_NONE;
default:
assert(!"unknown tex mipmap mode");
return GEN6_MIPFILTER_NONE;
@@ -69,24 +69,24 @@
}
}
-static int translate_compare_func(VkCompareFunc func)
+static int translate_compare_func(VkCompareOp func)
{
switch (func) {
- case VK_COMPARE_NEVER: return GEN6_COMPAREFUNCTION_NEVER;
- case VK_COMPARE_LESS: return GEN6_COMPAREFUNCTION_LESS;
- case VK_COMPARE_EQUAL: return GEN6_COMPAREFUNCTION_EQUAL;
- case VK_COMPARE_LESS_EQUAL: return GEN6_COMPAREFUNCTION_LEQUAL;
- case VK_COMPARE_GREATER: return GEN6_COMPAREFUNCTION_GREATER;
- case VK_COMPARE_NOT_EQUAL: return GEN6_COMPAREFUNCTION_NOTEQUAL;
- case VK_COMPARE_GREATER_EQUAL: return GEN6_COMPAREFUNCTION_GEQUAL;
- case VK_COMPARE_ALWAYS: return GEN6_COMPAREFUNCTION_ALWAYS;
+ case VK_COMPARE_OP_NEVER: return GEN6_COMPAREFUNCTION_NEVER;
+ case VK_COMPARE_OP_LESS: return GEN6_COMPAREFUNCTION_LESS;
+ case VK_COMPARE_OP_EQUAL: return GEN6_COMPAREFUNCTION_EQUAL;
+ case VK_COMPARE_OP_LESS_EQUAL: return GEN6_COMPAREFUNCTION_LEQUAL;
+ case VK_COMPARE_OP_GREATER: return GEN6_COMPAREFUNCTION_GREATER;
+ case VK_COMPARE_OP_NOT_EQUAL: return GEN6_COMPAREFUNCTION_NOTEQUAL;
+ case VK_COMPARE_OP_GREATER_EQUAL: return GEN6_COMPAREFUNCTION_GEQUAL;
+ case VK_COMPARE_OP_ALWAYS: return GEN6_COMPAREFUNCTION_ALWAYS;
default:
assert(!"unknown compare_func");
return GEN6_COMPAREFUNCTION_NEVER;
}
}
-static void translate_border_color(VkBorderColorType type, float rgba[4])
+static void translate_border_color(VkBorderColor type, float rgba[4])
{
switch (type) {
case VK_BORDER_COLOR_OPAQUE_WHITE:
@@ -253,7 +253,7 @@
* To achieve our goal, we just need to set MinLod to zero and set
* MagFilter to MinFilter when mipmapping is disabled.
*/
- if (info->mipMode == VK_TEX_MIPMAP_BASE && min_lod) {
+ if (info->mipMode == VK_TEX_MIPMAP_MODE_BASE && min_lod) {
min_lod = 0;
mag_filter = min_filter;
}
@@ -263,7 +263,7 @@
wrap_t = translate_tex_addr(info->addressV);
wrap_r = translate_tex_addr(info->addressW);
- translate_border_color(info->borderColorType, border_color);
+ translate_border_color(info->borderColor, border_color);
if (intel_gpu_gen(gpu) >= INTEL_GEN(7)) {
dw0 = 1 << 28 |
@@ -282,7 +282,7 @@
dw1 = min_lod << 20 |
max_lod << 8;
- dw1 |= translate_compare_func(info->compareFunc) << 1;
+ dw1 |= translate_compare_func(info->compareOp) << 1;
dw3 = max_aniso << 19;
@@ -313,7 +313,7 @@
mip_filter << 20 |
lod_bias << 3;
- dw0 |= translate_compare_func(info->compareFunc);
+ dw0 |= translate_compare_func(info->compareOp);
if (info->maxAnisotropy > 1) {
dw0 |= GEN6_MAPFILTER_ANISOTROPIC << 17 |
@@ -370,7 +370,7 @@
sampler = (struct intel_sampler *) intel_base_create(&dev->base.handle,
sizeof(*sampler), dev->base.dbg, VK_DBG_OBJECT_SAMPLER, info, 0);
if (!sampler)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
sampler->obj.destroy = sampler_destroy;
diff --git a/icd/intel/shader.c b/icd/intel/shader.c
index ce3f048..7009d41 100644
--- a/icd/intel/shader.c
+++ b/icd/intel/shader.c
@@ -50,7 +50,7 @@
sh = (struct intel_shader *) intel_base_create(&dev->base.handle,
sizeof(*sh), dev->base.dbg, VK_DBG_OBJECT_SHADER, info, 0);
if (!sh)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
if (info->codeSize < sizeof(*spv))
return VK_ERROR_INVALID_MEMORY_SIZE;
diff --git a/icd/intel/state.c b/icd/intel/state.c
index 05f927a..b064e4b 100644
--- a/icd/intel/state.c
+++ b/icd/intel/state.c
@@ -118,9 +118,9 @@
state->cmd_len += 2 * info->viewportAndScissorCount;
state->cmd = intel_alloc(state, sizeof(uint32_t) * state->cmd_len,
- 0, VK_SYSTEM_ALLOC_INTERNAL);
+ 0, VK_SYSTEM_ALLOC_TYPE_INTERNAL);
if (!state->cmd)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
return VK_SUCCESS;
}
@@ -232,7 +232,7 @@
sizeof(*state), dev->base.dbg, VK_DBG_OBJECT_VIEWPORT_STATE,
info, 0);
if (!state)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
state->obj.destroy = viewport_state_destroy;
@@ -270,7 +270,7 @@
sizeof(*state), dev->base.dbg, VK_DBG_OBJECT_RASTER_STATE,
info, 0);
if (!state)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
state->obj.destroy = raster_state_destroy;
state->rs_info = *info;
@@ -302,7 +302,7 @@
sizeof(*state), dev->base.dbg, VK_DBG_OBJECT_COLOR_BLEND_STATE,
info, 0);
if (!state)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
state->obj.destroy = blend_state_destroy;
state->cb_info = *info;
@@ -334,7 +334,7 @@
sizeof(*state), dev->base.dbg, VK_DBG_OBJECT_DEPTH_STENCIL_STATE,
info, 0);
if (!state)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
state->obj.destroy = ds_state_destroy;
diff --git a/icd/intel/view.c b/icd/intel/view.c
index ea6dc14..9a01463 100644
--- a/icd/intel/view.c
+++ b/icd/intel/view.c
@@ -203,20 +203,20 @@
static int img_type_to_view_type(VkImageType type)
{
switch (type) {
- case VK_IMAGE_1D: return VK_IMAGE_VIEW_1D;
- case VK_IMAGE_2D: return VK_IMAGE_VIEW_2D;
- case VK_IMAGE_3D: return VK_IMAGE_VIEW_3D;
- default: assert(!"unknown img type"); return VK_IMAGE_VIEW_1D;
+ case VK_IMAGE_TYPE_1D: return VK_IMAGE_VIEW_TYPE_1D;
+ case VK_IMAGE_TYPE_2D: return VK_IMAGE_VIEW_TYPE_2D;
+ case VK_IMAGE_TYPE_3D: return VK_IMAGE_VIEW_TYPE_3D;
+ default: assert(!"unknown img type"); return VK_IMAGE_VIEW_TYPE_1D;
}
}
static int view_type_to_surface_type(VkImageViewType type)
{
switch (type) {
- case VK_IMAGE_VIEW_1D: return GEN6_SURFTYPE_1D;
- case VK_IMAGE_VIEW_2D: return GEN6_SURFTYPE_2D;
- case VK_IMAGE_VIEW_3D: return GEN6_SURFTYPE_3D;
- case VK_IMAGE_VIEW_CUBE: return GEN6_SURFTYPE_CUBE;
+ case VK_IMAGE_VIEW_TYPE_1D: return GEN6_SURFTYPE_1D;
+ case VK_IMAGE_VIEW_TYPE_2D: return GEN6_SURFTYPE_2D;
+ case VK_IMAGE_VIEW_TYPE_3D: return GEN6_SURFTYPE_3D;
+ case VK_IMAGE_VIEW_TYPE_CUBE: return GEN6_SURFTYPE_CUBE;
default: assert(!"unknown view type"); return GEN6_SURFTYPE_NULL;
}
}
@@ -259,7 +259,7 @@
width = img->layout.width0;
height = img->layout.height0;
- depth = (type == VK_IMAGE_VIEW_3D) ?
+ depth = (type == VK_IMAGE_VIEW_TYPE_3D) ?
img->depth : num_layers;
pitch = img->layout.bo_stride;
@@ -592,7 +592,7 @@
width = img->layout.width0;
height = img->layout.height0;
- depth = (type == VK_IMAGE_VIEW_3D) ?
+ depth = (type == VK_IMAGE_VIEW_TYPE_3D) ?
img->depth : num_layers;
pitch = img->layout.bo_stride;
@@ -817,18 +817,18 @@
* As for GEN7+, separate_stencil is always true.
*/
switch (format) {
- case VK_FMT_D16_UNORM:
+ case VK_FORMAT_D16_UNORM:
info->format = GEN6_ZFORMAT_D16_UNORM;
break;
- case VK_FMT_D32_SFLOAT:
+ case VK_FORMAT_D32_SFLOAT:
info->format = GEN6_ZFORMAT_D32_FLOAT;
break;
- case VK_FMT_D32_SFLOAT_S8_UINT:
+ case VK_FORMAT_D32_SFLOAT_S8_UINT:
info->format = (separate_stencil) ?
GEN6_ZFORMAT_D32_FLOAT :
GEN6_ZFORMAT_D32_FLOAT_S8X24_UINT;
break;
- case VK_FMT_S8_UINT:
+ case VK_FORMAT_S8_UINT:
if (separate_stencil) {
info->format = GEN6_ZFORMAT_D32_FLOAT;
break;
@@ -841,7 +841,7 @@
break;
}
- if (format != VK_FMT_S8_UINT)
+ if (format != VK_FORMAT_S8_UINT)
info->zs.stride = img->layout.bo_stride;
if (img->s8_layout) {
@@ -866,7 +866,7 @@
intel_layout_pos_to_mem(img->s8_layout, x, y, &x, &y);
info->stencil.offset = intel_layout_mem_to_raw(img->s8_layout, x, y);
}
- } else if (format == VK_FMT_S8_UINT) {
+ } else if (format == VK_FORMAT_S8_UINT) {
info->stencil.stride = img->layout.bo_stride * 2;
}
@@ -880,7 +880,7 @@
info->width = img->layout.width0;
info->height = img->layout.height0;
- info->depth = (img->type == VK_IMAGE_3D) ?
+ info->depth = (img->type == VK_IMAGE_TYPE_3D) ?
img->depth : num_layers;
info->lod = level;
@@ -1074,7 +1074,7 @@
const bool will_write = (buf->usage & (VK_BUFFER_USAGE_STORAGE_TEXEL_BUFFER_BIT |
VK_BUFFER_USAGE_STORAGE_BUFFER_BIT));
VkFormat format;
- VkGpuSize stride;
+ VkDeviceSize stride;
uint32_t *cmd;
struct intel_buf_view *view;
int i;
@@ -1083,7 +1083,7 @@
sizeof(*view), dev->base.dbg, VK_DBG_OBJECT_BUFFER_VIEW,
info, 0);
if (!view)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
view->obj.destroy = buf_view_destroy;
@@ -1092,10 +1092,10 @@
/*
* The compiler expects uniform buffers to have pitch of
* 4 for fragment shaders, but 16 for other stages. The format
- * must be VK_FMT_R32G32B32A32_SFLOAT.
+ * must be VK_FORMAT_R32G32B32A32_SFLOAT.
*/
- if (info->viewType == VK_BUFFER_VIEW_RAW) {
- format = VK_FMT_R32G32B32A32_SFLOAT;
+ if (info->viewType == VK_BUFFER_VIEW_TYPE_RAW) {
+ format = VK_FORMAT_R32G32B32A32_SFLOAT;
stride = 16;
} else {
format = info->format;
@@ -1117,7 +1117,7 @@
}
/* switch to view->fs_cmd */
- if (info->viewType == VK_BUFFER_VIEW_RAW) {
+ if (info->viewType == VK_BUFFER_VIEW_TYPE_RAW) {
cmd = view->fs_cmd;
stride = 4;
} else {
@@ -1169,7 +1169,7 @@
view = (struct intel_img_view *) intel_base_create(&dev->base.handle,
sizeof(*view), dev->base.dbg, VK_DBG_OBJECT_IMAGE_VIEW, info, 0);
if (!view)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
view->obj.destroy = img_view_destroy;
@@ -1248,7 +1248,7 @@
sizeof(*view), dev->base.dbg, VK_DBG_OBJECT_COLOR_TARGET_VIEW,
info, 0);
if (!view)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
view->obj.destroy = rt_view_destroy;
@@ -1300,7 +1300,7 @@
sizeof(*view), dev->base.dbg, VK_DBG_OBJECT_DEPTH_STENCIL_VIEW,
info, 0);
if (!view)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
view->obj.destroy = ds_view_destroy;
diff --git a/icd/intel/wsi.h b/icd/intel/wsi.h
index f98cfb9..5115664 100644
--- a/icd/intel/wsi.h
+++ b/icd/intel/wsi.h
@@ -35,7 +35,7 @@
struct intel_img;
VkResult intel_wsi_gpu_get_info(struct intel_gpu *gpu,
- VkPhysicalGpuInfoType type,
+ VkPhysicalDeviceInfoType type,
size_t *size, void *data);
void intel_wsi_gpu_cleanup(struct intel_gpu *gpu);
diff --git a/icd/intel/wsi_x11.c b/icd/intel/wsi_x11.c
index 4436ae7..6443524 100644
--- a/icd/intel/wsi_x11.c
+++ b/icd/intel/wsi_x11.c
@@ -107,9 +107,9 @@
{
/* this is what DDX expects */
switch (format) {
- case VK_FMT_B5G6R5_UNORM:
- case VK_FMT_B8G8R8A8_UNORM:
- case VK_FMT_B8G8R8A8_SRGB:
+ case VK_FORMAT_B5G6R5_UNORM:
+ case VK_FORMAT_B8G8R8A8_UNORM:
+ case VK_FORMAT_B8G8R8A8_SRGB:
return true;
default:
return false;
@@ -361,7 +361,7 @@
/* create image */
memset(&img_info, 0, sizeof(img_info));
img_info.sType = VK_STRUCTURE_TYPE_IMAGE_CREATE_INFO;
- img_info.imageType = VK_IMAGE_2D;
+ img_info.imageType = VK_IMAGE_TYPE_2D;
img_info.format = info->format;
img_info.extent.width = info->extent.width;
img_info.extent.height = info->extent.height;
@@ -369,7 +369,7 @@
img_info.mipLevels = 1;
img_info.arraySize = 1;
img_info.samples = 1;
- img_info.tiling = VK_OPTIMAL_TILING;
+ img_info.tiling = VK_IMAGE_TILING_OPTIMAL;
img_info.usage = info->usage;
img_info.flags = 0;
@@ -549,7 +549,7 @@
return NULL;
}
- x11 = intel_alloc(gpu, sizeof(*x11), 0, VK_SYSTEM_ALLOC_API_OBJECT);
+ x11 = intel_alloc(gpu, sizeof(*x11), 0, VK_SYSTEM_ALLOC_TYPE_API_OBJECT);
if (!x11)
return NULL;
@@ -576,7 +576,7 @@
struct intel_wsi_x11 *x11 = (struct intel_wsi_x11 *) dev->gpu->wsi_data;
struct intel_x11_swap_chain *sc;
- sc = intel_alloc(dev, sizeof(*sc), 0, VK_SYSTEM_ALLOC_API_OBJECT);
+ sc = intel_alloc(dev, sizeof(*sc), 0, VK_SYSTEM_ALLOC_TYPE_API_OBJECT);
if (!sc)
return NULL;
@@ -643,7 +643,7 @@
return;
dpy->modes = intel_alloc(dpy, sizeof(dpy->modes[0]) * conn->count_modes,
- 0, VK_SYSTEM_ALLOC_INTERNAL);
+ 0, VK_SYSTEM_ALLOC_TYPE_INTERNAL);
if (!dpy->modes)
return;
@@ -709,7 +709,7 @@
struct intel_x11_display *dpy;
drmModeConnectorPtr conn;
- dpy = intel_alloc(gpu, sizeof(*dpy), 0, VK_SYSTEM_ALLOC_API_OBJECT);
+ dpy = intel_alloc(gpu, sizeof(*dpy), 0, VK_SYSTEM_ALLOC_TYPE_API_OBJECT);
if (!dpy)
return NULL;
@@ -749,7 +749,7 @@
return;
displays = intel_alloc(gpu, sizeof(*displays) * res->count_connectors,
- 0, VK_SYSTEM_ALLOC_INTERNAL);
+ 0, VK_SYSTEM_ALLOC_TYPE_INTERNAL);
if (!displays) {
drmModeFreeResources(res);
return;
@@ -768,7 +768,7 @@
}
VkResult intel_wsi_gpu_get_info(struct intel_gpu *gpu,
- VkPhysicalGpuInfoType type,
+ VkPhysicalDeviceInfoType type,
size_t *size, void *data)
{
if (false)
@@ -800,9 +800,9 @@
{
struct intel_x11_img_data *data;
- data = intel_alloc(img, sizeof(*data), 0, VK_SYSTEM_ALLOC_INTERNAL);
+ data = intel_alloc(img, sizeof(*data), 0, VK_SYSTEM_ALLOC_TYPE_INTERNAL);
if (!data)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
memset(data, 0, sizeof(*data));
@@ -829,9 +829,9 @@
{
struct intel_x11_fence_data *data;
- data = intel_alloc(fence, sizeof(*data), 0, VK_SYSTEM_ALLOC_INTERNAL);
+ data = intel_alloc(fence, sizeof(*data), 0, VK_SYSTEM_ALLOC_TYPE_INTERNAL);
if (!data)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
memset(data, 0, sizeof(*data));
@@ -866,7 +866,7 @@
}
ICD_EXPORT VkResult VKAPI vkWsiX11AssociateConnection(
- VkPhysicalGpu gpu_,
+ VkPhysicalDevice gpu_,
const VK_WSI_X11_CONNECTION_INFO* pConnectionInfo)
{
struct intel_gpu *gpu = intel_gpu(gpu_);
@@ -904,7 +904,7 @@
VkDevice device,
const VK_WSI_X11_PRESENTABLE_IMAGE_CREATE_INFO* pCreateInfo,
VkImage* pImage,
- VkGpuMemory* pMem)
+ VkDeviceMemory* pMem)
{
struct intel_dev *dev = intel_dev(device);
struct intel_wsi_x11 *x11 = (struct intel_wsi_x11 *) dev->gpu->wsi_data;
@@ -914,7 +914,7 @@
ret = wsi_x11_img_create(x11, dev, pCreateInfo, &img);
if (ret == VK_SUCCESS) {
*pImage = (VkImage) img;
- *pMem = (VkGpuMemory) img->obj.mem;
+ *pMem = (VkDeviceMemory) img->obj.mem;
}
return ret;
diff --git a/icd/nulldrv/nulldrv.c b/icd/nulldrv/nulldrv.c
index f611a75..8c44cc3 100644
--- a/icd/nulldrv/nulldrv.c
+++ b/icd/nulldrv/nulldrv.c
@@ -57,7 +57,7 @@
uint32_t *count;
switch (type) {
- case VK_INFO_TYPE_MEMORY_REQUIREMENTS:
+ case VK_OBJECT_INFO_TYPE_MEMORY_REQUIREMENTS:
{
s = sizeof(VkMemoryRequirements);
*size = s;
@@ -66,7 +66,7 @@
memset(data, 0, s);
break;
}
- case VK_INFO_TYPE_MEMORY_ALLOCATION_COUNT:
+ case VK_OBJECT_INFO_TYPE_MEMORY_ALLOCATION_COUNT:
*size = sizeof(uint32_t);
if (data == NULL)
return ret;
@@ -122,7 +122,7 @@
gpu = malloc(sizeof(*gpu));
if (!gpu)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
memset(gpu, 0, sizeof(*gpu));
// Initialize pointer to loader's dispatch table with ICD_LOADER_MAGIC
@@ -142,7 +142,7 @@
queue = (struct nulldrv_queue *) nulldrv_base_create(dev, sizeof(*queue),
VK_DBG_OBJECT_QUEUE);
if (!queue)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
queue->dev = dev;
@@ -202,7 +202,7 @@
ooxx = malloc(sizeof(*ooxx));
if (!ooxx)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
memset(ooxx, 0, sizeof(*ooxx));
@@ -225,7 +225,7 @@
dev = (struct nulldrv_dev *) nulldrv_base_create(NULL, sizeof(*dev),
VK_DBG_OBJECT_DEVICE);
if (!dev)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
for (i = 0; i < info->extensionCount; i++) {
const enum nulldrv_ext_type ext = nulldrv_gpu_lookup_extension(gpu,
@@ -253,7 +253,7 @@
return VK_SUCCESS;
}
-static struct nulldrv_gpu *nulldrv_gpu(VkPhysicalGpu gpu)
+static struct nulldrv_gpu *nulldrv_gpu(VkPhysicalDevice gpu)
{
return (struct nulldrv_gpu *) gpu;
}
@@ -267,7 +267,7 @@
view = (struct nulldrv_rt_view *) nulldrv_base_create(dev, sizeof(*view),
VK_DBG_OBJECT_COLOR_TARGET_VIEW);
if (!view)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
*view_ret = view;
@@ -283,7 +283,7 @@
fence = (struct nulldrv_fence *) nulldrv_base_create(dev, sizeof(*fence),
VK_DBG_OBJECT_FENCE);
if (!fence)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
*fence_ret = fence;
@@ -308,7 +308,7 @@
VkResult ret = VK_SUCCESS;
switch (type) {
- case VK_INFO_TYPE_MEMORY_REQUIREMENTS:
+ case VK_OBJECT_INFO_TYPE_MEMORY_REQUIREMENTS:
{
VkMemoryRequirements *mem_req = data;
@@ -337,7 +337,7 @@
img = (struct nulldrv_img *) nulldrv_base_create(dev, sizeof(*img),
VK_DBG_OBJECT_IMAGE);
if (!img)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
img->type = info->imageType;
img->depth = info->extent.depth;
@@ -367,11 +367,11 @@
mem = (struct nulldrv_mem *) nulldrv_base_create(dev, sizeof(*mem),
VK_DBG_OBJECT_GPU_MEMORY);
if (!mem)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
mem->bo = malloc(info->allocationSize);
if (!mem->bo) {
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
}
mem->size = info->allocationSize;
@@ -391,7 +391,7 @@
view = (struct nulldrv_ds_view *) nulldrv_base_create(dev, sizeof(*view),
VK_DBG_OBJECT_DEPTH_STENCIL_VIEW);
if (!view)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
view->img = img;
@@ -411,7 +411,7 @@
sampler = (struct nulldrv_sampler *) nulldrv_base_create(dev,
sizeof(*sampler), VK_DBG_OBJECT_SAMPLER);
if (!sampler)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
*sampler_ret = sampler;
@@ -428,7 +428,7 @@
view = (struct nulldrv_img_view *) nulldrv_base_create(dev, sizeof(*view),
VK_DBG_OBJECT_IMAGE_VIEW);
if (!view)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
view->img = img;
view->min_lod = info->minLod;
@@ -445,7 +445,7 @@
return mem->bo;
}
-static struct nulldrv_mem *nulldrv_mem(VkGpuMemory mem)
+static struct nulldrv_mem *nulldrv_mem(VkDeviceMemory mem)
{
return (struct nulldrv_mem *) mem;
}
@@ -462,7 +462,7 @@
VkResult ret = VK_SUCCESS;
switch (type) {
- case VK_INFO_TYPE_MEMORY_REQUIREMENTS:
+ case VK_OBJECT_INFO_TYPE_MEMORY_REQUIREMENTS:
{
VkMemoryRequirements *mem_req = data;
@@ -492,7 +492,7 @@
buf = (struct nulldrv_buf *) nulldrv_base_create(dev, sizeof(*buf),
VK_DBG_OBJECT_BUFFER);
if (!buf)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
buf->size = info->size;
buf->usage = info->usage;
@@ -514,7 +514,7 @@
nulldrv_base_create(dev, sizeof(*layout),
VK_DBG_OBJECT_DESCRIPTOR_SET_LAYOUT);
if (!layout)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
*layout_ret = layout;
@@ -532,7 +532,7 @@
nulldrv_base_create(dev, sizeof(*chain),
VK_DBG_OBJECT_DESCRIPTOR_SET_LAYOUT_CHAIN);
if (!chain)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
*chain_ret = chain;
@@ -553,7 +553,7 @@
sh = (struct nulldrv_shader *) nulldrv_base_create(dev, sizeof(*sh),
VK_DBG_OBJECT_SHADER);
if (!sh)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
*sh_ret = sh;
@@ -570,7 +570,7 @@
nulldrv_base_create(dev, sizeof(*pipeline),
VK_DBG_OBJECT_GRAPHICS_PIPELINE);
if (!pipeline)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
*pipeline_ret = pipeline;
@@ -586,7 +586,7 @@
state = (struct nulldrv_dynamic_vp *) nulldrv_base_create(dev,
sizeof(*state), VK_DBG_OBJECT_VIEWPORT_STATE);
if (!state)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
*state_ret = state;
@@ -602,7 +602,7 @@
state = (struct nulldrv_dynamic_rs *) nulldrv_base_create(dev,
sizeof(*state), VK_DBG_OBJECT_RASTER_STATE);
if (!state)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
*state_ret = state;
@@ -618,7 +618,7 @@
state = (struct nulldrv_dynamic_cb *) nulldrv_base_create(dev,
sizeof(*state), VK_DBG_OBJECT_COLOR_BLEND_STATE);
if (!state)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
*state_ret = state;
@@ -634,7 +634,7 @@
state = (struct nulldrv_dynamic_ds *) nulldrv_base_create(dev,
sizeof(*state), VK_DBG_OBJECT_DEPTH_STENCIL_STATE);
if (!state)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
*state_ret = state;
@@ -651,7 +651,7 @@
cmd = (struct nulldrv_cmd *) nulldrv_base_create(dev, sizeof(*cmd),
VK_DBG_OBJECT_CMD_BUFFER);
if (!cmd)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
*cmd_ret = cmd;
@@ -670,7 +670,7 @@
nulldrv_base_create(dev, sizeof(*pool),
VK_DBG_OBJECT_DESCRIPTOR_POOL);
if (!pool)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
pool->dev = dev;
@@ -691,7 +691,7 @@
nulldrv_base_create(dev, sizeof(*set),
VK_DBG_OBJECT_DESCRIPTOR_SET);
if (!set)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
set->ooxx = dev->desc_ooxx;
set->layout = layout;
@@ -714,7 +714,7 @@
fb = (struct nulldrv_framebuffer *) nulldrv_base_create(dev, sizeof(*fb),
VK_DBG_OBJECT_FRAMEBUFFER);
if (!fb)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
*fb_ret = fb;
@@ -730,7 +730,7 @@
rp = (struct nulldrv_render_pass *) nulldrv_base_create(dev, sizeof(*rp),
VK_DBG_OBJECT_RENDER_PASS);
if (!rp)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
*rp_ret = rp;
@@ -752,7 +752,7 @@
view = (struct nulldrv_buf_view *) nulldrv_base_create(dev, sizeof(*view),
VK_DBG_OBJECT_BUFFER_VIEW);
if (!view)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
view->buf = buf;
@@ -827,7 +827,7 @@
uint32_t startCounter,
uint32_t counterCount,
VkBuffer srcBuffer,
- VkGpuSize srcOffset)
+ VkDeviceSize srcOffset)
{
NULLDRV_LOG_FUNC;
}
@@ -838,7 +838,7 @@
uint32_t startCounter,
uint32_t counterCount,
VkBuffer destBuffer,
- VkGpuSize destOffset)
+ VkDeviceSize destOffset)
{
NULLDRV_LOG_FUNC;
}
@@ -925,8 +925,8 @@
ICD_EXPORT void VKAPI vkCmdUpdateBuffer(
VkCmdBuffer cmdBuffer,
VkBuffer destBuffer,
- VkGpuSize destOffset,
- VkGpuSize dataSize,
+ VkDeviceSize destOffset,
+ VkDeviceSize dataSize,
const uint32_t* pData)
{
NULLDRV_LOG_FUNC;
@@ -935,8 +935,8 @@
ICD_EXPORT void VKAPI vkCmdFillBuffer(
VkCmdBuffer cmdBuffer,
VkBuffer destBuffer,
- VkGpuSize destOffset,
- VkGpuSize fillSize,
+ VkDeviceSize destOffset,
+ VkDeviceSize fillSize,
uint32_t data)
{
NULLDRV_LOG_FUNC;
@@ -1025,8 +1025,8 @@
uint32_t startQuery,
uint32_t queryCount,
VkBuffer destBuffer,
- VkGpuSize destOffset,
- VkGpuSize destStride,
+ VkDeviceSize destOffset,
+ VkDeviceSize destStride,
VkFlags flags)
{
NULLDRV_LOG_FUNC;
@@ -1036,7 +1036,7 @@
VkCmdBuffer cmdBuffer,
VkTimestampType timestampType,
VkBuffer destBuffer,
- VkGpuSize destOffset)
+ VkDeviceSize destOffset)
{
NULLDRV_LOG_FUNC;
}
@@ -1074,7 +1074,7 @@
uint32_t startBinding,
uint32_t bindingCount,
const VkBuffer* pBuffers,
- const VkGpuSize* pOffsets)
+ const VkDeviceSize* pOffsets)
{
NULLDRV_LOG_FUNC;
}
@@ -1082,7 +1082,7 @@
ICD_EXPORT void VKAPI vkCmdBindIndexBuffer(
VkCmdBuffer cmdBuffer,
VkBuffer buffer,
- VkGpuSize offset,
+ VkDeviceSize offset,
VkIndexType indexType)
{
NULLDRV_LOG_FUNC;
@@ -1112,7 +1112,7 @@
ICD_EXPORT void VKAPI vkCmdDrawIndirect(
VkCmdBuffer cmdBuffer,
VkBuffer buffer,
- VkGpuSize offset,
+ VkDeviceSize offset,
uint32_t count,
uint32_t stride)
{
@@ -1122,7 +1122,7 @@
ICD_EXPORT void VKAPI vkCmdDrawIndexedIndirect(
VkCmdBuffer cmdBuffer,
VkBuffer buffer,
- VkGpuSize offset,
+ VkDeviceSize offset,
uint32_t count,
uint32_t stride)
{
@@ -1141,27 +1141,35 @@
ICD_EXPORT void VKAPI vkCmdDispatchIndirect(
VkCmdBuffer cmdBuffer,
VkBuffer buffer,
- VkGpuSize offset)
+ VkDeviceSize offset)
{
NULLDRV_LOG_FUNC;
}
-ICD_EXPORT void VKAPI vkCmdWaitEvents(
- VkCmdBuffer cmdBuffer,
- const VkEventWaitInfo* pWaitInfo)
+void VKAPI vkCmdWaitEvents(
+ VkCmdBuffer cmdBuffer,
+ VkWaitEvent waitEvent,
+ uint32_t eventCount,
+ const VkEvent* pEvents,
+ uint32_t memBarrierCount,
+ const void** ppMemBarriers)
{
NULLDRV_LOG_FUNC;
}
-ICD_EXPORT void VKAPI vkCmdPipelineBarrier(
- VkCmdBuffer cmdBuffer,
- const VkPipelineBarrier* pBarrier)
+void VKAPI vkCmdPipelineBarrier(
+ VkCmdBuffer cmdBuffer,
+ VkWaitEvent waitEvent,
+ uint32_t pipeEventCount,
+ const VkPipeEvent* pPipeEvents,
+ uint32_t memBarrierCount,
+ const void** ppMemBarriers)
{
NULLDRV_LOG_FUNC;
}
ICD_EXPORT VkResult VKAPI vkCreateDevice(
- VkPhysicalGpu gpu_,
+ VkPhysicalDevice gpu_,
const VkDeviceCreateInfo* pCreateInfo,
VkDevice* pDevice)
{
@@ -1303,9 +1311,9 @@
return VK_SUCCESS;
}
-ICD_EXPORT VkResult VKAPI vkGetGpuInfo(
- VkPhysicalGpu gpu_,
- VkPhysicalGpuInfoType infoType,
+ICD_EXPORT VkResult VKAPI vkGetPhysicalDeviceInfo(
+ VkPhysicalDevice gpu_,
+ VkPhysicalDeviceInfoType infoType,
size_t* pDataSize,
void* pData)
{
@@ -1347,7 +1355,7 @@
}
VkResult VKAPI vkGetPhysicalDeviceExtensionInfo(
- VkPhysicalGpu gpu,
+ VkPhysicalDevice gpu,
VkExtensionInfoType infoType,
uint32_t extensionIndex,
size_t* pDataSize,
@@ -1378,10 +1386,10 @@
return VK_SUCCESS;
}
-ICD_EXPORT VkResult VKAPI vkGetMultiGpuCompatibility(
- VkPhysicalGpu gpu0_,
- VkPhysicalGpu gpu1_,
- VkGpuCompatibilityInfo* pInfo)
+ICD_EXPORT VkResult VKAPI vkGetMultiDeviceCompatibility(
+ VkPhysicalDevice gpu0_,
+ VkPhysicalDevice gpu1_,
+ VkPhysicalDeviceCompatibilityInfo* pInfo)
{
NULLDRV_LOG_FUNC;
return VK_SUCCESS;
@@ -1391,7 +1399,7 @@
VkDevice device,
const VkPeerImageOpenInfo* pOpenInfo,
VkImage* pImage,
- VkGpuMemory* pMem)
+ VkDeviceMemory* pMem)
{
NULLDRV_LOG_FUNC;
return VK_SUCCESS;
@@ -1420,7 +1428,7 @@
VkResult ret = VK_SUCCESS;
switch (infoType) {
- case VK_INFO_TYPE_SUBRESOURCE_LAYOUT:
+ case VK_SUBRESOURCE_INFO_TYPE_LAYOUT:
{
VkSubresourceLayout *layout = (VkSubresourceLayout *) pData;
@@ -1445,7 +1453,7 @@
ICD_EXPORT VkResult VKAPI vkAllocMemory(
VkDevice device,
const VkMemoryAllocInfo* pAllocInfo,
- VkGpuMemory* pMem)
+ VkDeviceMemory* pMem)
{
NULLDRV_LOG_FUNC;
struct nulldrv_dev *dev = nulldrv_dev(device);
@@ -1454,14 +1462,14 @@
}
ICD_EXPORT VkResult VKAPI vkFreeMemory(
- VkGpuMemory mem_)
+ VkDeviceMemory mem_)
{
NULLDRV_LOG_FUNC;
return VK_SUCCESS;
}
ICD_EXPORT VkResult VKAPI vkSetMemoryPriority(
- VkGpuMemory mem_,
+ VkDeviceMemory mem_,
VkMemoryPriority priority)
{
NULLDRV_LOG_FUNC;
@@ -1469,7 +1477,7 @@
}
ICD_EXPORT VkResult VKAPI vkMapMemory(
- VkGpuMemory mem_,
+ VkDeviceMemory mem_,
VkFlags flags,
void** ppData)
{
@@ -1483,7 +1491,7 @@
}
ICD_EXPORT VkResult VKAPI vkUnmapMemory(
- VkGpuMemory mem_)
+ VkDeviceMemory mem_)
{
NULLDRV_LOG_FUNC;
return VK_SUCCESS;
@@ -1493,7 +1501,7 @@
VkDevice device,
const void* pSysMem,
size_t memSize,
- VkGpuMemory* pMem)
+ VkDeviceMemory* pMem)
{
NULLDRV_LOG_FUNC;
return VK_SUCCESS;
@@ -1502,7 +1510,7 @@
ICD_EXPORT VkResult VKAPI vkOpenSharedMemory(
VkDevice device,
const VkMemoryOpenInfo* pOpenInfo,
- VkGpuMemory* pMem)
+ VkDeviceMemory* pMem)
{
NULLDRV_LOG_FUNC;
return VK_SUCCESS;
@@ -1511,7 +1519,7 @@
ICD_EXPORT VkResult VKAPI vkOpenPeerMemory(
VkDevice device,
const VkPeerMemoryOpenInfo* pOpenInfo,
- VkGpuMemory* pMem)
+ VkDeviceMemory* pMem)
{
NULLDRV_LOG_FUNC;
return VK_SUCCESS;
@@ -1527,7 +1535,7 @@
inst = (struct nulldrv_instance *) nulldrv_base_create(NULL, sizeof(*inst),
VK_DBG_OBJECT_INSTANCE);
if (!inst)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
inst->obj.base.get_info = NULL;
@@ -1543,11 +1551,10 @@
return VK_SUCCESS;
}
-ICD_EXPORT VkResult VKAPI vkEnumerateGpus(
+ICD_EXPORT VkResult VKAPI vkEnumeratePhysicalDevices(
VkInstance instance,
- uint32_t maxGpus,
uint32_t* pGpuCount,
- VkPhysicalGpu* pGpus)
+ VkPhysicalDevice* pGpus)
{
NULLDRV_LOG_FUNC;
VkResult ret;
@@ -1555,12 +1562,12 @@
*pGpuCount = 1;
ret = nulldrv_gpu_add(0, 0, 0, &gpu);
if (ret == VK_SUCCESS)
- pGpus[0] = (VkPhysicalGpu) gpu;
+ pGpus[0] = (VkPhysicalDevice) gpu;
return ret;
}
ICD_EXPORT VkResult VKAPI vkEnumerateLayers(
- VkPhysicalGpu gpu,
+ VkPhysicalDevice gpu,
size_t maxLayerCount,
size_t maxStringSize,
size_t* pOutLayerCount,
@@ -1621,8 +1628,8 @@
VkQueue queue,
VkObject object,
uint32_t allocationIdx,
- VkGpuMemory mem_,
- VkGpuSize memOffset)
+ VkDeviceMemory mem_,
+ VkDeviceSize memOffset)
{
NULLDRV_LOG_FUNC;
return VK_SUCCESS;
@@ -1632,10 +1639,10 @@
VkQueue queue,
VkObject object,
uint32_t allocationIdx,
- VkGpuSize rangeOffset,
- VkGpuSize rangeSize,
- VkGpuMemory mem,
- VkGpuSize memOffset)
+ VkDeviceSize rangeOffset,
+ VkDeviceSize rangeSize,
+ VkDeviceMemory mem,
+ VkDeviceSize memOffset)
{
NULLDRV_LOG_FUNC;
return VK_SUCCESS;
@@ -1644,10 +1651,10 @@
ICD_EXPORT VkResult VKAPI vkQueueBindImageMemoryRange(
VkQueue queue,
VkImage image,
- uint32_t allocationIdx,
- const VkImageMemoryBindInfo* pBindInfo,
- VkGpuMemory mem,
- VkGpuSize memOffset)
+ uint32_t allocationIdx,
+ const VkImageMemoryBindInfo* pBindInfo,
+ VkDeviceMemory mem,
+ VkDeviceSize memOffset)
{
NULLDRV_LOG_FUNC;
return VK_SUCCESS;
@@ -1740,7 +1747,8 @@
uint32_t startQuery,
uint32_t queryCount,
size_t* pDataSize,
- void* pData)
+ void* pData,
+ VkQueryResultFlags flags)
{
NULLDRV_LOG_FUNC;
return VK_SUCCESS;
@@ -1749,7 +1757,7 @@
ICD_EXPORT VkResult VKAPI vkQueueAddMemReferences(
VkQueue queue,
uint32_t count,
- const VkGpuMemory* pMems)
+ const VkDeviceMemory* pMems)
{
NULLDRV_LOG_FUNC;
return VK_SUCCESS;
@@ -1758,7 +1766,7 @@
ICD_EXPORT VkResult VKAPI vkQueueRemoveMemReferences(
VkQueue queue,
uint32_t count,
- const VkGpuMemory* pMems)
+ const VkDeviceMemory* pMems)
{
NULLDRV_LOG_FUNC;
return VK_SUCCESS;
diff --git a/icd/nulldrv/nulldrv.h b/icd/nulldrv/nulldrv.h
index 4c01f4e..6f9895b 100644
--- a/icd/nulldrv/nulldrv.h
+++ b/icd/nulldrv/nulldrv.h
@@ -112,7 +112,7 @@
struct nulldrv_mem {
struct nulldrv_base base;
struct nulldrv_bo *bo;
- VkGpuSize size;
+ VkDeviceSize size;
};
struct nulldrv_ds_view {
@@ -134,7 +134,7 @@
struct nulldrv_buf {
struct nulldrv_obj obj;
- VkGpuSize size;
+ VkDeviceSize size;
VkFlags usage;
};
diff --git a/include/vkLayer.h b/include/vkLayer.h
index e9e8b03..793ace9 100644
--- a/include/vkLayer.h
+++ b/include/vkLayer.h
@@ -31,7 +31,7 @@
PFN_vkCreateInstance CreateInstance;
PFN_vkDestroyInstance DestroyInstance;
PFN_vkEnumeratePhysicalDevices EnumeratePhysicalDevices;
- PFN_vkGetGpuInfo GetGpuInfo;
+ PFN_vkGetPhysicalDeviceInfo GetPhysicalDeviceInfo;
PFN_vkCreateDevice CreateDevice;
PFN_vkDestroyDevice DestroyDevice;
PFN_vkGetGlobalExtensionInfo GetGlobalExtensionInfo;
@@ -49,7 +49,7 @@
PFN_vkMapMemory MapMemory;
PFN_vkUnmapMemory UnmapMemory;
PFN_vkPinSystemMemory PinSystemMemory;
- PFN_vkGetMultiGpuCompatibility GetMultiGpuCompatibility;
+ PFN_vkGetMultiDeviceCompatibility GetMultiDeviceCompatibility;
PFN_vkOpenSharedMemory OpenSharedMemory;
PFN_vkOpenSharedSemaphore OpenSharedSemaphore;
PFN_vkOpenPeerMemory OpenPeerMemory;
diff --git a/include/vkWsiX11Ext.h b/include/vkWsiX11Ext.h
index f260f9f..3b39c64 100644
--- a/include/vkWsiX11Ext.h
+++ b/include/vkWsiX11Ext.h
@@ -80,9 +80,9 @@
bool32_t flip;
} VK_WSI_X11_PRESENT_INFO;
-typedef VkResult (VKAPI *PFN_vkWsiX11AssociateConnection)(VkPhysicalGpu gpu, const VK_WSI_X11_CONNECTION_INFO* pConnectionInfo);
+typedef VkResult (VKAPI *PFN_vkWsiX11AssociateConnection)(VkPhysicalDevice gpu, const VK_WSI_X11_CONNECTION_INFO* pConnectionInfo);
typedef VkResult (VKAPI *PFN_vkWsiX11GetMSC)(VkDevice device, xcb_window_t window, xcb_randr_crtc_t crtc, uint64_t* pMsc);
-typedef VkResult (VKAPI *PFN_vkWsiX11CreatePresentableImage)(VkDevice device, const VK_WSI_X11_PRESENTABLE_IMAGE_CREATE_INFO* pCreateInfo, VkImage* pImage, VkGpuMemory* pMem);
+typedef VkResult (VKAPI *PFN_vkWsiX11CreatePresentableImage)(VkDevice device, const VK_WSI_X11_PRESENTABLE_IMAGE_CREATE_INFO* pCreateInfo, VkImage* pImage, VkDeviceMemory* pMem);
typedef VkResult (VKAPI *PFN_vkWsiX11QueuePresent)(VkQueue queue, const VK_WSI_X11_PRESENT_INFO* pPresentInfo, VkFence fence);
/**
@@ -98,7 +98,7 @@
* is supported.
*/
VkResult VKAPI vkWsiX11AssociateConnection(
- VkPhysicalGpu gpu,
+ VkPhysicalDevice gpu,
const VK_WSI_X11_CONNECTION_INFO* pConnectionInfo);
/**
@@ -113,7 +113,7 @@
uint64_t* pMsc);
/**
- * Create an VkImage that can be presented. An VkGpuMemory is created
+ * Create an VkImage that can be presented. An VkDeviceMemory is created
* and bound automatically. The memory returned can only be used in
* vkQueue[Add|Remove]MemReference. Destroying the memory or binding another memory to the
* image is not allowed.
@@ -122,7 +122,7 @@
VkDevice device,
const VK_WSI_X11_PRESENTABLE_IMAGE_CREATE_INFO* pCreateInfo,
VkImage* pImage,
- VkGpuMemory* pMem);
+ VkDeviceMemory* pMem);
/**
* Present an image to an X11 window. The presentation always occurs after
diff --git a/include/vk_platform.h b/include/vk_platform.h
index 26962d0..957c135 100644
--- a/include/vk_platform.h
+++ b/include/vk_platform.h
@@ -76,7 +76,7 @@
#endif
#endif // !defined(VK_NO_STDINT_H)
-typedef uint64_t VkGpuSize;
+typedef uint64_t VkDeviceSize;
typedef uint32_t bool32_t;
typedef uint32_t VkSampleMask;
diff --git a/include/vulkan.h b/include/vulkan.h
index 4e91b43..6a30234 100644
--- a/include/vulkan.h
+++ b/include/vulkan.h
@@ -55,11 +55,11 @@
#endif // __cplusplus
VK_DEFINE_HANDLE(VkInstance)
-VK_DEFINE_HANDLE(VkPhysicalGpu)
+VK_DEFINE_HANDLE(VkPhysicalDevice)
VK_DEFINE_HANDLE(VkBaseObject)
VK_DEFINE_SUBCLASS_HANDLE(VkDevice, VkBaseObject)
VK_DEFINE_SUBCLASS_HANDLE(VkQueue, VkBaseObject)
-VK_DEFINE_SUBCLASS_HANDLE(VkGpuMemory, VkBaseObject)
+VK_DEFINE_SUBCLASS_HANDLE(VkDeviceMemory, VkBaseObject)
VK_DEFINE_SUBCLASS_HANDLE(VkObject, VkBaseObject)
VK_DEFINE_SUBCLASS_HANDLE(VkBuffer, VkObject)
VK_DEFINE_SUBCLASS_HANDLE(VkBufferView, VkObject)
@@ -87,12 +87,14 @@
VK_DEFINE_SUBCLASS_HANDLE(VkFramebuffer, VkObject)
VK_DEFINE_SUBCLASS_HANDLE(VkRenderPass, VkObject)
-#define VK_MAX_PHYSICAL_GPU_NAME 256
+#define VK_MAX_PHYSICAL_DEVICE_NAME 256
#define VK_MAX_EXTENSION_NAME 256
#define VK_LOD_CLAMP_NONE MAX_FLOAT
#define VK_LAST_MIP_OR_SLICE 0xffffffff
+#define VK_WHOLE_SIZE UINT64_MAX
+
#define VK_TRUE 1
#define VK_FALSE 0
@@ -102,25 +104,34 @@
// to represent them. This may or may not be necessary on some compilers. The
// option to compile it out may allow compilers that warn about missing enumerants
// in switch statements to be silenced.
-#define VK_MAX_ENUM(T) T##_MAX_ENUM = 0x7FFFFFFF
+// Using this macro is not needed for flag bit enums because those aren't used
+// as storage type anywhere.
+#define VK_MAX_ENUM(Prefix) VK_##Prefix##_MAX_ENUM = 0x7FFFFFFF
+
+// This macro defines the BEGIN_RANGE, END_RANGE, NUM, and MAX_ENUM constants for
+// the enumerations.
+#define VK_ENUM_RANGE(Prefix, First, Last) \
+ VK_##Prefix##_BEGIN_RANGE = VK_##Prefix##_##First, \
+ VK_##Prefix##_END_RANGE = VK_##Prefix##_##Last, \
+ VK_NUM_##Prefix = (VK_##Prefix##_END_RANGE - VK_##Prefix##_BEGIN_RANGE + 1), \
+ VK_MAX_ENUM(Prefix)
+
+// This is a helper macro to define the value of flag bit enum values.
+#define VK_BIT(bit) (1 << (bit))
// ------------------------------------------------------------------------------------------------
// Enumerations
-
typedef enum VkMemoryPriority_
{
- VK_MEMORY_PRIORITY_UNUSED = 0x0,
- VK_MEMORY_PRIORITY_VERY_LOW = 0x1,
- VK_MEMORY_PRIORITY_LOW = 0x2,
- VK_MEMORY_PRIORITY_NORMAL = 0x3,
- VK_MEMORY_PRIORITY_HIGH = 0x4,
- VK_MEMORY_PRIORITY_VERY_HIGH = 0x5,
+ VK_MEMORY_PRIORITY_UNUSED = 0x00000000,
+ VK_MEMORY_PRIORITY_VERY_LOW = 0x00000001,
+ VK_MEMORY_PRIORITY_LOW = 0x00000002,
+ VK_MEMORY_PRIORITY_NORMAL = 0x00000003,
+ VK_MEMORY_PRIORITY_HIGH = 0x00000004,
+ VK_MEMORY_PRIORITY_VERY_HIGH = 0x00000005,
- VK_MEMORY_PRIORITY_BEGIN_RANGE = VK_MEMORY_PRIORITY_UNUSED,
- VK_MEMORY_PRIORITY_END_RANGE = VK_MEMORY_PRIORITY_VERY_HIGH,
- VK_NUM_MEMORY_PRIORITY = (VK_MEMORY_PRIORITY_END_RANGE - VK_MEMORY_PRIORITY_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkMemoryPriority)
+ VK_ENUM_RANGE(MEMORY_PRIORITY, UNUSED, VERY_HIGH)
} VkMemoryPriority;
typedef enum VkImageLayout_
@@ -134,75 +145,39 @@
VK_IMAGE_LAYOUT_CLEAR_OPTIMAL = 0x00000006, // Optimal layout when image is used only for clear operations
VK_IMAGE_LAYOUT_TRANSFER_SOURCE_OPTIMAL = 0x00000007, // Optimal layout when image is used only as source of transfer operations
VK_IMAGE_LAYOUT_TRANSFER_DESTINATION_OPTIMAL = 0x00000008, // Optimal layout when image is used only as destination of transfer operations
-
- VK_IMAGE_LAYOUT_BEGIN_RANGE = VK_IMAGE_LAYOUT_UNDEFINED,
- VK_IMAGE_LAYOUT_END_RANGE = VK_IMAGE_LAYOUT_TRANSFER_DESTINATION_OPTIMAL,
- VK_NUM_IMAGE_LAYOUT = (VK_IMAGE_LAYOUT_END_RANGE - VK_IMAGE_LAYOUT_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkImageLayout)
+
+ VK_ENUM_RANGE(IMAGE_LAYOUT, UNDEFINED, TRANSFER_DESTINATION_OPTIMAL)
} VkImageLayout;
typedef enum VkPipeEvent_
{
- VK_PIPE_EVENT_TOP_OF_PIPE = 0x00000001, // Set event before the GPU starts processing subsequent command
+ VK_PIPE_EVENT_TOP_OF_PIPE = 0x00000001, // Set event before the device starts processing subsequent command
VK_PIPE_EVENT_VERTEX_PROCESSING_COMPLETE = 0x00000002, // Set event when all pending vertex processing is complete
VK_PIPE_EVENT_LOCAL_FRAGMENT_PROCESSING_COMPLETE = 0x00000003, // Set event when all pending fragment shader executions are complete, within each fragment location
VK_PIPE_EVENT_FRAGMENT_PROCESSING_COMPLETE = 0x00000004, // Set event when all pending fragment shader executions are complete
VK_PIPE_EVENT_GRAPHICS_PIPELINE_COMPLETE = 0x00000005, // Set event when all pending graphics operations are complete
VK_PIPE_EVENT_COMPUTE_PIPELINE_COMPLETE = 0x00000006, // Set event when all pending compute operations are complete
VK_PIPE_EVENT_TRANSFER_COMPLETE = 0x00000007, // Set event when all pending transfer operations are complete
- VK_PIPE_EVENT_GPU_COMMANDS_COMPLETE = 0x00000008, // Set event when all pending GPU work is complete
+ VK_PIPE_EVENT_COMMANDS_COMPLETE = 0x00000008, // Set event when all pending work is complete
- VK_PIPE_EVENT_BEGIN_RANGE = VK_PIPE_EVENT_TOP_OF_PIPE,
- VK_PIPE_EVENT_END_RANGE = VK_PIPE_EVENT_GPU_COMMANDS_COMPLETE,
- VK_NUM_PIPE_EVENT = (VK_PIPE_EVENT_END_RANGE - VK_PIPE_EVENT_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkPipeEvent)
+ VK_ENUM_RANGE(PIPE_EVENT, TOP_OF_PIPE, COMMANDS_COMPLETE)
} VkPipeEvent;
typedef enum VkWaitEvent_
{
- VK_WAIT_EVENT_TOP_OF_PIPE = 0x00000001, // Wait event before the GPU starts processing subsequent commands
+ VK_WAIT_EVENT_TOP_OF_PIPE = 0x00000001, // Wait event before the device starts processing subsequent commands
VK_WAIT_EVENT_BEFORE_RASTERIZATION = 0x00000002, // Wait event before rasterizing subsequent primitives
- VK_WAIT_EVENT_BEGIN_RANGE = VK_WAIT_EVENT_TOP_OF_PIPE,
- VK_WAIT_EVENT_END_RANGE = VK_WAIT_EVENT_BEFORE_RASTERIZATION,
- VK_NUM_WAIT_EVENT = (VK_WAIT_EVENT_END_RANGE - VK_WAIT_EVENT_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkWaitEvent)
+ VK_ENUM_RANGE(WAIT_EVENT, TOP_OF_PIPE, BEFORE_RASTERIZATION)
} VkWaitEvent;
-typedef enum VkMemoryOutputFlags_
-{
- VK_MEMORY_OUTPUT_CPU_WRITE_BIT = 0x00000001, // Controls output coherency of CPU writes
- VK_MEMORY_OUTPUT_SHADER_WRITE_BIT = 0x00000002, // Controls output coherency of generic shader writes
- VK_MEMORY_OUTPUT_COLOR_ATTACHMENT_BIT = 0x00000004, // Controls output coherency of color attachment writes
- VK_MEMORY_OUTPUT_DEPTH_STENCIL_ATTACHMENT_BIT = 0x00000008, // Controls output coherency of depth/stencil attachment writes
- VK_MEMORY_OUTPUT_TRANSFER_BIT = 0x00000010, // Controls output coherency of transfer operations
- VK_MAX_ENUM(VkMemoryOutputFlags)
-} VkMemoryOutputFlags;
-
-typedef enum VkMemoryInputFlags_
-{
- VK_MEMORY_INPUT_CPU_READ_BIT = 0x00000001, // Controls input coherency of CPU reads
- VK_MEMORY_INPUT_INDIRECT_COMMAND_BIT = 0x00000002, // Controls input coherency of indirect command reads
- VK_MEMORY_INPUT_INDEX_FETCH_BIT = 0x00000004, // Controls input coherency of index fetches
- VK_MEMORY_INPUT_VERTEX_ATTRIBUTE_FETCH_BIT = 0x00000008, // Controls input coherency of vertex attribute fetches
- VK_MEMORY_INPUT_UNIFORM_READ_BIT = 0x00000010, // Controls input coherency of uniform buffer reads
- VK_MEMORY_INPUT_SHADER_READ_BIT = 0x00000020, // Controls input coherency of generic shader reads
- VK_MEMORY_INPUT_COLOR_ATTACHMENT_BIT = 0x00000040, // Controls input coherency of color attachment reads
- VK_MEMORY_INPUT_DEPTH_STENCIL_ATTACHMENT_BIT = 0x00000080, // Controls input coherency of depth/stencil attachment reads
- VK_MEMORY_INPUT_TRANSFER_BIT = 0x00000100, // Controls input coherency of transfer operations
- VK_MAX_ENUM(VkMemoryInputFlags)
-} VkMemoryInputFlags;
-
typedef enum VkAttachmentLoadOp_
{
VK_ATTACHMENT_LOAD_OP_LOAD = 0x00000000,
VK_ATTACHMENT_LOAD_OP_CLEAR = 0x00000001,
VK_ATTACHMENT_LOAD_OP_DONT_CARE = 0x00000002,
- VK_ATTACHMENT_LOAD_OP_BEGIN_RANGE = VK_ATTACHMENT_LOAD_OP_LOAD,
- VK_ATTACHMENT_LOAD_OP_END_RANGE = VK_ATTACHMENT_LOAD_OP_DONT_CARE,
- VK_NUM_ATTACHMENT_LOAD_OP = (VK_ATTACHMENT_LOAD_OP_END_RANGE - VK_ATTACHMENT_LOAD_OP_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkAttachmentLoadOp)
+ VK_ENUM_RANGE(ATTACHMENT_LOAD_OP, LOAD, DONT_CARE)
} VkAttachmentLoadOp;
typedef enum VkAttachmentStoreOp_
@@ -211,46 +186,34 @@
VK_ATTACHMENT_STORE_OP_RESOLVE_MSAA = 0x00000001,
VK_ATTACHMENT_STORE_OP_DONT_CARE = 0x00000002,
- VK_ATTACHMENT_STORE_OP_BEGIN_RANGE = VK_ATTACHMENT_STORE_OP_STORE,
- VK_ATTACHMENT_STORE_OP_END_RANGE = VK_ATTACHMENT_STORE_OP_DONT_CARE,
- VK_NUM_ATTACHMENT_STORE_OP = (VK_ATTACHMENT_STORE_OP_END_RANGE - VK_ATTACHMENT_STORE_OP_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkAttachmentStoreOp)
+ VK_ENUM_RANGE(ATTACHMENT_STORE_OP, STORE, DONT_CARE)
} VkAttachmentStoreOp;
typedef enum VkImageType_
{
- VK_IMAGE_1D = 0x00000000,
- VK_IMAGE_2D = 0x00000001,
- VK_IMAGE_3D = 0x00000002,
+ VK_IMAGE_TYPE_1D = 0x00000000,
+ VK_IMAGE_TYPE_2D = 0x00000001,
+ VK_IMAGE_TYPE_3D = 0x00000002,
- VK_IMAGE_TYPE_BEGIN_RANGE = VK_IMAGE_1D,
- VK_IMAGE_TYPE_END_RANGE = VK_IMAGE_3D,
- VK_NUM_IMAGE_TYPE = (VK_IMAGE_TYPE_END_RANGE - VK_IMAGE_TYPE_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkImageType)
+ VK_ENUM_RANGE(IMAGE_TYPE, 1D, 3D)
} VkImageType;
typedef enum VkImageTiling_
{
- VK_LINEAR_TILING = 0x00000000,
- VK_OPTIMAL_TILING = 0x00000001,
+ VK_IMAGE_TILING_LINEAR = 0x00000000,
+ VK_IMAGE_TILING_OPTIMAL = 0x00000001,
- VK_IMAGE_TILING_BEGIN_RANGE = VK_LINEAR_TILING,
- VK_IMAGE_TILING_END_RANGE = VK_OPTIMAL_TILING,
- VK_NUM_IMAGE_TILING = (VK_IMAGE_TILING_END_RANGE - VK_IMAGE_TILING_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkImageTiling)
+ VK_ENUM_RANGE(IMAGE_TILING, LINEAR, OPTIMAL)
} VkImageTiling;
typedef enum VkImageViewType_
{
- VK_IMAGE_VIEW_1D = 0x00000000,
- VK_IMAGE_VIEW_2D = 0x00000001,
- VK_IMAGE_VIEW_3D = 0x00000002,
- VK_IMAGE_VIEW_CUBE = 0x00000003,
+ VK_IMAGE_VIEW_TYPE_1D = 0x00000000,
+ VK_IMAGE_VIEW_TYPE_2D = 0x00000001,
+ VK_IMAGE_VIEW_TYPE_3D = 0x00000002,
+ VK_IMAGE_VIEW_TYPE_CUBE = 0x00000003,
- VK_IMAGE_VIEW_TYPE_BEGIN_RANGE = VK_IMAGE_VIEW_1D,
- VK_IMAGE_VIEW_TYPE_END_RANGE = VK_IMAGE_VIEW_CUBE,
- VK_NUM_IMAGE_VIEW_TYPE = (VK_IMAGE_VIEW_TYPE_END_RANGE - VK_IMAGE_VIEW_TYPE_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkImageViewType)
+ VK_ENUM_RANGE(IMAGE_VIEW_TYPE, 1D, CUBE)
} VkImageViewType;
typedef enum VkImageAspect_
@@ -259,12 +222,17 @@
VK_IMAGE_ASPECT_DEPTH = 0x00000001,
VK_IMAGE_ASPECT_STENCIL = 0x00000002,
- VK_IMAGE_ASPECT_BEGIN_RANGE = VK_IMAGE_ASPECT_COLOR,
- VK_IMAGE_ASPECT_END_RANGE = VK_IMAGE_ASPECT_STENCIL,
- VK_NUM_IMAGE_ASPECT = (VK_IMAGE_ASPECT_END_RANGE - VK_IMAGE_ASPECT_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkImageAspect)
+ VK_ENUM_RANGE(IMAGE_ASPECT, COLOR, STENCIL)
} VkImageAspect;
+typedef enum VkBufferViewType_
+{
+ VK_BUFFER_VIEW_TYPE_RAW = 0x00000000, // Raw buffer without special structure (UBO, SSBO)
+ VK_BUFFER_VIEW_TYPE_FORMATTED = 0x00000001, // Buffer with format (TBO, IBO)
+
+ VK_ENUM_RANGE(BUFFER_VIEW_TYPE, RAW, FORMATTED)
+} VkBufferViewType;
+
typedef enum VkChannelSwizzle_
{
VK_CHANNEL_SWIZZLE_ZERO = 0x00000000,
@@ -274,10 +242,7 @@
VK_CHANNEL_SWIZZLE_B = 0x00000004,
VK_CHANNEL_SWIZZLE_A = 0x00000005,
- VK_CHANNEL_SWIZZLE_BEGIN_RANGE = VK_CHANNEL_SWIZZLE_ZERO,
- VK_CHANNEL_SWIZZLE_END_RANGE = VK_CHANNEL_SWIZZLE_A,
- VK_NUM_CHANNEL_SWIZZLE = (VK_CHANNEL_SWIZZLE_END_RANGE - VK_CHANNEL_SWIZZLE_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkChannelSwizzle)
+ VK_ENUM_RANGE(CHANNEL_SWIZZLE, ZERO, A)
} VkChannelSwizzle;
typedef enum VkDescriptorType_
@@ -293,10 +258,7 @@
VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER_DYNAMIC = 0x00000008,
VK_DESCRIPTOR_TYPE_STORAGE_BUFFER_DYNAMIC = 0x00000009,
- VK_DESCRIPTOR_TYPE_BEGIN_RANGE = VK_DESCRIPTOR_TYPE_SAMPLER,
- VK_DESCRIPTOR_TYPE_END_RANGE = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER_DYNAMIC,
- VK_NUM_DESCRIPTOR_TYPE = (VK_DESCRIPTOR_TYPE_END_RANGE - VK_DESCRIPTOR_TYPE_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkDescriptorType)
+ VK_ENUM_RANGE(DESCRIPTOR_TYPE, SAMPLER, STORAGE_BUFFER_DYNAMIC)
} VkDescriptorType;
typedef enum VkDescriptorPoolUsage_
@@ -304,21 +266,15 @@
VK_DESCRIPTOR_POOL_USAGE_ONE_SHOT = 0x00000000,
VK_DESCRIPTOR_POOL_USAGE_DYNAMIC = 0x00000001,
- VK_DESCRIPTOR_POOL_USAGE_BEGIN_RANGE = VK_DESCRIPTOR_POOL_USAGE_ONE_SHOT,
- VK_DESCRIPTOR_POOL_USAGE_END_RANGE = VK_DESCRIPTOR_POOL_USAGE_DYNAMIC,
- VK_NUM_DESCRIPTOR_POOL_USAGE = (VK_DESCRIPTOR_POOL_USAGE_END_RANGE - VK_DESCRIPTOR_POOL_USAGE_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkDescriptorPoolUsage)
+ VK_ENUM_RANGE(DESCRIPTOR_POOL_USAGE, ONE_SHOT, DYNAMIC)
} VkDescriptorPoolUsage;
typedef enum VkDescriptorUpdateMode_
{
- VK_DESCRIPTOR_UDPATE_MODE_COPY = 0x00000000,
+ VK_DESCRIPTOR_UPDATE_MODE_COPY = 0x00000000,
VK_DESCRIPTOR_UPDATE_MODE_FASTEST = 0x00000001,
- VK_DESCRIPTOR_UPDATE_MODE_BEGIN_RANGE = VK_DESCRIPTOR_UDPATE_MODE_COPY,
- VK_DESCRIPTOR_UPDATE_MODE_END_RANGE = VK_DESCRIPTOR_UPDATE_MODE_FASTEST,
- VK_NUM_DESCRIPTOR_UPDATE_MODE = (VK_DESCRIPTOR_UPDATE_MODE_END_RANGE - VK_DESCRIPTOR_UPDATE_MODE_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkDescriptorUpdateMode)
+ VK_ENUM_RANGE(DESCRIPTOR_UPDATE_MODE, COPY, FASTEST)
} VkDescriptorUpdateMode;
typedef enum VkDescriptorSetUsage_
@@ -326,123 +282,93 @@
VK_DESCRIPTOR_SET_USAGE_ONE_SHOT = 0x00000000,
VK_DESCRIPTOR_SET_USAGE_STATIC = 0x00000001,
- VK_DESCRIPTOR_SET_USAGE_BEGIN_RANGE = VK_DESCRIPTOR_SET_USAGE_ONE_SHOT,
- VK_DESCRIPTOR_SET_USAGE_END_RANGE = VK_DESCRIPTOR_SET_USAGE_STATIC,
- VK_NUM_DESCRIPTOR_SET_USAGE = (VK_DESCRIPTOR_SET_USAGE_END_RANGE - VK_DESCRIPTOR_SET_USAGE_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkDescriptorSetUsage)
+ VK_ENUM_RANGE(DESCRIPTOR_SET_USAGE, ONE_SHOT, STATIC)
} VkDescriptorSetUsage;
typedef enum VkQueryType_
{
- VK_QUERY_OCCLUSION = 0x00000000,
- VK_QUERY_PIPELINE_STATISTICS = 0x00000001,
+ VK_QUERY_TYPE_OCCLUSION = 0x00000000,
+ VK_QUERY_TYPE_PIPELINE_STATISTICS = 0x00000001,
- VK_QUERY_TYPE_BEGIN_RANGE = VK_QUERY_OCCLUSION,
- VK_QUERY_TYPE_END_RANGE = VK_QUERY_PIPELINE_STATISTICS,
- VK_NUM_QUERY_TYPE = (VK_QUERY_TYPE_END_RANGE - VK_QUERY_TYPE_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkQueryType)
+ VK_ENUM_RANGE(QUERY_TYPE, OCCLUSION, PIPELINE_STATISTICS)
} VkQueryType;
typedef enum VkTimestampType_
{
- VK_TIMESTAMP_TOP = 0x00000000,
- VK_TIMESTAMP_BOTTOM = 0x00000001,
+ VK_TIMESTAMP_TYPE_TOP = 0x00000000,
+ VK_TIMESTAMP_TYPE_BOTTOM = 0x00000001,
- VK_TIMESTAMP_TYPE_BEGIN_RANGE = VK_TIMESTAMP_TOP,
- VK_TIMESTAMP_TYPE_END_RANGE = VK_TIMESTAMP_BOTTOM,
- VK_NUM_TIMESTAMP_TYPE = (VK_TIMESTAMP_TYPE_END_RANGE - VK_TIMESTAMP_TYPE_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkTimestampType)
+ VK_ENUM_RANGE(TIMESTAMP_TYPE, TOP, BOTTOM)
} VkTimestampType;
-typedef enum VkBorderColorType_
+typedef enum VkBorderColor_
{
VK_BORDER_COLOR_OPAQUE_WHITE = 0x00000000,
VK_BORDER_COLOR_TRANSPARENT_BLACK = 0x00000001,
VK_BORDER_COLOR_OPAQUE_BLACK = 0x00000002,
- VK_BORDER_COLOR_TYPE_BEGIN_RANGE = VK_BORDER_COLOR_OPAQUE_WHITE,
- VK_BORDER_COLOR_TYPE_END_RANGE = VK_BORDER_COLOR_OPAQUE_BLACK,
- VK_NUM_BORDER_COLOR_TYPE = (VK_BORDER_COLOR_TYPE_END_RANGE - VK_BORDER_COLOR_TYPE_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkBorderColorType)
-} VkBorderColorType;
+ VK_ENUM_RANGE(BORDER_COLOR, OPAQUE_WHITE, OPAQUE_BLACK)
+} VkBorderColor;
typedef enum VkPipelineBindPoint_
{
VK_PIPELINE_BIND_POINT_COMPUTE = 0x00000000,
VK_PIPELINE_BIND_POINT_GRAPHICS = 0x00000001,
- VK_PIPELINE_BIND_POINT_BEGIN_RANGE = VK_PIPELINE_BIND_POINT_COMPUTE,
- VK_PIPELINE_BIND_POINT_END_RANGE = VK_PIPELINE_BIND_POINT_GRAPHICS,
- VK_NUM_PIPELINE_BIND_POINT = (VK_PIPELINE_BIND_POINT_END_RANGE - VK_PIPELINE_BIND_POINT_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkPipelineBindPoint)
+ VK_ENUM_RANGE(PIPELINE_BIND_POINT, COMPUTE, GRAPHICS)
} VkPipelineBindPoint;
typedef enum VkStateBindPoint_
{
- VK_STATE_BIND_VIEWPORT = 0x00000000,
- VK_STATE_BIND_RASTER = 0x00000001,
- VK_STATE_BIND_COLOR_BLEND = 0x00000002,
- VK_STATE_BIND_DEPTH_STENCIL = 0x00000003,
+ VK_STATE_BIND_POINT_VIEWPORT = 0x00000000,
+ VK_STATE_BIND_POINT_RASTER = 0x00000001,
+ VK_STATE_BIND_POINT_COLOR_BLEND = 0x00000002,
+ VK_STATE_BIND_POINT_DEPTH_STENCIL = 0x00000003,
- VK_STATE_BIND_POINT_BEGIN_RANGE = VK_STATE_BIND_VIEWPORT,
- VK_STATE_BIND_POINT_END_RANGE = VK_STATE_BIND_DEPTH_STENCIL,
- VK_NUM_STATE_BIND_POINT = (VK_STATE_BIND_POINT_END_RANGE - VK_STATE_BIND_POINT_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkStateBindPoint)
+ VK_ENUM_RANGE(STATE_BIND_POINT, VIEWPORT, DEPTH_STENCIL)
} VkStateBindPoint;
typedef enum VkPrimitiveTopology_
{
- VK_TOPOLOGY_POINT_LIST = 0x00000000,
- VK_TOPOLOGY_LINE_LIST = 0x00000001,
- VK_TOPOLOGY_LINE_STRIP = 0x00000002,
- VK_TOPOLOGY_TRIANGLE_LIST = 0x00000003,
- VK_TOPOLOGY_TRIANGLE_STRIP = 0x00000004,
- VK_TOPOLOGY_TRIANGLE_FAN = 0x00000005,
- VK_TOPOLOGY_LINE_LIST_ADJ = 0x00000006,
- VK_TOPOLOGY_LINE_STRIP_ADJ = 0x00000007,
- VK_TOPOLOGY_TRIANGLE_LIST_ADJ = 0x00000008,
- VK_TOPOLOGY_TRIANGLE_STRIP_ADJ = 0x00000009,
- VK_TOPOLOGY_PATCH = 0x0000000a,
+ VK_PRIMITIVE_TOPOLOGY_POINT_LIST = 0x00000000,
+ VK_PRIMITIVE_TOPOLOGY_LINE_LIST = 0x00000001,
+ VK_PRIMITIVE_TOPOLOGY_LINE_STRIP = 0x00000002,
+ VK_PRIMITIVE_TOPOLOGY_TRIANGLE_LIST = 0x00000003,
+ VK_PRIMITIVE_TOPOLOGY_TRIANGLE_STRIP = 0x00000004,
+ VK_PRIMITIVE_TOPOLOGY_TRIANGLE_FAN = 0x00000005,
+ VK_PRIMITIVE_TOPOLOGY_LINE_LIST_ADJ = 0x00000006,
+ VK_PRIMITIVE_TOPOLOGY_LINE_STRIP_ADJ = 0x00000007,
+ VK_PRIMITIVE_TOPOLOGY_TRIANGLE_LIST_ADJ = 0x00000008,
+ VK_PRIMITIVE_TOPOLOGY_TRIANGLE_STRIP_ADJ = 0x00000009,
+ VK_PRIMITIVE_TOPOLOGY_PATCH = 0x0000000a,
- VK_PRIMITIVE_TOPOLOGY_BEGIN_RANGE = VK_TOPOLOGY_POINT_LIST,
- VK_PRIMITIVE_TOPOLOGY_END_RANGE = VK_TOPOLOGY_PATCH,
- VK_NUM_PRIMITIVE_TOPOLOGY = (VK_PRIMITIVE_TOPOLOGY_END_RANGE - VK_PRIMITIVE_TOPOLOGY_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkPrimitiveTopology)
+ VK_ENUM_RANGE(PRIMITIVE_TOPOLOGY, POINT_LIST, PATCH)
} VkPrimitiveTopology;
typedef enum VkIndexType_
{
- VK_INDEX_8 = 0x00000000,
- VK_INDEX_16 = 0x00000001,
- VK_INDEX_32 = 0x00000002,
+ VK_INDEX_TYPE_UINT8 = 0x00000000,
+ VK_INDEX_TYPE_UINT16 = 0x00000001,
+ VK_INDEX_TYPE_UINT32 = 0x00000002,
- VK_INDEX_TYPE_BEGIN_RANGE = VK_INDEX_8,
- VK_INDEX_TYPE_END_RANGE = VK_INDEX_32,
- VK_NUM_INDEX_TYPE = (VK_INDEX_TYPE_END_RANGE - VK_INDEX_TYPE_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkIndexType)
+ VK_ENUM_RANGE(INDEX_TYPE, UINT8, UINT32)
} VkIndexType;
typedef enum VkTexFilter_
{
- VK_TEX_FILTER_NEAREST = 0,
- VK_TEX_FILTER_LINEAR = 1,
+ VK_TEX_FILTER_NEAREST = 0x00000000,
+ VK_TEX_FILTER_LINEAR = 0x00000001,
- VK_TEX_FILTER_BEGIN_RANGE = VK_TEX_FILTER_NEAREST,
- VK_TEX_FILTER_END_RANGE = VK_TEX_FILTER_LINEAR,
- VK_NUM_TEX_FILTER = (VK_TEX_FILTER_END_RANGE - VK_TEX_FILTER_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkTexFilter)
+ VK_ENUM_RANGE(TEX_FILTER, NEAREST, LINEAR)
} VkTexFilter;
typedef enum VkTexMipmapMode_
{
- VK_TEX_MIPMAP_BASE = 0, // Always choose base level
- VK_TEX_MIPMAP_NEAREST = 1, // Choose nearest mip level
- VK_TEX_MIPMAP_LINEAR = 2, // Linear filter between mip levels
+ VK_TEX_MIPMAP_MODE_BASE = 0x00000000, // Always choose base level
+ VK_TEX_MIPMAP_MODE_NEAREST = 0x00000001, // Choose nearest mip level
+ VK_TEX_MIPMAP_MODE_LINEAR = 0x00000002, // Linear filter between mip levels
- VK_TEX_MIPMAP_BEGIN_RANGE = VK_TEX_MIPMAP_BASE,
- VK_TEX_MIPMAP_END_RANGE = VK_TEX_MIPMAP_LINEAR,
- VK_NUM_TEX_MIPMAP = (VK_TEX_MIPMAP_END_RANGE - VK_TEX_MIPMAP_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkTexMipmapMode)
+ VK_ENUM_RANGE(TEX_MIPMAP_MODE, BASE, LINEAR)
} VkTexMipmapMode;
typedef enum VkTexAddress_
@@ -453,85 +379,64 @@
VK_TEX_ADDRESS_MIRROR_ONCE = 0x00000003,
VK_TEX_ADDRESS_CLAMP_BORDER = 0x00000004,
- VK_TEX_ADDRESS_BEGIN_RANGE = VK_TEX_ADDRESS_WRAP,
- VK_TEX_ADDRESS_END_RANGE = VK_TEX_ADDRESS_CLAMP_BORDER,
- VK_NUM_TEX_ADDRESS = (VK_TEX_ADDRESS_END_RANGE - VK_TEX_ADDRESS_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkTexAddress)
+ VK_ENUM_RANGE(TEX_ADDRESS, WRAP, CLAMP_BORDER)
} VkTexAddress;
-typedef enum VkCompareFunc_
+typedef enum VkCompareOp_
{
- VK_COMPARE_NEVER = 0x00000000,
- VK_COMPARE_LESS = 0x00000001,
- VK_COMPARE_EQUAL = 0x00000002,
- VK_COMPARE_LESS_EQUAL = 0x00000003,
- VK_COMPARE_GREATER = 0x00000004,
- VK_COMPARE_NOT_EQUAL = 0x00000005,
- VK_COMPARE_GREATER_EQUAL = 0x00000006,
- VK_COMPARE_ALWAYS = 0x00000007,
+ VK_COMPARE_OP_NEVER = 0x00000000,
+ VK_COMPARE_OP_LESS = 0x00000001,
+ VK_COMPARE_OP_EQUAL = 0x00000002,
+ VK_COMPARE_OP_LESS_EQUAL = 0x00000003,
+ VK_COMPARE_OP_GREATER = 0x00000004,
+ VK_COMPARE_OP_NOT_EQUAL = 0x00000005,
+ VK_COMPARE_OP_GREATER_EQUAL = 0x00000006,
+ VK_COMPARE_OP_ALWAYS = 0x00000007,
- VK_COMPARE_FUNC_BEGIN_RANGE = VK_COMPARE_NEVER,
- VK_COMPARE_FUNC_END_RANGE = VK_COMPARE_ALWAYS,
- VK_NUM_COMPARE_FUNC = (VK_COMPARE_FUNC_END_RANGE - VK_COMPARE_FUNC_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkCompareFunc)
-} VkCompareFunc;
+ VK_ENUM_RANGE(COMPARE_OP, NEVER, ALWAYS)
+} VkCompareOp;
typedef enum VkFillMode_
{
- VK_FILL_POINTS = 0x00000000,
- VK_FILL_WIREFRAME = 0x00000001,
- VK_FILL_SOLID = 0x00000002,
+ VK_FILL_MODE_POINTS = 0x00000000,
+ VK_FILL_MODE_WIREFRAME = 0x00000001,
+ VK_FILL_MODE_SOLID = 0x00000002,
- VK_FILL_MODE_BEGIN_RANGE = VK_FILL_POINTS,
- VK_FILL_MODE_END_RANGE = VK_FILL_SOLID,
- VK_NUM_FILL_MODE = (VK_FILL_MODE_END_RANGE - VK_FILL_MODE_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkFillMode)
+ VK_ENUM_RANGE(FILL_MODE, POINTS, SOLID)
} VkFillMode;
typedef enum VkCullMode_
{
- VK_CULL_NONE = 0x00000000,
- VK_CULL_FRONT = 0x00000001,
- VK_CULL_BACK = 0x00000002,
- VK_CULL_FRONT_AND_BACK = 0x00000003,
+ VK_CULL_MODE_NONE = 0x00000000,
+ VK_CULL_MODE_FRONT = 0x00000001,
+ VK_CULL_MODE_BACK = 0x00000002,
+ VK_CULL_MODE_FRONT_AND_BACK = 0x00000003,
- VK_CULL_MODE_BEGIN_RANGE = VK_CULL_NONE,
- VK_CULL_MODE_END_RANGE = VK_CULL_FRONT_AND_BACK,
- VK_NUM_CULL_MODE = (VK_CULL_MODE_END_RANGE - VK_CULL_MODE_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkCullMode)
+ VK_ENUM_RANGE(CULL_MODE, NONE, FRONT_AND_BACK)
} VkCullMode;
-typedef enum VkFaceOrientation_
+typedef enum VkFrontFace_
{
VK_FRONT_FACE_CCW = 0x00000000,
VK_FRONT_FACE_CW = 0x00000001,
- VK_FACE_ORIENTATION_BEGIN_RANGE = VK_FRONT_FACE_CCW,
- VK_FACE_ORIENTATION_END_RANGE = VK_FRONT_FACE_CW,
- VK_NUM_FACE_ORIENTATION = (VK_FACE_ORIENTATION_END_RANGE - VK_FACE_ORIENTATION_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkFaceOrientation)
-} VkFaceOrientation;
+ VK_ENUM_RANGE(FRONT_FACE, CCW, CW)
+} VkFrontFace;
-typedef enum VkProvokingVertexConvention_
+typedef enum VkProvokingVertex_
{
VK_PROVOKING_VERTEX_FIRST = 0x00000000,
VK_PROVOKING_VERTEX_LAST = 0x00000001,
- VK_PROVOKING_VERTEX_BEGIN_RANGE = VK_PROVOKING_VERTEX_FIRST,
- VK_PROVOKING_VERTEX_END_RANGE = VK_PROVOKING_VERTEX_LAST,
- VK_NUM_PROVOKING_VERTEX_CONVENTION = (VK_PROVOKING_VERTEX_END_RANGE - VK_PROVOKING_VERTEX_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkProvokingVertexConvention)
-} VkProvokingVertexConvention;
+ VK_ENUM_RANGE(PROVOKING_VERTEX, FIRST, LAST)
+} VkProvokingVertex;
typedef enum VkCoordinateOrigin_
{
VK_COORDINATE_ORIGIN_UPPER_LEFT = 0x00000000,
VK_COORDINATE_ORIGIN_LOWER_LEFT = 0x00000001,
- VK_COORDINATE_ORIGIN_BEGIN_RANGE = VK_COORDINATE_ORIGIN_UPPER_LEFT,
- VK_COORDINATE_ORIGIN_END_RANGE = VK_COORDINATE_ORIGIN_LOWER_LEFT,
- VK_NUM_COORDINATE_ORIGIN = (VK_COORDINATE_ORIGIN_END_RANGE - VK_COORDINATE_ORIGIN_END_RANGE + 1),
- VK_MAX_ENUM(VkCoordinateOrigin)
+ VK_ENUM_RANGE(COORDINATE_ORIGIN, UPPER_LEFT, LOWER_LEFT)
} VkCoordinateOrigin;
typedef enum VkDepthMode_
@@ -539,10 +444,7 @@
VK_DEPTH_MODE_ZERO_TO_ONE = 0x00000000,
VK_DEPTH_MODE_NEGATIVE_ONE_TO_ONE = 0x00000001,
- VK_DEPTH_MODE_BEGIN_RANGE = VK_DEPTH_MODE_ZERO_TO_ONE,
- VK_DEPTH_MODE_END_RANGE = VK_DEPTH_MODE_NEGATIVE_ONE_TO_ONE,
- VK_NUM_DEPTH_MODE = (VK_DEPTH_MODE_END_RANGE - VK_DEPTH_MODE_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkDepthMode)
+ VK_ENUM_RANGE(DEPTH_MODE, ZERO_TO_ONE, NEGATIVE_ONE_TO_ONE)
} VkDepthMode;
typedef enum VkBlend_
@@ -567,25 +469,19 @@
VK_BLEND_SRC1_ALPHA = 0x00000011,
VK_BLEND_ONE_MINUS_SRC1_ALPHA = 0x00000012,
- VK_BLEND_BEGIN_RANGE = VK_BLEND_ZERO,
- VK_BLEND_END_RANGE = VK_BLEND_ONE_MINUS_SRC1_ALPHA,
- VK_NUM_BLEND = (VK_BLEND_END_RANGE - VK_BLEND_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkBlend)
+ VK_ENUM_RANGE(BLEND, ZERO, ONE_MINUS_SRC1_ALPHA)
} VkBlend;
-typedef enum VkBlendFunc_
+typedef enum VkBlendOp_
{
- VK_BLEND_FUNC_ADD = 0x00000000,
- VK_BLEND_FUNC_SUBTRACT = 0x00000001,
- VK_BLEND_FUNC_REVERSE_SUBTRACT = 0x00000002,
- VK_BLEND_FUNC_MIN = 0x00000003,
- VK_BLEND_FUNC_MAX = 0x00000004,
+ VK_BLEND_OP_ADD = 0x00000000,
+ VK_BLEND_OP_SUBTRACT = 0x00000001,
+ VK_BLEND_OP_REVERSE_SUBTRACT = 0x00000002,
+ VK_BLEND_OP_MIN = 0x00000003,
+ VK_BLEND_OP_MAX = 0x00000004,
- VK_BLEND_FUNC_BEGIN_RANGE = VK_BLEND_FUNC_ADD,
- VK_BLEND_FUNC_END_RANGE = VK_BLEND_FUNC_MAX,
- VK_NUM_BLEND_FUNC = (VK_BLEND_FUNC_END_RANGE - VK_BLEND_FUNC_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkBlendFunc)
-} VkBlendFunc;
+ VK_ENUM_RANGE(BLEND_OP, ADD, MAX)
+} VkBlendOp;
typedef enum VkStencilOp_
{
@@ -598,10 +494,7 @@
VK_STENCIL_OP_INC_WRAP = 0x00000006,
VK_STENCIL_OP_DEC_WRAP = 0x00000007,
- VK_STENCIL_OP_BEGIN_RANGE = VK_STENCIL_OP_KEEP,
- VK_STENCIL_OP_END_RANGE = VK_STENCIL_OP_DEC_WRAP,
- VK_NUM_STENCIL_OP = (VK_STENCIL_OP_END_RANGE - VK_STENCIL_OP_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkStencilOp)
+ VK_ENUM_RANGE(STENCIL_OP, KEEP, DEC_WRAP)
} VkStencilOp;
typedef enum VkLogicOp_
@@ -623,52 +516,41 @@
VK_LOGIC_OP_NAND = 0x0000000e,
VK_LOGIC_OP_SET = 0x0000000f,
- VK_LOGIC_OP_BEGIN_RANGE = VK_LOGIC_OP_COPY,
- VK_LOGIC_OP_END_RANGE = VK_LOGIC_OP_SET,
- VK_NUM_LOGIC_OP = (VK_LOGIC_OP_END_RANGE - VK_LOGIC_OP_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkLogicOp)
+ VK_ENUM_RANGE(LOGIC_OP, COPY, SET)
} VkLogicOp;
typedef enum VkSystemAllocType_
{
- VK_SYSTEM_ALLOC_API_OBJECT = 0x00000000,
- VK_SYSTEM_ALLOC_INTERNAL = 0x00000001,
- VK_SYSTEM_ALLOC_INTERNAL_TEMP = 0x00000002,
- VK_SYSTEM_ALLOC_INTERNAL_SHADER = 0x00000003,
- VK_SYSTEM_ALLOC_DEBUG = 0x00000004,
+ VK_SYSTEM_ALLOC_TYPE_API_OBJECT = 0x00000000,
+ VK_SYSTEM_ALLOC_TYPE_INTERNAL = 0x00000001,
+ VK_SYSTEM_ALLOC_TYPE_INTERNAL_TEMP = 0x00000002,
+ VK_SYSTEM_ALLOC_TYPE_INTERNAL_SHADER = 0x00000003,
+ VK_SYSTEM_ALLOC_TYPE_DEBUG = 0x00000004,
- VK_SYSTEM_ALLOC_BEGIN_RANGE = VK_SYSTEM_ALLOC_API_OBJECT,
- VK_SYSTEM_ALLOC_END_RANGE = VK_SYSTEM_ALLOC_DEBUG,
- VK_NUM_SYSTEM_ALLOC_TYPE = (VK_SYSTEM_ALLOC_END_RANGE - VK_SYSTEM_ALLOC_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkSystemAllocType)
+ VK_ENUM_RANGE(SYSTEM_ALLOC_TYPE, API_OBJECT, DEBUG)
} VkSystemAllocType;
-typedef enum VkPhysicalGpuType_
+typedef enum VkPhysicalDeviceType_
{
- VK_GPU_TYPE_OTHER = 0x00000000,
- VK_GPU_TYPE_INTEGRATED = 0x00000001,
- VK_GPU_TYPE_DISCRETE = 0x00000002,
- VK_GPU_TYPE_VIRTUAL = 0x00000003,
+ VK_PHYSICAL_DEVICE_TYPE_OTHER = 0x00000000,
+ VK_PHYSICAL_DEVICE_TYPE_INTEGRATED_GPU = 0x00000001,
+ VK_PHYSICAL_DEVICE_TYPE_DISCRETE_GPU = 0x00000002,
+ VK_PHYSICAL_DEVICE_TYPE_VIRTUAL_GPU = 0x00000003,
+ VK_PHYSICAL_DEVICE_TYPE_CPU = 0x00000004,
- VK_PHYSICAL_GPU_TYPE_BEGIN_RANGE = VK_GPU_TYPE_OTHER,
- VK_PHYSICAL_GPU_TYPE_END_RANGE = VK_GPU_TYPE_VIRTUAL,
- VK_NUM_PHYSICAL_GPU_TYPE = (VK_PHYSICAL_GPU_TYPE_END_RANGE - VK_PHYSICAL_GPU_TYPE_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkPhysicalGpuType)
-} VkPhysicalGpuType;
+ VK_ENUM_RANGE(PHYSICAL_DEVICE_TYPE, OTHER, CPU)
+} VkPhysicalDeviceType;
-typedef enum VkPhysicalGpuInfoType_
+typedef enum VkPhysicalDeviceInfoType_
{
- // Info type for vkGetGpuInfo()
- VK_INFO_TYPE_PHYSICAL_GPU_PROPERTIES = 0x00000000,
- VK_INFO_TYPE_PHYSICAL_GPU_PERFORMANCE = 0x00000001,
- VK_INFO_TYPE_PHYSICAL_GPU_QUEUE_PROPERTIES = 0x00000002,
- VK_INFO_TYPE_PHYSICAL_GPU_MEMORY_PROPERTIES = 0x00000003,
+ // Info type for vkGetPhysicalDeviceInfo()
+ VK_PHYSICAL_DEVICE_INFO_TYPE_PROPERTIES = 0x00000000,
+ VK_PHYSICAL_DEVICE_INFO_TYPE_PERFORMANCE = 0x00000001,
+ VK_PHYSICAL_DEVICE_INFO_TYPE_QUEUE_PROPERTIES = 0x00000002,
+ VK_PHYSICAL_DEVICE_INFO_TYPE_MEMORY_PROPERTIES = 0x00000003,
- VK_INFO_TYPE_PHYSICAL_GPU_BEGIN_RANGE = VK_INFO_TYPE_PHYSICAL_GPU_PROPERTIES,
- VK_INFO_TYPE_PHYSICAL_GPU_END_RANGE = VK_INFO_TYPE_PHYSICAL_GPU_MEMORY_PROPERTIES,
- VK_NUM_INFO_TYPE_PHYSICAL_GPU = (VK_INFO_TYPE_PHYSICAL_GPU_END_RANGE - VK_INFO_TYPE_PHYSICAL_GPU_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkPhysicalGpuInfoType)
-} VkPhysicalGpuInfoType;
+ VK_ENUM_RANGE(PHYSICAL_DEVICE_INFO_TYPE, PROPERTIES, MEMORY_PROPERTIES)
+} VkPhysicalDeviceInfoType;
typedef enum VkExtensionInfoType_
{
@@ -682,37 +564,290 @@
typedef enum VkFormatInfoType_
{
// Info type for vkGetFormatInfo()
- VK_INFO_TYPE_FORMAT_PROPERTIES = 0x00000000,
+ VK_FORMAT_INFO_TYPE_PROPERTIES = 0x00000000,
- VK_INFO_TYPE_FORMAT_BEGIN_RANGE = VK_INFO_TYPE_FORMAT_PROPERTIES,
- VK_INFO_TYPE_FORMAT_END_RANGE = VK_INFO_TYPE_FORMAT_PROPERTIES,
- VK_NUM_INFO_TYPE_FORMAT = (VK_INFO_TYPE_FORMAT_END_RANGE - VK_INFO_TYPE_FORMAT_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkFormatInfoType)
+ VK_ENUM_RANGE(FORMAT_INFO_TYPE, PROPERTIES, PROPERTIES)
} VkFormatInfoType;
typedef enum VkSubresourceInfoType_
{
// Info type for vkGetImageSubresourceInfo()
- VK_INFO_TYPE_SUBRESOURCE_LAYOUT = 0x00000000,
+ VK_SUBRESOURCE_INFO_TYPE_LAYOUT = 0x00000000,
- VK_INFO_TYPE_SUBRESOURCE_BEGIN_RANGE = VK_INFO_TYPE_SUBRESOURCE_LAYOUT,
- VK_INFO_TYPE_SUBRESOURCE_END_RANGE = VK_INFO_TYPE_SUBRESOURCE_LAYOUT,
- VK_NUM_INFO_TYPE_SUBRESOURCE = (VK_INFO_TYPE_SUBRESOURCE_END_RANGE - VK_INFO_TYPE_SUBRESOURCE_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkSubresourceInfoType)
+ VK_ENUM_RANGE(SUBRESOURCE_INFO_TYPE, LAYOUT, LAYOUT)
} VkSubresourceInfoType;
typedef enum VkObjectInfoType_
{
// Info type for vkGetObjectInfo()
- VK_INFO_TYPE_MEMORY_ALLOCATION_COUNT = 0x00000000,
- VK_INFO_TYPE_MEMORY_REQUIREMENTS = 0x00000001,
+ VK_OBJECT_INFO_TYPE_MEMORY_ALLOCATION_COUNT = 0x00000000,
+ VK_OBJECT_INFO_TYPE_MEMORY_REQUIREMENTS = 0x00000001,
- VK_INFO_TYPE_BEGIN_RANGE = VK_INFO_TYPE_MEMORY_ALLOCATION_COUNT,
- VK_INFO_TYPE_END_RANGE = VK_INFO_TYPE_MEMORY_REQUIREMENTS,
- VK_NUM_INFO_TYPE = (VK_INFO_TYPE_END_RANGE - VK_INFO_TYPE_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkObjectInfoType)
+ VK_ENUM_RANGE(OBJECT_INFO_TYPE, MEMORY_ALLOCATION_COUNT, MEMORY_REQUIREMENTS)
} VkObjectInfoType;
+typedef enum VkVertexInputStepRate_
+{
+ VK_VERTEX_INPUT_STEP_RATE_VERTEX = 0x0,
+ VK_VERTEX_INPUT_STEP_RATE_INSTANCE = 0x1,
+ VK_VERTEX_INPUT_STEP_RATE_DRAW = 0x2, //Optional
+
+ VK_ENUM_RANGE(VERTEX_INPUT_STEP_RATE, VERTEX, DRAW)
+} VkVertexInputStepRate;
+
+// ------------------------------------------------------------------------------------------------
+// Vulkan format definitions
+typedef enum VkFormat_
+{
+ VK_FORMAT_UNDEFINED = 0x00000000,
+ VK_FORMAT_R4G4_UNORM = 0x00000001,
+ VK_FORMAT_R4G4_USCALED = 0x00000002,
+ VK_FORMAT_R4G4B4A4_UNORM = 0x00000003,
+ VK_FORMAT_R4G4B4A4_USCALED = 0x00000004,
+ VK_FORMAT_R5G6B5_UNORM = 0x00000005,
+ VK_FORMAT_R5G6B5_USCALED = 0x00000006,
+ VK_FORMAT_R5G5B5A1_UNORM = 0x00000007,
+ VK_FORMAT_R5G5B5A1_USCALED = 0x00000008,
+ VK_FORMAT_R8_UNORM = 0x00000009,
+ VK_FORMAT_R8_SNORM = 0x0000000A,
+ VK_FORMAT_R8_USCALED = 0x0000000B,
+ VK_FORMAT_R8_SSCALED = 0x0000000C,
+ VK_FORMAT_R8_UINT = 0x0000000D,
+ VK_FORMAT_R8_SINT = 0x0000000E,
+ VK_FORMAT_R8_SRGB = 0x0000000F,
+ VK_FORMAT_R8G8_UNORM = 0x00000010,
+ VK_FORMAT_R8G8_SNORM = 0x00000011,
+ VK_FORMAT_R8G8_USCALED = 0x00000012,
+ VK_FORMAT_R8G8_SSCALED = 0x00000013,
+ VK_FORMAT_R8G8_UINT = 0x00000014,
+ VK_FORMAT_R8G8_SINT = 0x00000015,
+ VK_FORMAT_R8G8_SRGB = 0x00000016,
+ VK_FORMAT_R8G8B8_UNORM = 0x00000017,
+ VK_FORMAT_R8G8B8_SNORM = 0x00000018,
+ VK_FORMAT_R8G8B8_USCALED = 0x00000019,
+ VK_FORMAT_R8G8B8_SSCALED = 0x0000001A,
+ VK_FORMAT_R8G8B8_UINT = 0x0000001B,
+ VK_FORMAT_R8G8B8_SINT = 0x0000001C,
+ VK_FORMAT_R8G8B8_SRGB = 0x0000001D,
+ VK_FORMAT_R8G8B8A8_UNORM = 0x0000001E,
+ VK_FORMAT_R8G8B8A8_SNORM = 0x0000001F,
+ VK_FORMAT_R8G8B8A8_USCALED = 0x00000020,
+ VK_FORMAT_R8G8B8A8_SSCALED = 0x00000021,
+ VK_FORMAT_R8G8B8A8_UINT = 0x00000022,
+ VK_FORMAT_R8G8B8A8_SINT = 0x00000023,
+ VK_FORMAT_R8G8B8A8_SRGB = 0x00000024,
+ VK_FORMAT_R10G10B10A2_UNORM = 0x00000025,
+ VK_FORMAT_R10G10B10A2_SNORM = 0x00000026,
+ VK_FORMAT_R10G10B10A2_USCALED = 0x00000027,
+ VK_FORMAT_R10G10B10A2_SSCALED = 0x00000028,
+ VK_FORMAT_R10G10B10A2_UINT = 0x00000029,
+ VK_FORMAT_R10G10B10A2_SINT = 0x0000002A,
+ VK_FORMAT_R16_UNORM = 0x0000002B,
+ VK_FORMAT_R16_SNORM = 0x0000002C,
+ VK_FORMAT_R16_USCALED = 0x0000002D,
+ VK_FORMAT_R16_SSCALED = 0x0000002E,
+ VK_FORMAT_R16_UINT = 0x0000002F,
+ VK_FORMAT_R16_SINT = 0x00000030,
+ VK_FORMAT_R16_SFLOAT = 0x00000031,
+ VK_FORMAT_R16G16_UNORM = 0x00000032,
+ VK_FORMAT_R16G16_SNORM = 0x00000033,
+ VK_FORMAT_R16G16_USCALED = 0x00000034,
+ VK_FORMAT_R16G16_SSCALED = 0x00000035,
+ VK_FORMAT_R16G16_UINT = 0x00000036,
+ VK_FORMAT_R16G16_SINT = 0x00000037,
+ VK_FORMAT_R16G16_SFLOAT = 0x00000038,
+ VK_FORMAT_R16G16B16_UNORM = 0x00000039,
+ VK_FORMAT_R16G16B16_SNORM = 0x0000003A,
+ VK_FORMAT_R16G16B16_USCALED = 0x0000003B,
+ VK_FORMAT_R16G16B16_SSCALED = 0x0000003C,
+ VK_FORMAT_R16G16B16_UINT = 0x0000003D,
+ VK_FORMAT_R16G16B16_SINT = 0x0000003E,
+ VK_FORMAT_R16G16B16_SFLOAT = 0x0000003F,
+ VK_FORMAT_R16G16B16A16_UNORM = 0x00000040,
+ VK_FORMAT_R16G16B16A16_SNORM = 0x00000041,
+ VK_FORMAT_R16G16B16A16_USCALED = 0x00000042,
+ VK_FORMAT_R16G16B16A16_SSCALED = 0x00000043,
+ VK_FORMAT_R16G16B16A16_UINT = 0x00000044,
+ VK_FORMAT_R16G16B16A16_SINT = 0x00000045,
+ VK_FORMAT_R16G16B16A16_SFLOAT = 0x00000046,
+ VK_FORMAT_R32_UINT = 0x00000047,
+ VK_FORMAT_R32_SINT = 0x00000048,
+ VK_FORMAT_R32_SFLOAT = 0x00000049,
+ VK_FORMAT_R32G32_UINT = 0x0000004A,
+ VK_FORMAT_R32G32_SINT = 0x0000004B,
+ VK_FORMAT_R32G32_SFLOAT = 0x0000004C,
+ VK_FORMAT_R32G32B32_UINT = 0x0000004D,
+ VK_FORMAT_R32G32B32_SINT = 0x0000004E,
+ VK_FORMAT_R32G32B32_SFLOAT = 0x0000004F,
+ VK_FORMAT_R32G32B32A32_UINT = 0x00000050,
+ VK_FORMAT_R32G32B32A32_SINT = 0x00000051,
+ VK_FORMAT_R32G32B32A32_SFLOAT = 0x00000052,
+ VK_FORMAT_R64_SFLOAT = 0x00000053,
+ VK_FORMAT_R64G64_SFLOAT = 0x00000054,
+ VK_FORMAT_R64G64B64_SFLOAT = 0x00000055,
+ VK_FORMAT_R64G64B64A64_SFLOAT = 0x00000056,
+ VK_FORMAT_R11G11B10_UFLOAT = 0x00000057,
+ VK_FORMAT_R9G9B9E5_UFLOAT = 0x00000058,
+ VK_FORMAT_D16_UNORM = 0x00000059,
+ VK_FORMAT_D24_UNORM = 0x0000005A,
+ VK_FORMAT_D32_SFLOAT = 0x0000005B,
+ VK_FORMAT_S8_UINT = 0x0000005C,
+ VK_FORMAT_D16_UNORM_S8_UINT = 0x0000005D,
+ VK_FORMAT_D24_UNORM_S8_UINT = 0x0000005E,
+ VK_FORMAT_D32_SFLOAT_S8_UINT = 0x0000005F,
+ VK_FORMAT_BC1_RGB_UNORM = 0x00000060,
+ VK_FORMAT_BC1_RGB_SRGB = 0x00000061,
+ VK_FORMAT_BC1_RGBA_UNORM = 0x00000062,
+ VK_FORMAT_BC1_RGBA_SRGB = 0x00000063,
+ VK_FORMAT_BC2_UNORM = 0x00000064,
+ VK_FORMAT_BC2_SRGB = 0x00000065,
+ VK_FORMAT_BC3_UNORM = 0x00000066,
+ VK_FORMAT_BC3_SRGB = 0x00000067,
+ VK_FORMAT_BC4_UNORM = 0x00000068,
+ VK_FORMAT_BC4_SNORM = 0x00000069,
+ VK_FORMAT_BC5_UNORM = 0x0000006A,
+ VK_FORMAT_BC5_SNORM = 0x0000006B,
+ VK_FORMAT_BC6H_UFLOAT = 0x0000006C,
+ VK_FORMAT_BC6H_SFLOAT = 0x0000006D,
+ VK_FORMAT_BC7_UNORM = 0x0000006E,
+ VK_FORMAT_BC7_SRGB = 0x0000006F,
+ VK_FORMAT_ETC2_R8G8B8_UNORM = 0x00000070,
+ VK_FORMAT_ETC2_R8G8B8_SRGB = 0x00000071,
+ VK_FORMAT_ETC2_R8G8B8A1_UNORM = 0x00000072,
+ VK_FORMAT_ETC2_R8G8B8A1_SRGB = 0x00000073,
+ VK_FORMAT_ETC2_R8G8B8A8_UNORM = 0x00000074,
+ VK_FORMAT_ETC2_R8G8B8A8_SRGB = 0x00000075,
+ VK_FORMAT_EAC_R11_UNORM = 0x00000076,
+ VK_FORMAT_EAC_R11_SNORM = 0x00000077,
+ VK_FORMAT_EAC_R11G11_UNORM = 0x00000078,
+ VK_FORMAT_EAC_R11G11_SNORM = 0x00000079,
+ VK_FORMAT_ASTC_4x4_UNORM = 0x0000007A,
+ VK_FORMAT_ASTC_4x4_SRGB = 0x0000007B,
+ VK_FORMAT_ASTC_5x4_UNORM = 0x0000007C,
+ VK_FORMAT_ASTC_5x4_SRGB = 0x0000007D,
+ VK_FORMAT_ASTC_5x5_UNORM = 0x0000007E,
+ VK_FORMAT_ASTC_5x5_SRGB = 0x0000007F,
+ VK_FORMAT_ASTC_6x5_UNORM = 0x00000080,
+ VK_FORMAT_ASTC_6x5_SRGB = 0x00000081,
+ VK_FORMAT_ASTC_6x6_UNORM = 0x00000082,
+ VK_FORMAT_ASTC_6x6_SRGB = 0x00000083,
+ VK_FORMAT_ASTC_8x5_UNORM = 0x00000084,
+ VK_FORMAT_ASTC_8x5_SRGB = 0x00000085,
+ VK_FORMAT_ASTC_8x6_UNORM = 0x00000086,
+ VK_FORMAT_ASTC_8x6_SRGB = 0x00000087,
+ VK_FORMAT_ASTC_8x8_UNORM = 0x00000088,
+ VK_FORMAT_ASTC_8x8_SRGB = 0x00000089,
+ VK_FORMAT_ASTC_10x5_UNORM = 0x0000008A,
+ VK_FORMAT_ASTC_10x5_SRGB = 0x0000008B,
+ VK_FORMAT_ASTC_10x6_UNORM = 0x0000008C,
+ VK_FORMAT_ASTC_10x6_SRGB = 0x0000008D,
+ VK_FORMAT_ASTC_10x8_UNORM = 0x0000008E,
+ VK_FORMAT_ASTC_10x8_SRGB = 0x0000008F,
+ VK_FORMAT_ASTC_10x10_UNORM = 0x00000090,
+ VK_FORMAT_ASTC_10x10_SRGB = 0x00000091,
+ VK_FORMAT_ASTC_12x10_UNORM = 0x00000092,
+ VK_FORMAT_ASTC_12x10_SRGB = 0x00000093,
+ VK_FORMAT_ASTC_12x12_UNORM = 0x00000094,
+ VK_FORMAT_ASTC_12x12_SRGB = 0x00000095,
+ VK_FORMAT_B4G4R4A4_UNORM = 0x00000096,
+ VK_FORMAT_B5G5R5A1_UNORM = 0x00000097,
+ VK_FORMAT_B5G6R5_UNORM = 0x00000098,
+ VK_FORMAT_B5G6R5_USCALED = 0x00000099,
+ VK_FORMAT_B8G8R8_UNORM = 0x0000009A,
+ VK_FORMAT_B8G8R8_SNORM = 0x0000009B,
+ VK_FORMAT_B8G8R8_USCALED = 0x0000009C,
+ VK_FORMAT_B8G8R8_SSCALED = 0x0000009D,
+ VK_FORMAT_B8G8R8_UINT = 0x0000009E,
+ VK_FORMAT_B8G8R8_SINT = 0x0000009F,
+ VK_FORMAT_B8G8R8_SRGB = 0x000000A0,
+ VK_FORMAT_B8G8R8A8_UNORM = 0x000000A1,
+ VK_FORMAT_B8G8R8A8_SNORM = 0x000000A2,
+ VK_FORMAT_B8G8R8A8_USCALED = 0x000000A3,
+ VK_FORMAT_B8G8R8A8_SSCALED = 0x000000A4,
+ VK_FORMAT_B8G8R8A8_UINT = 0x000000A5,
+ VK_FORMAT_B8G8R8A8_SINT = 0x000000A6,
+ VK_FORMAT_B8G8R8A8_SRGB = 0x000000A7,
+ VK_FORMAT_B10G10R10A2_UNORM = 0x000000A8,
+ VK_FORMAT_B10G10R10A2_SNORM = 0x000000A9,
+ VK_FORMAT_B10G10R10A2_USCALED = 0x000000AA,
+ VK_FORMAT_B10G10R10A2_SSCALED = 0x000000AB,
+ VK_FORMAT_B10G10R10A2_UINT = 0x000000AC,
+ VK_FORMAT_B10G10R10A2_SINT = 0x000000AD,
+
+ VK_ENUM_RANGE(FORMAT, UNDEFINED, B10G10R10A2_SINT)
+} VkFormat;
+
+// Shader stage enumerant
+typedef enum VkShaderStage_
+{
+ VK_SHADER_STAGE_VERTEX = 0,
+ VK_SHADER_STAGE_TESS_CONTROL = 1,
+ VK_SHADER_STAGE_TESS_EVALUATION = 2,
+ VK_SHADER_STAGE_GEOMETRY = 3,
+ VK_SHADER_STAGE_FRAGMENT = 4,
+ VK_SHADER_STAGE_COMPUTE = 5,
+
+ VK_ENUM_RANGE(SHADER_STAGE, VERTEX, COMPUTE)
+} VkShaderStage;
+
+// Structure type enumerant
+typedef enum VkStructureType_
+{
+ VK_STRUCTURE_TYPE_APPLICATION_INFO = 0,
+ VK_STRUCTURE_TYPE_DEVICE_CREATE_INFO = 1,
+ VK_STRUCTURE_TYPE_MEMORY_ALLOC_INFO = 2,
+ VK_STRUCTURE_TYPE_MEMORY_OPEN_INFO = 3,
+ VK_STRUCTURE_TYPE_PEER_MEMORY_OPEN_INFO = 4,
+ VK_STRUCTURE_TYPE_BUFFER_VIEW_ATTACH_INFO = 5,
+ VK_STRUCTURE_TYPE_IMAGE_VIEW_ATTACH_INFO = 6,
+ VK_STRUCTURE_TYPE_IMAGE_VIEW_CREATE_INFO = 7,
+ VK_STRUCTURE_TYPE_COLOR_ATTACHMENT_VIEW_CREATE_INFO = 8,
+ VK_STRUCTURE_TYPE_DEPTH_STENCIL_VIEW_CREATE_INFO = 9,
+ VK_STRUCTURE_TYPE_SHADER_CREATE_INFO = 10,
+ VK_STRUCTURE_TYPE_COMPUTE_PIPELINE_CREATE_INFO = 11,
+ VK_STRUCTURE_TYPE_SAMPLER_CREATE_INFO = 12,
+ VK_STRUCTURE_TYPE_DESCRIPTOR_SET_LAYOUT_CREATE_INFO = 13,
+ VK_STRUCTURE_TYPE_DYNAMIC_VP_STATE_CREATE_INFO = 14,
+ VK_STRUCTURE_TYPE_DYNAMIC_RS_STATE_CREATE_INFO = 15,
+ VK_STRUCTURE_TYPE_DYNAMIC_CB_STATE_CREATE_INFO = 16,
+ VK_STRUCTURE_TYPE_DYNAMIC_DS_STATE_CREATE_INFO = 17,
+ VK_STRUCTURE_TYPE_CMD_BUFFER_CREATE_INFO = 18,
+ VK_STRUCTURE_TYPE_EVENT_CREATE_INFO = 19,
+ VK_STRUCTURE_TYPE_FENCE_CREATE_INFO = 20,
+ VK_STRUCTURE_TYPE_SEMAPHORE_CREATE_INFO = 21,
+ VK_STRUCTURE_TYPE_SEMAPHORE_OPEN_INFO = 22,
+ VK_STRUCTURE_TYPE_QUERY_POOL_CREATE_INFO = 23,
+ VK_STRUCTURE_TYPE_PIPELINE_SHADER_STAGE_CREATE_INFO = 24,
+ VK_STRUCTURE_TYPE_GRAPHICS_PIPELINE_CREATE_INFO = 25,
+ VK_STRUCTURE_TYPE_PIPELINE_VERTEX_INPUT_CREATE_INFO = 26,
+ VK_STRUCTURE_TYPE_PIPELINE_IA_STATE_CREATE_INFO = 27,
+ VK_STRUCTURE_TYPE_PIPELINE_TESS_STATE_CREATE_INFO = 28,
+ VK_STRUCTURE_TYPE_PIPELINE_VP_STATE_CREATE_INFO = 29,
+ VK_STRUCTURE_TYPE_PIPELINE_RS_STATE_CREATE_INFO = 30,
+ VK_STRUCTURE_TYPE_PIPELINE_MS_STATE_CREATE_INFO = 31,
+ VK_STRUCTURE_TYPE_PIPELINE_CB_STATE_CREATE_INFO = 32,
+ VK_STRUCTURE_TYPE_PIPELINE_DS_STATE_CREATE_INFO = 33,
+ VK_STRUCTURE_TYPE_IMAGE_CREATE_INFO = 34,
+ VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO = 35,
+ VK_STRUCTURE_TYPE_BUFFER_VIEW_CREATE_INFO = 36,
+ VK_STRUCTURE_TYPE_FRAMEBUFFER_CREATE_INFO = 37,
+ VK_STRUCTURE_TYPE_CMD_BUFFER_BEGIN_INFO = 38,
+ VK_STRUCTURE_TYPE_CMD_BUFFER_GRAPHICS_BEGIN_INFO = 39,
+ VK_STRUCTURE_TYPE_RENDER_PASS_CREATE_INFO = 40,
+ VK_STRUCTURE_TYPE_LAYER_CREATE_INFO = 41,
+ VK_STRUCTURE_TYPE_MEMORY_BARRIER = 42,
+ VK_STRUCTURE_TYPE_BUFFER_MEMORY_BARRIER = 43,
+ VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER = 44,
+ VK_STRUCTURE_TYPE_DESCRIPTOR_POOL_CREATE_INFO = 45,
+ VK_STRUCTURE_TYPE_UPDATE_SAMPLERS = 46,
+ VK_STRUCTURE_TYPE_UPDATE_SAMPLER_TEXTURES = 47,
+ VK_STRUCTURE_TYPE_UPDATE_IMAGES = 48,
+ VK_STRUCTURE_TYPE_UPDATE_BUFFERS = 49,
+ VK_STRUCTURE_TYPE_UPDATE_AS_COPY = 50,
+ VK_STRUCTURE_TYPE_INSTANCE_CREATE_INFO = 51,
+
+ VK_ENUM_RANGE(STRUCTURE_TYPE, APPLICATION_INFO, INSTANCE_CREATE_INFO)
+} VkStructureType;
// ------------------------------------------------------------------------------------------------
// Error and return codes
@@ -731,8 +866,8 @@
VK_ERROR_UNKNOWN = -(0x00000001),
VK_ERROR_UNAVAILABLE = -(0x00000002),
VK_ERROR_INITIALIZATION_FAILED = -(0x00000003),
- VK_ERROR_OUT_OF_MEMORY = -(0x00000004),
- VK_ERROR_OUT_OF_GPU_MEMORY = -(0x00000005),
+ VK_ERROR_OUT_OF_HOST_MEMORY = -(0x00000004),
+ VK_ERROR_OUT_OF_DEVICE_MEMORY = -(0x00000005),
VK_ERROR_DEVICE_ALREADY_CREATED = -(0x00000006),
VK_ERROR_DEVICE_LOST = -(0x00000007),
VK_ERROR_INVALID_POINTER = -(0x00000008),
@@ -762,485 +897,245 @@
VK_ERROR_MEMORY_NOT_BOUND = -(0x00000020),
VK_ERROR_INCOMPATIBLE_QUEUE = -(0x00000021),
VK_ERROR_NOT_SHAREABLE = -(0x00000022),
- VK_MAX_ENUM(VkResult)
+
+ VK_MAX_ENUM(RESULT)
} VkResult;
// ------------------------------------------------------------------------------------------------
-// Vulkan format definitions
-
-typedef enum VkVertexInputStepRate_
-{
- VK_VERTEX_INPUT_STEP_RATE_VERTEX = 0x0,
- VK_VERTEX_INPUT_STEP_RATE_INSTANCE = 0x1,
- VK_VERTEX_INPUT_STEP_RATE_DRAW = 0x2, //Optional
-
- VK_VERTEX_INPUT_STEP_RATE_BEGIN_RANGE = VK_VERTEX_INPUT_STEP_RATE_VERTEX,
- VK_VERTEX_INPUT_STEP_RATE_END_RANGE = VK_VERTEX_INPUT_STEP_RATE_DRAW,
- VK_NUM_VERTEX_INPUT_STEP_RATE = (VK_VERTEX_INPUT_STEP_RATE_END_RANGE - VK_VERTEX_INPUT_STEP_RATE_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkVertexInputStepRate)
-} VkVertexInputStepRate;
-
-typedef enum VkFormat_
-{
- VK_FMT_UNDEFINED = 0x00000000,
- VK_FMT_R4G4_UNORM = 0x00000001,
- VK_FMT_R4G4_USCALED = 0x00000002,
- VK_FMT_R4G4B4A4_UNORM = 0x00000003,
- VK_FMT_R4G4B4A4_USCALED = 0x00000004,
- VK_FMT_R5G6B5_UNORM = 0x00000005,
- VK_FMT_R5G6B5_USCALED = 0x00000006,
- VK_FMT_R5G5B5A1_UNORM = 0x00000007,
- VK_FMT_R5G5B5A1_USCALED = 0x00000008,
- VK_FMT_R8_UNORM = 0x00000009,
- VK_FMT_R8_SNORM = 0x0000000A,
- VK_FMT_R8_USCALED = 0x0000000B,
- VK_FMT_R8_SSCALED = 0x0000000C,
- VK_FMT_R8_UINT = 0x0000000D,
- VK_FMT_R8_SINT = 0x0000000E,
- VK_FMT_R8_SRGB = 0x0000000F,
- VK_FMT_R8G8_UNORM = 0x00000010,
- VK_FMT_R8G8_SNORM = 0x00000011,
- VK_FMT_R8G8_USCALED = 0x00000012,
- VK_FMT_R8G8_SSCALED = 0x00000013,
- VK_FMT_R8G8_UINT = 0x00000014,
- VK_FMT_R8G8_SINT = 0x00000015,
- VK_FMT_R8G8_SRGB = 0x00000016,
- VK_FMT_R8G8B8_UNORM = 0x00000017,
- VK_FMT_R8G8B8_SNORM = 0x00000018,
- VK_FMT_R8G8B8_USCALED = 0x00000019,
- VK_FMT_R8G8B8_SSCALED = 0x0000001A,
- VK_FMT_R8G8B8_UINT = 0x0000001B,
- VK_FMT_R8G8B8_SINT = 0x0000001C,
- VK_FMT_R8G8B8_SRGB = 0x0000001D,
- VK_FMT_R8G8B8A8_UNORM = 0x0000001E,
- VK_FMT_R8G8B8A8_SNORM = 0x0000001F,
- VK_FMT_R8G8B8A8_USCALED = 0x00000020,
- VK_FMT_R8G8B8A8_SSCALED = 0x00000021,
- VK_FMT_R8G8B8A8_UINT = 0x00000022,
- VK_FMT_R8G8B8A8_SINT = 0x00000023,
- VK_FMT_R8G8B8A8_SRGB = 0x00000024,
- VK_FMT_R10G10B10A2_UNORM = 0x00000025,
- VK_FMT_R10G10B10A2_SNORM = 0x00000026,
- VK_FMT_R10G10B10A2_USCALED = 0x00000027,
- VK_FMT_R10G10B10A2_SSCALED = 0x00000028,
- VK_FMT_R10G10B10A2_UINT = 0x00000029,
- VK_FMT_R10G10B10A2_SINT = 0x0000002A,
- VK_FMT_R16_UNORM = 0x0000002B,
- VK_FMT_R16_SNORM = 0x0000002C,
- VK_FMT_R16_USCALED = 0x0000002D,
- VK_FMT_R16_SSCALED = 0x0000002E,
- VK_FMT_R16_UINT = 0x0000002F,
- VK_FMT_R16_SINT = 0x00000030,
- VK_FMT_R16_SFLOAT = 0x00000031,
- VK_FMT_R16G16_UNORM = 0x00000032,
- VK_FMT_R16G16_SNORM = 0x00000033,
- VK_FMT_R16G16_USCALED = 0x00000034,
- VK_FMT_R16G16_SSCALED = 0x00000035,
- VK_FMT_R16G16_UINT = 0x00000036,
- VK_FMT_R16G16_SINT = 0x00000037,
- VK_FMT_R16G16_SFLOAT = 0x00000038,
- VK_FMT_R16G16B16_UNORM = 0x00000039,
- VK_FMT_R16G16B16_SNORM = 0x0000003A,
- VK_FMT_R16G16B16_USCALED = 0x0000003B,
- VK_FMT_R16G16B16_SSCALED = 0x0000003C,
- VK_FMT_R16G16B16_UINT = 0x0000003D,
- VK_FMT_R16G16B16_SINT = 0x0000003E,
- VK_FMT_R16G16B16_SFLOAT = 0x0000003F,
- VK_FMT_R16G16B16A16_UNORM = 0x00000040,
- VK_FMT_R16G16B16A16_SNORM = 0x00000041,
- VK_FMT_R16G16B16A16_USCALED = 0x00000042,
- VK_FMT_R16G16B16A16_SSCALED = 0x00000043,
- VK_FMT_R16G16B16A16_UINT = 0x00000044,
- VK_FMT_R16G16B16A16_SINT = 0x00000045,
- VK_FMT_R16G16B16A16_SFLOAT = 0x00000046,
- VK_FMT_R32_UINT = 0x00000047,
- VK_FMT_R32_SINT = 0x00000048,
- VK_FMT_R32_SFLOAT = 0x00000049,
- VK_FMT_R32G32_UINT = 0x0000004A,
- VK_FMT_R32G32_SINT = 0x0000004B,
- VK_FMT_R32G32_SFLOAT = 0x0000004C,
- VK_FMT_R32G32B32_UINT = 0x0000004D,
- VK_FMT_R32G32B32_SINT = 0x0000004E,
- VK_FMT_R32G32B32_SFLOAT = 0x0000004F,
- VK_FMT_R32G32B32A32_UINT = 0x00000050,
- VK_FMT_R32G32B32A32_SINT = 0x00000051,
- VK_FMT_R32G32B32A32_SFLOAT = 0x00000052,
- VK_FMT_R64_SFLOAT = 0x00000053,
- VK_FMT_R64G64_SFLOAT = 0x00000054,
- VK_FMT_R64G64B64_SFLOAT = 0x00000055,
- VK_FMT_R64G64B64A64_SFLOAT = 0x00000056,
- VK_FMT_R11G11B10_UFLOAT = 0x00000057,
- VK_FMT_R9G9B9E5_UFLOAT = 0x00000058,
- VK_FMT_D16_UNORM = 0x00000059,
- VK_FMT_D24_UNORM = 0x0000005A,
- VK_FMT_D32_SFLOAT = 0x0000005B,
- VK_FMT_S8_UINT = 0x0000005C,
- VK_FMT_D16_UNORM_S8_UINT = 0x0000005D,
- VK_FMT_D24_UNORM_S8_UINT = 0x0000005E,
- VK_FMT_D32_SFLOAT_S8_UINT = 0x0000005F,
- VK_FMT_BC1_RGB_UNORM = 0x00000060,
- VK_FMT_BC1_RGB_SRGB = 0x00000061,
- VK_FMT_BC1_RGBA_UNORM = 0x00000062,
- VK_FMT_BC1_RGBA_SRGB = 0x00000063,
- VK_FMT_BC2_UNORM = 0x00000064,
- VK_FMT_BC2_SRGB = 0x00000065,
- VK_FMT_BC3_UNORM = 0x00000066,
- VK_FMT_BC3_SRGB = 0x00000067,
- VK_FMT_BC4_UNORM = 0x00000068,
- VK_FMT_BC4_SNORM = 0x00000069,
- VK_FMT_BC5_UNORM = 0x0000006A,
- VK_FMT_BC5_SNORM = 0x0000006B,
- VK_FMT_BC6H_UFLOAT = 0x0000006C,
- VK_FMT_BC6H_SFLOAT = 0x0000006D,
- VK_FMT_BC7_UNORM = 0x0000006E,
- VK_FMT_BC7_SRGB = 0x0000006F,
- VK_FMT_ETC2_R8G8B8_UNORM = 0x00000070,
- VK_FMT_ETC2_R8G8B8_SRGB = 0x00000071,
- VK_FMT_ETC2_R8G8B8A1_UNORM = 0x00000072,
- VK_FMT_ETC2_R8G8B8A1_SRGB = 0x00000073,
- VK_FMT_ETC2_R8G8B8A8_UNORM = 0x00000074,
- VK_FMT_ETC2_R8G8B8A8_SRGB = 0x00000075,
- VK_FMT_EAC_R11_UNORM = 0x00000076,
- VK_FMT_EAC_R11_SNORM = 0x00000077,
- VK_FMT_EAC_R11G11_UNORM = 0x00000078,
- VK_FMT_EAC_R11G11_SNORM = 0x00000079,
- VK_FMT_ASTC_4x4_UNORM = 0x0000007A,
- VK_FMT_ASTC_4x4_SRGB = 0x0000007B,
- VK_FMT_ASTC_5x4_UNORM = 0x0000007C,
- VK_FMT_ASTC_5x4_SRGB = 0x0000007D,
- VK_FMT_ASTC_5x5_UNORM = 0x0000007E,
- VK_FMT_ASTC_5x5_SRGB = 0x0000007F,
- VK_FMT_ASTC_6x5_UNORM = 0x00000080,
- VK_FMT_ASTC_6x5_SRGB = 0x00000081,
- VK_FMT_ASTC_6x6_UNORM = 0x00000082,
- VK_FMT_ASTC_6x6_SRGB = 0x00000083,
- VK_FMT_ASTC_8x5_UNORM = 0x00000084,
- VK_FMT_ASTC_8x5_SRGB = 0x00000085,
- VK_FMT_ASTC_8x6_UNORM = 0x00000086,
- VK_FMT_ASTC_8x6_SRGB = 0x00000087,
- VK_FMT_ASTC_8x8_UNORM = 0x00000088,
- VK_FMT_ASTC_8x8_SRGB = 0x00000089,
- VK_FMT_ASTC_10x5_UNORM = 0x0000008A,
- VK_FMT_ASTC_10x5_SRGB = 0x0000008B,
- VK_FMT_ASTC_10x6_UNORM = 0x0000008C,
- VK_FMT_ASTC_10x6_SRGB = 0x0000008D,
- VK_FMT_ASTC_10x8_UNORM = 0x0000008E,
- VK_FMT_ASTC_10x8_SRGB = 0x0000008F,
- VK_FMT_ASTC_10x10_UNORM = 0x00000090,
- VK_FMT_ASTC_10x10_SRGB = 0x00000091,
- VK_FMT_ASTC_12x10_UNORM = 0x00000092,
- VK_FMT_ASTC_12x10_SRGB = 0x00000093,
- VK_FMT_ASTC_12x12_UNORM = 0x00000094,
- VK_FMT_ASTC_12x12_SRGB = 0x00000095,
- VK_FMT_B4G4R4A4_UNORM = 0x00000096,
- VK_FMT_B5G5R5A1_UNORM = 0x00000097,
- VK_FMT_B5G6R5_UNORM = 0x00000098,
- VK_FMT_B5G6R5_USCALED = 0x00000099,
- VK_FMT_B8G8R8_UNORM = 0x0000009A,
- VK_FMT_B8G8R8_SNORM = 0x0000009B,
- VK_FMT_B8G8R8_USCALED = 0x0000009C,
- VK_FMT_B8G8R8_SSCALED = 0x0000009D,
- VK_FMT_B8G8R8_UINT = 0x0000009E,
- VK_FMT_B8G8R8_SINT = 0x0000009F,
- VK_FMT_B8G8R8_SRGB = 0x000000A0,
- VK_FMT_B8G8R8A8_UNORM = 0x000000A1,
- VK_FMT_B8G8R8A8_SNORM = 0x000000A2,
- VK_FMT_B8G8R8A8_USCALED = 0x000000A3,
- VK_FMT_B8G8R8A8_SSCALED = 0x000000A4,
- VK_FMT_B8G8R8A8_UINT = 0x000000A5,
- VK_FMT_B8G8R8A8_SINT = 0x000000A6,
- VK_FMT_B8G8R8A8_SRGB = 0x000000A7,
- VK_FMT_B10G10R10A2_UNORM = 0x000000A8,
- VK_FMT_B10G10R10A2_SNORM = 0x000000A9,
- VK_FMT_B10G10R10A2_USCALED = 0x000000AA,
- VK_FMT_B10G10R10A2_SSCALED = 0x000000AB,
- VK_FMT_B10G10R10A2_UINT = 0x000000AC,
- VK_FMT_B10G10R10A2_SINT = 0x000000AD,
-
- VK_FMT_BEGIN_RANGE = VK_FMT_UNDEFINED,
- VK_FMT_END_RANGE = VK_FMT_B10G10R10A2_SINT,
- VK_NUM_FMT = (VK_FMT_END_RANGE - VK_FMT_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkFormat)
-} VkFormat;
-
-// Shader stage enumerant
-typedef enum VkPipelineShaderStage_
-{
- VK_SHADER_STAGE_VERTEX = 0,
- VK_SHADER_STAGE_TESS_CONTROL = 1,
- VK_SHADER_STAGE_TESS_EVALUATION = 2,
- VK_SHADER_STAGE_GEOMETRY = 3,
- VK_SHADER_STAGE_FRAGMENT = 4,
- VK_SHADER_STAGE_COMPUTE = 5,
-
- VK_SHADER_STAGE_BEGIN_RANGE = VK_SHADER_STAGE_VERTEX,
- VK_SHADER_STAGE_END_RANGE = VK_SHADER_STAGE_COMPUTE,
- VK_NUM_SHADER_STAGE = (VK_SHADER_STAGE_END_RANGE - VK_SHADER_STAGE_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkPipelineShaderStage)
-} VkPipelineShaderStage;
-
-typedef enum VkShaderStageFlags_
-{
- VK_SHADER_STAGE_FLAGS_VERTEX_BIT = 0x00000001,
- VK_SHADER_STAGE_FLAGS_TESS_CONTROL_BIT = 0x00000002,
- VK_SHADER_STAGE_FLAGS_TESS_EVALUATION_BIT = 0x00000004,
- VK_SHADER_STAGE_FLAGS_GEOMETRY_BIT = 0x00000008,
- VK_SHADER_STAGE_FLAGS_FRAGMENT_BIT = 0x00000010,
- VK_SHADER_STAGE_FLAGS_COMPUTE_BIT = 0x00000020,
-
- VK_SHADER_STAGE_FLAGS_ALL = 0x7FFFFFFF,
- VK_MAX_ENUM(VkShaderStageFlags)
-} VkShaderStageFlags;
-
-// Structure type enumerant
-typedef enum VkStructureType_
-{
- VK_STRUCTURE_TYPE_APPLICATION_INFO = 0,
- VK_STRUCTURE_TYPE_DEVICE_CREATE_INFO = 1,
- VK_STRUCTURE_TYPE_MEMORY_ALLOC_INFO = 2,
- VK_STRUCTURE_TYPE_MEMORY_OPEN_INFO = 4,
- VK_STRUCTURE_TYPE_PEER_MEMORY_OPEN_INFO = 5,
- VK_STRUCTURE_TYPE_BUFFER_VIEW_ATTACH_INFO = 6,
- VK_STRUCTURE_TYPE_IMAGE_VIEW_ATTACH_INFO = 7,
- VK_STRUCTURE_TYPE_EVENT_WAIT_INFO = 8,
- VK_STRUCTURE_TYPE_IMAGE_VIEW_CREATE_INFO = 9,
- VK_STRUCTURE_TYPE_COLOR_ATTACHMENT_VIEW_CREATE_INFO = 10,
- VK_STRUCTURE_TYPE_DEPTH_STENCIL_VIEW_CREATE_INFO = 11,
- VK_STRUCTURE_TYPE_SHADER_CREATE_INFO = 12,
- VK_STRUCTURE_TYPE_COMPUTE_PIPELINE_CREATE_INFO = 13,
- VK_STRUCTURE_TYPE_SAMPLER_CREATE_INFO = 14,
- VK_STRUCTURE_TYPE_DESCRIPTOR_SET_LAYOUT_CREATE_INFO = 15,
- VK_STRUCTURE_TYPE_DYNAMIC_VP_STATE_CREATE_INFO = 16,
- VK_STRUCTURE_TYPE_DYNAMIC_RS_STATE_CREATE_INFO = 17,
- VK_STRUCTURE_TYPE_DYNAMIC_CB_STATE_CREATE_INFO = 18,
- VK_STRUCTURE_TYPE_DYNAMIC_DS_STATE_CREATE_INFO = 19,
- VK_STRUCTURE_TYPE_CMD_BUFFER_CREATE_INFO = 20,
- VK_STRUCTURE_TYPE_EVENT_CREATE_INFO = 21,
- VK_STRUCTURE_TYPE_FENCE_CREATE_INFO = 22,
- VK_STRUCTURE_TYPE_SEMAPHORE_CREATE_INFO = 23,
- VK_STRUCTURE_TYPE_SEMAPHORE_OPEN_INFO = 24,
- VK_STRUCTURE_TYPE_QUERY_POOL_CREATE_INFO = 25,
- VK_STRUCTURE_TYPE_PIPELINE_SHADER_STAGE_CREATE_INFO = 26,
- VK_STRUCTURE_TYPE_GRAPHICS_PIPELINE_CREATE_INFO = 27,
- VK_STRUCTURE_TYPE_PIPELINE_VERTEX_INPUT_CREATE_INFO = 28,
- VK_STRUCTURE_TYPE_PIPELINE_IA_STATE_CREATE_INFO = 29,
- VK_STRUCTURE_TYPE_PIPELINE_TESS_STATE_CREATE_INFO = 30,
- VK_STRUCTURE_TYPE_PIPELINE_VP_STATE_CREATE_INFO = 31,
- VK_STRUCTURE_TYPE_PIPELINE_RS_STATE_CREATE_INFO = 32,
- VK_STRUCTURE_TYPE_PIPELINE_MS_STATE_CREATE_INFO = 33,
- VK_STRUCTURE_TYPE_PIPELINE_CB_STATE_CREATE_INFO = 34,
- VK_STRUCTURE_TYPE_PIPELINE_DS_STATE_CREATE_INFO = 35,
- VK_STRUCTURE_TYPE_IMAGE_CREATE_INFO = 36,
- VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO = 37,
- VK_STRUCTURE_TYPE_BUFFER_VIEW_CREATE_INFO = 38,
- VK_STRUCTURE_TYPE_FRAMEBUFFER_CREATE_INFO = 39,
- VK_STRUCTURE_TYPE_CMD_BUFFER_BEGIN_INFO = 40,
- VK_STRUCTURE_TYPE_CMD_BUFFER_GRAPHICS_BEGIN_INFO = 41,
- VK_STRUCTURE_TYPE_RENDER_PASS_CREATE_INFO = 42,
- VK_STRUCTURE_TYPE_LAYER_CREATE_INFO = 43,
- VK_STRUCTURE_TYPE_PIPELINE_BARRIER = 44,
- VK_STRUCTURE_TYPE_MEMORY_BARRIER = 45,
- VK_STRUCTURE_TYPE_BUFFER_MEMORY_BARRIER = 46,
- VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER = 47,
- VK_STRUCTURE_TYPE_DESCRIPTOR_POOL_CREATE_INFO = 48,
- VK_STRUCTURE_TYPE_UPDATE_SAMPLERS = 49,
- VK_STRUCTURE_TYPE_UPDATE_SAMPLER_TEXTURES = 50,
- VK_STRUCTURE_TYPE_UPDATE_IMAGES = 51,
- VK_STRUCTURE_TYPE_UPDATE_BUFFERS = 52,
- VK_STRUCTURE_TYPE_UPDATE_AS_COPY = 53,
- VK_STRUCTURE_TYPE_INSTANCE_CREATE_INFO = 54,
-
- VK_STRUCTURE_TYPE_BEGIN_RANGE = VK_STRUCTURE_TYPE_APPLICATION_INFO,
- VK_STRUCTURE_TYPE_END_RANGE = VK_STRUCTURE_TYPE_INSTANCE_CREATE_INFO,
-
- VK_NUM_STRUCTURE_TYPE = (VK_STRUCTURE_TYPE_END_RANGE - VK_STRUCTURE_TYPE_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkStructureType)
-} VkStructureType;
-
-// ------------------------------------------------------------------------------------------------
// Flags
// Device creation flags
-typedef enum VkDeviceCreateFlags_
+typedef VkFlags VkDeviceCreateFlags;
+typedef enum VkDeviceCreateFlagBits_
{
- VK_DEVICE_CREATE_VALIDATION_BIT = 0x00000001,
- VK_DEVICE_CREATE_MGPU_IQ_MATCH_BIT = 0x00000002,
- VK_MAX_ENUM(VkDeviceCreateFlags)
-} VkDeviceCreateFlags;
+ VK_DEVICE_CREATE_VALIDATION_BIT = VK_BIT(0),
+ VK_DEVICE_CREATE_MULTI_DEVICE_IQ_MATCH_BIT = VK_BIT(1),
+} VkDeviceCreateFlagBits;
// Queue capabilities
-typedef enum VkQueueFlags_
+typedef VkFlags VkQueueFlags;
+typedef enum VkQueueFlagBits_
{
- VK_QUEUE_GRAPHICS_BIT = 0x00000001, // Queue supports graphics operations
- VK_QUEUE_COMPUTE_BIT = 0x00000002, // Queue supports compute operations
- VK_QUEUE_DMA_BIT = 0x00000004, // Queue supports DMA operations
- VK_QUEUE_MEMMGR_BIT = 0x00000008, // Queue supports memory management operations
- VK_QUEUE_EXTENDED_BIT = 0x40000000, // Extended queue
- VK_MAX_ENUM(VkQueueFlags)
-} VkQueueFlags;
+ VK_QUEUE_GRAPHICS_BIT = VK_BIT(0), // Queue supports graphics operations
+ VK_QUEUE_COMPUTE_BIT = VK_BIT(1), // Queue supports compute operations
+ VK_QUEUE_DMA_BIT = VK_BIT(2), // Queue supports DMA operations
+ VK_QUEUE_MEMMGR_BIT = VK_BIT(3), // Queue supports memory management operations
+ VK_QUEUE_EXTENDED_BIT = VK_BIT(30), // Extended queue
+} VkQueueFlagBits;
-// memory properties passed into vkAllocMemory().
-typedef enum VkMemoryPropertyFlags_
+// Memory properties passed into vkAllocMemory().
+typedef VkFlags VkMemoryPropertyFlags;
+typedef enum VkMemoryPropertyFlagBits_
{
- VK_MEMORY_PROPERTY_GPU_ONLY = 0x00000000, // If not set, then allocate memory on device (GPU)
- VK_MEMORY_PROPERTY_CPU_VISIBLE_BIT = 0x00000001,
- VK_MEMORY_PROPERTY_CPU_GPU_COHERENT_BIT = 0x00000002,
- VK_MEMORY_PROPERTY_CPU_UNCACHED_BIT = 0x00000004,
- VK_MEMORY_PROPERTY_CPU_WRITE_COMBINED_BIT = 0x00000008,
- VK_MEMORY_PROPERTY_PREFER_CPU_LOCAL = 0x00000010, // all else being equal, prefer CPU access
- VK_MEMORY_PROPERTY_SHAREABLE_BIT = 0x00000020,
- VK_MAX_ENUM(VkMemoryPropertyFlags)
-} VkMemoryPropertyFlags;
+ VK_MEMORY_PROPERTY_DEVICE_ONLY = 0, // If otherwise stated, then allocate memory on device
+ VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT = VK_BIT(0), // Memory should be mappable by host
+ VK_MEMORY_PROPERTY_HOST_DEVICE_COHERENT_BIT = VK_BIT(1), // Memory should be coherent between host and device accesses
+ VK_MEMORY_PROPERTY_HOST_UNCACHED_BIT = VK_BIT(2), // Memory should not be cached by the host
+ VK_MEMORY_PROPERTY_HOST_WRITE_COMBINED_BIT = VK_BIT(3), // Memory should support host write combining
+ VK_MEMORY_PROPERTY_PREFER_HOST_LOCAL = VK_BIT(4), // If set, prefer host access
+ VK_MEMORY_PROPERTY_SHAREABLE_BIT = VK_BIT(5),
+} VkMemoryPropertyFlagBits;
-// Buffer and buffer allocation usage flags
-typedef enum VkBufferUsageFlags_
+// Memory output flags passed to resource transition commands
+typedef VkFlags VkMemoryOutputFlags;
+typedef enum VkMemoryOutputFlagBits_
{
- VK_BUFFER_USAGE_GENERAL = 0x00000000, // No special usage
- VK_BUFFER_USAGE_TRANSFER_SOURCE_BIT = 0x00000001, // Can be used as a source of transfer operations
- VK_BUFFER_USAGE_TRANSFER_DESTINATION_BIT = 0x00000002, // Can be used as a destination of transfer operations
- VK_BUFFER_USAGE_UNIFORM_TEXEL_BUFFER_BIT = 0x00000004, // Can be used as TBO
- VK_BUFFER_USAGE_STORAGE_TEXEL_BUFFER_BIT = 0x00000008, // Can be used as IBO
- VK_BUFFER_USAGE_UNIFORM_BUFFER_BIT = 0x00000010, // Can be used as UBO
- VK_BUFFER_USAGE_STORAGE_BUFFER_BIT = 0x00000020, // Can be used as SSBO
- VK_BUFFER_USAGE_INDEX_BUFFER_BIT = 0x00000040, // Can be used as source of fixed function index fetch (index buffer)
- VK_BUFFER_USAGE_VERTEX_BUFFER_BIT = 0x00000080, // Can be used as source of fixed function vertex fetch (VBO)
- VK_BUFFER_USAGE_INDIRECT_BUFFER_BIT = 0x00000100, // Can be the source of indirect parameters (e.g. indirect buffer, parameter buffer)
- VK_MAX_ENUM(VkBufferUsageFlags)
-} VkBufferUsageFlags;
+ VK_MEMORY_OUTPUT_CPU_WRITE_BIT = VK_BIT(0), // Controls output coherency of CPU writes
+ VK_MEMORY_OUTPUT_SHADER_WRITE_BIT = VK_BIT(1), // Controls output coherency of generic shader writes
+ VK_MEMORY_OUTPUT_COLOR_ATTACHMENT_BIT = VK_BIT(2), // Controls output coherency of color attachment writes
+ VK_MEMORY_OUTPUT_DEPTH_STENCIL_ATTACHMENT_BIT = VK_BIT(3), // Controls output coherency of depth/stencil attachment writes
+ VK_MEMORY_OUTPUT_TRANSFER_BIT = VK_BIT(4), // Controls output coherency of transfer operations
+} VkMemoryOutputFlagBits;
-// Buffer flags
-typedef enum VkBufferCreateFlags_
+// Memory input flags passed to resource transition commands
+typedef VkFlags VkMemoryInputFlags;
+typedef enum VkMemoryInputFlagBits_
{
- VK_BUFFER_CREATE_SHAREABLE_BIT = 0x00000001,
- VK_BUFFER_CREATE_SPARSE_BIT = 0x00000002,
- VK_MAX_ENUM(VkBufferCreateFlags)
-} VkBufferCreateFlags;
+ VK_MEMORY_INPUT_CPU_READ_BIT = VK_BIT(0), // Controls input coherency of CPU reads
+ VK_MEMORY_INPUT_INDIRECT_COMMAND_BIT = VK_BIT(1), // Controls input coherency of indirect command reads
+ VK_MEMORY_INPUT_INDEX_FETCH_BIT = VK_BIT(2), // Controls input coherency of index fetches
+ VK_MEMORY_INPUT_VERTEX_ATTRIBUTE_FETCH_BIT = VK_BIT(3), // Controls input coherency of vertex attribute fetches
+ VK_MEMORY_INPUT_UNIFORM_READ_BIT = VK_BIT(4), // Controls input coherency of uniform buffer reads
+ VK_MEMORY_INPUT_SHADER_READ_BIT = VK_BIT(5), // Controls input coherency of generic shader reads
+ VK_MEMORY_INPUT_COLOR_ATTACHMENT_BIT = VK_BIT(6), // Controls input coherency of color attachment reads
+ VK_MEMORY_INPUT_DEPTH_STENCIL_ATTACHMENT_BIT = VK_BIT(7), // Controls input coherency of depth/stencil attachment reads
+ VK_MEMORY_INPUT_TRANSFER_BIT = VK_BIT(8), // Controls input coherency of transfer operations
+} VkMemoryInputFlagBits;
-typedef enum VkBufferViewType_
+// Buffer usage flags
+typedef VkFlags VkBufferUsageFlags;
+typedef enum VkBufferUsageFlagBits_
{
- VK_BUFFER_VIEW_RAW = 0x00000000, // Raw buffer without special structure (UBO, SSBO)
- VK_BUFFER_VIEW_FORMATTED = 0x00000001, // Buffer with format (TBO, IBO)
+ VK_BUFFER_USAGE_GENERAL = 0, // No special usage
+ VK_BUFFER_USAGE_TRANSFER_SOURCE_BIT = VK_BIT(0), // Can be used as a source of transfer operations
+ VK_BUFFER_USAGE_TRANSFER_DESTINATION_BIT = VK_BIT(1), // Can be used as a destination of transfer operations
+ VK_BUFFER_USAGE_UNIFORM_TEXEL_BUFFER_BIT = VK_BIT(2), // Can be used as TBO
+ VK_BUFFER_USAGE_STORAGE_TEXEL_BUFFER_BIT = VK_BIT(3), // Can be used as IBO
+ VK_BUFFER_USAGE_UNIFORM_BUFFER_BIT = VK_BIT(4), // Can be used as UBO
+ VK_BUFFER_USAGE_STORAGE_BUFFER_BIT = VK_BIT(5), // Can be used as SSBO
+ VK_BUFFER_USAGE_INDEX_BUFFER_BIT = VK_BIT(6), // Can be used as source of fixed function index fetch (index buffer)
+ VK_BUFFER_USAGE_VERTEX_BUFFER_BIT = VK_BIT(7), // Can be used as source of fixed function vertex fetch (VBO)
+ VK_BUFFER_USAGE_INDIRECT_BUFFER_BIT = VK_BIT(8), // Can be the source of indirect parameters (e.g. indirect buffer, parameter buffer)
+} VkBufferUsageFlagBits;
- VK_BUFFER_VIEW_TYPE_BEGIN_RANGE = VK_BUFFER_VIEW_RAW,
- VK_BUFFER_VIEW_TYPE_END_RANGE = VK_BUFFER_VIEW_FORMATTED,
- VK_NUM_BUFFER_VIEW_TYPE = (VK_BUFFER_VIEW_TYPE_END_RANGE - VK_BUFFER_VIEW_TYPE_BEGIN_RANGE + 1),
- VK_MAX_ENUM(VkBufferViewType)
-} VkBufferViewType;
-
-// Image and image allocation usage flags
-typedef enum VkImageUsageFlags_
+// Buffer creation flags
+typedef VkFlags VkBufferCreateFlags;
+typedef enum VkBufferCreateFlagBits_
{
- VK_IMAGE_USAGE_GENERAL = 0x00000000, // no special usage
- VK_IMAGE_USAGE_TRANSFER_SOURCE_BIT = 0x00000001, // Can be used as a source of transfer operations
- VK_IMAGE_USAGE_TRANSFER_DESTINATION_BIT = 0x00000002, // Can be used as a destination of transfer operations
- VK_IMAGE_USAGE_SAMPLED_BIT = 0x00000004, // Can be sampled from (SAMPLED_IMAGE and COMBINED_IMAGE_SAMPLER descriptor types)
- VK_IMAGE_USAGE_STORAGE_BIT = 0x00000008, // Can be used as storage image (STORAGE_IMAGE descriptor type)
- VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT = 0x00000010, // Can be used as framebuffer color attachment
- VK_IMAGE_USAGE_DEPTH_STENCIL_BIT = 0x00000020, // Can be used as framebuffer depth/stencil attachment
- VK_IMAGE_USAGE_TRANSIENT_ATTACHMENT_BIT = 0x00000040, // Image data not needed outside of rendering
- VK_MAX_ENUM(VkImageUsageFlags)
-} VkImageUsageFlags;
+ VK_BUFFER_CREATE_SHAREABLE_BIT = VK_BIT(0), // Buffer should be shareable
+ VK_BUFFER_CREATE_SPARSE_BIT = VK_BIT(1), // Buffer should support sparse backing
+} VkBufferCreateFlagBits;
-// Image flags
-typedef enum VkImageCreateFlags_
+// Shader stage flags
+typedef VkFlags VkShaderStageFlags;
+typedef enum VkShaderStageFlagBits_
{
- VK_IMAGE_CREATE_INVARIANT_DATA_BIT = 0x00000001,
- VK_IMAGE_CREATE_CLONEABLE_BIT = 0x00000002,
- VK_IMAGE_CREATE_SHAREABLE_BIT = 0x00000004,
- VK_IMAGE_CREATE_SPARSE_BIT = 0x00000008,
- VK_IMAGE_CREATE_MUTABLE_FORMAT_BIT = 0x00000010, // Allows image views to have different format than the base image
- VK_IMAGE_CREATE_CUBE_COMPATIBLE_BIT = 0x00000020, // Allows creating image views with cube type from the created image
- VK_MAX_ENUM(VkImageCreateFlags)
-} VkImageCreateFlags;
+ VK_SHADER_STAGE_VERTEX_BIT = VK_BIT(0),
+ VK_SHADER_STAGE_TESS_CONTROL_BIT = VK_BIT(1),
+ VK_SHADER_STAGE_TESS_EVALUATION_BIT = VK_BIT(2),
+ VK_SHADER_STAGE_GEOMETRY_BIT = VK_BIT(3),
+ VK_SHADER_STAGE_FRAGMENT_BIT = VK_BIT(4),
+ VK_SHADER_STAGE_COMPUTE_BIT = VK_BIT(5),
+
+ VK_SHADER_STAGE_ALL = 0x7FFFFFFF,
+} VkShaderStageFlagBits;
+
+// Image usage flags
+typedef VkFlags VkImageUsageFlags;
+typedef enum VkImageUsageFlagBits_
+{
+ VK_IMAGE_USAGE_GENERAL = 0, // No special usage
+ VK_IMAGE_USAGE_TRANSFER_SOURCE_BIT = VK_BIT(0), // Can be used as a source of transfer operations
+ VK_IMAGE_USAGE_TRANSFER_DESTINATION_BIT = VK_BIT(1), // Can be used as a destination of transfer operations
+ VK_IMAGE_USAGE_SAMPLED_BIT = VK_BIT(2), // Can be sampled from (SAMPLED_IMAGE and COMBINED_IMAGE_SAMPLER descriptor types)
+ VK_IMAGE_USAGE_STORAGE_BIT = VK_BIT(3), // Can be used as storage image (STORAGE_IMAGE descriptor type)
+ VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT = VK_BIT(4), // Can be used as framebuffer color attachment
+ VK_IMAGE_USAGE_DEPTH_STENCIL_BIT = VK_BIT(5), // Can be used as framebuffer depth/stencil attachment
+ VK_IMAGE_USAGE_TRANSIENT_ATTACHMENT_BIT = VK_BIT(6), // Image data not needed outside of rendering
+} VkImageUsageFlagBits;
+
+// Image creation flags
+typedef VkFlags VkImageCreateFlags;
+typedef enum VkImageCreateFlagBits_
+{
+ VK_IMAGE_CREATE_INVARIANT_DATA_BIT = VK_BIT(0),
+ VK_IMAGE_CREATE_CLONEABLE_BIT = VK_BIT(1),
+ VK_IMAGE_CREATE_SHAREABLE_BIT = VK_BIT(2), // Image should be shareable
+ VK_IMAGE_CREATE_SPARSE_BIT = VK_BIT(3), // Image should support sparse backing
+ VK_IMAGE_CREATE_MUTABLE_FORMAT_BIT = VK_BIT(4), // Allows image views to have different format than the base image
+ VK_IMAGE_CREATE_CUBE_COMPATIBLE_BIT = VK_BIT(5), // Allows creating image views with cube type from the created image
+} VkImageCreateFlagBits;
// Depth-stencil view creation flags
-typedef enum VkDepthStencilViewCreateFlags_
+typedef VkFlags VkDepthStencilViewCreateFlags;
+typedef enum VkDepthStencilViewCreateFlagBits_
{
- VK_DEPTH_STENCIL_VIEW_CREATE_READ_ONLY_DEPTH_BIT = 0x00000001,
- VK_DEPTH_STENCIL_VIEW_CREATE_READ_ONLY_STENCIL_BIT = 0x00000002,
- VK_MAX_ENUM(VkDepthStencilViewCreateFlags)
-} VkDepthStencilViewCreateFlags;
+ VK_DEPTH_STENCIL_VIEW_CREATE_READ_ONLY_DEPTH_BIT = VK_BIT(0),
+ VK_DEPTH_STENCIL_VIEW_CREATE_READ_ONLY_STENCIL_BIT = VK_BIT(1),
+} VkDepthStencilViewCreateFlagBits;
// Pipeline creation flags
-typedef enum VkPipelineCreateFlags_
+typedef VkFlags VkPipelineCreateFlags;
+typedef enum VkPipelineCreateFlagBits_
{
- VK_PIPELINE_CREATE_DISABLE_OPTIMIZATION_BIT = 0x00000001,
- VK_PIPELINE_CREATE_ALLOW_DERIVATIVES_BIT = 0x00000002,
- VK_MAX_ENUM(VkPipelineCreateFlags)
-} VkPipelineCreateFlags;
+ VK_PIPELINE_CREATE_DISABLE_OPTIMIZATION_BIT = VK_BIT(0),
+ VK_PIPELINE_CREATE_ALLOW_DERIVATIVES_BIT = VK_BIT(1),
+} VkPipelineCreateFlagBits;
+
+// Channel flags
+typedef VkFlags VkChannelFlags;
+typedef enum VkChannelFlagBits_
+{
+ VK_CHANNEL_R_BIT = VK_BIT(0),
+ VK_CHANNEL_G_BIT = VK_BIT(1),
+ VK_CHANNEL_B_BIT = VK_BIT(2),
+ VK_CHANNEL_A_BIT = VK_BIT(3),
+} VkChannelFlagBits;
// Fence creation flags
-typedef enum VkFenceCreateFlags_
+typedef VkFlags VkFenceCreateFlags;
+typedef enum VkFenceCreateFlagBits_
{
- VK_FENCE_CREATE_SIGNALED_BIT = 0x00000001,
- VK_MAX_ENUM(VkFenceCreateFlags)
-} VkFenceCreateFlags;
+ VK_FENCE_CREATE_SIGNALED_BIT = VK_BIT(0),
+} VkFenceCreateFlagBits;
// Semaphore creation flags
-typedef enum VkSemaphoreCreateFlags_
+typedef VkFlags VkSemaphoreCreateFlags;
+typedef enum VkSemaphoreCreateFlagBits_
{
- VK_SEMAPHORE_CREATE_SHAREABLE_BIT = 0x00000001,
- VK_MAX_ENUM(VkSemaphoreCreateFlags)
-} VkSemaphoreCreateFlags;
+ VK_SEMAPHORE_CREATE_SHAREABLE_BIT = VK_BIT(0),
+} VkSemaphoreCreateFlagBits;
// Format capability flags
-typedef enum VkFormatFeatureFlags_
+typedef VkFlags VkFormatFeatureFlags;
+typedef enum VkFormatFeatureFlagBits_
{
- VK_FORMAT_SAMPLED_IMAGE_BIT = 0x00000001, // Format can be used for sampled images (SAMPLED_IMAGE and COMBINED_IMAGE_SAMPLER descriptor types)
- VK_FORMAT_STORAGE_IMAGE_BIT = 0x00000002, // Format can be used for storage images (STORAGE_IMAGE descriptor type)
- VK_FORMAT_STORAGE_IMAGE_ATOMIC_BIT = 0x00000004, // Format supports atomic operations in case it's used for storage images
- VK_FORMAT_UNIFORM_TEXEL_BUFFER_BIT = 0x00000008, // Format can be used for uniform texel buffers (TBOs)
- VK_FORMAT_STORAGE_TEXEL_BUFFER_BIT = 0x00000010, // Format can be used for storage texel buffers (IBOs)
- VK_FORMAT_STORAGE_TEXEL_BUFFER_ATOMIC_BIT = 0x00000020, // Format supports atomic operations in case it's used for storage texel buffers
- VK_FORMAT_VERTEX_BUFFER_BIT = 0x00000040, // Format can be used for vertex buffers (VBOs)
- VK_FORMAT_COLOR_ATTACHMENT_BIT = 0x00000080, // Format can be used for color attachment images
- VK_FORMAT_COLOR_ATTACHMENT_BLEND_BIT = 0x00000100, // Format supports blending in case it's used for color attachment images
- VK_FORMAT_DEPTH_STENCIL_ATTACHMENT_BIT = 0x00000200, // Format can be used for depth/stencil attachment images
- VK_FORMAT_CONVERSION_BIT = 0x00000400, // Format can be used as the source or destination of format converting blits
- VK_MAX_ENUM(VkFormatFeatureFlags)
-} VkFormatFeatureFlags;
+ VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT = VK_BIT(0), // Format can be used for sampled images (SAMPLED_IMAGE and COMBINED_IMAGE_SAMPLER descriptor types)
+ VK_FORMAT_FEATURE_STORAGE_IMAGE_BIT = VK_BIT(1), // Format can be used for storage images (STORAGE_IMAGE descriptor type)
+ VK_FORMAT_FEATURE_STORAGE_IMAGE_ATOMIC_BIT = VK_BIT(2), // Format supports atomic operations in case it's used for storage images
+ VK_FORMAT_FEATURE_UNIFORM_TEXEL_BUFFER_BIT = VK_BIT(3), // Format can be used for uniform texel buffers (TBOs)
+ VK_FORMAT_FEATURE_STORAGE_TEXEL_BUFFER_BIT = VK_BIT(4), // Format can be used for storage texel buffers (IBOs)
+ VK_FORMAT_FEATURE_STORAGE_TEXEL_BUFFER_ATOMIC_BIT = VK_BIT(5), // Format supports atomic operations in case it's used for storage texel buffers
+ VK_FORMAT_FEATURE_VERTEX_BUFFER_BIT = VK_BIT(6), // Format can be used for vertex buffers (VBOs)
+ VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT = VK_BIT(7), // Format can be used for color attachment images
+ VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BLEND_BIT = VK_BIT(8), // Format supports blending in case it's used for color attachment images
+ VK_FORMAT_FEATURE_DEPTH_STENCIL_ATTACHMENT_BIT = VK_BIT(9), // Format can be used for depth/stencil attachment images
+ VK_FORMAT_FEATURE_CONVERSION_BIT = VK_BIT(10), // Format can be used as the source or destination of format converting blits
+} VkFormatFeatureFlagBits;
-// Query flags
-typedef enum VkQueryControlFlags_
+// Query control flags
+typedef VkFlags VkQueryControlFlags;
+typedef enum VkQueryControlFlagBits_
{
- VK_QUERY_IMPRECISE_DATA_BIT = 0x00000001,
- VK_MAX_ENUM(VkQueryControlFlags)
-} VkQueryControlFlags;
+ VK_QUERY_CONTROL_CONSERVATIVE_BIT = VK_BIT(0), // Allow conservative results to be collected by the query
+} VkQueryControlFlagBits;
// Query result flags
-typedef enum VkQueryResultFlags_
+typedef VkFlags VkQueryResultFlags;
+typedef enum VkQueryResultFlagBits_
{
- VK_QUERY_RESULT_32_BIT = 0x00000000, // Results of the queries are written to the destination buffer as 32-bit values
- VK_QUERY_RESULT_64_BIT = 0x00000001, // Results of the queries are written to the destination buffer as 64-bit values
- // Duplicate enum result messes with validation
-// VK_QUERY_RESULT_NO_WAIT_BIT = 0x00000000, // Results of the queries aren't waited on before proceeding with the result copy
- VK_QUERY_RESULT_WAIT_BIT = 0x00000002, // Results of the queries are waited on before proceeding with the result copy
- VK_QUERY_RESULT_WITH_AVAILABILITY_BIT = 0x00000004, // Besides the results of the query, the availability of the results is also written
- VK_QUERY_RESULT_PARTIAL_BIT = 0x00000008 // Copy the partial results of the query even if the final results aren't available
-} VkQueryResultFlags;
+ VK_QUERY_RESULT_32_BIT = 0, // Results of the queries are written to the destination buffer as 32-bit values
+ VK_QUERY_RESULT_64_BIT = VK_BIT(0), // Results of the queries are written to the destination buffer as 64-bit values
+ // VK_QUERY_RESULT_NO_WAIT_BIT = 0, // Results of the queries aren't waited on before proceeding with the result copy
+ VK_QUERY_RESULT_WAIT_BIT = VK_BIT(1), // Results of the queries are waited on before proceeding with the result copy
+ VK_QUERY_RESULT_WITH_AVAILABILITY_BIT = VK_BIT(2), // Besides the results of the query, the availability of the results is also written
+ VK_QUERY_RESULT_PARTIAL_BIT = VK_BIT(3), // Copy the partial results of the query even if the final results aren't available
+} VkQueryResultFlagBits;
-// GPU compatibility flags
-typedef enum VkGpuCompatibilityFlags_
+// Physical device compatibility flags
+typedef VkFlags VkPhysicalDeviceCompatibilityFlags;
+typedef enum VkPhysicalDeviceCompatibilityFlagBits_
{
- VK_GPU_COMPAT_ASIC_FEATURES_BIT = 0x00000001,
- VK_GPU_COMPAT_IQ_MATCH_BIT = 0x00000002,
- VK_GPU_COMPAT_PEER_TRANSFER_BIT = 0x00000004,
- VK_GPU_COMPAT_SHARED_MEMORY_BIT = 0x00000008,
- VK_GPU_COMPAT_SHARED_SYNC_BIT = 0x00000010,
- VK_GPU_COMPAT_SHARED_GPU0_DISPLAY_BIT = 0x00000020,
- VK_GPU_COMPAT_SHARED_GPU1_DISPLAY_BIT = 0x00000040,
- VK_MAX_ENUM(VkGpuCompatibilityFlags)
-} VkGpuCompatibilityFlags;
+ VK_PHYSICAL_DEVICE_COMPATIBILITY_FEATURES_BIT = VK_BIT(0),
+ VK_PHYSICAL_DEVICE_COMPATIBILITY_IQ_MATCH_BIT = VK_BIT(1),
+ VK_PHYSICAL_DEVICE_COMPATIBILITY_PEER_TRANSFER_BIT = VK_BIT(2),
+ VK_PHYSICAL_DEVICE_COMPATIBILITY_SHARED_MEMORY_BIT = VK_BIT(3),
+ VK_PHYSICAL_DEVICE_COMPATIBILITY_SHARED_SYNC_BIT = VK_BIT(4),
+ VK_PHYSICAL_DEVICE_COMPATIBILITY_SHARED_DEVICE0_DISPLAY_BIT = VK_BIT(5),
+ VK_PHYSICAL_DEVICE_COMPATIBILITY_SHARED_DEVICE1_DISPLAY_BIT = VK_BIT(6),
+} VkPhysicalDeviceCompatibilityFlagBits;
-// Command buffer building flags
-typedef enum VkCmdBufferBuildFlags_
+// Shader creation flags
+typedef VkFlags VkShaderCreateFlags;
+
+// Event creation flags
+typedef VkFlags VkEventCreateFlags;
+
+// Command buffer creation flags
+typedef VkFlags VkCmdBufferCreateFlags;
+
+// Command buffer optimization flags
+typedef VkFlags VkCmdBufferOptimizeFlags;
+typedef enum VkCmdBufferOptimizeFlagBits_
{
- VK_CMD_BUFFER_OPTIMIZE_GPU_SMALL_BATCH_BIT = 0x00000001,
- VK_CMD_BUFFER_OPTIMIZE_PIPELINE_SWITCH_BIT = 0x00000002,
- VK_CMD_BUFFER_OPTIMIZE_ONE_TIME_SUBMIT_BIT = 0x00000004,
- VK_CMD_BUFFER_OPTIMIZE_DESCRIPTOR_SET_SWITCH_BIT = 0x00000008,
- VK_MAX_ENUM(VkCmdBufferBuildFlags)
-} VkCmdBufferBuildFlags;
+ VK_CMD_BUFFER_OPTIMIZE_SMALL_BATCH_BIT = VK_BIT(0),
+ VK_CMD_BUFFER_OPTIMIZE_PIPELINE_SWITCH_BIT = VK_BIT(1),
+ VK_CMD_BUFFER_OPTIMIZE_ONE_TIME_SUBMIT_BIT = VK_BIT(2),
+ VK_CMD_BUFFER_OPTIMIZE_DESCRIPTOR_SET_SWITCH_BIT = VK_BIT(3),
+} VkCmdBufferOptimizeFlagBits;
+
+// Memory mapping flags
+typedef VkFlags VkMemoryMapFlags;
// ------------------------------------------------------------------------------------------------
// Vulkan structures
@@ -1295,15 +1190,15 @@
VkChannelSwizzle a;
} VkChannelMapping;
-typedef struct VkPhysicalGpuProperties_
+typedef struct VkPhysicalDeviceProperties_
{
uint32_t apiVersion;
uint32_t driverVersion;
uint32_t vendorId;
uint32_t deviceId;
- VkPhysicalGpuType gpuType;
- char gpuName[VK_MAX_PHYSICAL_GPU_NAME];
- VkGpuSize maxInlineMemoryUpdateSize;
+ VkPhysicalDeviceType deviceType;
+ char deviceName[VK_MAX_PHYSICAL_DEVICE_NAME];
+ VkDeviceSize maxInlineMemoryUpdateSize;
uint32_t maxBoundDescriptorSets;
uint32_t maxThreadGroupSize;
uint64_t timestampFrequency;
@@ -1311,21 +1206,21 @@
uint32_t maxDescriptorSets; // at least 2?
uint32_t maxViewports; // at least 16?
uint32_t maxColorAttachments; // at least 8?
-} VkPhysicalGpuProperties;
+} VkPhysicalDeviceProperties;
-typedef struct VkPhysicalGpuPerformance_
+typedef struct VkPhysicalDevicePerformance_
{
- float maxGpuClock;
+ float maxDeviceClock;
float aluPerClock;
float texPerClock;
float primsPerClock;
float pixelsPerClock;
-} VkPhysicalGpuPerformance;
+} VkPhysicalDevicePerformance;
-typedef struct VkGpuCompatibilityInfo_
+typedef struct VkPhysicalDeviceCompatibilityInfo_
{
- VkFlags compatibilityFlags; // VkGpuCompatibilityFlags
-} VkGpuCompatibilityInfo;
+ VkPhysicalDeviceCompatibilityFlags compatibilityFlags;
+} VkPhysicalDeviceCompatibilityInfo;
typedef struct VkExtensionProperties_
{
@@ -1375,7 +1270,7 @@
const VkDeviceQueueCreateInfo* pRequestedQueues;
uint32_t extensionCount;
const char*const* ppEnabledExtensionNames;
- VkFlags flags; // VkDeviceCreateFlags
+ VkDeviceCreateFlags flags; // Device creation flags
} VkDeviceCreateInfo;
typedef struct VkInstanceCreateInfo_
@@ -1397,27 +1292,27 @@
const char *const* ppActiveLayerNames; // layer name from the layer's vkEnumerateLayers())
} VkLayerCreateInfo;
-typedef struct VkPhysicalGpuQueueProperties_
+typedef struct VkPhysicalDeviceQueueProperties_
{
- VkFlags queueFlags; // VkQueueFlags
+ VkQueueFlags queueFlags; // Queue flags
uint32_t queueCount;
uint32_t maxAtomicCounters;
bool32_t supportsTimestamps;
uint32_t maxMemReferences; // Tells how many memory references can be active for the given queue
-} VkPhysicalGpuQueueProperties;
+} VkPhysicalDeviceQueueProperties;
-typedef struct VkPhysicalGpuMemoryProperties_
+typedef struct VkPhysicalDeviceMemoryProperties_
{
bool32_t supportsMigration;
bool32_t supportsPinning;
-} VkPhysicalGpuMemoryProperties;
+} VkPhysicalDeviceMemoryProperties;
typedef struct VkMemoryAllocInfo_
{
VkStructureType sType; // Must be VK_STRUCTURE_TYPE_MEMORY_ALLOC_INFO
const void* pNext; // Pointer to next structure
- VkGpuSize allocationSize; // Size of memory allocation
- VkFlags memProps; // VkMemoryPropertyFlags
+ VkDeviceSize allocationSize; // Size of memory allocation
+ VkMemoryPropertyFlags memProps; // Memory property flags
VkMemoryPriority memPriority;
} VkMemoryAllocInfo;
@@ -1425,29 +1320,29 @@
{
VkStructureType sType; // Must be VK_STRUCTURE_TYPE_MEMORY_OPEN_INFO
const void* pNext; // Pointer to next structure
- VkGpuMemory sharedMem;
+ VkDeviceMemory sharedMem;
} VkMemoryOpenInfo;
typedef struct VkPeerMemoryOpenInfo_
{
VkStructureType sType; // Must be VK_STRUCTURE_TYPE_PEER_MEMORY_OPEN_INFO
const void* pNext; // Pointer to next structure
- VkGpuMemory originalMem;
+ VkDeviceMemory originalMem;
} VkPeerMemoryOpenInfo;
typedef struct VkMemoryRequirements_
{
- VkGpuSize size; // Specified in bytes
- VkGpuSize alignment; // Specified in bytes
- VkGpuSize granularity; // Granularity on which vkQueueBindObjectMemoryRange can bind sub-ranges of memory specified in bytes (usually the page size)
+ VkDeviceSize size; // Specified in bytes
+ VkDeviceSize alignment; // Specified in bytes
+ VkDeviceSize granularity; // Granularity on which vkBindObjectMemoryRange can bind sub-ranges of memory specified in bytes (usually the page size)
VkMemoryPropertyFlags memPropsAllowed; // Allowed memory property flags
VkMemoryPropertyFlags memPropsRequired; // Required memory property flags
} VkMemoryRequirements;
typedef struct VkFormatProperties_
{
- VkFlags linearTilingFeatures; // VkFormatFeatureFlags
- VkFlags optimalTilingFeatures; // VkFormatFeatureFlags
+ VkFormatFeatureFlags linearTilingFeatures; // Format features in case of linear tiling
+ VkFormatFeatureFlags optimalTilingFeatures; // Format features in case of optimal tiling
} VkFormatProperties;
typedef struct VkBufferViewAttachInfo_
@@ -1528,9 +1423,9 @@
{
VkStructureType sType; // Must be VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO
const void* pNext; // Pointer to next structure.
- VkGpuSize size; // Specified in bytes
- VkFlags usage; // VkBufferUsageFlags
- VkFlags flags; // VkBufferCreateFlags
+ VkDeviceSize size; // Specified in bytes
+ VkBufferUsageFlags usage; // Buffer usage flags
+ VkBufferCreateFlags flags; // Buffer creation flags
} VkBufferCreateInfo;
typedef struct VkBufferViewCreateInfo_
@@ -1540,8 +1435,8 @@
VkBuffer buffer;
VkBufferViewType viewType;
VkFormat format; // Optionally specifies format of elements
- VkGpuSize offset; // Specified in bytes
- VkGpuSize range; // View size specified in bytes
+ VkDeviceSize offset; // Specified in bytes
+ VkDeviceSize range; // View size specified in bytes
} VkBufferViewCreateInfo;
typedef struct VkImageSubresource_
@@ -1560,41 +1455,13 @@
uint32_t arraySize;
} VkImageSubresourceRange;
-typedef struct VkEventWaitInfo_
-{
- VkStructureType sType; // Must be VK_STRUCTURE_TYPE_EVENT_WAIT_INFO
- const void* pNext; // Pointer to next structure.
-
- uint32_t eventCount; // Number of events to wait on
- const VkEvent* pEvents; // Array of event objects to wait on
-
- VkWaitEvent waitEvent; // Pipeline event where the wait should happen
-
- uint32_t memBarrierCount; // Number of memory barriers
- const void** ppMemBarriers; // Array of pointers to memory barriers (any of them can be either VkMemoryBarrier, VkBufferMemoryBarrier, or VkImageMemoryBarrier)
-} VkEventWaitInfo;
-
-typedef struct VkPipelineBarrier_
-{
- VkStructureType sType; // Must be VK_STRUCTURE_TYPE_PIPELINE_BARRIER
- const void* pNext; // Pointer to next structure.
-
- uint32_t eventCount; // Number of events to wait on
- const VkPipeEvent* pEvents; // Array of pipeline events to wait on
-
- VkWaitEvent waitEvent; // Pipeline event where the wait should happen
-
- uint32_t memBarrierCount; // Number of memory barriers
- const void** ppMemBarriers; // Array of pointers to memory barriers (any of them can be either VkMemoryBarrier, VkBufferMemoryBarrier, or VkImageMemoryBarrier)
-} VkPipelineBarrier;
-
typedef struct VkMemoryBarrier_
{
VkStructureType sType; // Must be VK_STRUCTURE_TYPE_MEMORY_BARRIER
const void* pNext; // Pointer to next structure.
- VkFlags outputMask; // Outputs the barrier should sync (see VkMemoryOutputFlags)
- VkFlags inputMask; // Inputs the barrier should sync to (see VkMemoryInputFlags)
+ VkMemoryOutputFlags outputMask; // Outputs the barrier should sync
+ VkMemoryInputFlags inputMask; // Inputs the barrier should sync to
} VkMemoryBarrier;
typedef struct VkBufferMemoryBarrier_
@@ -1602,13 +1469,13 @@
VkStructureType sType; // Must be VK_STRUCTURE_TYPE_BUFFER_MEMORY_BARRIER
const void* pNext; // Pointer to next structure.
- VkFlags outputMask; // Outputs the barrier should sync (see VkMemoryOutputFlags)
- VkFlags inputMask; // Inputs the barrier should sync to (see VkMemoryInputFlags)
+ VkMemoryOutputFlags outputMask; // Outputs the barrier should sync
+ VkMemoryInputFlags inputMask; // Inputs the barrier should sync to
VkBuffer buffer; // Buffer to sync
- VkGpuSize offset; // Offset within the buffer to sync
- VkGpuSize size; // Amount of bytes to sync
+ VkDeviceSize offset; // Offset within the buffer to sync
+ VkDeviceSize size; // Amount of bytes to sync
} VkBufferMemoryBarrier;
typedef struct VkImageMemoryBarrier_
@@ -1616,8 +1483,8 @@
VkStructureType sType; // Must be VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER
const void* pNext; // Pointer to next structure.
- VkFlags outputMask; // Outputs the barrier should sync (see VkMemoryOutputFlags)
- VkFlags inputMask; // Inputs the barrier should sync to (see VkMemoryInputFlags)
+ VkMemoryOutputFlags outputMask; // Outputs the barrier should sync
+ VkMemoryInputFlags inputMask; // Inputs the barrier should sync to
VkImageLayout oldLayout; // Current layout of the image
VkImageLayout newLayout; // New layout to transition the image to
@@ -1638,8 +1505,8 @@
uint32_t arraySize;
uint32_t samples;
VkImageTiling tiling;
- VkFlags usage; // VkImageUsageFlags
- VkFlags flags; // VkImageCreateFlags
+ VkImageUsageFlags usage; // Image usage flags
+ VkImageCreateFlags flags; // Image creation flags
} VkImageCreateInfo;
typedef struct VkPeerImageOpenInfo_
@@ -1649,10 +1516,10 @@
typedef struct VkSubresourceLayout_
{
- VkGpuSize offset; // Specified in bytes
- VkGpuSize size; // Specified in bytes
- VkGpuSize rowPitch; // Specified in bytes
- VkGpuSize depthPitch; // Specified in bytes
+ VkDeviceSize offset; // Specified in bytes
+ VkDeviceSize size; // Specified in bytes
+ VkDeviceSize rowPitch; // Specified in bytes
+ VkDeviceSize depthPitch; // Specified in bytes
} VkSubresourceLayout;
typedef struct VkImageViewCreateInfo_
@@ -1690,7 +1557,7 @@
uint32_t arraySize;
VkImage msaaResolveImage;
VkImageSubresourceRange msaaResolveSubResource;
- VkFlags flags; // VkDepthStencilViewCreateFlags
+ VkDepthStencilViewCreateFlags flags; // Depth stencil attachment view flags
} VkDepthStencilViewCreateInfo;
typedef struct VkColorAttachmentBindInfo_
@@ -1707,9 +1574,9 @@
typedef struct VkBufferCopy_
{
- VkGpuSize srcOffset; // Specified in bytes
- VkGpuSize destOffset; // Specified in bytes
- VkGpuSize copySize; // Specified in bytes
+ VkDeviceSize srcOffset; // Specified in bytes
+ VkDeviceSize destOffset; // Specified in bytes
+ VkDeviceSize copySize; // Specified in bytes
} VkBufferCopy;
typedef struct VkImageMemoryBindInfo_
@@ -1740,7 +1607,7 @@
typedef struct VkBufferImageCopy_
{
- VkGpuSize bufferOffset; // Specified in bytes
+ VkDeviceSize bufferOffset; // Specified in bytes
VkImageSubresource imageSubresource;
VkOffset3D imageOffset; // Specified in pixels for both compressed and uncompressed images
VkExtent3D imageExtent; // Specified in pixels for both compressed and uncompressed images
@@ -1761,23 +1628,23 @@
const void* pNext; // Pointer to next structure
size_t codeSize; // Specified in bytes
const void* pCode;
- VkFlags flags; // Reserved
+ VkShaderCreateFlags flags; // Reserved
} VkShaderCreateInfo;
typedef struct VkDescriptorSetLayoutBinding_
{
- VkDescriptorType descriptorType;
- uint32_t count;
- VkFlags stageFlags; // VkShaderStageFlags
- const VkSampler* pImmutableSamplers;
+ VkDescriptorType descriptorType; // Type of the descriptors in this binding
+ uint32_t count; // Number of descriptors in this binding
+ VkShaderStageFlags stageFlags; // Shader stages this binding is visible to
+ const VkSampler* pImmutableSamplers; // Immutable samplers (used if descriptor type is SAMPLER or COMBINED_IMAGE_SAMPLER, is either NULL or contains <count> number of elements)
} VkDescriptorSetLayoutBinding;
typedef struct VkDescriptorSetLayoutCreateInfo_
{
- VkStructureType sType; // Must be VK_STRUCTURE_TYPE_DESCRIPTOR_SET_LAYOUT_CREATE_INFO
- const void* pNext; // Pointer to next structure
- uint32_t count; // Number of bindings in the descriptor set layout
- const VkDescriptorSetLayoutBinding* pBinding; // Array of descriptor set layout bindings
+ VkStructureType sType; // Must be VK_STRUCTURE_TYPE_DESCRIPTOR_SET_LAYOUT_CREATE_INFO
+ const void* pNext; // Pointer to next structure
+ uint32_t count; // Number of bindings in the descriptor set layout
+ const VkDescriptorSetLayoutBinding* pBinding; // Array of descriptor set layout bindings
} VkDescriptorSetLayoutCreateInfo;
typedef struct VkDescriptorTypeCount_
@@ -1816,7 +1683,7 @@
typedef struct VkPipelineShader_
{
- VkPipelineShaderStage stage;
+ VkShaderStage stage;
VkShader shader;
uint32_t linkConstBufferCount;
const VkLinkConstBuffer* pLinkConstBufferInfo;
@@ -1825,10 +1692,10 @@
typedef struct VkComputePipelineCreateInfo_
{
- VkStructureType sType; // Must be VK_STRUCTURE_TYPE_COMPUTE_PIPELINE_CREATE_INFO
- const void* pNext; // Pointer to next structure
+ VkStructureType sType; // Must be VK_STRUCTURE_TYPE_COMPUTE_PIPELINE_CREATE_INFO
+ const void* pNext; // Pointer to next structure
VkPipelineShader cs;
- VkFlags flags; // VkPipelineCreateFlags
+ VkPipelineCreateFlags flags; // Pipeline creation flags
VkDescriptorSetLayoutChain setLayoutChain;
uint32_t localSizeX;
uint32_t localSizeY;
@@ -1887,7 +1754,7 @@
{
VkStructureType sType; // Must be VK_STRUCTURE_TYPE_PIPELINE_VP_STATE_CREATE_INFO
const void* pNext; // Pointer to next structure
- uint32_t numViewports;
+ uint32_t viewportCount;
VkCoordinateOrigin clipOrigin; // optional (GL45)
VkDepthMode depthMode; // optional (GL45)
} VkPipelineVpStateCreateInfo;
@@ -1900,10 +1767,10 @@
bool32_t rasterizerDiscardEnable;
bool32_t programPointSize; // optional (GL45)
VkCoordinateOrigin pointOrigin; // optional (GL45)
- VkProvokingVertexConvention provokingVertex; // optional (GL45)
+ VkProvokingVertex provokingVertex; // optional (GL45)
VkFillMode fillMode; // optional (GL45)
VkCullMode cullMode;
- VkFaceOrientation frontFace;
+ VkFrontFace frontFace;
} VkPipelineRsStateCreateInfo;
typedef struct VkPipelineMsStateCreateInfo_
@@ -1923,11 +1790,11 @@
VkFormat format;
VkBlend srcBlendColor;
VkBlend destBlendColor;
- VkBlendFunc blendFuncColor;
+ VkBlendOp blendOpColor;
VkBlend srcBlendAlpha;
VkBlend destBlendAlpha;
- VkBlendFunc blendFuncAlpha;
- uint8_t channelWriteMask;
+ VkBlendOp blendOpAlpha;
+ VkChannelFlags channelWriteMask;
} VkPipelineCbAttachmentState;
typedef struct VkPipelineCbStateCreateInfo_
@@ -1946,7 +1813,7 @@
VkStencilOp stencilFailOp;
VkStencilOp stencilPassOp;
VkStencilOp stencilDepthFailOp;
- VkCompareFunc stencilFunc;
+ VkCompareOp stencilCompareOp;
} VkStencilOpState;
typedef struct VkPipelineDsStateCreateInfo_
@@ -1956,7 +1823,7 @@
VkFormat format;
bool32_t depthTestEnable;
bool32_t depthWriteEnable;
- VkCompareFunc depthFunc;
+ VkCompareOp depthCompareOp;
bool32_t depthBoundsEnable; // optional (depth_bounds_test)
bool32_t stencilTestEnable;
VkStencilOpState front;
@@ -1974,7 +1841,7 @@
{
VkStructureType sType; // Must be VK_STRUCTURE_TYPE_GRAPHICS_PIPELINE_CREATE_INFO
const void* pNext; // Pointer to next structure
- VkFlags flags; // VkPipelineCreateFlags
+ VkPipelineCreateFlags flags; // Pipeline creation flags
VkDescriptorSetLayoutChain pSetLayoutChain;
} VkGraphicsPipelineCreateInfo;
@@ -1990,10 +1857,10 @@
VkTexAddress addressW;
float mipLodBias;
uint32_t maxAnisotropy;
- VkCompareFunc compareFunc;
+ VkCompareOp compareOp;
float minLod;
float maxLod;
- VkBorderColorType borderColorType;
+ VkBorderColor borderColor;
} VkSamplerCreateInfo;
typedef struct VkDynamicVpStateCreateInfo_
@@ -2041,7 +1908,7 @@
VkStructureType sType; // Must be VK_STRUCTURE_TYPE_CMD_BUFFER_CREATE_INFO
const void* pNext; // Pointer to next structure
uint32_t queueNodeIndex;
- VkFlags flags;
+ VkCmdBufferCreateFlags flags; // Command buffer creation flags
} VkCmdBufferCreateInfo;
typedef struct VkCmdBufferBeginInfo_
@@ -2049,7 +1916,7 @@
VkStructureType sType; // Must be VK_STRUCTURE_TYPE_CMD_BUFFER_BEGIN_INFO
const void* pNext; // Pointer to next structure
- VkFlags flags; // VkCmdBufferBuildFlags
+ VkCmdBufferOptimizeFlags flags; // Command buffer optimization flags
} VkCmdBufferBeginInfo;
typedef struct VkRenderPassBegin_
@@ -2108,14 +1975,14 @@
{
VkStructureType sType; // Must be VK_STRUCTURE_TYPE_EVENT_CREATE_INFO
const void* pNext; // Pointer to next structure
- VkFlags flags; // Reserved
+ VkEventCreateFlags flags; // Event creation flags
} VkEventCreateInfo;
typedef struct VkFenceCreateInfo_
{
VkStructureType sType; // Must be VK_STRUCTURE_TYPE_FENCE_CREATE_INFO
const void* pNext; // Pointer to next structure
- VkFenceCreateFlags flags; // VkFenceCreateFlags
+ VkFenceCreateFlags flags; // Fence creation flags
} VkFenceCreateInfo;
typedef struct VkSemaphoreCreateInfo_
@@ -2123,7 +1990,7 @@
VkStructureType sType; // Must be VK_STRUCTURE_TYPE_SEMAPHORE_CREATE_INFO
const void* pNext; // Pointer to next structure
uint32_t initialCount;
- VkFlags flags; // VkSemaphoreCreateFlags
+ VkSemaphoreCreateFlags flags; // Semaphore creation flags
} VkSemaphoreCreateInfo;
typedef struct VkSemaphoreOpenInfo_
@@ -2199,36 +2066,37 @@
// API functions
typedef VkResult (VKAPI *PFN_vkCreateInstance)(const VkInstanceCreateInfo* pCreateInfo, VkInstance* pInstance);
typedef VkResult (VKAPI *PFN_vkDestroyInstance)(VkInstance instance);
-typedef VkResult (VKAPI *PFN_vkEnumeratePhysicalDevices)(VkInstance instance, uint32_t* pPhysicalDeviceCount, VkPhysicalGpu* pPhysicalDevices);
-typedef VkResult (VKAPI *PFN_vkGetGpuInfo)(VkPhysicalGpu gpu, VkPhysicalGpuInfoType infoType, size_t* pDataSize, void* pData);
-typedef void * (VKAPI *PFN_vkGetProcAddr)(VkPhysicalGpu gpu, const char * pName);
-typedef VkResult (VKAPI *PFN_vkCreateDevice)(VkPhysicalGpu gpu, const VkDeviceCreateInfo* pCreateInfo, VkDevice* pDevice);
+typedef VkResult (VKAPI *PFN_vkEnumeratePhysicalDevices)(VkInstance instance, uint32_t* pPhysicalDeviceCount, VkPhysicalDevice* pPhysicalDevices);
+typedef VkResult (VKAPI *PFN_vkGetPhysicalDeviceInfo)(VkPhysicalDevice physicalDevice, VkPhysicalDeviceInfoType infoType, size_t* pDataSize, void* pData);
+typedef void * (VKAPI *PFN_vkGetProcAddr)(VkPhysicalDevice physicalDevice, const char * pName);
+typedef VkResult (VKAPI *PFN_vkCreateDevice)(VkPhysicalDevice physicalDevice, const VkDeviceCreateInfo* pCreateInfo, VkDevice* pDevice);
typedef VkResult (VKAPI *PFN_vkDestroyDevice)(VkDevice device);
typedef VkResult (VKAPI *PFN_vkGetGlobalExtensionInfo)(VkExtensionInfoType infoType, uint32_t extensionIndex, size_t* pDataSize, void* pData);
-typedef VkResult (VKAPI *PFN_vkGetPhysicalDeviceExtensionInfo)(VkPhysicalGpu gpu, VkExtensionInfoType infoType, uint32_t extensionIndex, size_t* pDataSize, void* pData);
-typedef VkResult (VKAPI *PFN_vkEnumerateLayers)(VkPhysicalGpu gpu, size_t maxLayerCount, size_t maxStringSize, size_t* pOutLayerCount, char* const* pOutLayers, void* pReserved);
+typedef VkResult (VKAPI *PFN_vkGetPhysicalDeviceExtensionInfo)(VkPhysicalDevice gpu, VkExtensionInfoType infoType, uint32_t extensionIndex, size_t* pDataSize, void* pData);
+typedef VkResult (VKAPI *PFN_vkGetExtensionSupport)(VkPhysicalDevice physicalDevice, const char* pExtName);
+typedef VkResult (VKAPI *PFN_vkEnumerateLayers)(VkPhysicalDevice physicalDevice, size_t maxLayerCount, size_t maxStringSize, size_t* pOutLayerCount, char* const* pOutLayers, void* pReserved);
typedef VkResult (VKAPI *PFN_vkGetDeviceQueue)(VkDevice device, uint32_t queueNodeIndex, uint32_t queueIndex, VkQueue* pQueue);
typedef VkResult (VKAPI *PFN_vkQueueSubmit)(VkQueue queue, uint32_t cmdBufferCount, const VkCmdBuffer* pCmdBuffers, VkFence fence);
-typedef VkResult (VKAPI *PFN_vkQueueAddMemReferences)(VkQueue queue, uint32_t count, const VkGpuMemory* pMems);
-typedef VkResult (VKAPI *PFN_vkQueueRemoveMemReferences)(VkQueue queue, uint32_t count, const VkGpuMemory* pMems);
+typedef VkResult (VKAPI *PFN_vkQueueAddMemReferences)(VkQueue queue, uint32_t count, const VkDeviceMemory* pMems);
+typedef VkResult (VKAPI *PFN_vkQueueRemoveMemReferences)(VkQueue queue, uint32_t count, const VkDeviceMemory* pMems);
typedef VkResult (VKAPI *PFN_vkQueueWaitIdle)(VkQueue queue);
typedef VkResult (VKAPI *PFN_vkDeviceWaitIdle)(VkDevice device);
-typedef VkResult (VKAPI *PFN_vkAllocMemory)(VkDevice device, const VkMemoryAllocInfo* pAllocInfo, VkGpuMemory* pMem);
-typedef VkResult (VKAPI *PFN_vkFreeMemory)(VkGpuMemory mem);
-typedef VkResult (VKAPI *PFN_vkSetMemoryPriority)(VkGpuMemory mem, VkMemoryPriority priority);
-typedef VkResult (VKAPI *PFN_vkMapMemory)(VkGpuMemory mem, VkFlags flags, void** ppData);
-typedef VkResult (VKAPI *PFN_vkUnmapMemory)(VkGpuMemory mem);
-typedef VkResult (VKAPI *PFN_vkPinSystemMemory)(VkDevice device, const void* pSysMem, size_t memSize, VkGpuMemory* pMem);
-typedef VkResult (VKAPI *PFN_vkGetMultiGpuCompatibility)(VkPhysicalGpu gpu0, VkPhysicalGpu gpu1, VkGpuCompatibilityInfo* pInfo);
-typedef VkResult (VKAPI *PFN_vkOpenSharedMemory)(VkDevice device, const VkMemoryOpenInfo* pOpenInfo, VkGpuMemory* pMem);
+typedef VkResult (VKAPI *PFN_vkAllocMemory)(VkDevice device, const VkMemoryAllocInfo* pAllocInfo, VkDeviceMemory* pMem);
+typedef VkResult (VKAPI *PFN_vkFreeMemory)(VkDeviceMemory mem);
+typedef VkResult (VKAPI *PFN_vkSetMemoryPriority)(VkDeviceMemory mem, VkMemoryPriority priority);
+typedef VkResult (VKAPI *PFN_vkMapMemory)(VkDeviceMemory mem, VkFlags flags, void** ppData);
+typedef VkResult (VKAPI *PFN_vkUnmapMemory)(VkDeviceMemory mem);
+typedef VkResult (VKAPI *PFN_vkPinSystemMemory)(VkDevice device, const void* pSysMem, size_t memSize, VkDeviceMemory* pMem);
+typedef VkResult (VKAPI *PFN_vkGetMultiDeviceCompatibility)(VkPhysicalDevice physicalDevice0, VkPhysicalDevice physicalDevice1, VkPhysicalDeviceCompatibilityInfo* pInfo);
+typedef VkResult (VKAPI *PFN_vkOpenSharedMemory)(VkDevice device, const VkMemoryOpenInfo* pOpenInfo, VkDeviceMemory* pMem);
typedef VkResult (VKAPI *PFN_vkOpenSharedSemaphore)(VkDevice device, const VkSemaphoreOpenInfo* pOpenInfo, VkSemaphore* pSemaphore);
-typedef VkResult (VKAPI *PFN_vkOpenPeerMemory)(VkDevice device, const VkPeerMemoryOpenInfo* pOpenInfo, VkGpuMemory* pMem);
-typedef VkResult (VKAPI *PFN_vkOpenPeerImage)(VkDevice device, const VkPeerImageOpenInfo* pOpenInfo, VkImage* pImage, VkGpuMemory* pMem);
+typedef VkResult (VKAPI *PFN_vkOpenPeerMemory)(VkDevice device, const VkPeerMemoryOpenInfo* pOpenInfo, VkDeviceMemory* pMem);
+typedef VkResult (VKAPI *PFN_vkOpenPeerImage)(VkDevice device, const VkPeerImageOpenInfo* pOpenInfo, VkImage* pImage, VkDeviceMemory* pMem);
typedef VkResult (VKAPI *PFN_vkDestroyObject)(VkObject object);
typedef VkResult (VKAPI *PFN_vkGetObjectInfo)(VkBaseObject object, VkObjectInfoType infoType, size_t* pDataSize, void* pData);
-typedef VkResult (VKAPI *PFN_vkQueueBindObjectMemory)(VkQueue queue, VkObject object, uint32_t allocationIdx, VkGpuMemory mem, VkGpuSize offset);
-typedef VkResult (VKAPI *PFN_vkQueueBindObjectMemoryRange)(VkQueue queue, VkObject object, uint32_t allocationIdx, VkGpuSize rangeOffset,VkGpuSize rangeSize, VkGpuMemory mem, VkGpuSize memOffset);
-typedef VkResult (VKAPI *PFN_vkQueueBindImageMemoryRange)(VkQueue queue, VkImage image, uint32_t allocationIdx, const VkImageMemoryBindInfo* pBindInfo, VkGpuMemory mem, VkGpuSize memOffset);
+typedef VkResult (VKAPI *PFN_vkQueueBindObjectMemory)(VkQueue queue, VkObject object, uint32_t allocationIdx, VkDeviceMemory mem, VkDeviceSize offset);
+typedef VkResult (VKAPI *PFN_vkQueueBindObjectMemoryRange)(VkQueue queue, VkObject object, uint32_t allocationIdx, VkDeviceSize rangeOffset,VkDeviceSize rangeSize, VkDeviceMemory mem, VkDeviceSize memOffset);
+typedef VkResult (VKAPI *PFN_vkQueueBindImageMemoryRange)(VkQueue queue, VkImage image, uint32_t allocationIdx, const VkImageMemoryBindInfo* pBindInfo, VkDeviceMemory mem, VkDeviceSize memOffset);
typedef VkResult (VKAPI *PFN_vkCreateFence)(VkDevice device, const VkFenceCreateInfo* pCreateInfo, VkFence* pFence);
typedef VkResult (VKAPI *PFN_vkResetFences)(VkDevice device, uint32_t fenceCount, VkFence* pFences);
typedef VkResult (VKAPI *PFN_vkGetFenceStatus)(VkFence fence);
@@ -2241,7 +2109,7 @@
typedef VkResult (VKAPI *PFN_vkSetEvent)(VkEvent event);
typedef VkResult (VKAPI *PFN_vkResetEvent)(VkEvent event);
typedef VkResult (VKAPI *PFN_vkCreateQueryPool)(VkDevice device, const VkQueryPoolCreateInfo* pCreateInfo, VkQueryPool* pQueryPool);
-typedef VkResult (VKAPI *PFN_vkGetQueryPoolResults)(VkQueryPool queryPool, uint32_t startQuery, uint32_t queryCount, size_t* pDataSize, void* pData);
+typedef VkResult (VKAPI *PFN_vkGetQueryPoolResults)(VkQueryPool queryPool, uint32_t startQuery, uint32_t queryCount, size_t* pDataSize, void* pData, VkQueryResultFlags flags);
typedef VkResult (VKAPI *PFN_vkGetFormatInfo)(VkDevice device, VkFormat format, VkFormatInfoType infoType, size_t* pDataSize, void* pData);
typedef VkResult (VKAPI *PFN_vkCreateBuffer)(VkDevice device, const VkBufferCreateInfo* pCreateInfo, VkBuffer* pBuffer);
typedef VkResult (VKAPI *PFN_vkCreateBufferView)(VkDevice device, const VkBufferViewCreateInfo* pCreateInfo, VkBufferView* pView);
@@ -2266,7 +2134,7 @@
typedef VkResult (VKAPI *PFN_vkResetDescriptorPool)(VkDescriptorPool descriptorPool);
typedef VkResult (VKAPI *PFN_vkAllocDescriptorSets)(VkDescriptorPool descriptorPool, VkDescriptorSetUsage setUsage, uint32_t count, const VkDescriptorSetLayout* pSetLayouts, VkDescriptorSet* pDescriptorSets, uint32_t* pCount);
typedef void (VKAPI *PFN_vkClearDescriptorSets)(VkDescriptorPool descriptorPool, uint32_t count, const VkDescriptorSet* pDescriptorSets);
-typedef void (VKAPI *PFN_vkUpdateDescriptors)(VkDescriptorSet descriptorSet, uint32_t updateCount, const void** pUpdateArray);
+typedef void (VKAPI *PFN_vkUpdateDescriptors)(VkDescriptorSet descriptorSet, uint32_t updateCount, const void** ppUpdateArray);
typedef VkResult (VKAPI *PFN_vkCreateDynamicViewportState)(VkDevice device, const VkDynamicVpStateCreateInfo* pCreateInfo, VkDynamicVpState* pState);
typedef VkResult (VKAPI *PFN_vkCreateDynamicRasterState)(VkDevice device, const VkDynamicRsStateCreateInfo* pCreateInfo, VkDynamicRsState* pState);
typedef VkResult (VKAPI *PFN_vkCreateDynamicColorBlendState)(VkDevice device, const VkDynamicCbStateCreateInfo* pCreateInfo, VkDynamicCbState* pState);
@@ -2278,37 +2146,37 @@
typedef void (VKAPI *PFN_vkCmdBindPipeline)(VkCmdBuffer cmdBuffer, VkPipelineBindPoint pipelineBindPoint, VkPipeline pipeline);
typedef void (VKAPI *PFN_vkCmdBindDynamicStateObject)(VkCmdBuffer cmdBuffer, VkStateBindPoint stateBindPoint, VkDynamicStateObject state);
typedef void (VKAPI *PFN_vkCmdBindDescriptorSets)(VkCmdBuffer cmdBuffer, VkPipelineBindPoint pipelineBindPoint, VkDescriptorSetLayoutChain layoutChain, uint32_t layoutChainSlot, uint32_t count, const VkDescriptorSet* pDescriptorSets, const uint32_t* pUserData);
-typedef void (VKAPI *PFN_vkCmdBindIndexBuffer)(VkCmdBuffer cmdBuffer, VkBuffer buffer, VkGpuSize offset, VkIndexType indexType);
-typedef void (VKAPI *PFN_vkCmdBindVertexBuffers)(VkCmdBuffer cmdBuffer, uint32_t startBinding, uint32_t bindingCount, const VkBuffer* pBuffers, const VkGpuSize* pOffsets);
+typedef void (VKAPI *PFN_vkCmdBindIndexBuffer)(VkCmdBuffer cmdBuffer, VkBuffer buffer, VkDeviceSize offset, VkIndexType indexType);
+typedef void (VKAPI *PFN_vkCmdBindVertexBuffers)(VkCmdBuffer cmdBuffer, uint32_t startBinding, uint32_t bindingCount, const VkBuffer* pBuffers, const VkDeviceSize* pOffsets);
typedef void (VKAPI *PFN_vkCmdDraw)(VkCmdBuffer cmdBuffer, uint32_t firstVertex, uint32_t vertexCount, uint32_t firstInstance, uint32_t instanceCount);
typedef void (VKAPI *PFN_vkCmdDrawIndexed)(VkCmdBuffer cmdBuffer, uint32_t firstIndex, uint32_t indexCount, int32_t vertexOffset, uint32_t firstInstance, uint32_t instanceCount);
-typedef void (VKAPI *PFN_vkCmdDrawIndirect)(VkCmdBuffer cmdBuffer, VkBuffer buffer, VkGpuSize offset, uint32_t count, uint32_t stride);
-typedef void (VKAPI *PFN_vkCmdDrawIndexedIndirect)(VkCmdBuffer cmdBuffer, VkBuffer buffer, VkGpuSize offset, uint32_t count, uint32_t stride);
+typedef void (VKAPI *PFN_vkCmdDrawIndirect)(VkCmdBuffer cmdBuffer, VkBuffer buffer, VkDeviceSize offset, uint32_t count, uint32_t stride);
+typedef void (VKAPI *PFN_vkCmdDrawIndexedIndirect)(VkCmdBuffer cmdBuffer, VkBuffer buffer, VkDeviceSize offset, uint32_t count, uint32_t stride);
typedef void (VKAPI *PFN_vkCmdDispatch)(VkCmdBuffer cmdBuffer, uint32_t x, uint32_t y, uint32_t z);
-typedef void (VKAPI *PFN_vkCmdDispatchIndirect)(VkCmdBuffer cmdBuffer, VkBuffer buffer, VkGpuSize offset);
+typedef void (VKAPI *PFN_vkCmdDispatchIndirect)(VkCmdBuffer cmdBuffer, VkBuffer buffer, VkDeviceSize offset);
typedef void (VKAPI *PFN_vkCmdCopyBuffer)(VkCmdBuffer cmdBuffer, VkBuffer srcBuffer, VkBuffer destBuffer, uint32_t regionCount, const VkBufferCopy* pRegions);
typedef void (VKAPI *PFN_vkCmdCopyImage)(VkCmdBuffer cmdBuffer, VkImage srcImage, VkImageLayout srcImageLayout, VkImage destImage, VkImageLayout destImageLayout, uint32_t regionCount, const VkImageCopy* pRegions);
typedef void (VKAPI *PFN_vkCmdBlitImage)(VkCmdBuffer cmdBuffer, VkImage srcImage, VkImageLayout srcImageLayout, VkImage destImage, VkImageLayout destImageLayout, uint32_t regionCount, const VkImageBlit* pRegions);
typedef void (VKAPI *PFN_vkCmdCopyBufferToImage)(VkCmdBuffer cmdBuffer, VkBuffer srcBuffer, VkImage destImage, VkImageLayout destImageLayout, uint32_t regionCount, const VkBufferImageCopy* pRegions);
typedef void (VKAPI *PFN_vkCmdCopyImageToBuffer)(VkCmdBuffer cmdBuffer, VkImage srcImage, VkImageLayout srcImageLayout, VkBuffer destBuffer, uint32_t regionCount, const VkBufferImageCopy* pRegions);
typedef void (VKAPI *PFN_vkCmdCloneImageData)(VkCmdBuffer cmdBuffer, VkImage srcImage, VkImageLayout srcImageLayout, VkImage destImage, VkImageLayout destImageLayout);
-typedef void (VKAPI *PFN_vkCmdUpdateBuffer)(VkCmdBuffer cmdBuffer, VkBuffer destBuffer, VkGpuSize destOffset, VkGpuSize dataSize, const uint32_t* pData);
-typedef void (VKAPI *PFN_vkCmdFillBuffer)(VkCmdBuffer cmdBuffer, VkBuffer destBuffer, VkGpuSize destOffset, VkGpuSize fillSize, uint32_t data);
+typedef void (VKAPI *PFN_vkCmdUpdateBuffer)(VkCmdBuffer cmdBuffer, VkBuffer destBuffer, VkDeviceSize destOffset, VkDeviceSize dataSize, const uint32_t* pData);
+typedef void (VKAPI *PFN_vkCmdFillBuffer)(VkCmdBuffer cmdBuffer, VkBuffer destBuffer, VkDeviceSize destOffset, VkDeviceSize fillSize, uint32_t data);
typedef void (VKAPI *PFN_vkCmdClearColorImage)(VkCmdBuffer cmdBuffer, VkImage image, VkImageLayout imageLayout, VkClearColor color, uint32_t rangeCount, const VkImageSubresourceRange* pRanges);
typedef void (VKAPI *PFN_vkCmdClearDepthStencil)(VkCmdBuffer cmdBuffer, VkImage image, VkImageLayout imageLayout, float depth, uint32_t stencil, uint32_t rangeCount, const VkImageSubresourceRange* pRanges);
typedef void (VKAPI *PFN_vkCmdResolveImage)(VkCmdBuffer cmdBuffer, VkImage srcImage, VkImageLayout srcImageLayout, VkImage destImage, VkImageLayout destImageLayout, uint32_t regionCount, const VkImageResolve* pRegions);
typedef void (VKAPI *PFN_vkCmdSetEvent)(VkCmdBuffer cmdBuffer, VkEvent event, VkPipeEvent pipeEvent);
typedef void (VKAPI *PFN_vkCmdResetEvent)(VkCmdBuffer cmdBuffer, VkEvent event, VkPipeEvent pipeEvent);
-typedef void (VKAPI *PFN_vkCmdWaitEvents)(VkCmdBuffer cmdBuffer, const VkEventWaitInfo* pWaitInfo);
-typedef void (VKAPI *PFN_vkCmdPipelineBarrier)(VkCmdBuffer cmdBuffer, const VkPipelineBarrier* pBarrier);
-typedef void (VKAPI *PFN_vkCmdBeginQuery)(VkCmdBuffer cmdBuffer, VkQueryPool queryPool, uint32_t slot, VkFlags flags);
+typedef void (VKAPI *PFN_vkCmdWaitEvents)(VkCmdBuffer cmdBuffer, VkWaitEvent waitEvent, uint32_t eventCount, const VkEvent* pEvents, uint32_t memBarrierCount, const void** ppMemBarriers);
+typedef void (VKAPI *PFN_vkCmdPipelineBarrier)(VkCmdBuffer cmdBuffer, VkWaitEvent waitEvent, uint32_t pipeEventCount, const VkPipeEvent* pPipeEvents, uint32_t memBarrierCount, const void** ppMemBarriers);
+typedef void (VKAPI *PFN_vkCmdBeginQuery)(VkCmdBuffer cmdBuffer, VkQueryPool queryPool, uint32_t slot, VkQueryControlFlags flags);
typedef void (VKAPI *PFN_vkCmdEndQuery)(VkCmdBuffer cmdBuffer, VkQueryPool queryPool, uint32_t slot);
typedef void (VKAPI *PFN_vkCmdResetQueryPool)(VkCmdBuffer cmdBuffer, VkQueryPool queryPool, uint32_t startQuery, uint32_t queryCount);
-typedef void (VKAPI *PFN_vkCmdWriteTimestamp)(VkCmdBuffer cmdBuffer, VkTimestampType timestampType, VkBuffer destBuffer, VkGpuSize destOffset);
-typedef void (VKAPI *PFN_vkCmdCopyQueryPoolResults)(VkCmdBuffer cmdBuffer, VkQueryPool queryPool, uint32_t startQuery, uint32_t queryCount, VkBuffer destBuffer, VkGpuSize destOffset, VkGpuSize destStride, VkFlags flags);
+typedef void (VKAPI *PFN_vkCmdWriteTimestamp)(VkCmdBuffer cmdBuffer, VkTimestampType timestampType, VkBuffer destBuffer, VkDeviceSize destOffset);
+typedef void (VKAPI *PFN_vkCmdCopyQueryPoolResults)(VkCmdBuffer cmdBuffer, VkQueryPool queryPool, uint32_t startQuery, uint32_t queryCount, VkBuffer destBuffer, VkDeviceSize destOffset, VkDeviceSize destStride, VkFlags flags);
typedef void (VKAPI *PFN_vkCmdInitAtomicCounters)(VkCmdBuffer cmdBuffer, VkPipelineBindPoint pipelineBindPoint, uint32_t startCounter, uint32_t counterCount, const uint32_t* pData);
-typedef void (VKAPI *PFN_vkCmdLoadAtomicCounters)(VkCmdBuffer cmdBuffer, VkPipelineBindPoint pipelineBindPoint, uint32_t startCounter, uint32_t counterCount, VkBuffer srcBuffer, VkGpuSize srcOffset);
-typedef void (VKAPI *PFN_vkCmdSaveAtomicCounters)(VkCmdBuffer cmdBuffer, VkPipelineBindPoint pipelineBindPoint, uint32_t startCounter, uint32_t counterCount, VkBuffer destBuffer, VkGpuSize destOffset);
+typedef void (VKAPI *PFN_vkCmdLoadAtomicCounters)(VkCmdBuffer cmdBuffer, VkPipelineBindPoint pipelineBindPoint, uint32_t startCounter, uint32_t counterCount, VkBuffer srcBuffer, VkDeviceSize srcOffset);
+typedef void (VKAPI *PFN_vkCmdSaveAtomicCounters)(VkCmdBuffer cmdBuffer, VkPipelineBindPoint pipelineBindPoint, uint32_t startCounter, uint32_t counterCount, VkBuffer destBuffer, VkDeviceSize destOffset);
typedef VkResult (VKAPI *PFN_vkCreateFramebuffer)(VkDevice device, const VkFramebufferCreateInfo* pCreateInfo, VkFramebuffer* pFramebuffer);
typedef VkResult (VKAPI *PFN_vkCreateRenderPass)(VkDevice device, const VkRenderPassCreateInfo* pCreateInfo, VkRenderPass* pRenderPass);
typedef void (VKAPI *PFN_vkCmdBeginRenderPass)(VkCmdBuffer cmdBuffer, const VkRenderPassBegin* pRenderPassBegin);
@@ -2316,7 +2184,7 @@
#ifdef VK_PROTOTYPES
-// GPU initialization
+// Device initialization
VkResult VKAPI vkCreateInstance(
const VkInstanceCreateInfo* pCreateInfo,
@@ -2328,22 +2196,22 @@
VkResult VKAPI vkEnumeratePhysicalDevices(
VkInstance instance,
uint32_t* pPhysicalDeviceCount,
- VkPhysicalGpu* pPhysicalDevices);
+ VkPhysicalDevice* pPhysicalDevices);
-VkResult VKAPI vkGetGpuInfo(
- VkPhysicalGpu gpu,
- VkPhysicalGpuInfoType infoType,
+VkResult VKAPI vkGetPhysicalDeviceInfo(
+ VkPhysicalDevice physicalDevice,
+ VkPhysicalDeviceInfoType infoType,
size_t* pDataSize,
void* pData);
void * VKAPI vkGetProcAddr(
- VkPhysicalGpu gpu,
+ VkPhysicalDevice physicalDevice,
const char* pName);
// Device functions
VkResult VKAPI vkCreateDevice(
- VkPhysicalGpu gpu,
+ VkPhysicalDevice physicalDevice,
const VkDeviceCreateInfo* pCreateInfo,
VkDevice* pDevice);
@@ -2358,7 +2226,7 @@
void* pData);
VkResult VKAPI vkGetPhysicalDeviceExtensionInfo(
- VkPhysicalGpu gpu,
+ VkPhysicalDevice gpu,
VkExtensionInfoType infoType,
uint32_t extensionIndex,
size_t* pDataSize,
@@ -2367,7 +2235,7 @@
// Layer discovery functions
VkResult VKAPI vkEnumerateLayers(
- VkPhysicalGpu gpu,
+ VkPhysicalDevice physicalDevice,
size_t maxLayerCount,
size_t maxStringSize,
size_t* pOutLayerCount,
@@ -2391,12 +2259,12 @@
VkResult VKAPI vkQueueAddMemReferences(
VkQueue queue,
uint32_t count,
- const VkGpuMemory* pMems);
+ const VkDeviceMemory* pMems);
VkResult VKAPI vkQueueRemoveMemReferences(
VkQueue queue,
uint32_t count,
- const VkGpuMemory* pMems);
+ const VkDeviceMemory* pMems);
VkResult VKAPI vkQueueWaitIdle(
VkQueue queue);
@@ -2409,40 +2277,40 @@
VkResult VKAPI vkAllocMemory(
VkDevice device,
const VkMemoryAllocInfo* pAllocInfo,
- VkGpuMemory* pMem);
+ VkDeviceMemory* pMem);
VkResult VKAPI vkFreeMemory(
- VkGpuMemory mem);
+ VkDeviceMemory mem);
VkResult VKAPI vkSetMemoryPriority(
- VkGpuMemory mem,
+ VkDeviceMemory mem,
VkMemoryPriority priority);
VkResult VKAPI vkMapMemory(
- VkGpuMemory mem,
- VkFlags flags, // Reserved
+ VkDeviceMemory mem,
+ VkMemoryMapFlags flags,
void** ppData);
VkResult VKAPI vkUnmapMemory(
- VkGpuMemory mem);
+ VkDeviceMemory mem);
VkResult VKAPI vkPinSystemMemory(
VkDevice device,
const void* pSysMem,
size_t memSize,
- VkGpuMemory* pMem);
+ VkDeviceMemory* pMem);
// Multi-device functions
-VkResult VKAPI vkGetMultiGpuCompatibility(
- VkPhysicalGpu gpu0,
- VkPhysicalGpu gpu1,
- VkGpuCompatibilityInfo* pInfo);
+VkResult VKAPI vkGetMultiDeviceCompatibility(
+ VkPhysicalDevice physicalDevice0,
+ VkPhysicalDevice physicalDevice1,
+ VkPhysicalDeviceCompatibilityInfo* pInfo);
VkResult VKAPI vkOpenSharedMemory(
VkDevice device,
const VkMemoryOpenInfo* pOpenInfo,
- VkGpuMemory* pMem);
+ VkDeviceMemory* pMem);
VkResult VKAPI vkOpenSharedSemaphore(
VkDevice device,
@@ -2452,13 +2320,13 @@
VkResult VKAPI vkOpenPeerMemory(
VkDevice device,
const VkPeerMemoryOpenInfo* pOpenInfo,
- VkGpuMemory* pMem);
+ VkDeviceMemory* pMem);
VkResult VKAPI vkOpenPeerImage(
VkDevice device,
const VkPeerImageOpenInfo* pOpenInfo,
VkImage* pImage,
- VkGpuMemory* pMem);
+ VkDeviceMemory* pMem);
// Generic API object functions
@@ -2477,25 +2345,25 @@
VkQueue queue,
VkObject object,
uint32_t allocationIdx,
- VkGpuMemory mem,
- VkGpuSize memOffset);
+ VkDeviceMemory mem,
+ VkDeviceSize memOffset);
VkResult VKAPI vkQueueBindObjectMemoryRange(
VkQueue queue,
VkObject object,
uint32_t allocationIdx,
- VkGpuSize rangeOffset,
- VkGpuSize rangeSize,
- VkGpuMemory mem,
- VkGpuSize memOffset);
+ VkDeviceSize rangeOffset,
+ VkDeviceSize rangeSize,
+ VkDeviceMemory mem,
+ VkDeviceSize memOffset);
VkResult VKAPI vkQueueBindImageMemoryRange(
VkQueue queue,
VkImage image,
uint32_t allocationIdx,
const VkImageMemoryBindInfo* pBindInfo,
- VkGpuMemory mem,
- VkGpuSize memOffset);
+ VkDeviceMemory mem,
+ VkDeviceSize memOffset);
// Fence functions
@@ -2562,7 +2430,8 @@
uint32_t startQuery,
uint32_t queryCount,
size_t* pDataSize,
- void* pData);
+ void* pData,
+ VkQueryResultFlags flags);
// Format capabilities
@@ -2715,7 +2584,7 @@
void VKAPI vkUpdateDescriptors(
VkDescriptorSet descriptorSet,
uint32_t updateCount,
- const void** pUpdateArray);
+ const void** ppUpdateArray);
// State object functions
@@ -2775,12 +2644,12 @@
uint32_t layoutChainSlot,
uint32_t count,
const VkDescriptorSet* pDescriptorSets,
- const uint32_t* pUserData);
+ const uint32_t * pUserData);
void VKAPI vkCmdBindIndexBuffer(
VkCmdBuffer cmdBuffer,
VkBuffer buffer,
- VkGpuSize offset,
+ VkDeviceSize offset,
VkIndexType indexType);
void VKAPI vkCmdBindVertexBuffers(
@@ -2788,7 +2657,7 @@
uint32_t startBinding,
uint32_t bindingCount,
const VkBuffer* pBuffers,
- const VkGpuSize* pOffsets);
+ const VkDeviceSize* pOffsets);
void VKAPI vkCmdDraw(
VkCmdBuffer cmdBuffer,
@@ -2808,14 +2677,14 @@
void VKAPI vkCmdDrawIndirect(
VkCmdBuffer cmdBuffer,
VkBuffer buffer,
- VkGpuSize offset,
+ VkDeviceSize offset,
uint32_t count,
uint32_t stride);
void VKAPI vkCmdDrawIndexedIndirect(
VkCmdBuffer cmdBuffer,
VkBuffer buffer,
- VkGpuSize offset,
+ VkDeviceSize offset,
uint32_t count,
uint32_t stride);
@@ -2828,7 +2697,7 @@
void VKAPI vkCmdDispatchIndirect(
VkCmdBuffer cmdBuffer,
VkBuffer buffer,
- VkGpuSize offset);
+ VkDeviceSize offset);
void VKAPI vkCmdCopyBuffer(
VkCmdBuffer cmdBuffer,
@@ -2881,15 +2750,15 @@
void VKAPI vkCmdUpdateBuffer(
VkCmdBuffer cmdBuffer,
VkBuffer destBuffer,
- VkGpuSize destOffset,
- VkGpuSize dataSize,
+ VkDeviceSize destOffset,
+ VkDeviceSize dataSize,
const uint32_t* pData);
void VKAPI vkCmdFillBuffer(
VkCmdBuffer cmdBuffer,
VkBuffer destBuffer,
- VkGpuSize destOffset,
- VkGpuSize fillSize,
+ VkDeviceSize destOffset,
+ VkDeviceSize fillSize,
uint32_t data);
void VKAPI vkCmdClearColorImage(
@@ -2930,17 +2799,25 @@
void VKAPI vkCmdWaitEvents(
VkCmdBuffer cmdBuffer,
- const VkEventWaitInfo* pWaitInfo);
+ VkWaitEvent waitEvent,
+ uint32_t eventCount,
+ const VkEvent* pEvents,
+ uint32_t memBarrierCount,
+ const void** ppMemBarriers);
void VKAPI vkCmdPipelineBarrier(
VkCmdBuffer cmdBuffer,
- const VkPipelineBarrier* pBarrier);
+ VkWaitEvent waitEvent,
+ uint32_t pipeEventCount,
+ const VkPipeEvent* pPipeEvents,
+ uint32_t memBarrierCount,
+ const void** ppMemBarriers);
void VKAPI vkCmdBeginQuery(
VkCmdBuffer cmdBuffer,
VkQueryPool queryPool,
uint32_t slot,
- VkFlags flags);
+ VkQueryControlFlags flags);
void VKAPI vkCmdEndQuery(
VkCmdBuffer cmdBuffer,
@@ -2957,7 +2834,7 @@
VkCmdBuffer cmdBuffer,
VkTimestampType timestampType,
VkBuffer destBuffer,
- VkGpuSize destOffset);
+ VkDeviceSize destOffset);
void VKAPI vkCmdCopyQueryPoolResults(
VkCmdBuffer cmdBuffer,
@@ -2965,8 +2842,8 @@
uint32_t startQuery,
uint32_t queryCount,
VkBuffer destBuffer,
- VkGpuSize destOffset,
- VkGpuSize destStride,
+ VkDeviceSize destOffset,
+ VkDeviceSize destStride,
VkFlags flags); // VkQueryResultFlags
void VKAPI vkCmdInitAtomicCounters(
@@ -2982,7 +2859,7 @@
uint32_t startCounter,
uint32_t counterCount,
VkBuffer srcBuffer,
- VkGpuSize srcOffset);
+ VkDeviceSize srcOffset);
void VKAPI vkCmdSaveAtomicCounters(
VkCmdBuffer cmdBuffer,
@@ -2990,7 +2867,7 @@
uint32_t startCounter,
uint32_t counterCount,
VkBuffer destBuffer,
- VkGpuSize destOffset);
+ VkDeviceSize destOffset);
VkResult VKAPI vkCreateFramebuffer(
VkDevice device,
diff --git a/layers/basic.cpp b/layers/basic.cpp
index eddb9fc..b7ad6b7 100644
--- a/layers/basic.cpp
+++ b/layers/basic.cpp
@@ -49,7 +49,7 @@
return it->second;
}
- layer_initialize_dispatch_table(pTable, gpuw->pGPA, (VkPhysicalGpu) gpuw->nextObject);
+ layer_initialize_dispatch_table(pTable, gpuw->pGPA, (VkPhysicalDevice) gpuw->nextObject);
return pTable;
}
@@ -114,7 +114,7 @@
return VK_SUCCESS;
}
-VK_LAYER_EXPORT VkResult VKAPI vkCreateDevice(VkPhysicalGpu gpu, const VkDeviceCreateInfo* pCreateInfo, VkDevice* pDevice)
+VK_LAYER_EXPORT VkResult VKAPI vkCreateDevice(VkPhysicalDevice gpu, const VkDeviceCreateInfo* pCreateInfo, VkDevice* pDevice)
{
VkLayerDispatchTable* pTable = tableMap[gpu];
@@ -135,7 +135,7 @@
return result;
}
-VK_LAYER_EXPORT VkResult VKAPI vkEnumerateLayers(VkPhysicalGpu gpu, size_t maxLayerCount, size_t maxStringSize, size_t* pOutLayerCount, char* const* pOutLayers, void* pReserved)
+VK_LAYER_EXPORT VkResult VKAPI vkEnumerateLayers(VkPhysicalDevice gpu, size_t maxLayerCount, size_t maxStringSize, size_t* pOutLayerCount, char* const* pOutLayers, void* pReserved)
{
if (gpu != NULL)
{
@@ -151,11 +151,11 @@
// Example of a layer that is only compatible with Intel's GPUs
VkBaseLayerObject* gpuw = (VkBaseLayerObject*) pReserved;
- PFN_vkGetGpuInfo fpGetGpuInfo;
- VkPhysicalGpuProperties gpuProps;
- size_t dataSize = sizeof(VkPhysicalGpuProperties);
- fpGetGpuInfo = (PFN_vkGetGpuInfo) gpuw->pGPA((VkPhysicalGpu) gpuw->nextObject, "vkGetGpuInfo");
- fpGetGpuInfo((VkPhysicalGpu) gpuw->nextObject, VK_INFO_TYPE_PHYSICAL_GPU_PROPERTIES, &dataSize, &gpuProps);
+ PFN_vkGetPhysicalDeviceInfo fpGetGpuInfo;
+ VkPhysicalDeviceProperties gpuProps;
+ size_t dataSize = sizeof(VkPhysicalDeviceProperties);
+ fpGetGpuInfo = (PFN_vkGetPhysicalDeviceInfo) gpuw->pGPA((VkPhysicalDevice) gpuw->nextObject, "vkGetPhysicalDeviceInfo");
+ fpGetGpuInfo((VkPhysicalDevice) gpuw->nextObject, VK_PHYSICAL_DEVICE_INFO_TYPE_PROPERTIES, &dataSize, &gpuProps);
if (gpuProps.vendorId == 0x8086)
{
*pOutLayerCount = 1;
@@ -168,7 +168,7 @@
}
}
-VK_LAYER_EXPORT void * VKAPI vkGetProcAddr(VkPhysicalGpu gpu, const char* pName)
+VK_LAYER_EXPORT void * VKAPI vkGetProcAddr(VkPhysicalDevice gpu, const char* pName)
{
if (gpu == NULL)
return NULL;
@@ -189,6 +189,6 @@
VkBaseLayerObject* gpuw = (VkBaseLayerObject *) gpu;
if (gpuw->pGPA == NULL)
return NULL;
- return gpuw->pGPA((VkPhysicalGpu) gpuw->nextObject, pName);
+ return gpuw->pGPA((VkPhysicalDevice) gpuw->nextObject, pName);
}
}
diff --git a/layers/draw_state.cpp b/layers/draw_state.cpp
index 01e4e55..4d6620a 100644
--- a/layers/draw_state.cpp
+++ b/layers/draw_state.cpp
@@ -1424,7 +1424,7 @@
fpNextGPA = pCurObj->pGPA;
assert(fpNextGPA);
- layer_initialize_dispatch_table(&nextTable, fpNextGPA, (VkPhysicalGpu) pCurObj->nextObject);
+ layer_initialize_dispatch_table(&nextTable, fpNextGPA, (VkPhysicalDevice) pCurObj->nextObject);
if (!globalLockInitialized)
{
@@ -1438,7 +1438,7 @@
}
}
-VK_LAYER_EXPORT VkResult VKAPI vkCreateDevice(VkPhysicalGpu gpu, const VkDeviceCreateInfo* pCreateInfo, VkDevice* pDevice)
+VK_LAYER_EXPORT VkResult VKAPI vkCreateDevice(VkPhysicalDevice gpu, const VkDeviceCreateInfo* pCreateInfo, VkDevice* pDevice)
{
pCurObj = (VkBaseLayerObject *) gpu;
loader_platform_thread_once(&g_initOnce, initDrawState);
@@ -1519,7 +1519,7 @@
return VK_SUCCESS;
}
-VK_LAYER_EXPORT VkResult VKAPI vkEnumerateLayers(VkPhysicalGpu gpu, size_t maxLayerCount, size_t maxStringSize, size_t* pOutLayerCount, char* const* pOutLayers, void* pReserved)
+VK_LAYER_EXPORT VkResult VKAPI vkEnumerateLayers(VkPhysicalDevice gpu, size_t maxLayerCount, size_t maxStringSize, size_t* pOutLayerCount, char* const* pOutLayers, void* pReserved)
{
if (gpu != NULL)
{
@@ -1864,28 +1864,28 @@
VK_LAYER_EXPORT VkResult VKAPI vkCreateDynamicViewportState(VkDevice device, const VkDynamicVpStateCreateInfo* pCreateInfo, VkDynamicVpState* pState)
{
VkResult result = nextTable.CreateDynamicViewportState(device, pCreateInfo, pState);
- insertDynamicState(*pState, (GENERIC_HEADER*)pCreateInfo, VK_STATE_BIND_VIEWPORT);
+ insertDynamicState(*pState, (GENERIC_HEADER*)pCreateInfo, VK_STATE_BIND_POINT_VIEWPORT);
return result;
}
VK_LAYER_EXPORT VkResult VKAPI vkCreateDynamicRasterState(VkDevice device, const VkDynamicRsStateCreateInfo* pCreateInfo, VkDynamicRsState* pState)
{
VkResult result = nextTable.CreateDynamicRasterState(device, pCreateInfo, pState);
- insertDynamicState(*pState, (GENERIC_HEADER*)pCreateInfo, VK_STATE_BIND_RASTER);
+ insertDynamicState(*pState, (GENERIC_HEADER*)pCreateInfo, VK_STATE_BIND_POINT_RASTER);
return result;
}
VK_LAYER_EXPORT VkResult VKAPI vkCreateDynamicColorBlendState(VkDevice device, const VkDynamicCbStateCreateInfo* pCreateInfo, VkDynamicCbState* pState)
{
VkResult result = nextTable.CreateDynamicColorBlendState(device, pCreateInfo, pState);
- insertDynamicState(*pState, (GENERIC_HEADER*)pCreateInfo, VK_STATE_BIND_COLOR_BLEND);
+ insertDynamicState(*pState, (GENERIC_HEADER*)pCreateInfo, VK_STATE_BIND_POINT_COLOR_BLEND);
return result;
}
VK_LAYER_EXPORT VkResult VKAPI vkCreateDynamicDepthStencilState(VkDevice device, const VkDynamicDsStateCreateInfo* pCreateInfo, VkDynamicDsState* pState)
{
VkResult result = nextTable.CreateDynamicDepthStencilState(device, pCreateInfo, pState);
- insertDynamicState(*pState, (GENERIC_HEADER*)pCreateInfo, VK_STATE_BIND_DEPTH_STENCIL);
+ insertDynamicState(*pState, (GENERIC_HEADER*)pCreateInfo, VK_STATE_BIND_POINT_DEPTH_STENCIL);
return result;
}
@@ -2037,7 +2037,7 @@
nextTable.CmdBindDescriptorSets(cmdBuffer, pipelineBindPoint, layoutChain, layoutChainSlot, count, pDescriptorSets, pUserData);
}
-VK_LAYER_EXPORT void VKAPI vkCmdBindIndexBuffer(VkCmdBuffer cmdBuffer, VkBuffer buffer, VkGpuSize offset, VkIndexType indexType)
+VK_LAYER_EXPORT void VKAPI vkCmdBindIndexBuffer(VkCmdBuffer cmdBuffer, VkBuffer buffer, VkDeviceSize offset, VkIndexType indexType)
{
GLOBAL_CB_NODE* pCB = getCBNode(cmdBuffer);
if (pCB) {
@@ -2058,7 +2058,7 @@
uint32_t startBinding,
uint32_t bindingCount,
const VkBuffer* pBuffers,
- const VkGpuSize* pOffsets)
+ const VkDeviceSize* pOffsets)
{
GLOBAL_CB_NODE* pCB = getCBNode(cmdBuffer);
if (pCB) {
@@ -2115,7 +2115,7 @@
nextTable.CmdDrawIndexed(cmdBuffer, firstIndex, indexCount, vertexOffset, firstInstance, instanceCount);
}
-VK_LAYER_EXPORT void VKAPI vkCmdDrawIndirect(VkCmdBuffer cmdBuffer, VkBuffer buffer, VkGpuSize offset, uint32_t count, uint32_t stride)
+VK_LAYER_EXPORT void VKAPI vkCmdDrawIndirect(VkCmdBuffer cmdBuffer, VkBuffer buffer, VkDeviceSize offset, uint32_t count, uint32_t stride)
{
GLOBAL_CB_NODE* pCB = getCBNode(cmdBuffer);
if (pCB) {
@@ -2135,7 +2135,7 @@
nextTable.CmdDrawIndirect(cmdBuffer, buffer, offset, count, stride);
}
-VK_LAYER_EXPORT void VKAPI vkCmdDrawIndexedIndirect(VkCmdBuffer cmdBuffer, VkBuffer buffer, VkGpuSize offset, uint32_t count, uint32_t stride)
+VK_LAYER_EXPORT void VKAPI vkCmdDrawIndexedIndirect(VkCmdBuffer cmdBuffer, VkBuffer buffer, VkDeviceSize offset, uint32_t count, uint32_t stride)
{
GLOBAL_CB_NODE* pCB = getCBNode(cmdBuffer);
if (pCB) {
@@ -2170,7 +2170,7 @@
nextTable.CmdDispatch(cmdBuffer, x, y, z);
}
-VK_LAYER_EXPORT void VKAPI vkCmdDispatchIndirect(VkCmdBuffer cmdBuffer, VkBuffer buffer, VkGpuSize offset)
+VK_LAYER_EXPORT void VKAPI vkCmdDispatchIndirect(VkCmdBuffer cmdBuffer, VkBuffer buffer, VkDeviceSize offset)
{
GLOBAL_CB_NODE* pCB = getCBNode(cmdBuffer);
if (pCB) {
@@ -2289,7 +2289,7 @@
nextTable.CmdCloneImageData(cmdBuffer, srcImage, srcImageLayout, destImage, destImageLayout);
}
-VK_LAYER_EXPORT void VKAPI vkCmdUpdateBuffer(VkCmdBuffer cmdBuffer, VkBuffer destBuffer, VkGpuSize destOffset, VkGpuSize dataSize, const uint32_t* pData)
+VK_LAYER_EXPORT void VKAPI vkCmdUpdateBuffer(VkCmdBuffer cmdBuffer, VkBuffer destBuffer, VkDeviceSize destOffset, VkDeviceSize dataSize, const uint32_t* pData)
{
GLOBAL_CB_NODE* pCB = getCBNode(cmdBuffer);
if (pCB) {
@@ -2304,7 +2304,7 @@
nextTable.CmdUpdateBuffer(cmdBuffer, destBuffer, destOffset, dataSize, pData);
}
-VK_LAYER_EXPORT void VKAPI vkCmdFillBuffer(VkCmdBuffer cmdBuffer, VkBuffer destBuffer, VkGpuSize destOffset, VkGpuSize fillSize, uint32_t data)
+VK_LAYER_EXPORT void VKAPI vkCmdFillBuffer(VkCmdBuffer cmdBuffer, VkBuffer destBuffer, VkDeviceSize destOffset, VkDeviceSize fillSize, uint32_t data)
{
GLOBAL_CB_NODE* pCB = getCBNode(cmdBuffer);
if (pCB) {
@@ -2403,7 +2403,7 @@
nextTable.CmdResetEvent(cmdBuffer, event, pipeEvent);
}
-VK_LAYER_EXPORT void VKAPI vkCmdWaitEvents(VkCmdBuffer cmdBuffer, const VkEventWaitInfo* pWaitInfo)
+VK_LAYER_EXPORT void VKAPI vkCmdWaitEvents(VkCmdBuffer cmdBuffer, VkWaitEvent waitEvent, uint32_t eventCount, const VkEvent* pEvents, uint32_t memBarrierCount, const void** ppMemBarriers)
{
GLOBAL_CB_NODE* pCB = getCBNode(cmdBuffer);
if (pCB) {
@@ -2415,10 +2415,10 @@
sprintf(str, "Attempt to use CmdBuffer %p that doesn't exist!", (void*)cmdBuffer);
layerCbMsg(VK_DBG_MSG_ERROR, VK_VALIDATION_LEVEL_0, cmdBuffer, 0, DRAWSTATE_INVALID_CMD_BUFFER, "DS", str);
}
- nextTable.CmdWaitEvents(cmdBuffer, pWaitInfo);
+ nextTable.CmdWaitEvents(cmdBuffer, waitEvent, eventCount, pEvents, memBarrierCount, ppMemBarriers);
}
-VK_LAYER_EXPORT void VKAPI vkCmdPipelineBarrier(VkCmdBuffer cmdBuffer, const VkPipelineBarrier* pBarrier)
+VK_LAYER_EXPORT void VKAPI vkCmdPipelineBarrier(VkCmdBuffer cmdBuffer, VkWaitEvent waitEvent, uint32_t pipeEventCount, const VkPipeEvent* pPipeEvents, uint32_t memBarrierCount, const void** ppMemBarriers)
{
GLOBAL_CB_NODE* pCB = getCBNode(cmdBuffer);
if (pCB) {
@@ -2430,7 +2430,7 @@
sprintf(str, "Attempt to use CmdBuffer %p that doesn't exist!", (void*)cmdBuffer);
layerCbMsg(VK_DBG_MSG_ERROR, VK_VALIDATION_LEVEL_0, cmdBuffer, 0, DRAWSTATE_INVALID_CMD_BUFFER, "DS", str);
}
- nextTable.CmdPipelineBarrier(cmdBuffer, pBarrier);
+ nextTable.CmdPipelineBarrier(cmdBuffer, waitEvent, pipeEventCount, pPipeEvents, memBarrierCount, ppMemBarriers);
}
VK_LAYER_EXPORT void VKAPI vkCmdBeginQuery(VkCmdBuffer cmdBuffer, VkQueryPool queryPool, uint32_t slot, VkFlags flags)
@@ -2478,7 +2478,7 @@
nextTable.CmdResetQueryPool(cmdBuffer, queryPool, startQuery, queryCount);
}
-VK_LAYER_EXPORT void VKAPI vkCmdWriteTimestamp(VkCmdBuffer cmdBuffer, VkTimestampType timestampType, VkBuffer destBuffer, VkGpuSize destOffset)
+VK_LAYER_EXPORT void VKAPI vkCmdWriteTimestamp(VkCmdBuffer cmdBuffer, VkTimestampType timestampType, VkBuffer destBuffer, VkDeviceSize destOffset)
{
GLOBAL_CB_NODE* pCB = getCBNode(cmdBuffer);
if (pCB) {
@@ -2508,7 +2508,7 @@
nextTable.CmdInitAtomicCounters(cmdBuffer, pipelineBindPoint, startCounter, counterCount, pData);
}
-VK_LAYER_EXPORT void VKAPI vkCmdLoadAtomicCounters(VkCmdBuffer cmdBuffer, VkPipelineBindPoint pipelineBindPoint, uint32_t startCounter, uint32_t counterCount, VkBuffer srcBuffer, VkGpuSize srcOffset)
+VK_LAYER_EXPORT void VKAPI vkCmdLoadAtomicCounters(VkCmdBuffer cmdBuffer, VkPipelineBindPoint pipelineBindPoint, uint32_t startCounter, uint32_t counterCount, VkBuffer srcBuffer, VkDeviceSize srcOffset)
{
GLOBAL_CB_NODE* pCB = getCBNode(cmdBuffer);
if (pCB) {
@@ -2523,7 +2523,7 @@
nextTable.CmdLoadAtomicCounters(cmdBuffer, pipelineBindPoint, startCounter, counterCount, srcBuffer, srcOffset);
}
-VK_LAYER_EXPORT void VKAPI vkCmdSaveAtomicCounters(VkCmdBuffer cmdBuffer, VkPipelineBindPoint pipelineBindPoint, uint32_t startCounter, uint32_t counterCount, VkBuffer destBuffer, VkGpuSize destOffset)
+VK_LAYER_EXPORT void VKAPI vkCmdSaveAtomicCounters(VkCmdBuffer cmdBuffer, VkPipelineBindPoint pipelineBindPoint, uint32_t startCounter, uint32_t counterCount, VkBuffer destBuffer, VkDeviceSize destOffset)
{
GLOBAL_CB_NODE* pCB = getCBNode(cmdBuffer);
if (pCB) {
@@ -2620,7 +2620,7 @@
// This layer intercepts callbacks
VK_LAYER_DBG_FUNCTION_NODE* pNewDbgFuncNode = new VK_LAYER_DBG_FUNCTION_NODE;
if (!pNewDbgFuncNode)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
pNewDbgFuncNode->pfnMsgCallback = pfnMsgCallback;
pNewDbgFuncNode->pUserData = pUserData;
pNewDbgFuncNode->pNext = g_pDbgFunctionHead;
@@ -2726,7 +2726,7 @@
#endif // WIN32
}
-VK_LAYER_EXPORT void* VKAPI vkGetProcAddr(VkPhysicalGpu gpu, const char* funcName)
+VK_LAYER_EXPORT void* VKAPI vkGetProcAddr(VkPhysicalDevice gpu, const char* funcName)
{
VkBaseLayerObject* gpuw = (VkBaseLayerObject *) gpu;
@@ -2880,6 +2880,6 @@
else {
if (gpuw->pGPA == NULL)
return NULL;
- return gpuw->pGPA((VkPhysicalGpu)gpuw->nextObject, funcName);
+ return gpuw->pGPA((VkPhysicalDevice)gpuw->nextObject, funcName);
}
}
diff --git a/layers/mem_tracker.cpp b/layers/mem_tracker.cpp
index 5067578..207b408 100644
--- a/layers/mem_tracker.cpp
+++ b/layers/mem_tracker.cpp
@@ -51,7 +51,7 @@
#define MAX_BINDING 0xFFFFFFFF
map<VkCmdBuffer, MT_CB_INFO*> cbMap;
-map<VkGpuMemory, MT_MEM_OBJ_INFO*> memObjMap;
+map<VkDeviceMemory, MT_MEM_OBJ_INFO*> memObjMap;
map<VkObject, MT_OBJ_INFO*> objectMap;
map<uint64_t, MT_FENCE_INFO*> fenceMap; // Map fenceId to fence info
map<VkQueue, MT_QUEUE_INFO*> queueMap;
@@ -82,7 +82,7 @@
static void addCBInfo(const VkCmdBuffer cb)
{
MT_CB_INFO* pInfo = new MT_CB_INFO;
- memset(pInfo, 0, (sizeof(MT_CB_INFO) - sizeof(list<VkGpuMemory>)));
+ memset(pInfo, 0, (sizeof(MT_CB_INFO) - sizeof(list<VkDeviceMemory>)));
pInfo->cmdBuffer = cb;
cbMap[cb] = pInfo;
}
@@ -277,10 +277,10 @@
// Queue is validated by caller
static bool32_t checkMemRef(
VkQueue queue,
- VkGpuMemory mem)
+ VkDeviceMemory mem)
{
bool32_t result = VK_FALSE;
- list<VkGpuMemory>::iterator it;
+ list<VkDeviceMemory>::iterator it;
MT_QUEUE_INFO *pQueueInfo = queueMap[queue];
for (it = pQueueInfo->pMemRefList.begin(); it != pQueueInfo->pMemRefList.end(); ++it) {
if ((*it) == mem) {
@@ -316,7 +316,7 @@
result = VK_FALSE;
} else {
// Validate that all actual references are accounted for in pMemRefs
- for (list<VkGpuMemory>::iterator it = pCBInfo->pMemObjList.begin(); it != pCBInfo->pMemObjList.end(); ++it) {
+ for (list<VkDeviceMemory>::iterator it = pCBInfo->pMemObjList.begin(); it != pCBInfo->pMemObjList.end(); ++it) {
// Search for each memref in queues memreflist.
if (checkMemRef(queue, *it)) {
char str[1024];
@@ -346,7 +346,7 @@
// Return ptr to info in map container containing mem, or NULL if not found
// Calls to this function should be wrapped in mutex
-static MT_MEM_OBJ_INFO* getMemObjInfo(const VkGpuMemory mem)
+static MT_MEM_OBJ_INFO* getMemObjInfo(const VkDeviceMemory mem)
{
MT_MEM_OBJ_INFO* pMemObjInfo = NULL;
@@ -356,7 +356,7 @@
return pMemObjInfo;
}
-static void addMemObjInfo(const VkGpuMemory mem, const VkMemoryAllocInfo* pAllocInfo)
+static void addMemObjInfo(const VkDeviceMemory mem, const VkMemoryAllocInfo* pAllocInfo)
{
MT_MEM_OBJ_INFO* pInfo = new MT_MEM_OBJ_INFO;
pInfo->refCount = 0;
@@ -373,7 +373,7 @@
// Find CB Info and add mem binding to list container
// Find Mem Obj Info and add CB binding to list container
-static bool32_t updateCBBinding(const VkCmdBuffer cb, const VkGpuMemory mem)
+static bool32_t updateCBBinding(const VkCmdBuffer cb, const VkDeviceMemory mem)
{
bool32_t result = VK_TRUE;
// First update CB binding in MemObj mini CB list
@@ -408,7 +408,7 @@
} else {
// Search for memory object in cmd buffer's binding list
bool32_t found = VK_FALSE;
- for (list<VkGpuMemory>::iterator it = pCBInfo->pMemObjList.begin(); it != pCBInfo->pMemObjList.end(); ++it) {
+ for (list<VkDeviceMemory>::iterator it = pCBInfo->pMemObjList.begin(); it != pCBInfo->pMemObjList.end(); ++it) {
if ((*it) == mem) {
found = VK_TRUE;
break;
@@ -425,7 +425,7 @@
// Clear the CB Binding for mem
// Calls to this function should be wrapped in mutex
-static void clearCBBinding(const VkCmdBuffer cb, const VkGpuMemory mem)
+static void clearCBBinding(const VkCmdBuffer cb, const VkDeviceMemory mem)
{
MT_MEM_OBJ_INFO* pInfo = getMemObjInfo(mem);
// TODO : Having this check is not ideal, really if memInfo was deleted,
@@ -452,7 +452,7 @@
deleteFenceInfo(pCBInfo->fenceId);
}
- for (list<VkGpuMemory>::iterator it=pCBInfo->pMemObjList.begin(); it!=pCBInfo->pMemObjList.end(); ++it) {
+ for (list<VkDeviceMemory>::iterator it=pCBInfo->pMemObjList.begin(); it!=pCBInfo->pMemObjList.end(); ++it) {
clearCBBinding(cb, (*it));
}
pCBInfo->pMemObjList.clear();
@@ -522,7 +522,7 @@
}
}
-static void deleteMemObjInfo(VkGpuMemory mem)
+static void deleteMemObjInfo(VkDeviceMemory mem)
{
if (memObjMap.find(mem) != memObjMap.end()) {
MT_MEM_OBJ_INFO* pDelInfo = memObjMap[mem];
@@ -557,7 +557,7 @@
return result;
}
-static bool32_t freeMemObjInfo(VkGpuMemory mem, bool internal)
+static bool32_t freeMemObjInfo(VkDeviceMemory mem, bool internal)
{
bool32_t result = VK_TRUE;
// Parse global list to find info w/ mem
@@ -652,7 +652,7 @@
// Add reference from objectInfo to memoryInfo
// Add reference off of objInfo
// Return VK_TRUE if addition is successful, VK_FALSE otherwise
-static bool32_t updateObjectBinding(VkObject object, VkGpuMemory mem)
+static bool32_t updateObjectBinding(VkObject object, VkDeviceMemory mem)
{
bool32_t result = VK_FALSE;
// Handle NULL case separately, just clear previous binding & decrement reference
@@ -720,9 +720,9 @@
}
// For given Object, get 'mem' obj that it's bound to or NULL if no binding
-static VkGpuMemory getMemBindingFromObject(const VkObject object)
+static VkDeviceMemory getMemBindingFromObject(const VkObject object)
{
- VkGpuMemory mem = NULL;
+ VkDeviceMemory mem = NULL;
MT_OBJ_INFO* pObjInfo = getObjectInfo(object);
if (pObjInfo) {
if (pObjInfo->pMemObjInfo) {
@@ -753,7 +753,7 @@
sprintf(str, "MEM INFO : Details of Memory Object list of size %lu elements", memObjMap.size());
layerCbMsg(VK_DBG_MSG_UNKNOWN, VK_VALIDATION_LEVEL_0, NULL, 0, MEMTRACK_NONE, "MEM", str);
- for (map<VkGpuMemory, MT_MEM_OBJ_INFO*>::iterator ii=memObjMap.begin(); ii!=memObjMap.end(); ++ii) {
+ for (map<VkDeviceMemory, MT_MEM_OBJ_INFO*>::iterator ii=memObjMap.begin(); ii!=memObjMap.end(); ++ii) {
pInfo = (*ii).second;
sprintf(str, " ===MemObjInfo at %p===", (void*)pInfo);
@@ -802,7 +802,7 @@
(void*)getFenceFromId(pCBInfo->fenceId));
layerCbMsg(VK_DBG_MSG_UNKNOWN, VK_VALIDATION_LEVEL_0, NULL, 0, MEMTRACK_NONE, "MEM", str);
- for (list<VkGpuMemory>::iterator it = pCBInfo->pMemObjList.begin(); it != pCBInfo->pMemObjList.end(); ++it) {
+ for (list<VkDeviceMemory>::iterator it = pCBInfo->pMemObjList.begin(); it != pCBInfo->pMemObjList.end(); ++it) {
sprintf(str, " Mem obj %p", (*it));
layerCbMsg(VK_DBG_MSG_UNKNOWN, VK_VALIDATION_LEVEL_0, NULL, 0, MEMTRACK_NONE, "MEM", str);
}
@@ -833,7 +833,7 @@
fpNextGPA = pCurObj->pGPA;
assert(fpNextGPA);
- layer_initialize_dispatch_table(&nextTable, fpNextGPA, (VkPhysicalGpu) pCurObj->nextObject);
+ layer_initialize_dispatch_table(&nextTable, fpNextGPA, (VkPhysicalDevice) pCurObj->nextObject);
if (!globalLockInitialized)
{
@@ -847,7 +847,7 @@
}
}
-VK_LAYER_EXPORT VkResult VKAPI vkCreateDevice(VkPhysicalGpu gpu, const VkDeviceCreateInfo* pCreateInfo, VkDevice* pDevice)
+VK_LAYER_EXPORT VkResult VKAPI vkCreateDevice(VkPhysicalDevice gpu, const VkDeviceCreateInfo* pCreateInfo, VkDevice* pDevice)
{
pCurObj = (VkBaseLayerObject *) gpu;
loader_platform_thread_once(&g_initOnce, initMemTracker);
@@ -872,7 +872,7 @@
}
// Report any memory leaks
MT_MEM_OBJ_INFO* pInfo = NULL;
- for (map<VkGpuMemory, MT_MEM_OBJ_INFO*>::iterator ii=memObjMap.begin(); ii!=memObjMap.end(); ++ii) {
+ for (map<VkDeviceMemory, MT_MEM_OBJ_INFO*>::iterator ii=memObjMap.begin(); ii!=memObjMap.end(); ++ii) {
pInfo = (*ii).second;
if (pInfo->allocInfo.allocationSize != 0) {
@@ -943,7 +943,7 @@
return VK_SUCCESS;
}
-VK_LAYER_EXPORT VkResult VKAPI vkEnumerateLayers(VkPhysicalGpu gpu, size_t maxLayerCount,
+VK_LAYER_EXPORT VkResult VKAPI vkEnumerateLayers(VkPhysicalDevice gpu, size_t maxLayerCount,
size_t maxStringSize, size_t* pOutLayerCount, char* const* pOutLayers, void* pReserved)
{
if (gpu != NULL)
@@ -975,7 +975,7 @@
return result;
}
-VK_LAYER_EXPORT VkResult VKAPI vkQueueAddMemReferences(VkQueue queue, uint32_t count, const VkGpuMemory* pMems)
+VK_LAYER_EXPORT VkResult VKAPI vkQueueAddMemReferences(VkQueue queue, uint32_t count, const VkDeviceMemory* pMems)
{
VkResult result = nextTable.QueueAddMemReferences(queue, count, pMems);
if (result == VK_SUCCESS) {
@@ -1004,7 +1004,7 @@
return result;
}
-VK_LAYER_EXPORT VkResult VKAPI vkQueueRemoveMemReferences(VkQueue queue, uint32_t count, const VkGpuMemory* pMems)
+VK_LAYER_EXPORT VkResult VKAPI vkQueueRemoveMemReferences(VkQueue queue, uint32_t count, const VkDeviceMemory* pMems)
{
// TODO : Decrement ref count for this memory reference on this queue. Remove if ref count is zero.
VkResult result = nextTable.QueueRemoveMemReferences(queue, count, pMems);
@@ -1018,7 +1018,7 @@
layerCbMsg(VK_DBG_MSG_ERROR, VK_VALIDATION_LEVEL_0, queue, 0, MEMTRACK_INVALID_QUEUE, "MEM", str);
} else {
for (int i = 0; i < count; i++) {
- for (list<VkGpuMemory>::iterator it = pQueueInfo->pMemRefList.begin(); it != pQueueInfo->pMemRefList.end(); ++it) {
+ for (list<VkDeviceMemory>::iterator it = pQueueInfo->pMemRefList.begin(); it != pQueueInfo->pMemRefList.end(); ++it) {
if ((*it) == pMems[i]) {
it = pQueueInfo->pMemRefList.erase(it);
}
@@ -1059,7 +1059,7 @@
return result;
}
-VK_LAYER_EXPORT VkResult VKAPI vkAllocMemory(VkDevice device, const VkMemoryAllocInfo* pAllocInfo, VkGpuMemory* pMem)
+VK_LAYER_EXPORT VkResult VKAPI vkAllocMemory(VkDevice device, const VkMemoryAllocInfo* pAllocInfo, VkDeviceMemory* pMem)
{
VkResult result = nextTable.AllocMemory(device, pAllocInfo, pMem);
// TODO : Track allocations and overall size here
@@ -1070,7 +1070,7 @@
return result;
}
-VK_LAYER_EXPORT VkResult VKAPI vkFreeMemory(VkGpuMemory mem)
+VK_LAYER_EXPORT VkResult VKAPI vkFreeMemory(VkDeviceMemory mem)
{
/* From spec : A memory object is freed by calling vkFreeMemory() when it is no longer needed. Before
* freeing a memory object, an application must ensure the memory object is unbound from
@@ -1090,7 +1090,7 @@
return result;
}
-VK_LAYER_EXPORT VkResult VKAPI vkSetMemoryPriority(VkGpuMemory mem, VkMemoryPriority priority)
+VK_LAYER_EXPORT VkResult VKAPI vkSetMemoryPriority(VkDeviceMemory mem, VkMemoryPriority priority)
{
// TODO : Update tracking for this alloc
// Make sure memory is not pinned, which can't have priority set
@@ -1098,14 +1098,14 @@
return result;
}
-VK_LAYER_EXPORT VkResult VKAPI vkMapMemory(VkGpuMemory mem, VkFlags flags, void** ppData)
+VK_LAYER_EXPORT VkResult VKAPI vkMapMemory(VkDeviceMemory mem, VkFlags flags, void** ppData)
{
// TODO : Track when memory is mapped
loader_platform_thread_lock_mutex(&globalLock);
MT_MEM_OBJ_INFO *pMemObj = getMemObjInfo(mem);
- if ((pMemObj->allocInfo.memProps & VK_MEMORY_PROPERTY_CPU_VISIBLE_BIT) == 0) {
+ if ((pMemObj->allocInfo.memProps & VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT) == 0) {
char str[1024];
- sprintf(str, "Mapping Memory (%p) without VK_MEMORY_PROPERTY_CPU_VISIBLE_BIT set", (void*)mem);
+ sprintf(str, "Mapping Memory (%p) without VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT set", (void*)mem);
layerCbMsg(VK_DBG_MSG_ERROR, VK_VALIDATION_LEVEL_0, mem, 0, MEMTRACK_INVALID_STATE, "MEM", str);
}
loader_platform_thread_unlock_mutex(&globalLock);
@@ -1113,7 +1113,7 @@
return result;
}
-VK_LAYER_EXPORT VkResult VKAPI vkUnmapMemory(VkGpuMemory mem)
+VK_LAYER_EXPORT VkResult VKAPI vkUnmapMemory(VkDeviceMemory mem)
{
// TODO : Track as memory gets unmapped, do we want to check what changed following map?
// Make sure that memory was ever mapped to begin with
@@ -1121,7 +1121,7 @@
return result;
}
-VK_LAYER_EXPORT VkResult VKAPI vkPinSystemMemory(VkDevice device, const void* pSysMem, size_t memSize, VkGpuMemory* pMem)
+VK_LAYER_EXPORT VkResult VKAPI vkPinSystemMemory(VkDevice device, const void* pSysMem, size_t memSize, VkDeviceMemory* pMem)
{
// TODO : Track this
// Verify that memory is actually pinnable
@@ -1129,21 +1129,21 @@
return result;
}
-VK_LAYER_EXPORT VkResult VKAPI vkOpenSharedMemory(VkDevice device, const VkMemoryOpenInfo* pOpenInfo, VkGpuMemory* pMem)
+VK_LAYER_EXPORT VkResult VKAPI vkOpenSharedMemory(VkDevice device, const VkMemoryOpenInfo* pOpenInfo, VkDeviceMemory* pMem)
{
// TODO : Track this
VkResult result = nextTable.OpenSharedMemory(device, pOpenInfo, pMem);
return result;
}
-VK_LAYER_EXPORT VkResult VKAPI vkOpenPeerMemory(VkDevice device, const VkPeerMemoryOpenInfo* pOpenInfo, VkGpuMemory* pMem)
+VK_LAYER_EXPORT VkResult VKAPI vkOpenPeerMemory(VkDevice device, const VkPeerMemoryOpenInfo* pOpenInfo, VkDeviceMemory* pMem)
{
// TODO : Track this
VkResult result = nextTable.OpenPeerMemory(device, pOpenInfo, pMem);
return result;
}
-VK_LAYER_EXPORT VkResult VKAPI vkOpenPeerImage(VkDevice device, const VkPeerImageOpenInfo* pOpenInfo, VkImage* pImage, VkGpuMemory* pMem)
+VK_LAYER_EXPORT VkResult VKAPI vkOpenPeerImage(VkDevice device, const VkPeerImageOpenInfo* pOpenInfo, VkImage* pImage, VkDeviceMemory* pMem)
{
// TODO : Track this
VkResult result = nextTable.OpenPeerImage(device, pOpenInfo, pImage, pMem);
@@ -1164,7 +1164,7 @@
if (pDelInfo->pMemObjInfo) {
// Wsi allocated Memory is tied to image object so clear the binding and free that memory automatically
if (0 == pDelInfo->pMemObjInfo->allocInfo.allocationSize) { // Wsi allocated memory has NULL allocInfo w/ 0 size
- VkGpuMemory memToFree = pDelInfo->pMemObjInfo->mem;
+ VkDeviceMemory memToFree = pDelInfo->pMemObjInfo->mem;
clearObjectBinding(object);
freeMemObjInfo(memToFree, true);
}
@@ -1194,7 +1194,7 @@
return result;
}
-VK_LAYER_EXPORT VkResult VKAPI vkQueueBindObjectMemory(VkQueue queue, VkObject object, uint32_t allocationIdx, VkGpuMemory mem, VkGpuSize offset)
+VK_LAYER_EXPORT VkResult VKAPI vkQueueBindObjectMemory(VkQueue queue, VkObject object, uint32_t allocationIdx, VkDeviceMemory mem, VkDeviceSize offset)
{
VkResult result = nextTable.QueueBindObjectMemory(queue, object, allocationIdx, mem, offset);
loader_platform_thread_lock_mutex(&globalLock);
@@ -1621,20 +1621,20 @@
uint32_t startBinding,
uint32_t bindingCount,
const VkBuffer* pBuffers,
- const VkGpuSize* pOffsets)
+ const VkDeviceSize* pOffsets)
{
nextTable.CmdBindVertexBuffers(cmdBuffer, startBinding, bindingCount, pBuffers, pOffsets);
}
-VK_LAYER_EXPORT void VKAPI vkCmdBindIndexBuffer(VkCmdBuffer cmdBuffer, VkBuffer buffer, VkGpuSize offset, VkIndexType indexType)
+VK_LAYER_EXPORT void VKAPI vkCmdBindIndexBuffer(VkCmdBuffer cmdBuffer, VkBuffer buffer, VkDeviceSize offset, VkIndexType indexType)
{
nextTable.CmdBindIndexBuffer(cmdBuffer, buffer, offset, indexType);
}
-VK_LAYER_EXPORT void VKAPI vkCmdDrawIndirect(VkCmdBuffer cmdBuffer, VkBuffer buffer, VkGpuSize offset, uint32_t count, uint32_t stride)
+VK_LAYER_EXPORT void VKAPI vkCmdDrawIndirect(VkCmdBuffer cmdBuffer, VkBuffer buffer, VkDeviceSize offset, uint32_t count, uint32_t stride)
{
loader_platform_thread_lock_mutex(&globalLock);
- VkGpuMemory mem = getMemBindingFromObject(buffer);
+ VkDeviceMemory mem = getMemBindingFromObject(buffer);
if (VK_FALSE == updateCBBinding(cmdBuffer, mem)) {
char str[1024];
sprintf(str, "In vkCmdDrawIndirect() call unable to update binding of buffer %p to cmdBuffer %p", buffer, cmdBuffer);
@@ -1644,10 +1644,10 @@
nextTable.CmdDrawIndirect(cmdBuffer, buffer, offset, count, stride);
}
-VK_LAYER_EXPORT void VKAPI vkCmdDrawIndexedIndirect(VkCmdBuffer cmdBuffer, VkBuffer buffer, VkGpuSize offset, uint32_t count, uint32_t stride)
+VK_LAYER_EXPORT void VKAPI vkCmdDrawIndexedIndirect(VkCmdBuffer cmdBuffer, VkBuffer buffer, VkDeviceSize offset, uint32_t count, uint32_t stride)
{
loader_platform_thread_lock_mutex(&globalLock);
- VkGpuMemory mem = getMemBindingFromObject(buffer);
+ VkDeviceMemory mem = getMemBindingFromObject(buffer);
if (VK_FALSE == updateCBBinding(cmdBuffer, mem)) {
char str[1024];
sprintf(str, "In vkCmdDrawIndexedIndirect() call unable to update binding of buffer %p to cmdBuffer %p", buffer, cmdBuffer);
@@ -1657,10 +1657,10 @@
nextTable.CmdDrawIndexedIndirect(cmdBuffer, buffer, offset, count, stride);
}
-VK_LAYER_EXPORT void VKAPI vkCmdDispatchIndirect(VkCmdBuffer cmdBuffer, VkBuffer buffer, VkGpuSize offset)
+VK_LAYER_EXPORT void VKAPI vkCmdDispatchIndirect(VkCmdBuffer cmdBuffer, VkBuffer buffer, VkDeviceSize offset)
{
loader_platform_thread_lock_mutex(&globalLock);
- VkGpuMemory mem = getMemBindingFromObject(buffer);
+ VkDeviceMemory mem = getMemBindingFromObject(buffer);
if (VK_FALSE == updateCBBinding(cmdBuffer, mem)) {
char str[1024];
sprintf(str, "In vkCmdDispatchIndirect() call unable to update binding of buffer %p to cmdBuffer %p", buffer, cmdBuffer);
@@ -1674,7 +1674,7 @@
uint32_t regionCount, const VkBufferCopy* pRegions)
{
loader_platform_thread_lock_mutex(&globalLock);
- VkGpuMemory mem = getMemBindingFromObject(srcBuffer);
+ VkDeviceMemory mem = getMemBindingFromObject(srcBuffer);
if (VK_FALSE == updateCBBinding(cmdBuffer, mem)) {
char str[1024];
sprintf(str, "In vkCmdCopyBuffer() call unable to update binding of srcBuffer %p to cmdBuffer %p", srcBuffer, cmdBuffer);
@@ -1715,7 +1715,7 @@
{
// TODO : Track this
loader_platform_thread_lock_mutex(&globalLock);
- VkGpuMemory mem = getMemBindingFromObject(destImage);
+ VkDeviceMemory mem = getMemBindingFromObject(destImage);
if (VK_FALSE == updateCBBinding(cmdBuffer, mem)) {
char str[1024];
sprintf(str, "In vkCmdCopyMemoryToImage() call unable to update binding of destImage buffer %p to cmdBuffer %p", destImage, cmdBuffer);
@@ -1739,7 +1739,7 @@
{
// TODO : Track this
loader_platform_thread_lock_mutex(&globalLock);
- VkGpuMemory mem = getMemBindingFromObject(srcImage);
+ VkDeviceMemory mem = getMemBindingFromObject(srcImage);
if (VK_FALSE == updateCBBinding(cmdBuffer, mem)) {
char str[1024];
sprintf(str, "In vkCmdCopyImageToMemory() call unable to update binding of srcImage buffer %p to cmdBuffer %p", srcImage, cmdBuffer);
@@ -1760,7 +1760,7 @@
{
// TODO : Each image will have mem mapping so track them
loader_platform_thread_lock_mutex(&globalLock);
- VkGpuMemory mem = getMemBindingFromObject(srcImage);
+ VkDeviceMemory mem = getMemBindingFromObject(srcImage);
if (VK_FALSE == updateCBBinding(cmdBuffer, mem)) {
char str[1024];
sprintf(str, "In vkCmdCloneImageData() call unable to update binding of srcImage buffer %p to cmdBuffer %p", srcImage, cmdBuffer);
@@ -1776,10 +1776,10 @@
nextTable.CmdCloneImageData(cmdBuffer, srcImage, srcImageLayout, destImage, destImageLayout);
}
-VK_LAYER_EXPORT void VKAPI vkCmdUpdateBuffer(VkCmdBuffer cmdBuffer, VkBuffer destBuffer, VkGpuSize destOffset, VkGpuSize dataSize, const uint32_t* pData)
+VK_LAYER_EXPORT void VKAPI vkCmdUpdateBuffer(VkCmdBuffer cmdBuffer, VkBuffer destBuffer, VkDeviceSize destOffset, VkDeviceSize dataSize, const uint32_t* pData)
{
loader_platform_thread_lock_mutex(&globalLock);
- VkGpuMemory mem = getMemBindingFromObject(destBuffer);
+ VkDeviceMemory mem = getMemBindingFromObject(destBuffer);
if (VK_FALSE == updateCBBinding(cmdBuffer, mem)) {
char str[1024];
sprintf(str, "In vkCmdUpdateMemory() call unable to update binding of destBuffer %p to cmdBuffer %p", destBuffer, cmdBuffer);
@@ -1789,10 +1789,10 @@
nextTable.CmdUpdateBuffer(cmdBuffer, destBuffer, destOffset, dataSize, pData);
}
-VK_LAYER_EXPORT void VKAPI vkCmdFillBuffer(VkCmdBuffer cmdBuffer, VkBuffer destBuffer, VkGpuSize destOffset, VkGpuSize fillSize, uint32_t data)
+VK_LAYER_EXPORT void VKAPI vkCmdFillBuffer(VkCmdBuffer cmdBuffer, VkBuffer destBuffer, VkDeviceSize destOffset, VkDeviceSize fillSize, uint32_t data)
{
loader_platform_thread_lock_mutex(&globalLock);
- VkGpuMemory mem = getMemBindingFromObject(destBuffer);
+ VkDeviceMemory mem = getMemBindingFromObject(destBuffer);
if (VK_FALSE == updateCBBinding(cmdBuffer, mem)) {
char str[1024];
sprintf(str, "In vkCmdFillMemory() call unable to update binding of destBuffer %p to cmdBuffer %p", destBuffer, cmdBuffer);
@@ -1809,7 +1809,7 @@
{
// TODO : Verify memory is in VK_IMAGE_STATE_CLEAR state
loader_platform_thread_lock_mutex(&globalLock);
- VkGpuMemory mem = getMemBindingFromObject(image);
+ VkDeviceMemory mem = getMemBindingFromObject(image);
if (VK_FALSE == updateCBBinding(cmdBuffer, mem)) {
char str[1024];
sprintf(str, "In vkCmdClearColorImage() call unable to update binding of image buffer %p to cmdBuffer %p", image, cmdBuffer);
@@ -1826,7 +1826,7 @@
{
// TODO : Verify memory is in VK_IMAGE_STATE_CLEAR state
loader_platform_thread_lock_mutex(&globalLock);
- VkGpuMemory mem = getMemBindingFromObject(image);
+ VkDeviceMemory mem = getMemBindingFromObject(image);
if (VK_FALSE == updateCBBinding(cmdBuffer, mem)) {
char str[1024];
sprintf(str, "In vkCmdClearDepthStencil() call unable to update binding of image buffer %p to cmdBuffer %p", image, cmdBuffer);
@@ -1842,7 +1842,7 @@
uint32_t regionCount, const VkImageResolve* pRegions)
{
loader_platform_thread_lock_mutex(&globalLock);
- VkGpuMemory mem = getMemBindingFromObject(srcImage);
+ VkDeviceMemory mem = getMemBindingFromObject(srcImage);
if (VK_FALSE == updateCBBinding(cmdBuffer, mem)) {
char str[1024];
sprintf(str, "In vkCmdResolveImage() call unable to update binding of srcImage buffer %p to cmdBuffer %p", srcImage, cmdBuffer);
@@ -1861,7 +1861,7 @@
VK_LAYER_EXPORT void VKAPI vkCmdBeginQuery(VkCmdBuffer cmdBuffer, VkQueryPool queryPool, uint32_t slot, VkFlags flags)
{
loader_platform_thread_lock_mutex(&globalLock);
- VkGpuMemory mem = getMemBindingFromObject(queryPool);
+ VkDeviceMemory mem = getMemBindingFromObject(queryPool);
if (VK_FALSE == updateCBBinding(cmdBuffer, mem)) {
char str[1024];
sprintf(str, "In vkCmdBeginQuery() call unable to update binding of queryPool buffer %p to cmdBuffer %p", queryPool, cmdBuffer);
@@ -1874,7 +1874,7 @@
VK_LAYER_EXPORT void VKAPI vkCmdEndQuery(VkCmdBuffer cmdBuffer, VkQueryPool queryPool, uint32_t slot)
{
loader_platform_thread_lock_mutex(&globalLock);
- VkGpuMemory mem = getMemBindingFromObject(queryPool);
+ VkDeviceMemory mem = getMemBindingFromObject(queryPool);
if (VK_FALSE == updateCBBinding(cmdBuffer, mem)) {
char str[1024];
sprintf(str, "In vkCmdEndQuery() call unable to update binding of queryPool buffer %p to cmdBuffer %p", queryPool, cmdBuffer);
@@ -1887,7 +1887,7 @@
VK_LAYER_EXPORT void VKAPI vkCmdResetQueryPool(VkCmdBuffer cmdBuffer, VkQueryPool queryPool, uint32_t startQuery, uint32_t queryCount)
{
loader_platform_thread_lock_mutex(&globalLock);
- VkGpuMemory mem = getMemBindingFromObject(queryPool);
+ VkDeviceMemory mem = getMemBindingFromObject(queryPool);
if (VK_FALSE == updateCBBinding(cmdBuffer, mem)) {
char str[1024];
sprintf(str, "In vkCmdResetQueryPool() call unable to update binding of queryPool buffer %p to cmdBuffer %p", queryPool, cmdBuffer);
@@ -1902,7 +1902,7 @@
// This layer intercepts callbacks
VK_LAYER_DBG_FUNCTION_NODE *pNewDbgFuncNode = (VK_LAYER_DBG_FUNCTION_NODE*)malloc(sizeof(VK_LAYER_DBG_FUNCTION_NODE));
if (!pNewDbgFuncNode)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
pNewDbgFuncNode->pfnMsgCallback = pfnMsgCallback;
pNewDbgFuncNode->pUserData = pUserData;
pNewDbgFuncNode->pNext = g_pDbgFunctionHead;
@@ -1944,13 +1944,13 @@
#if !defined(WIN32)
VK_LAYER_EXPORT VkResult VKAPI vkWsiX11CreatePresentableImage(VkDevice device, const VK_WSI_X11_PRESENTABLE_IMAGE_CREATE_INFO* pCreateInfo,
- VkImage* pImage, VkGpuMemory* pMem)
+ VkImage* pImage, VkDeviceMemory* pMem)
{
VkResult result = nextTable.WsiX11CreatePresentableImage(device, pCreateInfo, pImage, pMem);
loader_platform_thread_lock_mutex(&globalLock);
if (VK_SUCCESS == result) {
// Add image object, then insert the new Mem Object and then bind it to created image
- addObjectInfo(*pImage, VkStructureType_MAX_ENUM, pCreateInfo, sizeof(VK_WSI_X11_PRESENTABLE_IMAGE_CREATE_INFO), "wsi_x11_image");
+ addObjectInfo(*pImage, VK_STRUCTURE_TYPE_MAX_ENUM, pCreateInfo, sizeof(VK_WSI_X11_PRESENTABLE_IMAGE_CREATE_INFO), "wsi_x11_image");
addMemObjInfo(*pMem, NULL);
if (VK_FALSE == updateObjectBinding(*pImage, *pMem)) {
char str[1024];
@@ -1977,7 +1977,7 @@
}
#endif // WIN32
-VK_LAYER_EXPORT void* VKAPI vkGetProcAddr(VkPhysicalGpu gpu, const char* funcName)
+VK_LAYER_EXPORT void* VKAPI vkGetProcAddr(VkPhysicalDevice gpu, const char* funcName)
{
VkBaseLayerObject* gpuw = (VkBaseLayerObject *) gpu;
@@ -2135,6 +2135,6 @@
else {
if (gpuw->pGPA == NULL)
return NULL;
- return gpuw->pGPA((VkPhysicalGpu)gpuw->nextObject, funcName);
+ return gpuw->pGPA((VkPhysicalDevice)gpuw->nextObject, funcName);
}
}
diff --git a/layers/mem_tracker.h b/layers/mem_tracker.h
index c13d95f..d1d2b82 100644
--- a/layers/mem_tracker.h
+++ b/layers/mem_tracker.h
@@ -85,7 +85,7 @@
// Data struct for tracking memory object
struct MT_MEM_OBJ_INFO {
uint32_t refCount; // Count of references (obj bindings or CB use)
- VkGpuMemory mem;
+ VkDeviceMemory mem;
VkMemoryAllocInfo allocInfo;
list<VkObject> pObjBindings; // list container of objects bound to this memory
list<VkCmdBuffer> pCmdBufferBindings; // list container of cmd buffers that reference this mem object
@@ -125,7 +125,7 @@
VkCmdBuffer cmdBuffer;
uint64_t fenceId;
// Order dependent, stl containers must be at end of struct
- list<VkGpuMemory> pMemObjList; // List container of Mem objs referenced by this CB
+ list<VkDeviceMemory> pMemObjList; // List container of Mem objs referenced by this CB
};
// Associate fenceId with a fence object
@@ -140,7 +140,7 @@
uint64_t lastRetiredId;
uint64_t lastSubmittedId;
list<VkCmdBuffer> pQueueCmdBuffers;
- list<VkGpuMemory> pMemRefList;
+ list<VkDeviceMemory> pMemRefList;
};
#ifdef __cplusplus
diff --git a/layers/multi.cpp b/layers/multi.cpp
index 785cfc7..3eb79c6 100644
--- a/layers/multi.cpp
+++ b/layers/multi.cpp
@@ -62,7 +62,7 @@
#endif
-VK_LAYER_EXPORT VkResult VKAPI multi1CreateDevice(VkPhysicalGpu gpu, const VkDeviceCreateInfo* pCreateInfo,
+VK_LAYER_EXPORT VkResult VKAPI multi1CreateDevice(VkPhysicalDevice gpu, const VkDeviceCreateInfo* pCreateInfo,
VkDevice* pDevice)
{
VkLayerDispatchTable* pTable = tableMap1[gpu];
@@ -97,7 +97,7 @@
return result;
}
-VK_LAYER_EXPORT VkResult VKAPI multi1EnumerateLayers(VkPhysicalGpu gpu, size_t maxLayerCount, size_t maxStringSize,
+VK_LAYER_EXPORT VkResult VKAPI multi1EnumerateLayers(VkPhysicalDevice gpu, size_t maxLayerCount, size_t maxStringSize,
size_t* pOutLayerCount, char* const* pOutLayers,
void* pReserved)
{
@@ -111,7 +111,7 @@
return result;
}
-VK_LAYER_EXPORT void * VKAPI multi1GetProcAddr(VkPhysicalGpu gpu, const char* pName)
+VK_LAYER_EXPORT void * VKAPI multi1GetProcAddr(VkPhysicalDevice gpu, const char* pName)
{
VkBaseLayerObject* gpuw = (VkBaseLayerObject *) gpu;
@@ -131,7 +131,7 @@
else {
if (gpuw->pGPA == NULL)
return NULL;
- return gpuw->pGPA((VkPhysicalGpu) gpuw->nextObject, pName);
+ return gpuw->pGPA((VkPhysicalDevice) gpuw->nextObject, pName);
}
}
@@ -157,7 +157,7 @@
}
}
-VK_LAYER_EXPORT VkResult VKAPI multi2CreateDevice(VkPhysicalGpu gpu, const VkDeviceCreateInfo* pCreateInfo,
+VK_LAYER_EXPORT VkResult VKAPI multi2CreateDevice(VkPhysicalDevice gpu, const VkDeviceCreateInfo* pCreateInfo,
VkDevice* pDevice)
{
VkLayerDispatchTable* pTable = tableMap2[gpu];
@@ -194,7 +194,7 @@
}
-VK_LAYER_EXPORT VkResult VKAPI multi2EnumerateLayers(VkPhysicalGpu gpu, size_t maxLayerCount, size_t maxStringSize,
+VK_LAYER_EXPORT VkResult VKAPI multi2EnumerateLayers(VkPhysicalDevice gpu, size_t maxLayerCount, size_t maxStringSize,
size_t* pOutLayerCount, char* const* pOutLayers,
void* pReserved)
{
@@ -209,7 +209,7 @@
return result;
}
-VK_LAYER_EXPORT void * VKAPI multi2GetProcAddr(VkPhysicalGpu gpu, const char* pName)
+VK_LAYER_EXPORT void * VKAPI multi2GetProcAddr(VkPhysicalDevice gpu, const char* pName)
{
VkBaseLayerObject* gpuw = (VkBaseLayerObject *) gpu;
@@ -229,12 +229,12 @@
else {
if (gpuw->pGPA == NULL)
return NULL;
- return gpuw->pGPA((VkPhysicalGpu) gpuw->nextObject, pName);
+ return gpuw->pGPA((VkPhysicalDevice) gpuw->nextObject, pName);
}
}
/********************************* Common functions ********************************/
-VK_LAYER_EXPORT VkResult VKAPI vkEnumerateLayers(VkPhysicalGpu gpu, size_t maxLayerCount, size_t maxStringSize,
+VK_LAYER_EXPORT VkResult VKAPI vkEnumerateLayers(VkPhysicalDevice gpu, size_t maxLayerCount, size_t maxStringSize,
size_t* pOutLayerCount, char* const* pOutLayers,
void* pReserved)
{
@@ -303,7 +303,7 @@
return VK_SUCCESS;
}
-VK_LAYER_EXPORT void * VKAPI vkGetProcAddr(VkPhysicalGpu gpu, const char* pName)
+VK_LAYER_EXPORT void * VKAPI vkGetProcAddr(VkPhysicalDevice gpu, const char* pName)
{
// to find each layers GPA routine Loader will search via "<layerName>GetProcAddr"
if (!strncmp("multi1GetProcAddr", pName, sizeof("multi1GetProcAddr")))
@@ -334,5 +334,5 @@
if (layerNum == 1 && layer2_first_activated == false)
layer1_first_activated = true;
- layer_initialize_dispatch_table(pTable, gpuw->pGPA, (VkPhysicalGpu) gpuw->nextObject);
+ layer_initialize_dispatch_table(pTable, gpuw->pGPA, (VkPhysicalDevice) gpuw->nextObject);
}
diff --git a/layers/object_track.h b/layers/object_track.h
index d24e0de..11dbeda 100644
--- a/layers/object_track.h
+++ b/layers/object_track.h
@@ -74,7 +74,7 @@
VkObjectTypeQueryPool,
VkObjectTypeEvent,
VkObjectTypeQueue,
- VkObjectTypePhysicalGpu,
+ VkObjectTypePhysicalDevice,
VkObjectTypeRenderPass,
VkObjectTypeFramebuffer,
VkObjectTypeImage,
@@ -83,7 +83,7 @@
VkObjectTypeInstance,
VkObjectTypeDynamicVpState,
VkObjectTypeColorAttachmentView,
- VkObjectTypeGpuMemory,
+ VkObjectTypeDeviceMemory,
VkObjectTypeDynamicRsState,
VkObjectTypeFence,
VkObjectTypeCmdBuffer,
@@ -107,8 +107,8 @@
return "DESCRIPTOR_SET_LAYOUT";
case VkObjectTypeDescriptorSetLayoutChain:
return "DESCRIPTOR_SET_LAYOUT_CHAIN";
- case VkObjectTypeGpuMemory:
- return "GPU_MEMORY";
+ case VkObjectTypeDeviceMemory:
+ return "DEVICE_MEMORY";
case VkObjectTypeQueue:
return "QUEUE";
case VkObjectTypeImage:
@@ -127,8 +127,8 @@
return "BUFFER_VIEW";
case VkObjectTypeDescriptorSet:
return "DESCRIPTOR_SET";
- case VkObjectTypePhysicalGpu:
- return "PHYSICAL_GPU";
+ case VkObjectTypePhysicalDevice:
+ return "PHYSICAL_DEVICE";
case VkObjectTypeImageView:
return "IMAGE_VIEW";
case VkObjectTypeBuffer:
diff --git a/layers/param_checker.cpp b/layers/param_checker.cpp
index 5408ee7..5635bfd 100644
--- a/layers/param_checker.cpp
+++ b/layers/param_checker.cpp
@@ -68,7 +68,7 @@
fpNextGPA = pCurObj->pGPA;
assert(fpNextGPA);
- layer_initialize_dispatch_table(&nextTable, fpNextGPA, (VkPhysicalGpu) pCurObj->nextObject);
+ layer_initialize_dispatch_table(&nextTable, fpNextGPA, (VkPhysicalDevice) pCurObj->nextObject);
}
void PreCreateInstance(const VkApplicationInfo* pAppInfo, const VkAllocCallbacks* pAllocCb)
@@ -139,24 +139,24 @@
return result;
}
-VK_LAYER_EXPORT VkResult VKAPI vkGetGpuInfo(VkPhysicalGpu gpu, VkPhysicalGpuInfoType infoType, size_t* pDataSize, void* pData)
+VK_LAYER_EXPORT VkResult VKAPI vkGetPhysicalDeviceInfo(VkPhysicalDevice gpu, VkPhysicalDeviceInfoType infoType, size_t* pDataSize, void* pData)
{
pCurObj = (VkBaseLayerObject *) gpu;
loader_platform_thread_once(&tabOnce, initParamChecker);
char str[1024];
- if (!validate_VkPhysicalGpuInfoType(infoType)) {
- sprintf(str, "Parameter infoType to function GetGpuInfo has invalid value of %i.", (int)infoType);
+ if (!validate_VkPhysicalDeviceInfoType(infoType)) {
+ sprintf(str, "Parameter infoType to function GetPhysicalDeviceInfo has invalid value of %i.", (int)infoType);
layerCbMsg(VK_DBG_MSG_ERROR, VK_VALIDATION_LEVEL_0, NULL, 0, 1, "PARAMCHECK", str);
}
- VkResult result = nextTable.GetGpuInfo(gpu, infoType, pDataSize, pData);
+ VkResult result = nextTable.GetPhysicalDeviceInfo(gpu, infoType, pDataSize, pData);
return result;
}
-void PreCreateDevice(VkPhysicalGpu gpu, const VkDeviceCreateInfo* pCreateInfo)
+void PreCreateDevice(VkPhysicalDevice gpu, const VkDeviceCreateInfo* pCreateInfo)
{
if(gpu == nullptr)
{
- char const str[] = "vkCreateDevice parameter, VkPhysicalGpu gpu, is nullptr "\
+ char const str[] = "vkCreateDevice parameter, VkPhysicalDevice gpu, is nullptr "\
"(precondition).";
layerCbMsg(VK_DBG_MSG_UNKNOWN, VK_VALIDATION_LEVEL_0, NULL, 0, 1, "PARAMCHECK", str);
return;
@@ -226,7 +226,7 @@
}
}
-VK_LAYER_EXPORT VkResult VKAPI vkCreateDevice(VkPhysicalGpu gpu, const VkDeviceCreateInfo* pCreateInfo, VkDevice* pDevice)
+VK_LAYER_EXPORT VkResult VKAPI vkCreateDevice(VkPhysicalDevice gpu, const VkDeviceCreateInfo* pCreateInfo, VkDevice* pDevice)
{
pCurObj = (VkBaseLayerObject *) gpu;
loader_platform_thread_once(&tabOnce, initParamChecker);
@@ -297,7 +297,7 @@
return VK_SUCCESS;
}
-VK_LAYER_EXPORT VkResult VKAPI vkEnumerateLayers(VkPhysicalGpu gpu, size_t maxLayerCount, size_t maxStringSize, size_t* pOutLayerCount, char* const* pOutLayers, void* pReserved)
+VK_LAYER_EXPORT VkResult VKAPI vkEnumerateLayers(VkPhysicalDevice gpu, size_t maxLayerCount, size_t maxStringSize, size_t* pOutLayerCount, char* const* pOutLayers, void* pReserved)
{
char str[1024];
if (gpu != NULL) {
@@ -333,13 +333,13 @@
return result;
}
-VK_LAYER_EXPORT VkResult VKAPI vkQueueAddMemReferences(VkQueue queue, uint32_t count, const VkGpuMemory* pMems)
+VK_LAYER_EXPORT VkResult VKAPI vkQueueAddMemReferences(VkQueue queue, uint32_t count, const VkDeviceMemory* pMems)
{
VkResult result = nextTable.QueueAddMemReferences(queue, count, pMems);
return result;
}
-VK_LAYER_EXPORT VkResult VKAPI vkQueueRemoveMemReferences(VkQueue queue, uint32_t count, const VkGpuMemory* pMems)
+VK_LAYER_EXPORT VkResult VKAPI vkQueueRemoveMemReferences(VkQueue queue, uint32_t count, const VkDeviceMemory* pMems)
{
VkResult result = nextTable.QueueRemoveMemReferences(queue, count, pMems);
return result;
@@ -358,7 +358,7 @@
return result;
}
-VK_LAYER_EXPORT VkResult VKAPI vkAllocMemory(VkDevice device, const VkMemoryAllocInfo* pAllocInfo, VkGpuMemory* pMem)
+VK_LAYER_EXPORT VkResult VKAPI vkAllocMemory(VkDevice device, const VkMemoryAllocInfo* pAllocInfo, VkDeviceMemory* pMem)
{
char str[1024];
if (!pAllocInfo) {
@@ -372,14 +372,14 @@
return result;
}
-VK_LAYER_EXPORT VkResult VKAPI vkFreeMemory(VkGpuMemory mem)
+VK_LAYER_EXPORT VkResult VKAPI vkFreeMemory(VkDeviceMemory mem)
{
VkResult result = nextTable.FreeMemory(mem);
return result;
}
-VK_LAYER_EXPORT VkResult VKAPI vkSetMemoryPriority(VkGpuMemory mem, VkMemoryPriority priority)
+VK_LAYER_EXPORT VkResult VKAPI vkSetMemoryPriority(VkDeviceMemory mem, VkMemoryPriority priority)
{
char str[1024];
if (!validate_VkMemoryPriority(priority)) {
@@ -390,37 +390,37 @@
return result;
}
-VK_LAYER_EXPORT VkResult VKAPI vkMapMemory(VkGpuMemory mem, VkFlags flags, void** ppData)
+VK_LAYER_EXPORT VkResult VKAPI vkMapMemory(VkDeviceMemory mem, VkFlags flags, void** ppData)
{
VkResult result = nextTable.MapMemory(mem, flags, ppData);
return result;
}
-VK_LAYER_EXPORT VkResult VKAPI vkUnmapMemory(VkGpuMemory mem)
+VK_LAYER_EXPORT VkResult VKAPI vkUnmapMemory(VkDeviceMemory mem)
{
VkResult result = nextTable.UnmapMemory(mem);
return result;
}
-VK_LAYER_EXPORT VkResult VKAPI vkPinSystemMemory(VkDevice device, const void* pSysMem, size_t memSize, VkGpuMemory* pMem)
+VK_LAYER_EXPORT VkResult VKAPI vkPinSystemMemory(VkDevice device, const void* pSysMem, size_t memSize, VkDeviceMemory* pMem)
{
VkResult result = nextTable.PinSystemMemory(device, pSysMem, memSize, pMem);
return result;
}
-VK_LAYER_EXPORT VkResult VKAPI vkGetMultiGpuCompatibility(VkPhysicalGpu gpu0, VkPhysicalGpu gpu1, VkGpuCompatibilityInfo* pInfo)
+VK_LAYER_EXPORT VkResult VKAPI vkGetMultiDeviceCompatibility(VkPhysicalDevice gpu0, VkPhysicalDevice gpu1, VkPhysicalDeviceCompatibilityInfo* pInfo)
{
pCurObj = (VkBaseLayerObject *) gpu0;
loader_platform_thread_once(&tabOnce, initParamChecker);
- VkResult result = nextTable.GetMultiGpuCompatibility(gpu0, gpu1, pInfo);
+ VkResult result = nextTable.GetMultiDeviceCompatibility(gpu0, gpu1, pInfo);
return result;
}
-VK_LAYER_EXPORT VkResult VKAPI vkOpenSharedMemory(VkDevice device, const VkMemoryOpenInfo* pOpenInfo, VkGpuMemory* pMem)
+VK_LAYER_EXPORT VkResult VKAPI vkOpenSharedMemory(VkDevice device, const VkMemoryOpenInfo* pOpenInfo, VkDeviceMemory* pMem)
{
char str[1024];
if (!pOpenInfo) {
@@ -450,7 +450,7 @@
return result;
}
-VK_LAYER_EXPORT VkResult VKAPI vkOpenPeerMemory(VkDevice device, const VkPeerMemoryOpenInfo* pOpenInfo, VkGpuMemory* pMem)
+VK_LAYER_EXPORT VkResult VKAPI vkOpenPeerMemory(VkDevice device, const VkPeerMemoryOpenInfo* pOpenInfo, VkDeviceMemory* pMem)
{
char str[1024];
if (!pOpenInfo) {
@@ -465,7 +465,7 @@
return result;
}
-VK_LAYER_EXPORT VkResult VKAPI vkOpenPeerImage(VkDevice device, const VkPeerImageOpenInfo* pOpenInfo, VkImage* pImage, VkGpuMemory* pMem)
+VK_LAYER_EXPORT VkResult VKAPI vkOpenPeerImage(VkDevice device, const VkPeerImageOpenInfo* pOpenInfo, VkImage* pImage, VkDeviceMemory* pMem)
{
char str[1024];
if (!pOpenInfo) {
@@ -498,21 +498,21 @@
return result;
}
-VK_LAYER_EXPORT VkResult VKAPI vkQueueBindObjectMemory(VkQueue queue, VkObject object, uint32_t allocationIdx, VkGpuMemory mem, VkGpuSize offset)
+VK_LAYER_EXPORT VkResult VKAPI vkQueueBindObjectMemory(VkQueue queue, VkObject object, uint32_t allocationIdx, VkDeviceMemory mem, VkDeviceSize offset)
{
VkResult result = nextTable.QueueBindObjectMemory(queue, object, allocationIdx, mem, offset);
return result;
}
-VK_LAYER_EXPORT VkResult VKAPI vkQueueBindObjectMemoryRange(VkQueue queue, VkObject object, uint32_t allocationIdx, VkGpuSize rangeOffset, VkGpuSize rangeSize, VkGpuMemory mem, VkGpuSize memOffset)
+VK_LAYER_EXPORT VkResult VKAPI vkQueueBindObjectMemoryRange(VkQueue queue, VkObject object, uint32_t allocationIdx, VkDeviceSize rangeOffset, VkDeviceSize rangeSize, VkDeviceMemory mem, VkDeviceSize memOffset)
{
VkResult result = nextTable.QueueBindObjectMemoryRange(queue, object, allocationIdx, rangeOffset, rangeSize, mem, memOffset);
return result;
}
-VK_LAYER_EXPORT VkResult VKAPI vkQueueBindImageMemoryRange(VkQueue queue, VkImage image, uint32_t allocationIdx, const VkImageMemoryBindInfo* pBindInfo, VkGpuMemory mem, VkGpuSize memOffset)
+VK_LAYER_EXPORT VkResult VKAPI vkQueueBindImageMemoryRange(VkQueue queue, VkImage image, uint32_t allocationIdx, const VkImageMemoryBindInfo* pBindInfo, VkDeviceMemory mem, VkDeviceSize memOffset)
{
char str[1024];
if (!pBindInfo) {
@@ -643,10 +643,10 @@
return result;
}
-VK_LAYER_EXPORT VkResult VKAPI vkGetQueryPoolResults(VkQueryPool queryPool, uint32_t startQuery, uint32_t queryCount, size_t* pDataSize, void* pData)
+VK_LAYER_EXPORT VkResult VKAPI vkGetQueryPoolResults(VkQueryPool queryPool, uint32_t startQuery, uint32_t queryCount, size_t* pDataSize, void* pData, VkQueryResultFlags flags)
{
- VkResult result = nextTable.GetQueryPoolResults(queryPool, startQuery, queryCount, pDataSize, pData);
+ VkResult result = nextTable.GetQueryPoolResults(queryPool, startQuery, queryCount, pDataSize, pData, flags);
return result;
}
@@ -732,7 +732,7 @@
VkFormatProperties properties;
size_t size = sizeof(properties);
VkResult result = nextTable.GetFormatInfo(device, pCreateInfo->format,
- VK_INFO_TYPE_FORMAT_PROPERTIES, &size, &properties);
+ VK_FORMAT_INFO_TYPE_PROPERTIES, &size, &properties);
if(result != VK_SUCCESS)
{
char const str[] = "vkCreateImage parameter, VkFormat pCreateInfo->format, cannot be "\
@@ -1219,12 +1219,12 @@
uint32_t startBinding,
uint32_t bindingCount,
const VkBuffer* pBuffers,
- const VkGpuSize* pOffsets)
+ const VkDeviceSize* pOffsets)
{
nextTable.CmdBindVertexBuffers(cmdBuffer, startBinding, bindingCount, pBuffers, pOffsets);
}
-VK_LAYER_EXPORT void VKAPI vkCmdBindIndexBuffer(VkCmdBuffer cmdBuffer, VkBuffer buffer, VkGpuSize offset, VkIndexType indexType)
+VK_LAYER_EXPORT void VKAPI vkCmdBindIndexBuffer(VkCmdBuffer cmdBuffer, VkBuffer buffer, VkDeviceSize offset, VkIndexType indexType)
{
char str[1024];
if (!validate_VkIndexType(indexType)) {
@@ -1246,13 +1246,13 @@
nextTable.CmdDrawIndexed(cmdBuffer, firstIndex, indexCount, vertexOffset, firstInstance, instanceCount);
}
-VK_LAYER_EXPORT void VKAPI vkCmdDrawIndirect(VkCmdBuffer cmdBuffer, VkBuffer buffer, VkGpuSize offset, uint32_t count, uint32_t stride)
+VK_LAYER_EXPORT void VKAPI vkCmdDrawIndirect(VkCmdBuffer cmdBuffer, VkBuffer buffer, VkDeviceSize offset, uint32_t count, uint32_t stride)
{
nextTable.CmdDrawIndirect(cmdBuffer, buffer, offset, count, stride);
}
-VK_LAYER_EXPORT void VKAPI vkCmdDrawIndexedIndirect(VkCmdBuffer cmdBuffer, VkBuffer buffer, VkGpuSize offset, uint32_t count, uint32_t stride)
+VK_LAYER_EXPORT void VKAPI vkCmdDrawIndexedIndirect(VkCmdBuffer cmdBuffer, VkBuffer buffer, VkDeviceSize offset, uint32_t count, uint32_t stride)
{
nextTable.CmdDrawIndexedIndirect(cmdBuffer, buffer, offset, count, stride);
@@ -1264,7 +1264,7 @@
nextTable.CmdDispatch(cmdBuffer, x, y, z);
}
-VK_LAYER_EXPORT void VKAPI vkCmdDispatchIndirect(VkCmdBuffer cmdBuffer, VkBuffer buffer, VkGpuSize offset)
+VK_LAYER_EXPORT void VKAPI vkCmdDispatchIndirect(VkCmdBuffer cmdBuffer, VkBuffer buffer, VkDeviceSize offset)
{
nextTable.CmdDispatchIndirect(cmdBuffer, buffer, offset);
@@ -1373,13 +1373,13 @@
nextTable.CmdCloneImageData(cmdBuffer, srcImage, srcImageLayout, destImage, destImageLayout);
}
-VK_LAYER_EXPORT void VKAPI vkCmdUpdateBuffer(VkCmdBuffer cmdBuffer, VkBuffer destBuffer, VkGpuSize destOffset, VkGpuSize dataSize, const uint32_t* pData)
+VK_LAYER_EXPORT void VKAPI vkCmdUpdateBuffer(VkCmdBuffer cmdBuffer, VkBuffer destBuffer, VkDeviceSize destOffset, VkDeviceSize dataSize, const uint32_t* pData)
{
nextTable.CmdUpdateBuffer(cmdBuffer, destBuffer, destOffset, dataSize, pData);
}
-VK_LAYER_EXPORT void VKAPI vkCmdFillBuffer(VkCmdBuffer cmdBuffer, VkBuffer destBuffer, VkGpuSize destOffset, VkGpuSize fillSize, uint32_t data)
+VK_LAYER_EXPORT void VKAPI vkCmdFillBuffer(VkCmdBuffer cmdBuffer, VkBuffer destBuffer, VkDeviceSize destOffset, VkDeviceSize fillSize, uint32_t data)
{
nextTable.CmdFillBuffer(cmdBuffer, destBuffer, destOffset, fillSize, data);
@@ -1460,32 +1460,14 @@
nextTable.CmdResetEvent(cmdBuffer, event, pipeEvent);
}
-VK_LAYER_EXPORT void VKAPI vkCmdWaitEvents(VkCmdBuffer cmdBuffer, const VkEventWaitInfo* pWaitInfo)
+VK_LAYER_EXPORT void VKAPI vkCmdWaitEvents(VkCmdBuffer cmdBuffer, VkWaitEvent waitEvent, uint32_t eventCount, const VkEvent* pEvents, uint32_t memBarrierCount, const void** ppMemBarriers)
{
- char str[1024];
- if (!pWaitInfo) {
- sprintf(str, "Struct ptr parameter pWaitInfo to function CmdWaitEvents is NULL.");
- layerCbMsg(VK_DBG_MSG_UNKNOWN, VK_VALIDATION_LEVEL_0, NULL, 0, 1, "PARAMCHECK", str);
- }
- else if (!vk_validate_vkeventwaitinfo(pWaitInfo)) {
- sprintf(str, "Parameter pWaitInfo to function CmdWaitEvents contains an invalid value.");
- layerCbMsg(VK_DBG_MSG_ERROR, VK_VALIDATION_LEVEL_0, NULL, 0, 1, "PARAMCHECK", str);
- }
- nextTable.CmdWaitEvents(cmdBuffer, pWaitInfo);
+ nextTable.CmdWaitEvents(cmdBuffer, waitEvent, eventCount, pEvents, memBarrierCount, ppMemBarriers);
}
-VK_LAYER_EXPORT void VKAPI vkCmdPipelineBarrier(VkCmdBuffer cmdBuffer, const VkPipelineBarrier* pBarrier)
+VK_LAYER_EXPORT void VKAPI vkCmdPipelineBarrier(VkCmdBuffer cmdBuffer, VkWaitEvent waitEvent, uint32_t pipeEventCount, const VkPipeEvent* pPipeEvents, uint32_t memBarrierCount, const void** ppMemBarriers)
{
- char str[1024];
- if (!pBarrier) {
- sprintf(str, "Struct ptr parameter pBarrier to function CmdPipelineBarrier is NULL.");
- layerCbMsg(VK_DBG_MSG_UNKNOWN, VK_VALIDATION_LEVEL_0, NULL, 0, 1, "PARAMCHECK", str);
- }
- else if (!vk_validate_vkpipelinebarrier(pBarrier)) {
- sprintf(str, "Parameter pBarrier to function CmdPipelineBarrier contains an invalid value.");
- layerCbMsg(VK_DBG_MSG_ERROR, VK_VALIDATION_LEVEL_0, NULL, 0, 1, "PARAMCHECK", str);
- }
- nextTable.CmdPipelineBarrier(cmdBuffer, pBarrier);
+ nextTable.CmdPipelineBarrier(cmdBuffer, waitEvent, pipeEventCount, pPipeEvents, memBarrierCount, ppMemBarriers);
}
VK_LAYER_EXPORT void VKAPI vkCmdBeginQuery(VkCmdBuffer cmdBuffer, VkQueryPool queryPool, uint32_t slot, VkFlags flags)
@@ -1506,7 +1488,7 @@
nextTable.CmdResetQueryPool(cmdBuffer, queryPool, startQuery, queryCount);
}
-VK_LAYER_EXPORT void VKAPI vkCmdWriteTimestamp(VkCmdBuffer cmdBuffer, VkTimestampType timestampType, VkBuffer destBuffer, VkGpuSize destOffset)
+VK_LAYER_EXPORT void VKAPI vkCmdWriteTimestamp(VkCmdBuffer cmdBuffer, VkTimestampType timestampType, VkBuffer destBuffer, VkDeviceSize destOffset)
{
char str[1024];
if (!validate_VkTimestampType(timestampType)) {
@@ -1526,7 +1508,7 @@
nextTable.CmdInitAtomicCounters(cmdBuffer, pipelineBindPoint, startCounter, counterCount, pData);
}
-VK_LAYER_EXPORT void VKAPI vkCmdLoadAtomicCounters(VkCmdBuffer cmdBuffer, VkPipelineBindPoint pipelineBindPoint, uint32_t startCounter, uint32_t counterCount, VkBuffer srcBuffer, VkGpuSize srcOffset)
+VK_LAYER_EXPORT void VKAPI vkCmdLoadAtomicCounters(VkCmdBuffer cmdBuffer, VkPipelineBindPoint pipelineBindPoint, uint32_t startCounter, uint32_t counterCount, VkBuffer srcBuffer, VkDeviceSize srcOffset)
{
char str[1024];
if (!validate_VkPipelineBindPoint(pipelineBindPoint)) {
@@ -1536,7 +1518,7 @@
nextTable.CmdLoadAtomicCounters(cmdBuffer, pipelineBindPoint, startCounter, counterCount, srcBuffer, srcOffset);
}
-VK_LAYER_EXPORT void VKAPI vkCmdSaveAtomicCounters(VkCmdBuffer cmdBuffer, VkPipelineBindPoint pipelineBindPoint, uint32_t startCounter, uint32_t counterCount, VkBuffer destBuffer, VkGpuSize destOffset)
+VK_LAYER_EXPORT void VKAPI vkCmdSaveAtomicCounters(VkCmdBuffer cmdBuffer, VkPipelineBindPoint pipelineBindPoint, uint32_t startCounter, uint32_t counterCount, VkBuffer destBuffer, VkDeviceSize destOffset)
{
char str[1024];
if (!validate_VkPipelineBindPoint(pipelineBindPoint)) {
@@ -1618,7 +1600,7 @@
VkFormatProperties properties;
size_t size = sizeof(properties);
VkResult result = nextTable.GetFormatInfo(device, pCreateInfo->pColorFormats[i],
- VK_INFO_TYPE_FORMAT_PROPERTIES, &size, &properties);
+ VK_FORMAT_INFO_TYPE_PROPERTIES, &size, &properties);
if(result != VK_SUCCESS)
{
std::stringstream ss;
@@ -1746,7 +1728,7 @@
VkFormatProperties properties;
size_t size = sizeof(properties);
VkResult result = nextTable.GetFormatInfo(device, pCreateInfo->depthStencilFormat,
- VK_INFO_TYPE_FORMAT_PROPERTIES, &size, &properties);
+ VK_FORMAT_INFO_TYPE_PROPERTIES, &size, &properties);
if(result != VK_SUCCESS)
{
char const str[] = "vkCreateRenderPass parameter, VkFormat pCreateInfo->"\
@@ -1861,7 +1843,7 @@
// This layer intercepts callbacks
VK_LAYER_DBG_FUNCTION_NODE *pNewDbgFuncNode = (VK_LAYER_DBG_FUNCTION_NODE*)malloc(sizeof(VK_LAYER_DBG_FUNCTION_NODE));
if (!pNewDbgFuncNode)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
pNewDbgFuncNode->pfnMsgCallback = pfnMsgCallback;
pNewDbgFuncNode->pUserData = pUserData;
pNewDbgFuncNode->pNext = g_pDbgFunctionHead;
@@ -1942,7 +1924,7 @@
#if defined(__linux__) || defined(XCB_NVIDIA)
-VK_LAYER_EXPORT VkResult VKAPI vkWsiX11AssociateConnection(VkPhysicalGpu gpu, const VK_WSI_X11_CONNECTION_INFO* pConnectionInfo)
+VK_LAYER_EXPORT VkResult VKAPI vkWsiX11AssociateConnection(VkPhysicalDevice gpu, const VK_WSI_X11_CONNECTION_INFO* pConnectionInfo)
{
pCurObj = (VkBaseLayerObject *) gpu;
loader_platform_thread_once(&tabOnce, initParamChecker);
@@ -1958,7 +1940,7 @@
return result;
}
-VK_LAYER_EXPORT VkResult VKAPI vkWsiX11CreatePresentableImage(VkDevice device, const VK_WSI_X11_PRESENTABLE_IMAGE_CREATE_INFO* pCreateInfo, VkImage* pImage, VkGpuMemory* pMem)
+VK_LAYER_EXPORT VkResult VKAPI vkWsiX11CreatePresentableImage(VkDevice device, const VK_WSI_X11_PRESENTABLE_IMAGE_CREATE_INFO* pCreateInfo, VkImage* pImage, VkDeviceMemory* pMem)
{
VkResult result = nextTable.WsiX11CreatePresentableImage(device, pCreateInfo, pImage, pMem);
@@ -1975,7 +1957,7 @@
#endif
#include "vk_generic_intercept_proc_helper.h"
-VK_LAYER_EXPORT void* VKAPI vkGetProcAddr(VkPhysicalGpu gpu, const char* funcName)
+VK_LAYER_EXPORT void* VKAPI vkGetProcAddr(VkPhysicalDevice gpu, const char* funcName)
{
VkBaseLayerObject* gpuw = (VkBaseLayerObject *) gpu;
void* addr;
@@ -1990,7 +1972,7 @@
else {
if (gpuw->pGPA == NULL)
return NULL;
- return gpuw->pGPA((VkPhysicalGpu)gpuw->nextObject, funcName);
+ return gpuw->pGPA((VkPhysicalDevice)gpuw->nextObject, funcName);
}
}
diff --git a/loader/loader.c b/loader/loader.c
index b71a39a..086867e 100644
--- a/loader/loader.c
+++ b/loader/loader.c
@@ -716,7 +716,7 @@
loader.layer_scanned = true;
}
-static void loader_init_dispatch_table(VkLayerDispatchTable *tab, PFN_vkGetProcAddr fpGPA, VkPhysicalGpu gpu)
+static void loader_init_dispatch_table(VkLayerDispatchTable *tab, PFN_vkGetProcAddr fpGPA, VkPhysicalDevice gpu)
{
loader_initialize_dispatch_table(tab, fpGPA, gpu);
@@ -724,7 +724,7 @@
tab->EnumerateLayers = vkEnumerateLayers;
}
-static void *loader_gpa_internal(VkPhysicalGpu gpu, const char * pName)
+static void *loader_gpa_internal(VkPhysicalDevice gpu, const char * pName)
{
if (gpu == NULL) {
return NULL;;
@@ -1099,20 +1099,20 @@
ptr_instance = (struct loader_instance*) malloc(sizeof(struct loader_instance));
if (ptr_instance == NULL) {
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
}
memset(ptr_instance, 0, sizeof(struct loader_instance));
ptr_instance->extension_count = pCreateInfo->extensionCount;
ptr_instance->extension_names = (ptr_instance->extension_count > 0) ?
malloc(sizeof (char *) * ptr_instance->extension_count) : NULL;
if (ptr_instance->extension_names == NULL && (ptr_instance->extension_count > 0))
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
for (i = 0; i < ptr_instance->extension_count; i++) {
if (!loader_is_extension_scanned(pCreateInfo->ppEnabledExtensionNames[i]))
return VK_ERROR_INVALID_EXTENSION;
ptr_instance->extension_names[i] = malloc(strlen(pCreateInfo->ppEnabledExtensionNames[i]) + 1);
if (ptr_instance->extension_names[i] == NULL)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
strcpy(ptr_instance->extension_names[i], pCreateInfo->ppEnabledExtensionNames[i]);
}
ptr_instance->next = loader.instances;
@@ -1198,7 +1198,7 @@
VkInstance instance,
uint32_t* pPhysicalDeviceCount,
- VkPhysicalGpu* pPhysicalDevices)
+ VkPhysicalDevice* pPhysicalDevices)
{
struct loader_instance *ptr_instance = (struct loader_instance *) instance;
struct loader_icd *icd;
@@ -1223,12 +1223,12 @@
} else
{
- VkPhysicalGpu* gpus;
+ VkPhysicalDevice* gpus;
if (*pPhysicalDeviceCount < ptr_instance->total_gpu_count)
return VK_ERROR_INVALID_VALUE;
- gpus = malloc( sizeof(VkPhysicalGpu) * *pPhysicalDeviceCount);
+ gpus = malloc( sizeof(VkPhysicalDevice) * *pPhysicalDeviceCount);
if (!gpus)
- return VK_ERROR_OUT_OF_MEMORY;
+ return VK_ERROR_OUT_OF_HOST_MEMORY;
while (icd) {
VkBaseLayerObject * wrapped_gpus;
PFN_vkGetProcAddr get_proc_addr = icd->scanned_icds->GetProcAddr;
@@ -1282,7 +1282,7 @@
return (count > 0) ? VK_SUCCESS : res;
}
-LOADER_EXPORT void * VKAPI vkGetProcAddr(VkPhysicalGpu gpu, const char * pName)
+LOADER_EXPORT void * VKAPI vkGetProcAddr(VkPhysicalDevice gpu, const char * pName)
{
if (gpu == NULL) {
@@ -1369,7 +1369,7 @@
return VK_SUCCESS;
}
-LOADER_EXPORT VkResult VKAPI vkEnumerateLayers(VkPhysicalGpu gpu, size_t maxLayerCount, size_t maxStringSize, size_t* pOutLayerCount, char* const* pOutLayers, void* pReserved)
+LOADER_EXPORT VkResult VKAPI vkEnumerateLayers(VkPhysicalDevice gpu, size_t maxLayerCount, size_t maxStringSize, size_t* pOutLayerCount, char* const* pOutLayers, void* pReserved)
{
uint32_t gpu_index;
size_t count = 0;
diff --git a/tests/blit_tests.cpp b/tests/blit_tests.cpp
index 756e1b3..c92c080 100644
--- a/tests/blit_tests.cpp
+++ b/tests/blit_tests.cpp
@@ -41,7 +41,7 @@
void set_solid_pattern(const std::vector<uint8_t> &solid);
- VkGpuSize buffer_size() const;
+ VkDeviceSize buffer_size() const;
bool fill(Buffer &buf) const { return walk(FILL, buf); }
bool fill(Image &img) const { return walk(FILL, img); }
bool check(Buffer &buf) const { return walk(CHECK, buf); }
@@ -87,7 +87,7 @@
: info_(info), regions_(), pattern_(HASH)
{
// create a region for every mip level in array slice 0
- VkGpuSize offset = 0;
+ VkDeviceSize offset = 0;
for (uint32_t lv = 0; lv < info_.mipLevels; lv++) {
VkBufferImageCopy region = {};
@@ -97,14 +97,14 @@
region.imageExtent = Image::extent(info_.extent, lv);
if (info_.usage & VK_IMAGE_USAGE_DEPTH_STENCIL_BIT) {
- if (info_.format != VK_FMT_S8_UINT) {
+ if (info_.format != VK_FORMAT_S8_UINT) {
region.imageSubresource.aspect = VK_IMAGE_ASPECT_DEPTH;
regions_.push_back(region);
}
- if (info_.format == VK_FMT_D16_UNORM_S8_UINT ||
- info_.format == VK_FMT_D32_SFLOAT_S8_UINT ||
- info_.format == VK_FMT_S8_UINT) {
+ if (info_.format == VK_FORMAT_D16_UNORM_S8_UINT ||
+ info_.format == VK_FORMAT_D32_SFLOAT_S8_UINT ||
+ info_.format == VK_FORMAT_S8_UINT) {
region.imageSubresource.aspect = VK_IMAGE_ASPECT_STENCIL;
regions_.push_back(region);
}
@@ -119,7 +119,7 @@
// arraySize should be limited in our tests. If this proves to be an
// issue, we can store only the regions for array slice 0 and be smart.
if (info_.arraySize > 1) {
- const VkGpuSize slice_pitch = offset;
+ const VkDeviceSize slice_pitch = offset;
const uint32_t slice_region_count = regions_.size();
regions_.reserve(slice_region_count * info_.arraySize);
@@ -139,7 +139,7 @@
ImageChecker::ImageChecker(const VkImageCreateInfo &info, const std::vector<VkImageSubresourceRange> &ranges)
: info_(info), regions_(), pattern_(HASH)
{
- VkGpuSize offset = 0;
+ VkDeviceSize offset = 0;
for (std::vector<VkImageSubresourceRange>::const_iterator it = ranges.begin();
it != ranges.end(); it++) {
for (uint32_t lv = 0; lv < it->mipLevels; lv++) {
@@ -182,9 +182,9 @@
return layout;
}
-VkGpuSize ImageChecker::buffer_size() const
+VkDeviceSize ImageChecker::buffer_size() const
{
- VkGpuSize size = 0;
+ VkDeviceSize size = 0;
for (std::vector<VkBufferImageCopy>::const_iterator it = regions_.begin();
it != regions_.end(); it++) {
@@ -301,182 +301,182 @@
static const struct format_info {
size_t size;
uint32_t channel_count;
- } format_table[VK_NUM_FMT] = {
- [VK_FMT_UNDEFINED] = { 0, 0 },
- [VK_FMT_R4G4_UNORM] = { 1, 2 },
- [VK_FMT_R4G4_USCALED] = { 1, 2 },
- [VK_FMT_R4G4B4A4_UNORM] = { 2, 4 },
- [VK_FMT_R4G4B4A4_USCALED] = { 2, 4 },
- [VK_FMT_R5G6B5_UNORM] = { 2, 3 },
- [VK_FMT_R5G6B5_USCALED] = { 2, 3 },
- [VK_FMT_R5G5B5A1_UNORM] = { 2, 4 },
- [VK_FMT_R5G5B5A1_USCALED] = { 2, 4 },
- [VK_FMT_R8_UNORM] = { 1, 1 },
- [VK_FMT_R8_SNORM] = { 1, 1 },
- [VK_FMT_R8_USCALED] = { 1, 1 },
- [VK_FMT_R8_SSCALED] = { 1, 1 },
- [VK_FMT_R8_UINT] = { 1, 1 },
- [VK_FMT_R8_SINT] = { 1, 1 },
- [VK_FMT_R8_SRGB] = { 1, 1 },
- [VK_FMT_R8G8_UNORM] = { 2, 2 },
- [VK_FMT_R8G8_SNORM] = { 2, 2 },
- [VK_FMT_R8G8_USCALED] = { 2, 2 },
- [VK_FMT_R8G8_SSCALED] = { 2, 2 },
- [VK_FMT_R8G8_UINT] = { 2, 2 },
- [VK_FMT_R8G8_SINT] = { 2, 2 },
- [VK_FMT_R8G8_SRGB] = { 2, 2 },
- [VK_FMT_R8G8B8_UNORM] = { 3, 3 },
- [VK_FMT_R8G8B8_SNORM] = { 3, 3 },
- [VK_FMT_R8G8B8_USCALED] = { 3, 3 },
- [VK_FMT_R8G8B8_SSCALED] = { 3, 3 },
- [VK_FMT_R8G8B8_UINT] = { 3, 3 },
- [VK_FMT_R8G8B8_SINT] = { 3, 3 },
- [VK_FMT_R8G8B8_SRGB] = { 3, 3 },
- [VK_FMT_R8G8B8A8_UNORM] = { 4, 4 },
- [VK_FMT_R8G8B8A8_SNORM] = { 4, 4 },
- [VK_FMT_R8G8B8A8_USCALED] = { 4, 4 },
- [VK_FMT_R8G8B8A8_SSCALED] = { 4, 4 },
- [VK_FMT_R8G8B8A8_UINT] = { 4, 4 },
- [VK_FMT_R8G8B8A8_SINT] = { 4, 4 },
- [VK_FMT_R8G8B8A8_SRGB] = { 4, 4 },
- [VK_FMT_R10G10B10A2_UNORM] = { 4, 4 },
- [VK_FMT_R10G10B10A2_SNORM] = { 4, 4 },
- [VK_FMT_R10G10B10A2_USCALED] = { 4, 4 },
- [VK_FMT_R10G10B10A2_SSCALED] = { 4, 4 },
- [VK_FMT_R10G10B10A2_UINT] = { 4, 4 },
- [VK_FMT_R10G10B10A2_SINT] = { 4, 4 },
- [VK_FMT_R16_UNORM] = { 2, 1 },
- [VK_FMT_R16_SNORM] = { 2, 1 },
- [VK_FMT_R16_USCALED] = { 2, 1 },
- [VK_FMT_R16_SSCALED] = { 2, 1 },
- [VK_FMT_R16_UINT] = { 2, 1 },
- [VK_FMT_R16_SINT] = { 2, 1 },
- [VK_FMT_R16_SFLOAT] = { 2, 1 },
- [VK_FMT_R16G16_UNORM] = { 4, 2 },
- [VK_FMT_R16G16_SNORM] = { 4, 2 },
- [VK_FMT_R16G16_USCALED] = { 4, 2 },
- [VK_FMT_R16G16_SSCALED] = { 4, 2 },
- [VK_FMT_R16G16_UINT] = { 4, 2 },
- [VK_FMT_R16G16_SINT] = { 4, 2 },
- [VK_FMT_R16G16_SFLOAT] = { 4, 2 },
- [VK_FMT_R16G16B16_UNORM] = { 6, 3 },
- [VK_FMT_R16G16B16_SNORM] = { 6, 3 },
- [VK_FMT_R16G16B16_USCALED] = { 6, 3 },
- [VK_FMT_R16G16B16_SSCALED] = { 6, 3 },
- [VK_FMT_R16G16B16_UINT] = { 6, 3 },
- [VK_FMT_R16G16B16_SINT] = { 6, 3 },
- [VK_FMT_R16G16B16_SFLOAT] = { 6, 3 },
- [VK_FMT_R16G16B16A16_UNORM] = { 8, 4 },
- [VK_FMT_R16G16B16A16_SNORM] = { 8, 4 },
- [VK_FMT_R16G16B16A16_USCALED] = { 8, 4 },
- [VK_FMT_R16G16B16A16_SSCALED] = { 8, 4 },
- [VK_FMT_R16G16B16A16_UINT] = { 8, 4 },
- [VK_FMT_R16G16B16A16_SINT] = { 8, 4 },
- [VK_FMT_R16G16B16A16_SFLOAT] = { 8, 4 },
- [VK_FMT_R32_UINT] = { 4, 1 },
- [VK_FMT_R32_SINT] = { 4, 1 },
- [VK_FMT_R32_SFLOAT] = { 4, 1 },
- [VK_FMT_R32G32_UINT] = { 8, 2 },
- [VK_FMT_R32G32_SINT] = { 8, 2 },
- [VK_FMT_R32G32_SFLOAT] = { 8, 2 },
- [VK_FMT_R32G32B32_UINT] = { 12, 3 },
- [VK_FMT_R32G32B32_SINT] = { 12, 3 },
- [VK_FMT_R32G32B32_SFLOAT] = { 12, 3 },
- [VK_FMT_R32G32B32A32_UINT] = { 16, 4 },
- [VK_FMT_R32G32B32A32_SINT] = { 16, 4 },
- [VK_FMT_R32G32B32A32_SFLOAT] = { 16, 4 },
- [VK_FMT_R64_SFLOAT] = { 8, 1 },
- [VK_FMT_R64G64_SFLOAT] = { 16, 2 },
- [VK_FMT_R64G64B64_SFLOAT] = { 24, 3 },
- [VK_FMT_R64G64B64A64_SFLOAT] = { 32, 4 },
- [VK_FMT_R11G11B10_UFLOAT] = { 4, 3 },
- [VK_FMT_R9G9B9E5_UFLOAT] = { 4, 3 },
- [VK_FMT_D16_UNORM] = { 2, 1 },
- [VK_FMT_D24_UNORM] = { 3, 1 },
- [VK_FMT_D32_SFLOAT] = { 4, 1 },
- [VK_FMT_S8_UINT] = { 1, 1 },
- [VK_FMT_D16_UNORM_S8_UINT] = { 3, 2 },
- [VK_FMT_D24_UNORM_S8_UINT] = { 4, 2 },
- [VK_FMT_D32_SFLOAT_S8_UINT] = { 4, 2 },
- [VK_FMT_BC1_RGB_UNORM] = { 8, 4 },
- [VK_FMT_BC1_RGB_SRGB] = { 8, 4 },
- [VK_FMT_BC1_RGBA_UNORM] = { 8, 4 },
- [VK_FMT_BC1_RGBA_SRGB] = { 8, 4 },
- [VK_FMT_BC2_UNORM] = { 16, 4 },
- [VK_FMT_BC2_SRGB] = { 16, 4 },
- [VK_FMT_BC3_UNORM] = { 16, 4 },
- [VK_FMT_BC3_SRGB] = { 16, 4 },
- [VK_FMT_BC4_UNORM] = { 8, 4 },
- [VK_FMT_BC4_SNORM] = { 8, 4 },
- [VK_FMT_BC5_UNORM] = { 16, 4 },
- [VK_FMT_BC5_SNORM] = { 16, 4 },
- [VK_FMT_BC6H_UFLOAT] = { 16, 4 },
- [VK_FMT_BC6H_SFLOAT] = { 16, 4 },
- [VK_FMT_BC7_UNORM] = { 16, 4 },
- [VK_FMT_BC7_SRGB] = { 16, 4 },
+ } format_table[VK_NUM_FORMAT] = {
+ [VK_FORMAT_UNDEFINED] = { 0, 0 },
+ [VK_FORMAT_R4G4_UNORM] = { 1, 2 },
+ [VK_FORMAT_R4G4_USCALED] = { 1, 2 },
+ [VK_FORMAT_R4G4B4A4_UNORM] = { 2, 4 },
+ [VK_FORMAT_R4G4B4A4_USCALED] = { 2, 4 },
+ [VK_FORMAT_R5G6B5_UNORM] = { 2, 3 },
+ [VK_FORMAT_R5G6B5_USCALED] = { 2, 3 },
+ [VK_FORMAT_R5G5B5A1_UNORM] = { 2, 4 },
+ [VK_FORMAT_R5G5B5A1_USCALED] = { 2, 4 },
+ [VK_FORMAT_R8_UNORM] = { 1, 1 },
+ [VK_FORMAT_R8_SNORM] = { 1, 1 },
+ [VK_FORMAT_R8_USCALED] = { 1, 1 },
+ [VK_FORMAT_R8_SSCALED] = { 1, 1 },
+ [VK_FORMAT_R8_UINT] = { 1, 1 },
+ [VK_FORMAT_R8_SINT] = { 1, 1 },
+ [VK_FORMAT_R8_SRGB] = { 1, 1 },
+ [VK_FORMAT_R8G8_UNORM] = { 2, 2 },
+ [VK_FORMAT_R8G8_SNORM] = { 2, 2 },
+ [VK_FORMAT_R8G8_USCALED] = { 2, 2 },
+ [VK_FORMAT_R8G8_SSCALED] = { 2, 2 },
+ [VK_FORMAT_R8G8_UINT] = { 2, 2 },
+ [VK_FORMAT_R8G8_SINT] = { 2, 2 },
+ [VK_FORMAT_R8G8_SRGB] = { 2, 2 },
+ [VK_FORMAT_R8G8B8_UNORM] = { 3, 3 },
+ [VK_FORMAT_R8G8B8_SNORM] = { 3, 3 },
+ [VK_FORMAT_R8G8B8_USCALED] = { 3, 3 },
+ [VK_FORMAT_R8G8B8_SSCALED] = { 3, 3 },
+ [VK_FORMAT_R8G8B8_UINT] = { 3, 3 },
+ [VK_FORMAT_R8G8B8_SINT] = { 3, 3 },
+ [VK_FORMAT_R8G8B8_SRGB] = { 3, 3 },
+ [VK_FORMAT_R8G8B8A8_UNORM] = { 4, 4 },
+ [VK_FORMAT_R8G8B8A8_SNORM] = { 4, 4 },
+ [VK_FORMAT_R8G8B8A8_USCALED] = { 4, 4 },
+ [VK_FORMAT_R8G8B8A8_SSCALED] = { 4, 4 },
+ [VK_FORMAT_R8G8B8A8_UINT] = { 4, 4 },
+ [VK_FORMAT_R8G8B8A8_SINT] = { 4, 4 },
+ [VK_FORMAT_R8G8B8A8_SRGB] = { 4, 4 },
+ [VK_FORMAT_R10G10B10A2_UNORM] = { 4, 4 },
+ [VK_FORMAT_R10G10B10A2_SNORM] = { 4, 4 },
+ [VK_FORMAT_R10G10B10A2_USCALED] = { 4, 4 },
+ [VK_FORMAT_R10G10B10A2_SSCALED] = { 4, 4 },
+ [VK_FORMAT_R10G10B10A2_UINT] = { 4, 4 },
+ [VK_FORMAT_R10G10B10A2_SINT] = { 4, 4 },
+ [VK_FORMAT_R16_UNORM] = { 2, 1 },
+ [VK_FORMAT_R16_SNORM] = { 2, 1 },
+ [VK_FORMAT_R16_USCALED] = { 2, 1 },
+ [VK_FORMAT_R16_SSCALED] = { 2, 1 },
+ [VK_FORMAT_R16_UINT] = { 2, 1 },
+ [VK_FORMAT_R16_SINT] = { 2, 1 },
+ [VK_FORMAT_R16_SFLOAT] = { 2, 1 },
+ [VK_FORMAT_R16G16_UNORM] = { 4, 2 },
+ [VK_FORMAT_R16G16_SNORM] = { 4, 2 },
+ [VK_FORMAT_R16G16_USCALED] = { 4, 2 },
+ [VK_FORMAT_R16G16_SSCALED] = { 4, 2 },
+ [VK_FORMAT_R16G16_UINT] = { 4, 2 },
+ [VK_FORMAT_R16G16_SINT] = { 4, 2 },
+ [VK_FORMAT_R16G16_SFLOAT] = { 4, 2 },
+ [VK_FORMAT_R16G16B16_UNORM] = { 6, 3 },
+ [VK_FORMAT_R16G16B16_SNORM] = { 6, 3 },
+ [VK_FORMAT_R16G16B16_USCALED] = { 6, 3 },
+ [VK_FORMAT_R16G16B16_SSCALED] = { 6, 3 },
+ [VK_FORMAT_R16G16B16_UINT] = { 6, 3 },
+ [VK_FORMAT_R16G16B16_SINT] = { 6, 3 },
+ [VK_FORMAT_R16G16B16_SFLOAT] = { 6, 3 },
+ [VK_FORMAT_R16G16B16A16_UNORM] = { 8, 4 },
+ [VK_FORMAT_R16G16B16A16_SNORM] = { 8, 4 },
+ [VK_FORMAT_R16G16B16A16_USCALED] = { 8, 4 },
+ [VK_FORMAT_R16G16B16A16_SSCALED] = { 8, 4 },
+ [VK_FORMAT_R16G16B16A16_UINT] = { 8, 4 },
+ [VK_FORMAT_R16G16B16A16_SINT] = { 8, 4 },
+ [VK_FORMAT_R16G16B16A16_SFLOAT] = { 8, 4 },
+ [VK_FORMAT_R32_UINT] = { 4, 1 },
+ [VK_FORMAT_R32_SINT] = { 4, 1 },
+ [VK_FORMAT_R32_SFLOAT] = { 4, 1 },
+ [VK_FORMAT_R32G32_UINT] = { 8, 2 },
+ [VK_FORMAT_R32G32_SINT] = { 8, 2 },
+ [VK_FORMAT_R32G32_SFLOAT] = { 8, 2 },
+ [VK_FORMAT_R32G32B32_UINT] = { 12, 3 },
+ [VK_FORMAT_R32G32B32_SINT] = { 12, 3 },
+ [VK_FORMAT_R32G32B32_SFLOAT] = { 12, 3 },
+ [VK_FORMAT_R32G32B32A32_UINT] = { 16, 4 },
+ [VK_FORMAT_R32G32B32A32_SINT] = { 16, 4 },
+ [VK_FORMAT_R32G32B32A32_SFLOAT] = { 16, 4 },
+ [VK_FORMAT_R64_SFLOAT] = { 8, 1 },
+ [VK_FORMAT_R64G64_SFLOAT] = { 16, 2 },
+ [VK_FORMAT_R64G64B64_SFLOAT] = { 24, 3 },
+ [VK_FORMAT_R64G64B64A64_SFLOAT] = { 32, 4 },
+ [VK_FORMAT_R11G11B10_UFLOAT] = { 4, 3 },
+ [VK_FORMAT_R9G9B9E5_UFLOAT] = { 4, 3 },
+ [VK_FORMAT_D16_UNORM] = { 2, 1 },
+ [VK_FORMAT_D24_UNORM] = { 3, 1 },
+ [VK_FORMAT_D32_SFLOAT] = { 4, 1 },
+ [VK_FORMAT_S8_UINT] = { 1, 1 },
+ [VK_FORMAT_D16_UNORM_S8_UINT] = { 3, 2 },
+ [VK_FORMAT_D24_UNORM_S8_UINT] = { 4, 2 },
+ [VK_FORMAT_D32_SFLOAT_S8_UINT] = { 4, 2 },
+ [VK_FORMAT_BC1_RGB_UNORM] = { 8, 4 },
+ [VK_FORMAT_BC1_RGB_SRGB] = { 8, 4 },
+ [VK_FORMAT_BC1_RGBA_UNORM] = { 8, 4 },
+ [VK_FORMAT_BC1_RGBA_SRGB] = { 8, 4 },
+ [VK_FORMAT_BC2_UNORM] = { 16, 4 },
+ [VK_FORMAT_BC2_SRGB] = { 16, 4 },
+ [VK_FORMAT_BC3_UNORM] = { 16, 4 },
+ [VK_FORMAT_BC3_SRGB] = { 16, 4 },
+ [VK_FORMAT_BC4_UNORM] = { 8, 4 },
+ [VK_FORMAT_BC4_SNORM] = { 8, 4 },
+ [VK_FORMAT_BC5_UNORM] = { 16, 4 },
+ [VK_FORMAT_BC5_SNORM] = { 16, 4 },
+ [VK_FORMAT_BC6H_UFLOAT] = { 16, 4 },
+ [VK_FORMAT_BC6H_SFLOAT] = { 16, 4 },
+ [VK_FORMAT_BC7_UNORM] = { 16, 4 },
+ [VK_FORMAT_BC7_SRGB] = { 16, 4 },
// TODO: Initialize remaining compressed formats.
- [VK_FMT_ETC2_R8G8B8_UNORM] = { 0, 0 },
- [VK_FMT_ETC2_R8G8B8_SRGB] = { 0, 0 },
- [VK_FMT_ETC2_R8G8B8A1_UNORM] = { 0, 0 },
- [VK_FMT_ETC2_R8G8B8A1_SRGB] = { 0, 0 },
- [VK_FMT_ETC2_R8G8B8A8_UNORM] = { 0, 0 },
- [VK_FMT_ETC2_R8G8B8A8_SRGB] = { 0, 0 },
- [VK_FMT_EAC_R11_UNORM] = { 0, 0 },
- [VK_FMT_EAC_R11_SNORM] = { 0, 0 },
- [VK_FMT_EAC_R11G11_UNORM] = { 0, 0 },
- [VK_FMT_EAC_R11G11_SNORM] = { 0, 0 },
- [VK_FMT_ASTC_4x4_UNORM] = { 0, 0 },
- [VK_FMT_ASTC_4x4_SRGB] = { 0, 0 },
- [VK_FMT_ASTC_5x4_UNORM] = { 0, 0 },
- [VK_FMT_ASTC_5x4_SRGB] = { 0, 0 },
- [VK_FMT_ASTC_5x5_UNORM] = { 0, 0 },
- [VK_FMT_ASTC_5x5_SRGB] = { 0, 0 },
- [VK_FMT_ASTC_6x5_UNORM] = { 0, 0 },
- [VK_FMT_ASTC_6x5_SRGB] = { 0, 0 },
- [VK_FMT_ASTC_6x6_UNORM] = { 0, 0 },
- [VK_FMT_ASTC_6x6_SRGB] = { 0, 0 },
- [VK_FMT_ASTC_8x5_UNORM] = { 0, 0 },
- [VK_FMT_ASTC_8x5_SRGB] = { 0, 0 },
- [VK_FMT_ASTC_8x6_UNORM] = { 0, 0 },
- [VK_FMT_ASTC_8x6_SRGB] = { 0, 0 },
- [VK_FMT_ASTC_8x8_UNORM] = { 0, 0 },
- [VK_FMT_ASTC_8x8_SRGB] = { 0, 0 },
- [VK_FMT_ASTC_10x5_UNORM] = { 0, 0 },
- [VK_FMT_ASTC_10x5_SRGB] = { 0, 0 },
- [VK_FMT_ASTC_10x6_UNORM] = { 0, 0 },
- [VK_FMT_ASTC_10x6_SRGB] = { 0, 0 },
- [VK_FMT_ASTC_10x8_UNORM] = { 0, 0 },
- [VK_FMT_ASTC_10x8_SRGB] = { 0, 0 },
- [VK_FMT_ASTC_10x10_UNORM] = { 0, 0 },
- [VK_FMT_ASTC_10x10_SRGB] = { 0, 0 },
- [VK_FMT_ASTC_12x10_UNORM] = { 0, 0 },
- [VK_FMT_ASTC_12x10_SRGB] = { 0, 0 },
- [VK_FMT_ASTC_12x12_UNORM] = { 0, 0 },
- [VK_FMT_ASTC_12x12_SRGB] = { 0, 0 },
- [VK_FMT_B4G4R4A4_UNORM] = { 2, 4 },
- [VK_FMT_B5G5R5A1_UNORM] = { 2, 4 },
- [VK_FMT_B5G6R5_UNORM] = { 2, 3 },
- [VK_FMT_B5G6R5_USCALED] = { 2, 3 },
- [VK_FMT_B8G8R8_UNORM] = { 3, 3 },
- [VK_FMT_B8G8R8_SNORM] = { 3, 3 },
- [VK_FMT_B8G8R8_USCALED] = { 3, 3 },
- [VK_FMT_B8G8R8_SSCALED] = { 3, 3 },
- [VK_FMT_B8G8R8_UINT] = { 3, 3 },
- [VK_FMT_B8G8R8_SINT] = { 3, 3 },
- [VK_FMT_B8G8R8_SRGB] = { 3, 3 },
- [VK_FMT_B8G8R8A8_UNORM] = { 4, 4 },
- [VK_FMT_B8G8R8A8_SNORM] = { 4, 4 },
- [VK_FMT_B8G8R8A8_USCALED] = { 4, 4 },
- [VK_FMT_B8G8R8A8_SSCALED] = { 4, 4 },
- [VK_FMT_B8G8R8A8_UINT] = { 4, 4 },
- [VK_FMT_B8G8R8A8_SINT] = { 4, 4 },
- [VK_FMT_B8G8R8A8_SRGB] = { 4, 4 },
- [VK_FMT_B10G10R10A2_UNORM] = { 4, 4 },
- [VK_FMT_B10G10R10A2_SNORM] = { 4, 4 },
- [VK_FMT_B10G10R10A2_USCALED] = { 4, 4 },
- [VK_FMT_B10G10R10A2_SSCALED] = { 4, 4 },
- [VK_FMT_B10G10R10A2_UINT] = { 4, 4 },
- [VK_FMT_B10G10R10A2_SINT] = { 4, 4 },
+ [VK_FORMAT_ETC2_R8G8B8_UNORM] = { 0, 0 },
+ [VK_FORMAT_ETC2_R8G8B8_SRGB] = { 0, 0 },
+ [VK_FORMAT_ETC2_R8G8B8A1_UNORM] = { 0, 0 },
+ [VK_FORMAT_ETC2_R8G8B8A1_SRGB] = { 0, 0 },
+ [VK_FORMAT_ETC2_R8G8B8A8_UNORM] = { 0, 0 },
+ [VK_FORMAT_ETC2_R8G8B8A8_SRGB] = { 0, 0 },
+ [VK_FORMAT_EAC_R11_UNORM] = { 0, 0 },
+ [VK_FORMAT_EAC_R11_SNORM] = { 0, 0 },
+ [VK_FORMAT_EAC_R11G11_UNORM] = { 0, 0 },
+ [VK_FORMAT_EAC_R11G11_SNORM] = { 0, 0 },
+ [VK_FORMAT_ASTC_4x4_UNORM] = { 0, 0 },
+ [VK_FORMAT_ASTC_4x4_SRGB] = { 0, 0 },
+ [VK_FORMAT_ASTC_5x4_UNORM] = { 0, 0 },
+ [VK_FORMAT_ASTC_5x4_SRGB] = { 0, 0 },
+ [VK_FORMAT_ASTC_5x5_UNORM] = { 0, 0 },
+ [VK_FORMAT_ASTC_5x5_SRGB] = { 0, 0 },
+ [VK_FORMAT_ASTC_6x5_UNORM] = { 0, 0 },
+ [VK_FORMAT_ASTC_6x5_SRGB] = { 0, 0 },
+ [VK_FORMAT_ASTC_6x6_UNORM] = { 0, 0 },
+ [VK_FORMAT_ASTC_6x6_SRGB] = { 0, 0 },
+ [VK_FORMAT_ASTC_8x5_UNORM] = { 0, 0 },
+ [VK_FORMAT_ASTC_8x5_SRGB] = { 0, 0 },
+ [VK_FORMAT_ASTC_8x6_UNORM] = { 0, 0 },
+ [VK_FORMAT_ASTC_8x6_SRGB] = { 0, 0 },
+ [VK_FORMAT_ASTC_8x8_UNORM] = { 0, 0 },
+ [VK_FORMAT_ASTC_8x8_SRGB] = { 0, 0 },
+ [VK_FORMAT_ASTC_10x5_UNORM] = { 0, 0 },
+ [VK_FORMAT_ASTC_10x5_SRGB] = { 0, 0 },
+ [VK_FORMAT_ASTC_10x6_UNORM] = { 0, 0 },
+ [VK_FORMAT_ASTC_10x6_SRGB] = { 0, 0 },
+ [VK_FORMAT_ASTC_10x8_UNORM] = { 0, 0 },
+ [VK_FORMAT_ASTC_10x8_SRGB] = { 0, 0 },
+ [VK_FORMAT_ASTC_10x10_UNORM] = { 0, 0 },
+ [VK_FORMAT_ASTC_10x10_SRGB] = { 0, 0 },
+ [VK_FORMAT_ASTC_12x10_UNORM] = { 0, 0 },
+ [VK_FORMAT_ASTC_12x10_SRGB] = { 0, 0 },
+ [VK_FORMAT_ASTC_12x12_UNORM] = { 0, 0 },
+ [VK_FORMAT_ASTC_12x12_SRGB] = { 0, 0 },
+ [VK_FORMAT_B4G4R4A4_UNORM] = { 2, 4 },
+ [VK_FORMAT_B5G5R5A1_UNORM] = { 2, 4 },
+ [VK_FORMAT_B5G6R5_UNORM] = { 2, 3 },
+ [VK_FORMAT_B5G6R5_USCALED] = { 2, 3 },
+ [VK_FORMAT_B8G8R8_UNORM] = { 3, 3 },
+ [VK_FORMAT_B8G8R8_SNORM] = { 3, 3 },
+ [VK_FORMAT_B8G8R8_USCALED] = { 3, 3 },
+ [VK_FORMAT_B8G8R8_SSCALED] = { 3, 3 },
+ [VK_FORMAT_B8G8R8_UINT] = { 3, 3 },
+ [VK_FORMAT_B8G8R8_SINT] = { 3, 3 },
+ [VK_FORMAT_B8G8R8_SRGB] = { 3, 3 },
+ [VK_FORMAT_B8G8R8A8_UNORM] = { 4, 4 },
+ [VK_FORMAT_B8G8R8A8_SNORM] = { 4, 4 },
+ [VK_FORMAT_B8G8R8A8_USCALED] = { 4, 4 },
+ [VK_FORMAT_B8G8R8A8_SSCALED] = { 4, 4 },
+ [VK_FORMAT_B8G8R8A8_UINT] = { 4, 4 },
+ [VK_FORMAT_B8G8R8A8_SINT] = { 4, 4 },
+ [VK_FORMAT_B8G8R8A8_SRGB] = { 4, 4 },
+ [VK_FORMAT_B10G10R10A2_UNORM] = { 4, 4 },
+ [VK_FORMAT_B10G10R10A2_SNORM] = { 4, 4 },
+ [VK_FORMAT_B10G10R10A2_USCALED] = { 4, 4 },
+ [VK_FORMAT_B10G10R10A2_SSCALED] = { 4, 4 },
+ [VK_FORMAT_B10G10R10A2_UINT] = { 4, 4 },
+ [VK_FORMAT_B10G10R10A2_SINT] = { 4, 4 },
};
return format_table[format].size;
@@ -522,9 +522,9 @@
void add_memory_ref(const vk_testing::Object &obj)
{
- const std::vector<VkGpuMemory> mems = obj.memories();
- for (std::vector<VkGpuMemory>::const_iterator it = mems.begin(); it != mems.end(); it++) {
- std::vector<VkGpuMemory>::iterator ref;
+ const std::vector<VkDeviceMemory> mems = obj.memories();
+ for (std::vector<VkDeviceMemory>::const_iterator it = mems.begin(); it != mems.end(); it++) {
+ std::vector<VkDeviceMemory>::iterator ref;
for (ref = mem_refs_.begin(); ref != mem_refs_.end(); ref++) {
if (*ref == *it)
break;
@@ -540,7 +540,7 @@
vk_testing::Queue &queue_;
vk_testing::CmdBuffer cmd_;
- std::vector<VkGpuMemory> mem_refs_;
+ std::vector<VkDeviceMemory> mem_refs_;
};
typedef VkCmdBlitTest VkCmdFillBufferTest;
@@ -570,7 +570,7 @@
TEST_F(VkCmdFillBufferTest, Large)
{
- const VkGpuSize size = 32 * 1024 * 1024;
+ const VkDeviceSize size = 32 * 1024 * 1024;
vk_testing::Buffer buf;
buf.init(dev_, size);
@@ -584,7 +584,7 @@
submit_and_done();
const uint32_t *data = static_cast<const uint32_t *>(buf.map());
- VkGpuSize offset;
+ VkDeviceSize offset;
for (offset = 0; offset < size / 2; offset += 4)
EXPECT_EQ(0x11111111, data[offset / 4]) << "Offset is: " << offset;
for (; offset < size; offset += 4)
@@ -607,7 +607,7 @@
submit_and_done();
const uint32_t *data = static_cast<const uint32_t *>(buf.map());
- VkGpuSize offset;
+ VkDeviceSize offset;
for (offset = 0; offset < 32; offset += 4)
EXPECT_EQ(0x11111111, data[offset / 4]) << "Offset is: " << offset;
for (; offset < 64; offset += 4)
@@ -618,7 +618,7 @@
TEST_F(VkCmdFillBufferTest, MultiAlignments)
{
vk_testing::Buffer bufs[9];
- VkGpuSize size = 4;
+ VkDeviceSize size = 4;
cmd_.begin();
for (int i = 0; i < ARRAY_SIZE(bufs); i++) {
@@ -634,7 +634,7 @@
size = 4;
for (int i = 0; i < ARRAY_SIZE(bufs); i++) {
const uint32_t *data = static_cast<const uint32_t *>(bufs[i].map());
- VkGpuSize offset;
+ VkDeviceSize offset;
for (offset = 0; offset < size; offset += 4)
EXPECT_EQ(0x11111111, data[offset / 4]) << "Buffser is: " << i << "\n" <<
"Offset is: " << offset;
@@ -674,12 +674,12 @@
TEST_F(VkCmdCopyBufferTest, Large)
{
- const VkGpuSize size = 32 * 1024 * 1024;
+ const VkDeviceSize size = 32 * 1024 * 1024;
vk_testing::Buffer src, dst;
src.init(dev_, size);
uint32_t *data = static_cast<uint32_t *>(src.map());
- VkGpuSize offset;
+ VkDeviceSize offset;
for (offset = 0; offset < size; offset += 4)
data[offset / 4] = offset;
src.unmap();
@@ -772,16 +772,16 @@
err = vkCreateEvent(dev_.obj(), &event_info, &event);
ASSERT_VK_SUCCESS(err);
- err = vkGetObjectInfo(event, VK_INFO_TYPE_MEMORY_REQUIREMENTS,
+ err = vkGetObjectInfo(event, VK_OBJECT_INFO_TYPE_MEMORY_REQUIREMENTS,
&data_size, &mem_req);
ASSERT_VK_SUCCESS(err);
// VkResult VKAPI vkAllocMemory(
// VkDevice device,
// const VkMemoryAllocInfo* pAllocInfo,
- // VkGpuMemory* pMem);
+ // VkDeviceMemory* pMem);
VkMemoryAllocInfo mem_info;
- VkGpuMemory event_mem;
+ VkDeviceMemory event_mem;
ASSERT_NE(0, mem_req.size) << "vkGetObjectInfo (Event): Failed - expect events to require memory";
@@ -817,14 +817,7 @@
VkBufferMemoryBarrier *pmemory_barrier = &memory_barrier;
VkPipeEvent set_events[] = { VK_PIPE_EVENT_TRANSFER_COMPLETE };
- VkPipelineBarrier pipeline_barrier = {};
- pipeline_barrier.sType = VK_STRUCTURE_TYPE_PIPELINE_BARRIER;
- pipeline_barrier.eventCount = 1;
- pipeline_barrier.pEvents = set_events;
- pipeline_barrier.waitEvent = VK_WAIT_EVENT_TOP_OF_PIPE;
- pipeline_barrier.memBarrierCount = 1;
- pipeline_barrier.ppMemBarriers = (const void **)&pmemory_barrier;
- vkCmdPipelineBarrier(cmd_.obj(), &pipeline_barrier);
+ vkCmdPipelineBarrier(cmd_.obj(), VK_WAIT_EVENT_TOP_OF_PIPE, 1, set_events, 1, (const void **)&pmemory_barrier);
VkBufferCopy region = {};
region.copySize = 4;
@@ -833,13 +826,7 @@
memory_barrier = bufs[1].buffer_memory_barrier(
VK_MEMORY_OUTPUT_TRANSFER_BIT, VK_MEMORY_INPUT_TRANSFER_BIT, 0, 4);
pmemory_barrier = &memory_barrier;
- pipeline_barrier.sType = VK_STRUCTURE_TYPE_PIPELINE_BARRIER;
- pipeline_barrier.eventCount = 1;
- pipeline_barrier.pEvents = set_events;
- pipeline_barrier.waitEvent = VK_WAIT_EVENT_TOP_OF_PIPE;
- pipeline_barrier.memBarrierCount = 1;
- pipeline_barrier.ppMemBarriers = (const void **)&pmemory_barrier;
- vkCmdPipelineBarrier(cmd_.obj(), &pipeline_barrier);
+ vkCmdPipelineBarrier(cmd_.obj(), VK_WAIT_EVENT_TOP_OF_PIPE, 1, set_events, 1, (const void **)&pmemory_barrier);
vkCmdCopyBuffer(cmd_.obj(), bufs[1].obj(), bufs[2].obj(), 1, ®ion);
@@ -853,14 +840,7 @@
memory_barrier = bufs[1].buffer_memory_barrier(
VK_MEMORY_OUTPUT_TRANSFER_BIT, VK_MEMORY_INPUT_CPU_READ_BIT, 0, 4);
pmemory_barrier = &memory_barrier;
- VkEventWaitInfo wait_info = {};
- wait_info.sType = VK_STRUCTURE_TYPE_EVENT_WAIT_INFO;
- wait_info.eventCount = 1;
- wait_info.pEvents = &event;
- wait_info.waitEvent = VK_WAIT_EVENT_TOP_OF_PIPE;
- wait_info.memBarrierCount = 1;
- wait_info.ppMemBarriers = (const void **)&pmemory_barrier;
- vkCmdWaitEvents(cmd_.obj(), &wait_info);
+ vkCmdWaitEvents(cmd_.obj(), VK_WAIT_EVENT_TOP_OF_PIPE, 1, &event, 1, (const void **)&pmemory_barrier);
cmd_.end();
@@ -885,19 +865,19 @@
protected:
void init_test_formats(VkFlags features)
{
- first_linear_format_ = VK_FMT_UNDEFINED;
- first_optimal_format_ = VK_FMT_UNDEFINED;
+ first_linear_format_ = VK_FORMAT_UNDEFINED;
+ first_optimal_format_ = VK_FORMAT_UNDEFINED;
for (std::vector<vk_testing::Device::Format>::const_iterator it = dev_.formats().begin();
it != dev_.formats().end(); it++) {
if (it->features & features) {
test_formats_.push_back(*it);
- if (it->tiling == VK_LINEAR_TILING &&
- first_linear_format_ == VK_FMT_UNDEFINED)
+ if (it->tiling == VK_IMAGE_TILING_LINEAR &&
+ first_linear_format_ == VK_FORMAT_UNDEFINED)
first_linear_format_ = it->format;
- if (it->tiling == VK_OPTIMAL_TILING &&
- first_optimal_format_ == VK_FMT_UNDEFINED)
+ if (it->tiling == VK_IMAGE_TILING_OPTIMAL &&
+ first_optimal_format_ == VK_FORMAT_UNDEFINED)
first_optimal_format_ = it->format;
}
}
@@ -972,7 +952,7 @@
virtual void SetUp()
{
VkCmdBlitTest::SetUp();
- init_test_formats(VK_FORMAT_COLOR_ATTACHMENT_BIT);
+ init_test_formats(VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT);
ASSERT_NE(true, test_formats_.empty());
}
@@ -1019,13 +999,13 @@
it != test_formats_.end(); it++) {
// not sure what to do here
- if (it->format == VK_FMT_UNDEFINED ||
- (it->format >= VK_FMT_B8G8R8_UNORM &&
- it->format <= VK_FMT_B8G8R8_SRGB))
+ if (it->format == VK_FORMAT_UNDEFINED ||
+ (it->format >= VK_FORMAT_B8G8R8_UNORM &&
+ it->format <= VK_FORMAT_B8G8R8_SRGB))
continue;
VkImageCreateInfo img_info = vk_testing::Image::create_info();
- img_info.imageType = VK_IMAGE_2D;
+ img_info.imageType = VK_IMAGE_TYPE_2D;
img_info.format = it->format;
img_info.extent.width = 64;
img_info.extent.height = 64;
@@ -1040,7 +1020,7 @@
virtual void SetUp()
{
VkCmdBlitTest::SetUp();
- init_test_formats(VK_FORMAT_COLOR_ATTACHMENT_BIT);
+ init_test_formats(VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT);
ASSERT_NE(true, test_formats_.empty());
}
@@ -1087,13 +1067,13 @@
it != test_formats_.end(); it++) {
// not sure what to do here
- if (it->format == VK_FMT_UNDEFINED ||
- (it->format >= VK_FMT_B8G8R8_UNORM &&
- it->format <= VK_FMT_B8G8R8_SRGB))
+ if (it->format == VK_FORMAT_UNDEFINED ||
+ (it->format >= VK_FORMAT_B8G8R8_UNORM &&
+ it->format <= VK_FORMAT_B8G8R8_SRGB))
continue;
VkImageCreateInfo img_info = vk_testing::Image::create_info();
- img_info.imageType = VK_IMAGE_2D;
+ img_info.imageType = VK_IMAGE_TYPE_2D;
img_info.format = it->format;
img_info.extent.width = 64;
img_info.extent.height = 64;
@@ -1108,7 +1088,7 @@
virtual void SetUp()
{
VkCmdBlitTest::SetUp();
- init_test_formats(VK_FORMAT_COLOR_ATTACHMENT_BIT);
+ init_test_formats(VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT);
ASSERT_NE(true, test_formats_.empty());
}
@@ -1117,7 +1097,7 @@
{
// convert VkImageCopy to two sets of VkBufferImageCopy
std::vector<VkBufferImageCopy> src_regions, dst_regions;
- VkGpuSize src_offset = 0, dst_offset = 0;
+ VkDeviceSize src_offset = 0, dst_offset = 0;
for (std::vector<VkImageCopy>::const_iterator it = copies.begin(); it != copies.end(); it++) {
VkBufferImageCopy src_region = {}, dst_region = {};
@@ -1133,7 +1113,7 @@
dst_region.imageExtent = it->extent;
dst_regions.push_back(dst_region);
- const VkGpuSize size = it->extent.width * it->extent.height * it->extent.depth;
+ const VkDeviceSize size = it->extent.width * it->extent.height * it->extent.depth;
src_offset += vk_testing::get_format_size(src_info.format) * size;
dst_offset += vk_testing::get_format_size(dst_info.format) * size;
}
@@ -1169,13 +1149,13 @@
it != test_formats_.end(); it++) {
// not sure what to do here
- if (it->format == VK_FMT_UNDEFINED ||
- (it->format >= VK_FMT_B8G8R8_UNORM &&
- it->format <= VK_FMT_B8G8R8_SRGB))
+ if (it->format == VK_FORMAT_UNDEFINED ||
+ (it->format >= VK_FORMAT_B8G8R8_UNORM &&
+ it->format <= VK_FORMAT_B8G8R8_SRGB))
continue;
VkImageCreateInfo img_info = vk_testing::Image::create_info();
- img_info.imageType = VK_IMAGE_2D;
+ img_info.imageType = VK_IMAGE_TYPE_2D;
img_info.format = it->format;
img_info.extent.width = 64;
img_info.extent.height = 64;
@@ -1233,21 +1213,21 @@
for (std::vector<vk_testing::Device::Format>::const_iterator it = test_formats_.begin();
it != test_formats_.end(); it++) {
// not sure what to do here
- if (it->format == VK_FMT_UNDEFINED ||
- (it->format >= VK_FMT_R32G32B32_UINT &&
- it->format <= VK_FMT_R32G32B32_SFLOAT) ||
- (it->format >= VK_FMT_B8G8R8_UNORM &&
- it->format <= VK_FMT_B8G8R8_SRGB) ||
- (it->format >= VK_FMT_BC1_RGB_UNORM &&
- it->format <= VK_FMT_ASTC_12x12_SRGB) ||
- (it->format >= VK_FMT_D16_UNORM &&
- it->format <= VK_FMT_D32_SFLOAT_S8_UINT) ||
- it->format == VK_FMT_R64G64B64_SFLOAT ||
- it->format == VK_FMT_R64G64B64A64_SFLOAT)
+ if (it->format == VK_FORMAT_UNDEFINED ||
+ (it->format >= VK_FORMAT_R32G32B32_UINT &&
+ it->format <= VK_FORMAT_R32G32B32_SFLOAT) ||
+ (it->format >= VK_FORMAT_B8G8R8_UNORM &&
+ it->format <= VK_FORMAT_B8G8R8_SRGB) ||
+ (it->format >= VK_FORMAT_BC1_RGB_UNORM &&
+ it->format <= VK_FORMAT_ASTC_12x12_SRGB) ||
+ (it->format >= VK_FORMAT_D16_UNORM &&
+ it->format <= VK_FORMAT_D32_SFLOAT_S8_UINT) ||
+ it->format == VK_FORMAT_R64G64B64_SFLOAT ||
+ it->format == VK_FORMAT_R64G64B64A64_SFLOAT)
continue;
VkImageCreateInfo img_info = vk_testing::Image::create_info();
- img_info.imageType = VK_IMAGE_2D;
+ img_info.imageType = VK_IMAGE_TYPE_2D;
img_info.format = it->format;
img_info.extent.width = 64;
img_info.extent.height = 64;
@@ -1274,7 +1254,7 @@
if (test_raw_)
init_test_formats();
else
- init_test_formats(VK_FORMAT_CONVERSION_BIT);
+ init_test_formats(VK_FORMAT_FEATURE_CONVERSION_BIT);
ASSERT_NE(true, test_formats_.empty());
}
@@ -1292,13 +1272,13 @@
// TODO support all formats
switch (format) {
- case VK_FMT_R8G8B8A8_UNORM:
+ case VK_FORMAT_R8G8B8A8_UNORM:
raw.push_back(color[0] * 255.0f);
raw.push_back(color[1] * 255.0f);
raw.push_back(color[2] * 255.0f);
raw.push_back(color[3] * 255.0f);
break;
- case VK_FMT_B8G8R8A8_UNORM:
+ case VK_FORMAT_B8G8R8A8_UNORM:
raw.push_back(color[2] * 255.0f);
raw.push_back(color[1] * 255.0f);
raw.push_back(color[0] * 255.0f);
@@ -1317,13 +1297,13 @@
// TODO support all formats
switch (format) {
- case VK_FMT_R8G8B8A8_UNORM:
+ case VK_FORMAT_R8G8B8A8_UNORM:
raw.push_back(static_cast<uint8_t>(color[0]));
raw.push_back(static_cast<uint8_t>(color[1]));
raw.push_back(static_cast<uint8_t>(color[2]));
raw.push_back(static_cast<uint8_t>(color[3]));
break;
- case VK_FMT_B8G8R8A8_UNORM:
+ case VK_FORMAT_B8G8R8A8_UNORM:
raw.push_back(static_cast<uint8_t>(color[2]));
raw.push_back(static_cast<uint8_t>(color[1]));
raw.push_back(static_cast<uint8_t>(color[0]));
@@ -1388,27 +1368,14 @@
cmd_.begin();
- VkPipeEvent set_events[] = { VK_PIPE_EVENT_GPU_COMMANDS_COMPLETE };
- VkPipelineBarrier pipeline_barrier = {};
- pipeline_barrier.sType = VK_STRUCTURE_TYPE_PIPELINE_BARRIER;
- pipeline_barrier.eventCount = 1;
- pipeline_barrier.pEvents = set_events;
- pipeline_barrier.waitEvent = VK_WAIT_EVENT_TOP_OF_PIPE;
- pipeline_barrier.memBarrierCount = to_clear.size();
- pipeline_barrier.ppMemBarriers = (const void **)&p_to_clear[0];
- vkCmdPipelineBarrier(cmd_.obj(), &pipeline_barrier);
+ VkPipeEvent set_events[] = { VK_PIPE_EVENT_COMMANDS_COMPLETE };
+ vkCmdPipelineBarrier(cmd_.obj(), VK_WAIT_EVENT_TOP_OF_PIPE, 1, set_events, 1, (const void **)&p_to_clear[0]);
vkCmdClearColorImage(cmd_.obj(),
img.obj(), VK_IMAGE_LAYOUT_CLEAR_OPTIMAL,
clear_color, ranges.size(), &ranges[0]);
- pipeline_barrier.sType = VK_STRUCTURE_TYPE_PIPELINE_BARRIER;
- pipeline_barrier.eventCount = 1;
- pipeline_barrier.pEvents = set_events;
- pipeline_barrier.waitEvent = VK_WAIT_EVENT_TOP_OF_PIPE;
- pipeline_barrier.memBarrierCount = to_xfer.size();
- pipeline_barrier.ppMemBarriers = (const void **)&p_to_xfer[0];
- vkCmdPipelineBarrier(cmd_.obj(), &pipeline_barrier);
+ vkCmdPipelineBarrier(cmd_.obj(), VK_WAIT_EVENT_TOP_OF_PIPE, 1, set_events, 1, (const void **)&p_to_xfer[0]);
cmd_.end();
@@ -1445,7 +1412,7 @@
const float color[4] = { 0.0f, 1.0f, 0.0f, 1.0f };
VkImageCreateInfo img_info = vk_testing::Image::create_info();
- img_info.imageType = VK_IMAGE_2D;
+ img_info.imageType = VK_IMAGE_TYPE_2D;
img_info.format = it->format;
img_info.extent.width = 64;
img_info.extent.height = 64;
@@ -1481,23 +1448,23 @@
const uint32_t color[4] = { 0x11111111, 0x22222222, 0x33333333, 0x44444444 };
// not sure what to do here
- if (it->format == VK_FMT_UNDEFINED ||
- (it->format >= VK_FMT_R8G8B8_UNORM &&
- it->format <= VK_FMT_R8G8B8_SRGB) ||
- (it->format >= VK_FMT_B8G8R8_UNORM &&
- it->format <= VK_FMT_B8G8R8_SRGB) ||
- (it->format >= VK_FMT_R16G16B16_UNORM &&
- it->format <= VK_FMT_R16G16B16_SFLOAT) ||
- (it->format >= VK_FMT_R32G32B32_UINT &&
- it->format <= VK_FMT_R32G32B32_SFLOAT) ||
- it->format == VK_FMT_R64G64B64_SFLOAT ||
- it->format == VK_FMT_R64G64B64A64_SFLOAT ||
- (it->format >= VK_FMT_D16_UNORM &&
- it->format <= VK_FMT_D32_SFLOAT_S8_UINT))
+ if (it->format == VK_FORMAT_UNDEFINED ||
+ (it->format >= VK_FORMAT_R8G8B8_UNORM &&
+ it->format <= VK_FORMAT_R8G8B8_SRGB) ||
+ (it->format >= VK_FORMAT_B8G8R8_UNORM &&
+ it->format <= VK_FORMAT_B8G8R8_SRGB) ||
+ (it->format >= VK_FORMAT_R16G16B16_UNORM &&
+ it->format <= VK_FORMAT_R16G16B16_SFLOAT) ||
+ (it->format >= VK_FORMAT_R32G32B32_UINT &&
+ it->format <= VK_FORMAT_R32G32B32_SFLOAT) ||
+ it->format == VK_FORMAT_R64G64B64_SFLOAT ||
+ it->format == VK_FORMAT_R64G64B64A64_SFLOAT ||
+ (it->format >= VK_FORMAT_D16_UNORM &&
+ it->format <= VK_FORMAT_D32_SFLOAT_S8_UINT))
continue;
VkImageCreateInfo img_info = vk_testing::Image::create_info();
- img_info.imageType = VK_IMAGE_2D;
+ img_info.imageType = VK_IMAGE_TYPE_2D;
img_info.format = it->format;
img_info.extent.width = 64;
img_info.extent.height = 64;
@@ -1516,7 +1483,7 @@
virtual void SetUp()
{
VkCmdBlitTest::SetUp();
- init_test_formats(VK_FORMAT_DEPTH_STENCIL_ATTACHMENT_BIT);
+ init_test_formats(VK_FORMAT_FEATURE_DEPTH_STENCIL_ATTACHMENT_BIT);
ASSERT_NE(true, test_formats_.empty());
}
@@ -1526,16 +1493,16 @@
// depth
switch (format) {
- case VK_FMT_D16_UNORM:
- case VK_FMT_D16_UNORM_S8_UINT:
+ case VK_FORMAT_D16_UNORM:
+ case VK_FORMAT_D16_UNORM_S8_UINT:
{
const uint16_t unorm = depth * 65535.0f;
raw.push_back(unorm & 0xff);
raw.push_back(unorm >> 8);
}
break;
- case VK_FMT_D32_SFLOAT:
- case VK_FMT_D32_SFLOAT_S8_UINT:
+ case VK_FORMAT_D32_SFLOAT:
+ case VK_FORMAT_D32_SFLOAT_S8_UINT:
{
const union {
float depth;
@@ -1554,14 +1521,14 @@
// stencil
switch (format) {
- case VK_FMT_S8_UINT:
+ case VK_FORMAT_S8_UINT:
raw.push_back(stencil);
break;
- case VK_FMT_D16_UNORM_S8_UINT:
+ case VK_FORMAT_D16_UNORM_S8_UINT:
raw.push_back(stencil);
raw.push_back(0);
break;
- case VK_FMT_D32_SFLOAT_S8_UINT:
+ case VK_FORMAT_D32_SFLOAT_S8_UINT:
raw.push_back(stencil);
raw.push_back(0);
raw.push_back(0);
@@ -1618,28 +1585,15 @@
cmd_.begin();
- VkPipeEvent set_events[] = { VK_PIPE_EVENT_GPU_COMMANDS_COMPLETE };
- VkPipelineBarrier pipeline_barrier = {};
- pipeline_barrier.sType = VK_STRUCTURE_TYPE_PIPELINE_BARRIER;
- pipeline_barrier.eventCount = 1;
- pipeline_barrier.pEvents = set_events;
- pipeline_barrier.waitEvent = VK_WAIT_EVENT_TOP_OF_PIPE;
- pipeline_barrier.memBarrierCount = to_clear.size();
- pipeline_barrier.ppMemBarriers = (const void **)&p_to_clear[0];
- vkCmdPipelineBarrier(cmd_.obj(), &pipeline_barrier);
+ VkPipeEvent set_events[] = { VK_PIPE_EVENT_COMMANDS_COMPLETE };
+ vkCmdPipelineBarrier(cmd_.obj(), VK_WAIT_EVENT_TOP_OF_PIPE, 1, set_events, to_clear.size(), (const void **)&p_to_clear[0]);
vkCmdClearDepthStencil(cmd_.obj(),
img.obj(), VK_IMAGE_LAYOUT_CLEAR_OPTIMAL,
depth, stencil,
ranges.size(), &ranges[0]);
- pipeline_barrier.sType = VK_STRUCTURE_TYPE_PIPELINE_BARRIER;
- pipeline_barrier.eventCount = 1;
- pipeline_barrier.pEvents = set_events;
- pipeline_barrier.waitEvent = VK_WAIT_EVENT_TOP_OF_PIPE;
- pipeline_barrier.memBarrierCount = to_xfer.size();
- pipeline_barrier.ppMemBarriers = (const void **)&p_to_xfer[0];
- vkCmdPipelineBarrier(cmd_.obj(), &pipeline_barrier);
+ vkCmdPipelineBarrier(cmd_.obj(), VK_WAIT_EVENT_TOP_OF_PIPE, 1, set_events, to_xfer.size(), (const void **)&p_to_xfer[0]);
cmd_.end();
@@ -1661,14 +1615,14 @@
for (std::vector<vk_testing::Device::Format>::const_iterator it = test_formats_.begin();
it != test_formats_.end(); it++) {
// known driver issues
- if (it->format == VK_FMT_S8_UINT ||
- it->format == VK_FMT_D24_UNORM ||
- it->format == VK_FMT_D16_UNORM_S8_UINT ||
- it->format == VK_FMT_D24_UNORM_S8_UINT)
+ if (it->format == VK_FORMAT_S8_UINT ||
+ it->format == VK_FORMAT_D24_UNORM ||
+ it->format == VK_FORMAT_D16_UNORM_S8_UINT ||
+ it->format == VK_FORMAT_D24_UNORM_S8_UINT)
continue;
VkImageCreateInfo img_info = vk_testing::Image::create_info();
- img_info.imageType = VK_IMAGE_2D;
+ img_info.imageType = VK_IMAGE_TYPE_2D;
img_info.format = it->format;
img_info.extent.width = 64;
img_info.extent.height = 64;
diff --git a/tests/image_tests.cpp b/tests/image_tests.cpp
index 326a8c8..736daea 100644
--- a/tests/image_tests.cpp
+++ b/tests/image_tests.cpp
@@ -78,11 +78,11 @@
protected:
vk_testing::Device *m_device;
VkApplicationInfo app_info;
- VkPhysicalGpu objs[16];
+ VkPhysicalDevice objs[16];
uint32_t gpu_count;
VkInstance inst;
VkImage m_image;
- VkGpuMemory *m_image_mem;
+ VkDeviceMemory *m_image_mem;
uint32_t m_num_mem;
virtual void SetUp() {
@@ -140,11 +140,11 @@
mipCount++;
}
- fmt = VK_FMT_R8G8B8A8_UINT;
+ fmt = VK_FORMAT_R8G8B8A8_UINT;
// TODO: Pick known good format rather than just expect common format
/*
* XXX: What should happen if given NULL HANDLE for the pData argument?
- * We're not requesting VK_INFO_TYPE_MEMORY_REQUIREMENTS so there is
+ * We're not requesting VK_OBJECT_INFO_TYPE_MEMORY_REQUIREMENTS so there is
* an expectation that pData is a valid pointer.
* However, why include a returned size value? That implies that the
* amount of data may vary and that doesn't work well for using a
@@ -152,7 +152,7 @@
*/
size = sizeof(image_fmt);
err = vkGetFormatInfo(this->device(), fmt,
- VK_INFO_TYPE_FORMAT_PROPERTIES,
+ VK_FORMAT_INFO_TYPE_PROPERTIES,
&size, &image_fmt);
ASSERT_VK_SUCCESS(err);
@@ -174,7 +174,7 @@
VkImageCreateInfo imageCreateInfo = {};
imageCreateInfo.sType = VK_STRUCTURE_TYPE_IMAGE_CREATE_INFO;
- imageCreateInfo.imageType = VK_IMAGE_2D;
+ imageCreateInfo.imageType = VK_IMAGE_TYPE_2D;
imageCreateInfo.format = fmt;
imageCreateInfo.arraySize = 1;
imageCreateInfo.extent.width = w;
@@ -182,11 +182,11 @@
imageCreateInfo.extent.depth = 1;
imageCreateInfo.mipLevels = mipCount;
imageCreateInfo.samples = 1;
- if (image_fmt.linearTilingFeatures & VK_FORMAT_SAMPLED_IMAGE_BIT) {
- imageCreateInfo.tiling = VK_LINEAR_TILING;
+ if (image_fmt.linearTilingFeatures & VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT) {
+ imageCreateInfo.tiling = VK_IMAGE_TILING_LINEAR;
}
- else if (image_fmt.optimalTilingFeatures & VK_FORMAT_SAMPLED_IMAGE_BIT) {
- imageCreateInfo.tiling = VK_OPTIMAL_TILING;
+ else if (image_fmt.optimalTilingFeatures & VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT) {
+ imageCreateInfo.tiling = VK_IMAGE_TILING_OPTIMAL;
}
else {
ASSERT_TRUE(false) << "Cannot find supported tiling format - Exiting";
@@ -223,15 +223,15 @@
mem_info.sType = VK_STRUCTURE_TYPE_MEMORY_ALLOC_INFO;
mem_info.pNext = NULL;
- err = vkGetObjectInfo(m_image, VK_INFO_TYPE_MEMORY_ALLOCATION_COUNT,
+ err = vkGetObjectInfo(m_image, VK_OBJECT_INFO_TYPE_MEMORY_ALLOCATION_COUNT,
&num_alloc_size, &num_allocations);
ASSERT_VK_SUCCESS(err);
ASSERT_EQ(num_alloc_size,sizeof(num_allocations));
mem_req = (VkMemoryRequirements *) malloc(num_allocations * sizeof(VkMemoryRequirements));
- m_image_mem = (VkGpuMemory *) malloc(num_allocations * sizeof(VkGpuMemory));
+ m_image_mem = (VkDeviceMemory *) malloc(num_allocations * sizeof(VkDeviceMemory));
m_num_mem = num_allocations;
err = vkGetObjectInfo(m_image,
- VK_INFO_TYPE_MEMORY_REQUIREMENTS,
+ VK_OBJECT_INFO_TYPE_MEMORY_REQUIREMENTS,
&mem_reqs_size, mem_req);
ASSERT_VK_SUCCESS(err);
ASSERT_EQ(mem_reqs_size, num_allocations * sizeof(VkMemoryRequirements));
@@ -285,7 +285,7 @@
VkFormat fmt;
VkImageView imageView;
- fmt = VK_FMT_R8G8B8A8_UINT;
+ fmt = VK_FORMAT_R8G8B8A8_UINT;
CreateImage(512, 256);
@@ -302,7 +302,7 @@
// } VkImageViewCreateInfo;
VkImageViewCreateInfo viewInfo = {};
viewInfo.sType = VK_STRUCTURE_TYPE_IMAGE_VIEW_CREATE_INFO;
- viewInfo.viewType = VK_IMAGE_VIEW_2D;
+ viewInfo.viewType = VK_IMAGE_VIEW_TYPE_2D;
viewInfo.format = fmt;
viewInfo.channels.r = VK_CHANNEL_SWIZZLE_R;
diff --git a/tests/init.cpp b/tests/init.cpp
index de02b34..a0c44d1 100644
--- a/tests/init.cpp
+++ b/tests/init.cpp
@@ -80,13 +80,13 @@
protected:
VkApplicationInfo app_info;
VkInstance inst;
- VkPhysicalGpu objs[16];
+ VkPhysicalDevice objs[16];
uint32_t gpu_count;
uint32_t m_device_id;
vk_testing::Device *m_device;
- VkPhysicalGpuProperties props;
- std::vector<VkPhysicalGpuQueueProperties> queue_props;
+ VkPhysicalDeviceProperties props;
+ std::vector<VkPhysicalDeviceQueueProperties> queue_props;
uint32_t graphics_queue_node_index;
virtual void SetUp() {
@@ -137,10 +137,10 @@
}
};
-TEST(Initialization, vkEnumerateGpus) {
+TEST(Initialization, vkEnumeratePhysicalDevices) {
VkApplicationInfo app_info = {};
VkInstance inst;
- VkPhysicalGpu objs[16];
+ VkPhysicalDevice objs[16];
uint32_t gpu_count;
VkResult err;
vk_testing::PhysicalGpu *gpu;
@@ -192,13 +192,13 @@
TEST_F(XglTest, AllocMemory) {
VkResult err;
VkMemoryAllocInfo alloc_info = {};
- VkGpuMemory gpu_mem;
+ VkDeviceMemory gpu_mem;
uint8_t *pData;
alloc_info.sType = VK_STRUCTURE_TYPE_MEMORY_ALLOC_INFO;
alloc_info.allocationSize = 1024 * 1024; // 1MB
alloc_info.memProps = VK_MEMORY_PROPERTY_SHAREABLE_BIT |
- VK_MEMORY_PROPERTY_CPU_VISIBLE_BIT;
+ VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT;
// TODO: Try variety of memory priorities
@@ -239,16 +239,16 @@
err = vkCreateEvent(device(), &event_info, &event);
ASSERT_VK_SUCCESS(err);
- err = vkGetObjectInfo(event, VK_INFO_TYPE_MEMORY_REQUIREMENTS,
+ err = vkGetObjectInfo(event, VK_OBJECT_INFO_TYPE_MEMORY_REQUIREMENTS,
&data_size, &mem_req);
ASSERT_VK_SUCCESS(err);
// VkResult VKAPI vkAllocMemory(
// VkDevice device,
// const VkMemoryAllocInfo* pAllocInfo,
- // VkGpuMemory* pMem);
+ // VkDeviceMemory* pMem);
VkMemoryAllocInfo mem_info;
- VkGpuMemory event_mem;
+ VkDeviceMemory event_mem;
ASSERT_NE(0, mem_req.size) << "vkGetObjectInfo (Event): Failed - expect events to require memory";
@@ -339,11 +339,11 @@
// typedef enum VkQueryType_
// {
- // VK_QUERY_OCCLUSION = 0x00000000,
- // VK_QUERY_PIPELINE_STATISTICS = 0x00000001,
+ // VK_QUERY_TYPE_OCCLUSION = 0x00000000,
+ // VK_QUERY_TYPE_PIPELINE_STATISTICS = 0x00000001,
- // VK_QUERY_TYPE_BEGIN_RANGE = VK_QUERY_OCCLUSION,
- // VK_QUERY_TYPE_END_RANGE = VK_QUERY_PIPELINE_STATISTICS,
+ // VK_QUERY_TYPE_BEGIN_RANGE = VK_QUERY_TYPE_OCCLUSION,
+ // VK_QUERY_TYPE_END_RANGE = VK_QUERY_TYPE_PIPELINE_STATISTICS,
// VK_NUM_QUERY_TYPE = (VK_QUERY_TYPE_END_RANGE - VK_QUERY_TYPE_BEGIN_RANGE + 1),
// VK_MAX_ENUM(VkQueryType_)
// } VkQueryType;
@@ -358,7 +358,7 @@
memset(&query_info, 0, sizeof(query_info));
query_info.sType = VK_STRUCTURE_TYPE_QUERY_POOL_CREATE_INFO;
- query_info.queryType = VK_QUERY_OCCLUSION;
+ query_info.queryType = VK_QUERY_TYPE_OCCLUSION;
query_info.slots = MAX_QUERY_SLOTS;
// VkResult VKAPI vkCreateQueryPool(
@@ -370,7 +370,7 @@
ASSERT_VK_SUCCESS(err);
data_size = sizeof(mem_req);
- err = vkGetObjectInfo(query_pool, VK_INFO_TYPE_MEMORY_REQUIREMENTS,
+ err = vkGetObjectInfo(query_pool, VK_OBJECT_INFO_TYPE_MEMORY_REQUIREMENTS,
&data_size, &mem_req);
ASSERT_VK_SUCCESS(err);
ASSERT_NE(0, data_size) << "Invalid data_size";
@@ -378,9 +378,9 @@
// VkResult VKAPI vkAllocMemory(
// VkDevice device,
// const VkMemoryAllocInfo* pAllocInfo,
- // VkGpuMemory* pMem);
+ // VkDeviceMemory* pMem);
VkMemoryAllocInfo mem_info;
- VkGpuMemory query_mem;
+ VkDeviceMemory query_mem;
memset(&mem_info, 0, sizeof(mem_info));
mem_info.sType = VK_STRUCTURE_TYPE_MEMORY_ALLOC_INFO;
@@ -404,13 +404,13 @@
// TOOD: vkCmdEndQuery
err = vkGetQueryPoolResults(query_pool, 0, MAX_QUERY_SLOTS,
- &query_result_size, VK_NULL_HANDLE);
+ &query_result_size, VK_NULL_HANDLE, 0);
ASSERT_VK_SUCCESS(err);
if (query_result_size > 0) {
query_result_data = new uint32_t [query_result_size];
err = vkGetQueryPoolResults(query_pool, 0, MAX_QUERY_SLOTS,
- &query_result_size, query_result_data);
+ &query_result_size, query_result_data, 0);
ASSERT_VK_SUCCESS(err);
// TODO: Test Query result data.
@@ -431,7 +431,7 @@
VkResult err;
VkQueue queue;
- const VkPhysicalGpuQueueProperties props = device->gpu().queue_properties()[queue_node_index];
+ const VkPhysicalDeviceQueueProperties props = device->gpu().queue_properties()[queue_node_index];
for (que_idx = 0; que_idx < props.queueCount; que_idx++) {
// TODO: Need to add support for separate MEMMGR and work queues, including synchronization
err = vkGetDeviceQueue(device->obj(), queue_node_index, que_idx, &queue);
@@ -443,12 +443,12 @@
void print_queue_info(vk_testing::Device *device, uint32_t queue_node_index)
{
uint32_t que_idx;
- VkPhysicalGpuQueueProperties queue_props;
- VkPhysicalGpuProperties props;
+ VkPhysicalDeviceQueueProperties queue_props;
+ VkPhysicalDeviceProperties props;
props = device->gpu().properties();
queue_props = device->gpu().queue_properties()[queue_node_index];
- ASSERT_NE(0, queue_props.queueCount) << "No Queues available at Node Index #" << queue_node_index << " GPU: " << props.gpuName;
+ ASSERT_NE(0, queue_props.queueCount) << "No Queues available at Node Index #" << queue_node_index << " GPU: " << props.deviceName;
// VkResult VKAPI vkGetDeviceQueue(
// VkDevice device,
@@ -459,7 +459,7 @@
* queue handles are retrieved from the device by calling
* vkGetDeviceQueue() with a queue node index and a requested logical
* queue ID. The queue node index is the index into the array of
- * VkPhysicalGpuQueueProperties returned by GetGpuInfo. Each
+ * VkPhysicalDeviceQueueProperties returned by GetPhysicalDeviceInfo. Each
* queue node index has different attributes specified by the VkQueueFlags property.
* The logical queue ID is a sequential number starting from zero
* and referencing up to the number of queues supported of that node index
@@ -533,11 +533,11 @@
mipCount++;
}
- fmt = VK_FMT_R8G8B8A8_UINT;
+ fmt = VK_FORMAT_R8G8B8A8_UINT;
// TODO: Pick known good format rather than just expect common format
/*
* XXX: What should happen if given NULL HANDLE for the pData argument?
- * We're not requesting VK_INFO_TYPE_MEMORY_REQUIREMENTS so there is
+ * We're not requesting VK_OBJECT_INFO_TYPE_MEMORY_REQUIREMENTS so there is
* an expectation that pData is a valid pointer.
* However, why include a returned size value? That implies that the
* amount of data may vary and that doesn't work well for using a
@@ -546,7 +546,7 @@
size = sizeof(image_fmt);
err = vkGetFormatInfo(device(), fmt,
- VK_INFO_TYPE_FORMAT_PROPERTIES,
+ VK_FORMAT_INFO_TYPE_PROPERTIES,
&size, &image_fmt);
ASSERT_VK_SUCCESS(err);
@@ -568,7 +568,7 @@
VkImageCreateInfo imageCreateInfo = {};
imageCreateInfo.sType = VK_STRUCTURE_TYPE_IMAGE_CREATE_INFO;
- imageCreateInfo.imageType = VK_IMAGE_2D;
+ imageCreateInfo.imageType = VK_IMAGE_TYPE_2D;
imageCreateInfo.format = fmt;
imageCreateInfo.arraySize = 1;
imageCreateInfo.extent.width = w;
@@ -576,7 +576,7 @@
imageCreateInfo.extent.depth = 1;
imageCreateInfo.mipLevels = mipCount;
imageCreateInfo.samples = 1;
- imageCreateInfo.tiling = VK_LINEAR_TILING;
+ imageCreateInfo.tiling = VK_IMAGE_TILING_LINEAR;
imageCreateInfo.usage = VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT;
@@ -596,10 +596,10 @@
// void* pData);
// typedef struct VkSubresourceLayout_
// {
-// VkGpuSize offset; // Specified in bytes
-// VkGpuSize size; // Specified in bytes
-// VkGpuSize rowPitch; // Specified in bytes
-// VkGpuSize depthPitch; // Specified in bytes
+// VkDeviceSize offset; // Specified in bytes
+// VkDeviceSize size; // Specified in bytes
+// VkDeviceSize rowPitch; // Specified in bytes
+// VkDeviceSize depthPitch; // Specified in bytes
// } VkSubresourceLayout;
// typedef struct VkImageSubresource_
@@ -611,7 +611,7 @@
// typedef enum VkSubresourceInfoType_
// {
// // Info type for vkGetImageSubresourceInfo()
-// VK_INFO_TYPE_SUBRESOURCE_LAYOUT = 0x00000000,
+// VK_SUBRESOURCE_INFO_TYPE_LAYOUT = 0x00000000,
// VK_MAX_ENUM(VkSubresourceInfoType_)
// } VkSubresourceInfoType;
@@ -625,7 +625,7 @@
{
VkSubresourceLayout layout = {};
data_size = sizeof(layout);
- err = vkGetImageSubresourceInfo(image, &subresource, VK_INFO_TYPE_SUBRESOURCE_LAYOUT,
+ err = vkGetImageSubresourceInfo(image, &subresource, VK_SUBRESOURCE_INFO_TYPE_LAYOUT,
&data_size, &layout);
ASSERT_VK_SUCCESS(err);
ASSERT_EQ(sizeof(VkSubresourceLayout), data_size) << "Invalid structure (VkSubresourceLayout) size";
@@ -640,7 +640,7 @@
VkMemoryRequirements mem_req;
data_size = sizeof(mem_req);
- err = vkGetObjectInfo(image, VK_INFO_TYPE_MEMORY_REQUIREMENTS,
+ err = vkGetObjectInfo(image, VK_OBJECT_INFO_TYPE_MEMORY_REQUIREMENTS,
&data_size, &mem_req);
ASSERT_VK_SUCCESS(err);
ASSERT_EQ(data_size, sizeof(mem_req));
@@ -648,9 +648,9 @@
// VkResult VKAPI vkAllocMemory(
// VkDevice device,
// const VkMemoryAllocInfo* pAllocInfo,
- // VkGpuMemory* pMem);
+ // VkDeviceMemory* pMem);
VkMemoryAllocInfo mem_info = {};
- VkGpuMemory image_mem;
+ VkDeviceMemory image_mem;
mem_info.sType = VK_STRUCTURE_TYPE_MEMORY_ALLOC_INFO;
mem_info.pNext = NULL;
@@ -679,7 +679,7 @@
VkImageView view;
viewInfo.sType = VK_STRUCTURE_TYPE_IMAGE_VIEW_CREATE_INFO;
viewInfo.image = image;
- viewInfo.viewType = VK_IMAGE_VIEW_2D;
+ viewInfo.viewType = VK_IMAGE_VIEW_TYPE_2D;
viewInfo.format = fmt;
viewInfo.channels.r = VK_CHANNEL_SWIZZLE_R;
diff --git a/tests/render_tests.cpp b/tests/render_tests.cpp
index 89cc5a6..1f657fa 100644
--- a/tests/render_tests.cpp
+++ b/tests/render_tests.cpp
@@ -226,7 +226,7 @@
VkImage m_texture;
VkImageView m_textureView;
VkImageViewAttachInfo m_textureViewInfo;
- VkGpuMemory m_textureMem;
+ VkDeviceMemory m_textureMem;
VkSampler m_sampler;
@@ -291,10 +291,10 @@
}
cmdBuffer->PrepareAttachments();
- cmdBuffer->BindStateObject(VK_STATE_BIND_RASTER, m_stateRaster);
- cmdBuffer->BindStateObject(VK_STATE_BIND_VIEWPORT, m_stateViewport);
- cmdBuffer->BindStateObject(VK_STATE_BIND_COLOR_BLEND, m_colorBlend);
- cmdBuffer->BindStateObject(VK_STATE_BIND_DEPTH_STENCIL, m_stateDepthStencil);
+ cmdBuffer->BindStateObject(VK_STATE_BIND_POINT_RASTER, m_stateRaster);
+ cmdBuffer->BindStateObject(VK_STATE_BIND_POINT_VIEWPORT, m_stateViewport);
+ cmdBuffer->BindStateObject(VK_STATE_BIND_POINT_COLOR_BLEND, m_colorBlend);
+ cmdBuffer->BindStateObject(VK_STATE_BIND_POINT_DEPTH_STENCIL, m_stateDepthStencil);
descriptorSet.CreateVKDescriptorSet(cmdBuffer);
pipelineobj.CreateVKPipeline(descriptorSet);
cmdBuffer->BindPipeline(pipelineobj);
@@ -701,11 +701,11 @@
VkVertexInputAttributeDescription vi_attribs[2];
vi_attribs[0].binding = MESH_BIND_ID; // Binding ID
vi_attribs[0].location = 0; // location, position
- vi_attribs[0].format = VK_FMT_R32G32B32A32_SFLOAT; // format of source data
+ vi_attribs[0].format = VK_FORMAT_R32G32B32A32_SFLOAT; // format of source data
vi_attribs[0].offsetInBytes = 0; // Offset of first element in bytes from base of vertex
vi_attribs[1].binding = MESH_BIND_ID; // Binding ID
vi_attribs[1].location = 1; // location, color
- vi_attribs[1].format = VK_FMT_R32G32B32A32_SFLOAT; // format of source data
+ vi_attribs[1].format = VK_FORMAT_R32G32B32A32_SFLOAT; // format of source data
vi_attribs[1].offsetInBytes = 1*sizeof(float)*4; // Offset of first element in bytes from base of vertex
pipelineobj.AddVertexInputAttribs(vi_attribs,2);
@@ -779,7 +779,7 @@
VkVertexInputAttributeDescription vi_attrib;
vi_attrib.binding = MESH_BUF_ID; // index into vertexBindingDescriptions
vi_attrib.location = 0;
- vi_attrib.format = VK_FMT_R32G32_SFLOAT; // format of source data
+ vi_attrib.format = VK_FORMAT_R32G32_SFLOAT; // format of source data
vi_attrib.offsetInBytes = 0; // Offset of first element in bytes from base of vertex
pipelineobj.AddVertexInputAttribs(&vi_attrib, 1);
@@ -871,7 +871,7 @@
meshBuffer.BufferMemoryBarrier();
VkIndexBufferObj indexBuffer(m_device);
- indexBuffer.CreateAndInitBuffer(sizeof(g_idxData)/sizeof(g_idxData[0]), VK_INDEX_16, g_idxData);
+ indexBuffer.CreateAndInitBuffer(sizeof(g_idxData)/sizeof(g_idxData[0]), VK_INDEX_TYPE_UINT16, g_idxData);
indexBuffer.BufferMemoryBarrier();
VkShaderObj vs(m_device,vertShaderText,VK_SHADER_STAGE_VERTEX, this);
@@ -896,11 +896,11 @@
VkVertexInputAttributeDescription vi_attribs[2];
vi_attribs[0].binding = MESH_BIND_ID; // binding ID from BINDING_DESCRIPTION array to use for this attribute
vi_attribs[0].location = 0; // layout location of vertex attribute
- vi_attribs[0].format = VK_FMT_R32G32B32A32_SFLOAT; // format of source data
+ vi_attribs[0].format = VK_FORMAT_R32G32B32A32_SFLOAT; // format of source data
vi_attribs[0].offsetInBytes = 0; // Offset of first element in bytes from base of vertex
vi_attribs[1].binding = MESH_BIND_ID; // binding ID from BINDING_DESCRIPTION array to use for this attribute
vi_attribs[1].location = 1; // layout location of vertex attribute
- vi_attribs[1].format = VK_FMT_R32G32B32A32_SFLOAT; // format of source data
+ vi_attribs[1].format = VK_FORMAT_R32G32B32A32_SFLOAT; // format of source data
vi_attribs[1].offsetInBytes = 16; // Offset of first element in bytes from base of vertex
pipelineobj.AddVertexInputAttribs(vi_attribs,2);
@@ -990,7 +990,7 @@
VkVertexInputAttributeDescription vi_attribs[1];
vi_attribs[0].binding = MESH_BIND_ID; // binding ID
vi_attribs[0].location = 0;
- vi_attribs[0].format = VK_FMT_R32G32B32A32_SFLOAT; // format of source data
+ vi_attribs[0].format = VK_FORMAT_R32G32B32A32_SFLOAT; // format of source data
vi_attribs[0].offsetInBytes = 0; // Offset of first element in bytes from base of vertex
pipelineobj.AddVertexInputAttribs(vi_attribs,1);
@@ -1080,7 +1080,7 @@
VkVertexInputAttributeDescription vi_attribs[1];
vi_attribs[0].binding = MESH_BIND_ID; // binding ID
vi_attribs[0].location = 0;
- vi_attribs[0].format = VK_FMT_R32G32B32A32_SFLOAT; // format of source data
+ vi_attribs[0].format = VK_FORMAT_R32G32B32A32_SFLOAT; // format of source data
vi_attribs[0].offsetInBytes = 0; // Offset of first element in bytes from base of vertex
pipelineobj.AddVertexInputAttribs(vi_attribs,1);
@@ -1180,7 +1180,7 @@
VkVertexInputAttributeDescription vi_attribs[1];
vi_attribs[0].binding = MESH_BIND_ID; // binding ID
vi_attribs[0].location = 0;
- vi_attribs[0].format = VK_FMT_R32G32B32A32_SFLOAT; // format of source data
+ vi_attribs[0].format = VK_FORMAT_R32G32B32A32_SFLOAT; // format of source data
vi_attribs[0].offsetInBytes = 0; // Offset of first element in bytes from base of vertex
pipelineobj.AddVertexInputAttribs(vi_attribs,1);
@@ -1271,7 +1271,7 @@
VkVertexInputAttributeDescription vi_attribs[1];
vi_attribs[0].binding = MESH_BIND_ID; // binding ID
vi_attribs[0].location = 0;
- vi_attribs[0].format = VK_FMT_R32G32B32A32_SFLOAT; // format of source data
+ vi_attribs[0].format = VK_FORMAT_R32G32B32A32_SFLOAT; // format of source data
vi_attribs[0].offsetInBytes = 0; // Offset of first element in bytes from base of vertex
pipelineobj.AddVertexInputAttribs(vi_attribs,1);
@@ -1505,11 +1505,11 @@
VkVertexInputAttributeDescription vi_attribs[2];
vi_attribs[0].binding = MESH_BUF_ID; // binding ID
vi_attribs[0].location = 0;
- vi_attribs[0].format = VK_FMT_R32G32_SFLOAT; // format of source data
+ vi_attribs[0].format = VK_FORMAT_R32G32_SFLOAT; // format of source data
vi_attribs[0].offsetInBytes = 0; // Offset of first element in bytes from base of vertex
vi_attribs[1].binding = MESH_BUF_ID; // binding ID
vi_attribs[1].location = 1;
- vi_attribs[1].format = VK_FMT_R32G32_SFLOAT; // format of source data
+ vi_attribs[1].format = VK_FORMAT_R32G32_SFLOAT; // format of source data
vi_attribs[1].offsetInBytes = 16; // Offset of first element in bytes from base of vertex
pipelineobj.AddVertexInputAttribs(vi_attribs, 2);
@@ -1610,11 +1610,11 @@
VkVertexInputAttributeDescription vi_attribs[2];
vi_attribs[0].binding = MESH_BUF_ID; // binding ID
vi_attribs[0].location = 4;
- vi_attribs[0].format = VK_FMT_R32G32B32A32_SFLOAT; // format of source data
+ vi_attribs[0].format = VK_FORMAT_R32G32B32A32_SFLOAT; // format of source data
vi_attribs[0].offsetInBytes = sizeof(float) * 4 * 2; // Offset of first element in bytes from base of vertex
vi_attribs[1].binding = MESH_BUF_ID; // binding ID
vi_attribs[1].location = 1;
- vi_attribs[1].format = VK_FMT_R32G32B32A32_SFLOAT; // format of source data
+ vi_attribs[1].format = VK_FORMAT_R32G32B32A32_SFLOAT; // format of source data
vi_attribs[1].offsetInBytes = sizeof(float) * 4 * 1; // Offset of first element in bytes from base of vertex
pipelineobj.AddVertexInputAttribs(vi_attribs, 2);
@@ -1705,11 +1705,11 @@
VkVertexInputAttributeDescription vi_attribs[2];
vi_attribs[0].binding = MESH_BUF_ID; // binding ID
vi_attribs[0].location = 0;
- vi_attribs[0].format = VK_FMT_R32G32B32A32_SFLOAT; // format of source data
+ vi_attribs[0].format = VK_FORMAT_R32G32B32A32_SFLOAT; // format of source data
vi_attribs[0].offsetInBytes = 0; // Offset of first element in bytes from base of vertex
vi_attribs[1].binding = MESH_BUF_ID; // binding ID
vi_attribs[1].location = 1;
- vi_attribs[1].format = VK_FMT_R32G32B32A32_SFLOAT; // format of source data
+ vi_attribs[1].format = VK_FORMAT_R32G32B32A32_SFLOAT; // format of source data
vi_attribs[1].offsetInBytes = 16; // Offset of first element in bytes from base of vertex
pipelineobj.AddVertexInputAttribs(vi_attribs, 2);
@@ -1795,14 +1795,14 @@
VkPipelineDsStateCreateInfo ds_state;
ds_state.depthTestEnable = VK_TRUE;
ds_state.depthWriteEnable = VK_TRUE;
- ds_state.depthFunc = VK_COMPARE_LESS_EQUAL;
+ ds_state.depthCompareOp = VK_COMPARE_OP_LESS_EQUAL;
ds_state.depthBoundsEnable = VK_FALSE;
ds_state.stencilTestEnable = VK_FALSE;
ds_state.back.stencilDepthFailOp = VK_STENCIL_OP_KEEP;
ds_state.back.stencilFailOp = VK_STENCIL_OP_KEEP;
ds_state.back.stencilPassOp = VK_STENCIL_OP_KEEP;
- ds_state.back.stencilFunc = VK_COMPARE_ALWAYS;
- ds_state.format = VK_FMT_D32_SFLOAT;
+ ds_state.back.stencilCompareOp = VK_COMPARE_OP_ALWAYS;
+ ds_state.format = VK_FORMAT_D32_SFLOAT;
ds_state.front = ds_state.back;
pipelineobj.SetDepthStencil(&ds_state);
@@ -1819,11 +1819,11 @@
VkVertexInputAttributeDescription vi_attribs[2];
vi_attribs[0].binding = MESH_BUF_ID; // binding ID
vi_attribs[0].location = 0;
- vi_attribs[0].format = VK_FMT_R32G32B32A32_SFLOAT; // format of source data
+ vi_attribs[0].format = VK_FORMAT_R32G32B32A32_SFLOAT; // format of source data
vi_attribs[0].offsetInBytes = 0; // Offset of first element in bytes from base of vertex
vi_attribs[1].binding = MESH_BUF_ID; // binding ID
vi_attribs[1].location = 1;
- vi_attribs[1].format = VK_FMT_R32G32B32A32_SFLOAT; // format of source data
+ vi_attribs[1].format = VK_FORMAT_R32G32B32A32_SFLOAT; // format of source data
vi_attribs[1].offsetInBytes = 16; // Offset of first element in bytes from base of vertex
pipelineobj.AddVertexInputAttribs(vi_attribs, 2);
@@ -2582,11 +2582,11 @@
VkVertexInputAttributeDescription vi_attribs[2];
vi_attribs[0].binding = MESH_BIND_ID; // Binding ID
vi_attribs[0].location = 0; // location
- vi_attribs[0].format = VK_FMT_R32G32B32A32_SFLOAT; // format of source data
+ vi_attribs[0].format = VK_FORMAT_R32G32B32A32_SFLOAT; // format of source data
vi_attribs[0].offsetInBytes = 0; // Offset of first element in bytes from base of vertex
vi_attribs[1].binding = MESH_BIND_ID; // Binding ID
vi_attribs[1].location = 1; // location
- vi_attribs[1].format = VK_FMT_R32G32B32A32_SFLOAT; // format of source data
+ vi_attribs[1].format = VK_FORMAT_R32G32B32A32_SFLOAT; // format of source data
vi_attribs[1].offsetInBytes = 16; // Offset of first element in bytes from base of vertex
pipelineobj.AddVertexInputAttribs(vi_attribs,2);
@@ -2596,14 +2596,14 @@
VkPipelineDsStateCreateInfo ds_state;
ds_state.depthTestEnable = VK_TRUE;
ds_state.depthWriteEnable = VK_TRUE;
- ds_state.depthFunc = VK_COMPARE_LESS_EQUAL;
+ ds_state.depthCompareOp = VK_COMPARE_OP_LESS_EQUAL;
ds_state.depthBoundsEnable = VK_FALSE;
ds_state.stencilTestEnable = VK_FALSE;
ds_state.back.stencilDepthFailOp = VK_STENCIL_OP_KEEP;
ds_state.back.stencilFailOp = VK_STENCIL_OP_KEEP;
ds_state.back.stencilPassOp = VK_STENCIL_OP_KEEP;
- ds_state.back.stencilFunc = VK_COMPARE_ALWAYS;
- ds_state.format = VK_FMT_D32_SFLOAT;
+ ds_state.back.stencilCompareOp = VK_COMPARE_OP_ALWAYS;
+ ds_state.format = VK_FORMAT_D32_SFLOAT;
ds_state.front = ds_state.back;
pipelineobj.SetDepthStencil(&ds_state);
diff --git a/tests/test_common.h b/tests/test_common.h
index 2683597..fc00882 100644
--- a/tests/test_common.h
+++ b/tests/test_common.h
@@ -28,8 +28,8 @@
STR(VK_ERROR_UNKNOWN);
STR(VK_ERROR_UNAVAILABLE);
STR(VK_ERROR_INITIALIZATION_FAILED);
- STR(VK_ERROR_OUT_OF_MEMORY);
- STR(VK_ERROR_OUT_OF_GPU_MEMORY);
+ STR(VK_ERROR_OUT_OF_HOST_MEMORY);
+ STR(VK_ERROR_OUT_OF_DEVICE_MEMORY);
STR(VK_ERROR_DEVICE_ALREADY_CREATED);
STR(VK_ERROR_DEVICE_LOST);
STR(VK_ERROR_INVALID_POINTER);
diff --git a/tests/test_environment.h b/tests/test_environment.h
index 2081992..43530a7 100644
--- a/tests/test_environment.h
+++ b/tests/test_environment.h
@@ -18,7 +18,7 @@
const std::vector<Device *> &devices() { return devs_; }
Device &default_device() { return *(devs_[default_dev_]); }
- VkPhysicalGpu gpus[16];
+ VkPhysicalDevice gpus[16];
private:
VkApplicationInfo app_;
diff --git a/tests/vkrenderframework.cpp b/tests/vkrenderframework.cpp
index 6c1de3b..a81ae89 100644
--- a/tests/vkrenderframework.cpp
+++ b/tests/vkrenderframework.cpp
@@ -39,8 +39,8 @@
m_stateDepthStencil( VK_NULL_HANDLE ),
m_width( 256.0 ), // default window width
m_height( 256.0 ), // default window height
- m_render_target_fmt( VK_FMT_R8G8B8A8_UNORM ),
- m_depth_stencil_fmt( VK_FMT_UNDEFINED ),
+ m_render_target_fmt( VK_FORMAT_R8G8B8A8_UNORM ),
+ m_depth_stencil_fmt( VK_FORMAT_UNDEFINED ),
m_depth_clear_color( 1.0 ),
m_stencil_clear_color( 0 )
{
@@ -114,7 +114,7 @@
{
VkResult err;
- m_render_target_fmt = VK_FMT_B8G8R8A8_UNORM;
+ m_render_target_fmt = VK_FORMAT_B8G8R8A8_UNORM;
// create a raster state (solid, back-face culling)
VkDynamicRsStateCreateInfo raster = {};
@@ -256,7 +256,7 @@
-VkDeviceObj::VkDeviceObj(uint32_t id, VkPhysicalGpu obj) :
+VkDeviceObj::VkDeviceObj(uint32_t id, VkPhysicalDevice obj) :
vk_testing::Device(obj), id(id)
{
init();
@@ -356,7 +356,7 @@
for (int i = 0; i < m_type_counts.size(); i++) {
bindings[i].descriptorType = m_type_counts[i].type;
bindings[i].count = m_type_counts[i].count;
- bindings[i].stageFlags = VK_SHADER_STAGE_FLAGS_ALL;
+ bindings[i].stageFlags = VK_SHADER_STAGE_ALL;
bindings[i].pImmutableSamplers = NULL;
}
@@ -428,18 +428,10 @@
VkImageMemoryBarrier *pmemory_barrier = &barrier;
- VkPipeEvent pipe_events[] = { VK_PIPE_EVENT_GPU_COMMANDS_COMPLETE };
- VkPipelineBarrier pipeline_barrier = {};
- pipeline_barrier.sType = VK_STRUCTURE_TYPE_PIPELINE_BARRIER;
- pipeline_barrier.pNext = NULL;
- pipeline_barrier.eventCount = 1;
- pipeline_barrier.pEvents = pipe_events;
- pipeline_barrier.waitEvent = VK_WAIT_EVENT_TOP_OF_PIPE;
- pipeline_barrier.memBarrierCount = 1;
- pipeline_barrier.ppMemBarriers = (const void **)&pmemory_barrier;
+ VkPipeEvent pipe_events[] = { VK_PIPE_EVENT_COMMANDS_COMPLETE };
// write barrier to the command buffer
- vkCmdPipelineBarrier(cmd_buf->obj(), &pipeline_barrier);
+ vkCmdPipelineBarrier(cmd_buf->obj(), VK_WAIT_EVENT_TOP_OF_PIPE, 1, pipe_events, 1, (const void **)&pmemory_barrier);
}
void VkImageObj::SetLayout(VkCommandBufferObj *cmd_buf,
@@ -518,7 +510,7 @@
bool VkImageObj::IsCompatible(VkFlags usage, VkFlags features)
{
if ((usage & VK_IMAGE_USAGE_SAMPLED_BIT) &&
- !(features & VK_FORMAT_SAMPLED_IMAGE_BIT))
+ !(features & VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT))
return false;
return true;
@@ -547,28 +539,28 @@
size = sizeof(image_fmt);
err = vkGetFormatInfo(m_device->obj(), fmt,
- VK_INFO_TYPE_FORMAT_PROPERTIES,
+ VK_FORMAT_INFO_TYPE_PROPERTIES,
&size, &image_fmt);
ASSERT_VK_SUCCESS(err);
- if (requested_tiling == VK_LINEAR_TILING) {
+ if (requested_tiling == VK_IMAGE_TILING_LINEAR) {
if (IsCompatible(usage, image_fmt.linearTilingFeatures)) {
- tiling = VK_LINEAR_TILING;
+ tiling = VK_IMAGE_TILING_LINEAR;
} else if (IsCompatible(usage, image_fmt.optimalTilingFeatures)) {
- tiling = VK_OPTIMAL_TILING;
+ tiling = VK_IMAGE_TILING_OPTIMAL;
} else {
ASSERT_TRUE(false) << "Error: Cannot find requested tiling configuration";
}
} else if (IsCompatible(usage, image_fmt.optimalTilingFeatures)) {
- tiling = VK_OPTIMAL_TILING;
+ tiling = VK_IMAGE_TILING_OPTIMAL;
} else if (IsCompatible(usage, image_fmt.linearTilingFeatures)) {
- tiling = VK_LINEAR_TILING;
+ tiling = VK_IMAGE_TILING_LINEAR;
} else {
ASSERT_TRUE(false) << "Error: Cannot find requested tiling configuration";
}
VkImageCreateInfo imageCreateInfo = vk_testing::Image::create_info();
- imageCreateInfo.imageType = VK_IMAGE_2D;
+ imageCreateInfo.imageType = VK_IMAGE_TYPE_2D;
imageCreateInfo.format = fmt;
imageCreateInfo.extent.width = w;
imageCreateInfo.extent.height = h;
@@ -653,13 +645,13 @@
:VkImageObj(device)
{
m_device = device;
- const VkFormat tex_format = VK_FMT_B8G8R8A8_UNORM;
+ const VkFormat tex_format = VK_FORMAT_B8G8R8A8_UNORM;
uint32_t tex_colors[2] = { 0xffff0000, 0xff00ff00 };
void *data;
int32_t x, y;
VkImageObj stagingImage(device);
- stagingImage.init(16, 16, tex_format, 0, VK_LINEAR_TILING);
+ stagingImage.init(16, 16, tex_format, 0, VK_IMAGE_TILING_LINEAR);
VkSubresourceLayout layout = stagingImage.subresource_layout(subresource(VK_IMAGE_ASPECT_COLOR, 0, 0));
if (colors == NULL)
@@ -673,7 +665,7 @@
view.sType = VK_STRUCTURE_TYPE_IMAGE_VIEW_CREATE_INFO;
view.pNext = NULL;
view.image = VK_NULL_HANDLE;
- view.viewType = VK_IMAGE_VIEW_2D;
+ view.viewType = VK_IMAGE_VIEW_TYPE_2D;
view.format = tex_format;
view.channels.r = VK_CHANNEL_SWIZZLE_R;
view.channels.g = VK_CHANNEL_SWIZZLE_G;
@@ -687,7 +679,7 @@
view.minLod = 0.0f;
/* create image */
- init(16, 16, tex_format, VK_IMAGE_USAGE_SAMPLED_BIT, VK_OPTIMAL_TILING);
+ init(16, 16, tex_format, VK_IMAGE_USAGE_SAMPLED_BIT, VK_IMAGE_TILING_OPTIMAL);
/* create image view */
view.image = obj();
@@ -714,16 +706,16 @@
samplerCreateInfo.sType = VK_STRUCTURE_TYPE_SAMPLER_CREATE_INFO;
samplerCreateInfo.magFilter = VK_TEX_FILTER_NEAREST;
samplerCreateInfo.minFilter = VK_TEX_FILTER_NEAREST;
- samplerCreateInfo.mipMode = VK_TEX_MIPMAP_BASE;
+ samplerCreateInfo.mipMode = VK_TEX_MIPMAP_MODE_BASE;
samplerCreateInfo.addressU = VK_TEX_ADDRESS_WRAP;
samplerCreateInfo.addressV = VK_TEX_ADDRESS_WRAP;
samplerCreateInfo.addressW = VK_TEX_ADDRESS_WRAP;
samplerCreateInfo.mipLodBias = 0.0;
samplerCreateInfo.maxAnisotropy = 0.0;
- samplerCreateInfo.compareFunc = VK_COMPARE_NEVER;
+ samplerCreateInfo.compareOp = VK_COMPARE_OP_NEVER;
samplerCreateInfo.minLod = 0.0;
samplerCreateInfo.maxLod = 0.0;
- samplerCreateInfo.borderColorType = VK_BORDER_COLOR_OPAQUE_WHITE;
+ samplerCreateInfo.borderColor = VK_BORDER_COLOR_OPAQUE_WHITE;
init(*m_device, samplerCreateInfo);
}
@@ -767,7 +759,7 @@
VkBufferViewCreateInfo view_info = {};
view_info.sType = VK_STRUCTURE_TYPE_BUFFER_VIEW_CREATE_INFO;
view_info.buffer = obj();
- view_info.viewType = VK_BUFFER_VIEW_RAW;
+ view_info.viewType = VK_BUFFER_VIEW_TYPE_RAW;
view_info.offset = 0;
view_info.range = allocationSize;
m_bufferView.init(*m_device, view_info);
@@ -776,7 +768,7 @@
this->m_bufferViewInfo.view = m_bufferView.obj();
}
-void VkConstantBufferObj::Bind(VkCmdBuffer cmdBuffer, VkGpuSize offset, uint32_t binding)
+void VkConstantBufferObj::Bind(VkCmdBuffer cmdBuffer, VkDeviceSize offset, uint32_t binding)
{
vkCmdBindVertexBuffers(cmdBuffer, binding, 1, &obj(), &offset);
}
@@ -827,17 +819,10 @@
buffer_memory_barrier(outputMask, inputMask, 0, m_numVertices * m_stride);
VkBufferMemoryBarrier *pmemory_barrier = &memory_barrier;
- VkPipeEvent set_events[] = { VK_PIPE_EVENT_GPU_COMMANDS_COMPLETE };
- VkPipelineBarrier pipeline_barrier = {};
- pipeline_barrier.sType = VK_STRUCTURE_TYPE_PIPELINE_BARRIER;
- pipeline_barrier.eventCount = 1;
- pipeline_barrier.pEvents = set_events;
- pipeline_barrier.waitEvent = VK_WAIT_EVENT_TOP_OF_PIPE;
- pipeline_barrier.memBarrierCount = 1;
- pipeline_barrier.ppMemBarriers = (const void **)&pmemory_barrier;
+ VkPipeEvent set_events[] = { VK_PIPE_EVENT_COMMANDS_COMPLETE };
// write barrier to the command buffer
- m_commandBuffer->PipelineBarrier(&pipeline_barrier);
+ m_commandBuffer->PipelineBarrier(VK_WAIT_EVENT_TOP_OF_PIPE, 1, set_events, 1, (const void **)&pmemory_barrier);
// finish recording the command buffer
err = m_commandBuffer->EndCommandBuffer();
@@ -870,17 +855,17 @@
m_numVertices = numIndexes;
m_indexType = indexType;
switch (indexType) {
- case VK_INDEX_8:
+ case VK_INDEX_TYPE_UINT8:
m_stride = 1;
- viewFormat = VK_FMT_R8_UINT;
+ viewFormat = VK_FORMAT_R8_UINT;
break;
- case VK_INDEX_16:
+ case VK_INDEX_TYPE_UINT16:
m_stride = 2;
- viewFormat = VK_FMT_R16_UINT;
+ viewFormat = VK_FORMAT_R16_UINT;
break;
- case VK_INDEX_32:
+ case VK_INDEX_TYPE_UINT32:
m_stride = 4;
- viewFormat = VK_FMT_R32_UINT;
+ viewFormat = VK_FORMAT_R32_UINT;
break;
default:
assert(!"unknown index type");
@@ -898,7 +883,7 @@
VkBufferViewCreateInfo view_info = {};
view_info.sType = VK_STRUCTURE_TYPE_BUFFER_VIEW_CREATE_INFO;
view_info.buffer = obj();
- view_info.viewType = VK_BUFFER_VIEW_FORMATTED;
+ view_info.viewType = VK_BUFFER_VIEW_TYPE_FORMATTED;
view_info.format = viewFormat;
view_info.offset = 0;
view_info.range = allocationSize;
@@ -908,7 +893,7 @@
this->m_bufferViewInfo.view = m_bufferView.obj();
}
-void VkIndexBufferObj::Bind(VkCmdBuffer cmdBuffer, VkGpuSize offset)
+void VkIndexBufferObj::Bind(VkCmdBuffer cmdBuffer, VkDeviceSize offset)
{
vkCmdBindIndexBuffer(cmdBuffer, obj(), offset, m_indexType);
}
@@ -930,7 +915,7 @@
return stageInfo;
}
-VkShaderObj::VkShaderObj(VkDeviceObj *device, const char * shader_code, VkPipelineShaderStage stage, VkRenderFramework *framework)
+VkShaderObj::VkShaderObj(VkDeviceObj *device, const char * shader_code, VkShaderStage stage, VkRenderFramework *framework)
{
VkResult err = VK_SUCCESS;
std::vector<unsigned int> spv;
@@ -950,7 +935,7 @@
createInfo.pCode = malloc(createInfo.codeSize);
createInfo.flags = 0;
- /* try version 0 first: VkPipelineShaderStage followed by GLSL */
+ /* try version 0 first: VkShaderStage followed by GLSL */
((uint32_t *) createInfo.pCode)[0] = ICD_SPV_MAGIC;
((uint32_t *) createInfo.pCode)[1] = 0;
((uint32_t *) createInfo.pCode)[2] = stage;
@@ -981,7 +966,7 @@
m_ia_state.sType = VK_STRUCTURE_TYPE_PIPELINE_IA_STATE_CREATE_INFO;
m_ia_state.pNext = VK_NULL_HANDLE;
- m_ia_state.topology = VK_TOPOLOGY_TRIANGLE_LIST;
+ m_ia_state.topology = VK_PRIMITIVE_TOPOLOGY_TRIANGLE_LIST;
m_ia_state.disableVertexReuse = VK_FALSE;
m_ia_state.primitiveRestartEnable = VK_FALSE;
m_ia_state.primitiveRestartIndex = 0;
@@ -993,8 +978,8 @@
m_rs_state.programPointSize = VK_FALSE;
m_rs_state.pointOrigin = VK_COORDINATE_ORIGIN_UPPER_LEFT;
m_rs_state.provokingVertex = VK_PROVOKING_VERTEX_LAST;
- m_rs_state.fillMode = VK_FILL_SOLID;
- m_rs_state.cullMode = VK_CULL_NONE;
+ m_rs_state.fillMode = VK_FILL_MODE_SOLID;
+ m_rs_state.cullMode = VK_CULL_MODE_NONE;
m_rs_state.frontFace = VK_FRONT_FACE_CCW;
memset(&m_cb_state,0,sizeof(m_cb_state));
@@ -1013,21 +998,21 @@
m_ds_state.sType = VK_STRUCTURE_TYPE_PIPELINE_DS_STATE_CREATE_INFO;
m_ds_state.pNext = &m_ms_state,
- m_ds_state.format = VK_FMT_D32_SFLOAT;
+ m_ds_state.format = VK_FORMAT_D32_SFLOAT;
m_ds_state.depthTestEnable = VK_FALSE;
m_ds_state.depthWriteEnable = VK_FALSE;
m_ds_state.depthBoundsEnable = VK_FALSE;
- m_ds_state.depthFunc = VK_COMPARE_LESS_EQUAL;
+ m_ds_state.depthCompareOp = VK_COMPARE_OP_LESS_EQUAL;
m_ds_state.back.stencilDepthFailOp = VK_STENCIL_OP_KEEP;
m_ds_state.back.stencilFailOp = VK_STENCIL_OP_KEEP;
m_ds_state.back.stencilPassOp = VK_STENCIL_OP_KEEP;
- m_ds_state.back.stencilFunc = VK_COMPARE_ALWAYS;
+ m_ds_state.back.stencilCompareOp = VK_COMPARE_OP_ALWAYS;
m_ds_state.stencilTestEnable = VK_FALSE;
m_ds_state.front = m_ds_state.back;
VkPipelineCbAttachmentState att = {};
att.blendEnable = VK_FALSE;
- att.format = VK_FMT_B8G8R8A8_UNORM;
+ att.format = VK_FORMAT_B8G8R8A8_UNORM;
att.channelWriteMask = 0xf;
AddColorAttachment(0, &att);
@@ -1072,7 +1057,7 @@
m_ds_state.depthTestEnable = ds_state->depthTestEnable;
m_ds_state.depthWriteEnable = ds_state->depthWriteEnable;
m_ds_state.depthBoundsEnable = ds_state->depthBoundsEnable;
- m_ds_state.depthFunc = ds_state->depthFunc;
+ m_ds_state.depthCompareOp = ds_state->depthCompareOp;
m_ds_state.stencilTestEnable = ds_state->stencilTestEnable;
m_ds_state.back = ds_state->back;
m_ds_state.front = ds_state->front;
@@ -1110,9 +1095,9 @@
init(*m_device, info);
}
-vector<VkGpuMemory> VkMemoryRefManager::mem_refs() const
+vector<VkDeviceMemory> VkMemoryRefManager::mem_refs() const
{
- std::vector<VkGpuMemory> mems;
+ std::vector<VkDeviceMemory> mems;
if (this->mem_refs_.size()) {
mems.reserve(this->mem_refs_.size());
for (uint32_t i = 0; i < this->mem_refs_.size(); i++)
@@ -1124,11 +1109,11 @@
void VkMemoryRefManager::AddMemoryRefs(vk_testing::Object &vkObject)
{
- const std::vector<VkGpuMemory> mems = vkObject.memories();
+ const std::vector<VkDeviceMemory> mems = vkObject.memories();
AddMemoryRefs(mems);
}
-void VkMemoryRefManager::AddMemoryRefs(vector<VkGpuMemory> mem)
+void VkMemoryRefManager::AddMemoryRefs(vector<VkDeviceMemory> mem)
{
for (size_t i = 0; i < mem.size(); i++) {
if (mem[i] != NULL) {
@@ -1182,9 +1167,9 @@
return VK_SUCCESS;
}
-void VkCommandBufferObj::PipelineBarrier(VkPipelineBarrier *barrierPtr)
+void VkCommandBufferObj::PipelineBarrier(VkWaitEvent waitEvent, uint32_t pipeEventCount, const VkPipeEvent* pPipeEvents, uint32_t memBarrierCount, const void** ppMemBarriers)
{
- vkCmdPipelineBarrier(obj(), barrierPtr);
+ vkCmdPipelineBarrier(obj(), waitEvent, pipeEventCount, pPipeEvents, memBarrierCount, ppMemBarriers);
}
void VkCommandBufferObj::ClearAllBuffers(VkClearColor clear_color, float depth_clear_color, uint32_t stencil_clear_color,
@@ -1215,19 +1200,12 @@
memory_barrier.subresourceRange = srRange;
VkImageMemoryBarrier *pmemory_barrier = &memory_barrier;
- VkPipeEvent set_events[] = { VK_PIPE_EVENT_GPU_COMMANDS_COMPLETE };
- VkPipelineBarrier pipeline_barrier = {};
- pipeline_barrier.sType = VK_STRUCTURE_TYPE_PIPELINE_BARRIER;
- pipeline_barrier.eventCount = 1;
- pipeline_barrier.pEvents = set_events;
- pipeline_barrier.waitEvent = VK_WAIT_EVENT_TOP_OF_PIPE;
- pipeline_barrier.memBarrierCount = 1;
- pipeline_barrier.ppMemBarriers = (const void **)&pmemory_barrier;
+ VkPipeEvent set_events[] = { VK_PIPE_EVENT_COMMANDS_COMPLETE };
for (i = 0; i < m_renderTargets.size(); i++) {
memory_barrier.image = m_renderTargets[i]->image();
memory_barrier.oldLayout = m_renderTargets[i]->layout();
- vkCmdPipelineBarrier( obj(), &pipeline_barrier);
+ vkCmdPipelineBarrier( obj(), VK_WAIT_EVENT_TOP_OF_PIPE, 1, set_events, 1, (const void **)&pmemory_barrier);
m_renderTargets[i]->layout(memory_barrier.newLayout);
vkCmdClearColorImage(obj(),
@@ -1253,7 +1231,7 @@
memory_barrier.image = depthStencilObj->obj();
memory_barrier.subresourceRange = dsRange;
- vkCmdPipelineBarrier( obj(), &pipeline_barrier);
+ vkCmdPipelineBarrier( obj(), VK_WAIT_EVENT_TOP_OF_PIPE, 1, set_events, 1, (const void **)&pmemory_barrier);
vkCmdClearDepthStencil(obj(),
depthStencilObj->obj(), VK_IMAGE_LAYOUT_CLEAR_OPTIMAL,
@@ -1266,7 +1244,7 @@
memory_barrier.oldLayout = VK_IMAGE_LAYOUT_CLEAR_OPTIMAL;
memory_barrier.newLayout = depthStencilObj->BindInfo()->layout;
memory_barrier.subresourceRange = dsRange;
- vkCmdPipelineBarrier( obj(), &pipeline_barrier);
+ vkCmdPipelineBarrier( obj(), VK_WAIT_EVENT_TOP_OF_PIPE, 1, set_events, 1, (const void **)&pmemory_barrier);
}
}
@@ -1305,20 +1283,13 @@
memory_barrier.subresourceRange = srRange;
VkImageMemoryBarrier *pmemory_barrier = &memory_barrier;
- VkPipeEvent set_events[] = { VK_PIPE_EVENT_GPU_COMMANDS_COMPLETE };
- VkPipelineBarrier pipeline_barrier = {};
- pipeline_barrier.sType = VK_STRUCTURE_TYPE_PIPELINE_BARRIER;
- pipeline_barrier.eventCount = 1;
- pipeline_barrier.pEvents = set_events;
- pipeline_barrier.waitEvent = VK_WAIT_EVENT_TOP_OF_PIPE;
- pipeline_barrier.memBarrierCount = 1;
- pipeline_barrier.ppMemBarriers = (const void **)&pmemory_barrier;
+ VkPipeEvent set_events[] = { VK_PIPE_EVENT_COMMANDS_COMPLETE };
for(i=0; i<m_renderTargets.size(); i++)
{
memory_barrier.image = m_renderTargets[i]->image();
memory_barrier.oldLayout = m_renderTargets[i]->layout();
- vkCmdPipelineBarrier( obj(), &pipeline_barrier);
+ vkCmdPipelineBarrier( obj(), VK_WAIT_EVENT_TOP_OF_PIPE, 1, set_events, 1, (const void **)&pmemory_barrier);
m_renderTargets[i]->layout(memory_barrier.newLayout);
}
}
@@ -1411,7 +1382,7 @@
mem_ref_mgr.AddMemoryRefs(*indexBuffer);
}
-void VkCommandBufferObj::BindVertexBuffer(VkConstantBufferObj *vertexBuffer, VkGpuSize offset, uint32_t binding)
+void VkCommandBufferObj::BindVertexBuffer(VkConstantBufferObj *vertexBuffer, VkDeviceSize offset, uint32_t binding)
{
vkCmdBindVertexBuffers(obj(), binding, 1, &vertexBuffer->obj(), &offset);
mem_ref_mgr.AddMemoryRefs(*vertexBuffer);
@@ -1438,11 +1409,11 @@
m_device = device;
m_initialized = true;
- m_depth_stencil_fmt = VK_FMT_D16_UNORM;
+ m_depth_stencil_fmt = VK_FORMAT_D16_UNORM;
image_info.sType = VK_STRUCTURE_TYPE_IMAGE_CREATE_INFO;
image_info.pNext = NULL;
- image_info.imageType = VK_IMAGE_2D;
+ image_info.imageType = VK_IMAGE_TYPE_2D;
image_info.format = m_depth_stencil_fmt;
image_info.extent.width = width;
image_info.extent.height = height;
@@ -1450,7 +1421,7 @@
image_info.mipLevels = 1;
image_info.arraySize = 1;
image_info.samples = 1;
- image_info.tiling = VK_OPTIMAL_TILING;
+ image_info.tiling = VK_IMAGE_TILING_OPTIMAL;
image_info.usage = VK_IMAGE_USAGE_DEPTH_STENCIL_BIT;
image_info.flags = 0;
init(*m_device, image_info);
diff --git a/tests/vkrenderframework.h b/tests/vkrenderframework.h
index 55a4769..09c07dc 100644
--- a/tests/vkrenderframework.h
+++ b/tests/vkrenderframework.h
@@ -34,14 +34,14 @@
class VkDeviceObj : public vk_testing::Device
{
public:
- VkDeviceObj(uint32_t id, VkPhysicalGpu obj);
+ VkDeviceObj(uint32_t id, VkPhysicalDevice obj);
VkDevice device() { return obj(); }
void get_device_queue();
uint32_t id;
- VkPhysicalGpuProperties props;
- const VkPhysicalGpuQueueProperties *queue_props;
+ VkPhysicalDeviceProperties props;
+ const VkPhysicalDeviceQueueProperties *queue_props;
VkQueue m_queue;
};
@@ -50,13 +50,13 @@
{
public:
void AddMemoryRefs(vk_testing::Object &vkObject);
- void AddMemoryRefs(vector<VkGpuMemory> mem);
+ void AddMemoryRefs(vector<VkDeviceMemory> mem);
void EmitAddMemoryRefs(VkQueue queue);
void EmitRemoveMemoryRefs(VkQueue queue);
- vector<VkGpuMemory> mem_refs() const;
+ vector<VkDeviceMemory> mem_refs() const;
protected:
- vector<VkGpuMemory> mem_refs_;
+ vector<VkDeviceMemory> mem_refs_;
};
@@ -83,7 +83,7 @@
~VkRenderFramework();
VkDevice device() {return m_device->device();}
- VkPhysicalGpu gpu() {return objs[0];}
+ VkPhysicalDevice gpu() {return objs[0];}
VkRenderPass renderPass() {return m_renderPass;}
VkFramebuffer framebuffer() {return m_framebuffer;}
void InitViewport(float width, float height);
@@ -100,7 +100,7 @@
protected:
VkApplicationInfo app_info;
VkInstance inst;
- VkPhysicalGpu objs[16];
+ VkPhysicalDevice objs[16];
uint32_t gpu_count;
VkDeviceObj *m_device;
VkCmdBuffer m_cmdBuffer;
@@ -157,17 +157,17 @@
VkResult BeginCommandBuffer(VkCmdBufferBeginInfo *pInfo);
VkResult BeginCommandBuffer(VkRenderPass renderpass_obj, VkFramebuffer framebuffer_obj);
VkResult EndCommandBuffer();
- void PipelineBarrier(VkPipelineBarrier *barrierPtr);
+ void PipelineBarrier(VkWaitEvent waitEvent, uint32_t pipeEventCount, const VkPipeEvent* pPipeEvents, uint32_t memBarrierCount, const void** ppMemBarriers);
void AddRenderTarget(VkImageObj *renderTarget);
void AddDepthStencil();
void ClearAllBuffers(VkClearColor clear_color, float depth_clear_color, uint32_t stencil_clear_color, VkDepthStencilObj *depthStencilObj);
void PrepareAttachments();
void AddMemoryRefs(vk_testing::Object &vkObject);
- void AddMemoryRefs(uint32_t ref_count, const VkGpuMemory *mem);
+ void AddMemoryRefs(uint32_t ref_count, const VkDeviceMemory *mem);
void AddMemoryRefs(vector<vk_testing::Object *> images);
void BindPipeline(VkPipelineObj &pipeline);
void BindDescriptorSet(VkDescriptorSetObj &descriptorSet);
- void BindVertexBuffer(VkConstantBufferObj *vertexBuffer, VkGpuSize offset, uint32_t binding);
+ void BindVertexBuffer(VkConstantBufferObj *vertexBuffer, VkDeviceSize offset, uint32_t binding);
void BindIndexBuffer(VkIndexBufferObj *indexBuffer, uint32_t offset);
void BindStateObject(VkStateBindPoint stateBindPoint, VkDynamicStateObject stateObject);
void BeginRenderPass(VkRenderPass renderpass, VkFramebuffer framebuffer);
@@ -208,7 +208,7 @@
VK_MEMORY_INPUT_DEPTH_STENCIL_ATTACHMENT_BIT |
VK_MEMORY_INPUT_TRANSFER_BIT);
- void Bind(VkCmdBuffer cmdBuffer, VkGpuSize offset, uint32_t binding);
+ void Bind(VkCmdBuffer cmdBuffer, VkDeviceSize offset, uint32_t binding);
VkBufferViewAttachInfo m_bufferViewInfo;
@@ -226,7 +226,7 @@
public:
VkIndexBufferObj(VkDeviceObj *device);
void CreateAndInitBuffer(int numIndexes, VkIndexType dataFormat, const void* data);
- void Bind(VkCmdBuffer cmdBuffer, VkGpuSize offset);
+ void Bind(VkCmdBuffer cmdBuffer, VkDeviceSize offset);
VkIndexType GetIndexType();
protected:
@@ -242,7 +242,7 @@
public:
void init(uint32_t w, uint32_t h,
VkFormat fmt, VkFlags usage,
- VkImageTiling tiling=VK_LINEAR_TILING);
+ VkImageTiling tiling=VK_IMAGE_TILING_LINEAR);
// void clear( CommandBuffer*, uint32_t[4] );
@@ -251,9 +251,9 @@
m_imageInfo.layout = layout;
}
- VkGpuMemory memory() const
+ VkDeviceMemory memory() const
{
- const std::vector<VkGpuMemory> mems = memories();
+ const std::vector<VkDeviceMemory> mems = memories();
return mems.empty() ? VK_NULL_HANDLE : mems[0];
}
@@ -278,7 +278,7 @@
VK_STRUCTURE_TYPE_COLOR_ATTACHMENT_VIEW_CREATE_INFO,
VK_NULL_HANDLE,
obj(),
- VK_FMT_B8G8R8A8_UNORM,
+ VK_FORMAT_B8G8R8A8_UNORM,
0,
0,
1
@@ -328,7 +328,7 @@
protected:
VkDeviceObj *m_device;
vk_testing::ImageView m_textureView;
- VkGpuSize m_rowPitch;
+ VkDeviceSize m_rowPitch;
};
class VkSamplerObj : public vk_testing::Sampler
@@ -376,12 +376,12 @@
class VkShaderObj : public vk_testing::Shader
{
public:
- VkShaderObj(VkDeviceObj *device, const char * shaderText, VkPipelineShaderStage stage, VkRenderFramework *framework);
+ VkShaderObj(VkDeviceObj *device, const char * shaderText, VkShaderStage stage, VkRenderFramework *framework);
VkPipelineShaderStageCreateInfo* GetStageCreateInfo();
protected:
VkPipelineShaderStageCreateInfo stage_info;
- VkPipelineShaderStage m_stage;
+ VkShaderStage m_stage;
VkDeviceObj *m_device;
};
diff --git a/tests/vktestbinding.cpp b/tests/vktestbinding.cpp
index d9cdc9b..a3e134d 100644
--- a/tests/vktestbinding.cpp
+++ b/tests/vktestbinding.cpp
@@ -63,13 +63,13 @@
}
template<typename T>
-std::vector<T> get_info(VkPhysicalGpu gpu, VkPhysicalGpuInfoType type, size_t min_elems)
+std::vector<T> get_info(VkPhysicalDevice gpu, VkPhysicalDeviceInfoType type, size_t min_elems)
{
std::vector<T> info;
size_t size;
- if (EXPECT(vkGetGpuInfo(gpu, type, &size, NULL) == VK_SUCCESS && size % sizeof(T) == 0)) {
+ if (EXPECT(vkGetPhysicalDeviceInfo(gpu, type, &size, NULL) == VK_SUCCESS && size % sizeof(T) == 0)) {
info.resize(size / sizeof(T));
- if (!EXPECT(vkGetGpuInfo(gpu, type, &size, &info[0]) == VK_SUCCESS && size == info.size() * sizeof(T)))
+ if (!EXPECT(vkGetPhysicalDeviceInfo(gpu, type, &size, &info[0]) == VK_SUCCESS && size == info.size() * sizeof(T)))
info.clear();
}
@@ -105,24 +105,24 @@
error_callback = callback;
}
-VkPhysicalGpuProperties PhysicalGpu::properties() const
+VkPhysicalDeviceProperties PhysicalGpu::properties() const
{
- return get_info<VkPhysicalGpuProperties>(gpu_, VK_INFO_TYPE_PHYSICAL_GPU_PROPERTIES, 1)[0];
+ return get_info<VkPhysicalDeviceProperties>(gpu_, VK_PHYSICAL_DEVICE_INFO_TYPE_PROPERTIES, 1)[0];
}
-VkPhysicalGpuPerformance PhysicalGpu::performance() const
+VkPhysicalDevicePerformance PhysicalGpu::performance() const
{
- return get_info<VkPhysicalGpuPerformance>(gpu_, VK_INFO_TYPE_PHYSICAL_GPU_PERFORMANCE, 1)[0];
+ return get_info<VkPhysicalDevicePerformance>(gpu_, VK_PHYSICAL_DEVICE_INFO_TYPE_PERFORMANCE, 1)[0];
}
-std::vector<VkPhysicalGpuQueueProperties> PhysicalGpu::queue_properties() const
+std::vector<VkPhysicalDeviceQueueProperties> PhysicalGpu::queue_properties() const
{
- return get_info<VkPhysicalGpuQueueProperties>(gpu_, VK_INFO_TYPE_PHYSICAL_GPU_QUEUE_PROPERTIES, 0);
+ return get_info<VkPhysicalDeviceQueueProperties>(gpu_, VK_PHYSICAL_DEVICE_INFO_TYPE_QUEUE_PROPERTIES, 0);
}
-VkPhysicalGpuMemoryProperties PhysicalGpu::memory_properties() const
+VkPhysicalDeviceMemoryProperties PhysicalGpu::memory_properties() const
{
- return get_info<VkPhysicalGpuMemoryProperties>(gpu_, VK_INFO_TYPE_PHYSICAL_GPU_MEMORY_PROPERTIES, 1)[0];
+ return get_info<VkPhysicalDeviceMemoryProperties>(gpu_, VK_PHYSICAL_DEVICE_INFO_TYPE_MEMORY_PROPERTIES, 1)[0];
}
std::vector<const char *> PhysicalGpu::layers(std::vector<char> &buf) const
@@ -164,10 +164,10 @@
return exts;
}
-VkGpuCompatibilityInfo PhysicalGpu::compatibility(const PhysicalGpu &other) const
+VkPhysicalDeviceCompatibilityInfo PhysicalGpu::compatibility(const PhysicalGpu &other) const
{
- VkGpuCompatibilityInfo data;
- if (!EXPECT(vkGetMultiGpuCompatibility(gpu_, other.gpu_, &data) == VK_SUCCESS))
+ VkPhysicalDeviceCompatibilityInfo data;
+ if (!EXPECT(vkGetMultiDeviceCompatibility(gpu_, other.gpu_, &data) == VK_SUCCESS))
memset(&data, 0, sizeof(data));
return data;
@@ -187,7 +187,7 @@
uint32_t BaseObject::memory_allocation_count() const
{
- return get_info<uint32_t>(obj_, VK_INFO_TYPE_MEMORY_ALLOCATION_COUNT, 1)[0];
+ return get_info<uint32_t>(obj_, VK_OBJECT_INFO_TYPE_MEMORY_ALLOCATION_COUNT, 1)[0];
}
std::vector<VkMemoryRequirements> BaseObject::memory_requirements() const
@@ -195,11 +195,11 @@
VkResult err;
uint32_t num_allocations = 0;
size_t num_alloc_size = sizeof(num_allocations);
- err = vkGetObjectInfo(obj_, VK_INFO_TYPE_MEMORY_ALLOCATION_COUNT,
+ err = vkGetObjectInfo(obj_, VK_OBJECT_INFO_TYPE_MEMORY_ALLOCATION_COUNT,
&num_alloc_size, &num_allocations);
EXPECT(err == VK_SUCCESS && num_alloc_size == sizeof(num_allocations));
std::vector<VkMemoryRequirements> info =
- get_info<VkMemoryRequirements>(obj_, VK_INFO_TYPE_MEMORY_REQUIREMENTS, 0);
+ get_info<VkMemoryRequirements>(obj_, VK_OBJECT_INFO_TYPE_MEMORY_REQUIREMENTS, 0);
EXPECT(info.size() == num_allocations);
if (info.size() == 1 && !info[0].size)
info.clear();
@@ -241,15 +241,15 @@
EXPECT(vkDestroyObject(obj()) == VK_SUCCESS);
}
-void Object::bind_memory(const Device &dev, uint32_t alloc_idx, const GpuMemory &mem, VkGpuSize mem_offset)
+void Object::bind_memory(const Device &dev, uint32_t alloc_idx, const GpuMemory &mem, VkDeviceSize mem_offset)
{
bound = true;
VkQueue queue = dev.graphics_queues()[0]->obj();
EXPECT(vkQueueBindObjectMemory(queue, obj(), alloc_idx, mem.obj(), mem_offset) == VK_SUCCESS);
}
-void Object::bind_memory(const Device &dev, uint32_t alloc_idx, VkGpuSize offset, VkGpuSize size,
- const GpuMemory &mem, VkGpuSize mem_offset)
+void Object::bind_memory(const Device &dev, uint32_t alloc_idx, VkDeviceSize offset, VkDeviceSize size,
+ const GpuMemory &mem, VkDeviceSize mem_offset)
{
bound = true;
VkQueue queue = dev.graphics_queues()[0]->obj();
@@ -286,7 +286,7 @@
}
}
-void Object::alloc_memory(const Device &dev, const std::vector<VkGpuMemory> &mems)
+void Object::alloc_memory(const Device &dev, const std::vector<VkDeviceMemory> &mems)
{
if (!EXPECT(!internal_mems_) || !mem_alloc_count_)
return;
@@ -305,9 +305,9 @@
}
}
-std::vector<VkGpuMemory> Object::memories() const
+std::vector<VkDeviceMemory> Object::memories() const
{
- std::vector<VkGpuMemory> mems;
+ std::vector<VkDeviceMemory> mems;
if (internal_mems_) {
mems.reserve(mem_alloc_count_);
for (uint32_t i = 0; i < mem_alloc_count_; i++)
@@ -334,7 +334,7 @@
void Device::init(bool enable_layers)
{
// request all queues
- const std::vector<VkPhysicalGpuQueueProperties> queue_props = gpu_.queue_properties();
+ const std::vector<VkPhysicalDeviceQueueProperties> queue_props = gpu_.queue_properties();
std::vector<VkDeviceQueueCreateInfo> queue_info;
queue_info.reserve(queue_props.size());
for (int i = 0; i < queue_props.size(); i++) {
@@ -387,16 +387,16 @@
size_t data_size;
uint32_t queue_node_count;
- err = vkGetGpuInfo(gpu_.obj(), VK_INFO_TYPE_PHYSICAL_GPU_QUEUE_PROPERTIES,
+ err = vkGetPhysicalDeviceInfo(gpu_.obj(), VK_PHYSICAL_DEVICE_INFO_TYPE_QUEUE_PROPERTIES,
&data_size, NULL);
EXPECT(err == VK_SUCCESS);
- queue_node_count = data_size / sizeof(VkPhysicalGpuQueueProperties);
+ queue_node_count = data_size / sizeof(VkPhysicalDeviceQueueProperties);
EXPECT(queue_node_count >= 1);
- VkPhysicalGpuQueueProperties queue_props[queue_node_count];
+ VkPhysicalDeviceQueueProperties queue_props[queue_node_count];
- err = vkGetGpuInfo(gpu_.obj(), VK_INFO_TYPE_PHYSICAL_GPU_QUEUE_PROPERTIES,
+ err = vkGetPhysicalDeviceInfo(gpu_.obj(), VK_PHYSICAL_DEVICE_INFO_TYPE_QUEUE_PROPERTIES,
&data_size, queue_props);
EXPECT(err == VK_SUCCESS);
@@ -427,17 +427,17 @@
void Device::init_formats()
{
- for (int f = VK_FMT_BEGIN_RANGE; f <= VK_FMT_END_RANGE; f++) {
+ for (int f = VK_FORMAT_BEGIN_RANGE; f <= VK_FORMAT_END_RANGE; f++) {
const VkFormat fmt = static_cast<VkFormat>(f);
const VkFormatProperties props = format_properties(fmt);
if (props.linearTilingFeatures) {
- const Format tmp = { fmt, VK_LINEAR_TILING, props.linearTilingFeatures };
+ const Format tmp = { fmt, VK_IMAGE_TILING_LINEAR, props.linearTilingFeatures };
formats_.push_back(tmp);
}
if (props.optimalTilingFeatures) {
- const Format tmp = { fmt, VK_OPTIMAL_TILING, props.optimalTilingFeatures };
+ const Format tmp = { fmt, VK_IMAGE_TILING_OPTIMAL, props.optimalTilingFeatures };
formats_.push_back(tmp);
}
}
@@ -447,7 +447,7 @@
VkFormatProperties Device::format_properties(VkFormat format)
{
- const VkFormatInfoType type = VK_INFO_TYPE_FORMAT_PROPERTIES;
+ const VkFormatInfoType type = VK_FORMAT_INFO_TYPE_PROPERTIES;
VkFormatProperties data;
size_t size = sizeof(data);
if (!EXPECT(vkGetFormatInfo(obj(), format, type, &size, &data) == VK_SUCCESS && size == sizeof(data)))
@@ -497,12 +497,12 @@
submit(cmd, fence);
}
-void Queue::add_mem_references(const std::vector<VkGpuMemory> &mem_refs)
+void Queue::add_mem_references(const std::vector<VkDeviceMemory> &mem_refs)
{
EXPECT(vkQueueAddMemReferences(obj(), mem_refs.size(), &mem_refs[0]) == VK_SUCCESS);
}
-void Queue::remove_mem_references(const std::vector<VkGpuMemory> &mem_refs)
+void Queue::remove_mem_references(const std::vector<VkDeviceMemory> &mem_refs)
{
EXPECT(vkQueueRemoveMemReferences(obj(), mem_refs.size(), &mem_refs[0]) == VK_SUCCESS);
}
@@ -622,7 +622,7 @@
VkResult QueryPool::results(uint32_t start, uint32_t count, size_t size, void *data)
{
size_t tmp = size;
- VkResult err = vkGetQueryPoolResults(obj(), start, count, &tmp, data);
+ VkResult err = vkGetQueryPoolResults(obj(), start, count, &tmp, data, 0);
if (err == VK_SUCCESS) {
if (!EXPECT(tmp == size))
memset(data, 0, size);
@@ -668,12 +668,12 @@
void Image::init(const Device &dev, const VkPeerImageOpenInfo &info, const VkImageCreateInfo &original_info)
{
VkImage img;
- VkGpuMemory mem;
+ VkDeviceMemory mem;
EXPECT(vkOpenPeerImage(dev.obj(), &info, &img, &mem) == VK_SUCCESS);
Object::init(img);
init_info(dev, original_info);
- alloc_memory(dev, std::vector<VkGpuMemory>(1, mem));
+ alloc_memory(dev, std::vector<VkDeviceMemory>(1, mem));
}
void Image::init_info(const Device &dev, const VkImageCreateInfo &info)
@@ -689,7 +689,7 @@
}
void Image::bind_memory(const Device &dev, uint32_t alloc_idx, const VkImageMemoryBindInfo &info,
- const GpuMemory &mem, VkGpuSize mem_offset)
+ const GpuMemory &mem, VkDeviceSize mem_offset)
{
VkQueue queue = dev.graphics_queues()[0]->obj();
EXPECT(!alloc_idx && vkQueueBindImageMemoryRange(queue, obj(), 0, &info, mem.obj(), mem_offset) == VK_SUCCESS);
@@ -697,7 +697,7 @@
VkSubresourceLayout Image::subresource_layout(const VkImageSubresource &subres) const
{
- const VkSubresourceInfoType type = VK_INFO_TYPE_SUBRESOURCE_LAYOUT;
+ const VkSubresourceInfoType type = VK_SUBRESOURCE_INFO_TYPE_LAYOUT;
VkSubresourceLayout data;
size_t size = sizeof(data);
if (!EXPECT(vkGetImageSubresourceInfo(obj(), &subres, type, &size, &data) == VK_SUCCESS && size == sizeof(data)))
@@ -708,7 +708,7 @@
bool Image::transparent() const
{
- return (create_info_.tiling == VK_LINEAR_TILING &&
+ return (create_info_.tiling == VK_IMAGE_TILING_LINEAR &&
create_info_.samples == 1 &&
!(create_info_.usage & (VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT |
VK_IMAGE_USAGE_DEPTH_STENCIL_BIT)));
@@ -924,7 +924,7 @@
graphics_cmd_buf_info.renderPassContinue.renderPass = renderpass_obj;
graphics_cmd_buf_info.renderPassContinue.framebuffer = framebuffer_obj;
- info.flags = VK_CMD_BUFFER_OPTIMIZE_GPU_SMALL_BATCH_BIT |
+ info.flags = VK_CMD_BUFFER_OPTIMIZE_SMALL_BATCH_BIT |
VK_CMD_BUFFER_OPTIMIZE_ONE_TIME_SUBMIT_BIT;
info.sType = VK_STRUCTURE_TYPE_CMD_BUFFER_BEGIN_INFO;
info.pNext = &graphics_cmd_buf_info;
@@ -935,7 +935,7 @@
void CmdBuffer::begin()
{
VkCmdBufferBeginInfo info = {};
- info.flags = VK_CMD_BUFFER_OPTIMIZE_GPU_SMALL_BATCH_BIT |
+ info.flags = VK_CMD_BUFFER_OPTIMIZE_SMALL_BATCH_BIT |
VK_CMD_BUFFER_OPTIMIZE_ONE_TIME_SUBMIT_BIT;
info.sType = VK_STRUCTURE_TYPE_CMD_BUFFER_BEGIN_INFO;
diff --git a/tests/vktestbinding.h b/tests/vktestbinding.h
index 18de306..a8eb41b 100644
--- a/tests/vktestbinding.h
+++ b/tests/vktestbinding.h
@@ -66,15 +66,15 @@
class PhysicalGpu {
public:
- explicit PhysicalGpu(VkPhysicalGpu gpu) : gpu_(gpu) {}
+ explicit PhysicalGpu(VkPhysicalDevice gpu) : gpu_(gpu) {}
- const VkPhysicalGpu &obj() const { return gpu_; }
+ const VkPhysicalDevice &obj() const { return gpu_; }
- // vkGetGpuInfo()
- VkPhysicalGpuProperties properties() const;
- VkPhysicalGpuPerformance performance() const;
- VkPhysicalGpuMemoryProperties memory_properties() const;
- std::vector<VkPhysicalGpuQueueProperties> queue_properties() const;
+ // vkGetPhysicalDeviceInfo()
+ VkPhysicalDeviceProperties properties() const;
+ VkPhysicalDevicePerformance performance() const;
+ VkPhysicalDeviceMemoryProperties memory_properties() const;
+ std::vector<VkPhysicalDeviceQueueProperties> queue_properties() const;
// vkGetProcAddr()
void *get_proc(const char *name) const { return vkGetProcAddr(gpu_, name); }
@@ -85,11 +85,11 @@
// vkEnumerateLayers()
std::vector<const char *> layers(std::vector<char> &buf) const;
- // vkGetMultiGpuCompatibility()
- VkGpuCompatibilityInfo compatibility(const PhysicalGpu &other) const;
+ // vkGetMultiDeviceCompatibility()
+ VkPhysicalDeviceCompatibilityInfo compatibility(const PhysicalGpu &other) const;
private:
- VkPhysicalGpu gpu_;
+ VkPhysicalDevice gpu_;
};
class BaseObject {
@@ -127,18 +127,18 @@
const VkObject &obj() const { return reinterpret_cast<const VkObject &>(BaseObject::obj()); }
// vkQueueBindObjectMemory()
- void bind_memory(const Device &dev, uint32_t alloc_idx, const GpuMemory &mem, VkGpuSize mem_offset);
+ void bind_memory(const Device &dev, uint32_t alloc_idx, const GpuMemory &mem, VkDeviceSize mem_offset);
void unbind_memory(const Device &dev, uint32_t alloc_idx);
void unbind_memory(const Device &dev);
// vkQueueBindObjectMemoryRange()
- void bind_memory(const Device &dev, uint32_t alloc_idx, VkGpuSize offset, VkGpuSize size,
- const GpuMemory &mem, VkGpuSize mem_offset);
+ void bind_memory(const Device &dev, uint32_t alloc_idx, VkDeviceSize offset, VkDeviceSize size,
+ const GpuMemory &mem, VkDeviceSize mem_offset);
// Unless an object is initialized with init_no_mem(), memories are
// automatically allocated and bound. These methods can be used to get
// the memories (for vkQueueAddMemReferences), or to map/unmap the primary memory.
- std::vector<VkGpuMemory> memories() const;
+ std::vector<VkDeviceMemory> memories() const;
const void *map(VkFlags flags) const;
void *map(VkFlags flags);
@@ -161,7 +161,7 @@
// allocate and bind internal memories
void alloc_memory(const Device &dev);
- void alloc_memory(const Device &dev, const std::vector<VkGpuMemory> &mems);
+ void alloc_memory(const Device &dev, const std::vector<VkDeviceMemory> &mems);
private:
void cleanup();
@@ -196,7 +196,7 @@
class Device : public DerivedObject<VkDevice, BaseObject> {
public:
- explicit Device(VkPhysicalGpu gpu) : gpu_(gpu) {}
+ explicit Device(VkPhysicalDevice gpu) : gpu_(gpu) {}
~Device();
// vkCreateDevice()
@@ -261,8 +261,8 @@
// vkQueueAddMemReferences()
// vkQueueRemoveMemReferences()
- void add_mem_references(const std::vector<VkGpuMemory> &mem_refs);
- void remove_mem_references(const std::vector<VkGpuMemory> &mem_refs);
+ void add_mem_references(const std::vector<VkDeviceMemory> &mem_refs);
+ void remove_mem_references(const std::vector<VkDeviceMemory> &mem_refs);
// vkQueueWaitIdle()
void wait();
@@ -273,7 +273,7 @@
void wait_semaphore(Semaphore &sem);
};
-class GpuMemory : public DerivedObject<VkGpuMemory, BaseObject> {
+class GpuMemory : public DerivedObject<VkDeviceMemory, BaseObject> {
public:
~GpuMemory();
@@ -286,7 +286,7 @@
// vkOpenPeerMemory()
void init(const Device &dev, const VkPeerMemoryOpenInfo &info);
- void init(VkGpuMemory mem) { BaseObject::init(mem, false); }
+ void init(VkDeviceMemory mem) { BaseObject::init(mem, false); }
// vkSetMemoryPriority()
void set_priority(VkMemoryPriority priority);
@@ -356,17 +356,17 @@
public:
explicit Buffer() {}
explicit Buffer(const Device &dev, const VkBufferCreateInfo &info) { init(dev, info); }
- explicit Buffer(const Device &dev, VkGpuSize size) { init(dev, size); }
+ explicit Buffer(const Device &dev, VkDeviceSize size) { init(dev, size); }
// vkCreateBuffer()
void init(const Device &dev, const VkBufferCreateInfo &info);
- void init(const Device &dev, VkGpuSize size) { init(dev, create_info(size, 0)); }
+ void init(const Device &dev, VkDeviceSize size) { init(dev, create_info(size, 0)); }
void init_no_mem(const Device &dev, const VkBufferCreateInfo &info);
- static VkBufferCreateInfo create_info(VkGpuSize size, VkFlags usage);
+ static VkBufferCreateInfo create_info(VkDeviceSize size, VkFlags usage);
VkBufferMemoryBarrier buffer_memory_barrier(VkFlags output_mask, VkFlags input_mask,
- VkGpuSize offset, VkGpuSize size) const
+ VkDeviceSize offset, VkDeviceSize size) const
{
VkBufferMemoryBarrier barrier = {};
barrier.sType = VK_STRUCTURE_TYPE_BUFFER_MEMORY_BARRIER;
@@ -400,13 +400,13 @@
// vkQueueBindImageMemoryRange()
void bind_memory(const Device &dev, uint32_t alloc_idx, const VkImageMemoryBindInfo &info,
- const GpuMemory &mem, VkGpuSize mem_offset);
+ const GpuMemory &mem, VkDeviceSize mem_offset);
// vkGetImageSubresourceInfo()
VkSubresourceLayout subresource_layout(const VkImageSubresource &subres) const;
bool transparent() const;
- bool copyable() const { return (format_features_ & VK_FORMAT_SAMPLED_IMAGE_BIT); }
+ bool copyable() const { return (format_features_ & VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT); }
VkImageSubresourceRange subresource_range(VkImageAspect aspect) const { return subresource_range(create_info_, aspect); }
VkExtent3D extent() const { return create_info_.extent; }
@@ -632,7 +632,7 @@
return info;
}
-inline VkBufferCreateInfo Buffer::create_info(VkGpuSize size, VkFlags usage)
+inline VkBufferCreateInfo Buffer::create_info(VkDeviceSize size, VkFlags usage)
{
VkBufferCreateInfo info = {};
info.sType = VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO;
diff --git a/tests/vktestframework.cpp b/tests/vktestframework.cpp
index de59fd4..8f9f360 100644
--- a/tests/vktestframework.cpp
+++ b/tests/vktestframework.cpp
@@ -184,7 +184,7 @@
int x, y;
VkImageObj displayImage(image->device());
- displayImage.init(image->extent().width, image->extent().height, image->format(), 0, VK_LINEAR_TILING);
+ displayImage.init(image->extent().width, image->extent().height, image->format(), 0, VK_IMAGE_TILING_LINEAR);
displayImage.CopyImage(*image);
filename.append(basename);
@@ -197,7 +197,7 @@
size_t data_size = sizeof(sr_layout);
err = vkGetImageSubresourceInfo( image->image(), &sr,
- VK_INFO_TYPE_SUBRESOURCE_LAYOUT,
+ VK_SUBRESOURCE_INFO_TYPE_LAYOUT,
&data_size, &sr_layout);
ASSERT_VK_SUCCESS( err );
ASSERT_EQ(data_size, sizeof(sr_layout));
@@ -218,7 +218,7 @@
const int *row = (const int *) ptr;
int swapped;
- if (displayImage.format() == VK_FMT_B8G8R8A8_UNORM)
+ if (displayImage.format() == VK_FORMAT_B8G8R8A8_UNORM)
{
for (x = 0; x < displayImage.width(); x++) {
swapped = (*row & 0xff00ff00) | (*row & 0x000000ff) << 16 | (*row & 0x00ff0000) >> 16;
@@ -226,7 +226,7 @@
row++;
}
}
- else if (displayImage.format() == VK_FMT_R8G8B8A8_UNORM)
+ else if (displayImage.format() == VK_FORMAT_R8G8B8A8_UNORM)
{
for (x = 0; x < displayImage.width(); x++) {
file.write((char *) row, 3);
@@ -311,7 +311,7 @@
if (!m_show_images) return;
- err = vkGetImageSubresourceInfo( image->image(), &sr, VK_INFO_TYPE_SUBRESOURCE_LAYOUT,
+ err = vkGetImageSubresourceInfo( image->image(), &sr, VK_SUBRESOURCE_INFO_TYPE_LAYOUT,
&data_size, &sr_layout);
ASSERT_VK_SUCCESS( err );
ASSERT_EQ(data_size, sizeof(sr_layout));
@@ -492,7 +492,7 @@
for (int x=0; x < m_images.size(); x++)
{
VK_WSI_X11_PRESENTABLE_IMAGE_CREATE_INFO presentable_image_info = {};
- presentable_image_info.format = VK_FMT_B8G8R8A8_UNORM;
+ presentable_image_info.format = VK_FORMAT_B8G8R8A8_UNORM;
presentable_image_info.usage = VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT;
presentable_image_info.extent.width = m_display_image->m_width;
presentable_image_info.extent.height = m_display_image->m_height;
@@ -505,7 +505,7 @@
assert(!err);
vk_testing::Buffer buf;
- buf.init(m_device, (VkGpuSize) m_display_image->m_data_size);
+ buf.init(m_device, (VkDeviceSize) m_display_image->m_data_size);
dest_ptr = buf.map();
memcpy(dest_ptr,m_display_image->m_data, m_display_image->m_data_size);
buf.unmap();
@@ -1081,7 +1081,7 @@
//
// Convert VK shader type to compiler's
//
-EShLanguage VkTestFramework::FindLanguage(const VkPipelineShaderStage shader_type)
+EShLanguage VkTestFramework::FindLanguage(const VkShaderStage shader_type)
{
switch (shader_type) {
case VK_SHADER_STAGE_VERTEX:
@@ -1112,7 +1112,7 @@
// Compile a given string containing GLSL into SPV for use by VK
// Return value of false means an error was encountered.
//
-bool VkTestFramework::GLSLtoSPV(const VkPipelineShaderStage shader_type,
+bool VkTestFramework::GLSLtoSPV(const VkShaderStage shader_type,
const char *pshader,
std::vector<unsigned int> &spv)
{
diff --git a/tests/vktestframework.h b/tests/vktestframework.h
index 76f02a1..ba53923 100644
--- a/tests/vktestframework.h
+++ b/tests/vktestframework.h
@@ -63,7 +63,7 @@
int m_height;
void *m_data;
VkImage m_presentableImage;
- VkGpuMemory m_presentableMemory;
+ VkDeviceMemory m_presentableMemory;
unsigned m_data_size;
};
@@ -81,7 +81,7 @@
void Compare(const char *comment, VkImageObj *image);
void RecordImage(VkImageObj * image);
void RecordImages(vector<VkImageObj *> image);
- bool GLSLtoSPV(const VkPipelineShaderStage shader_type,
+ bool GLSLtoSPV(const VkShaderStage shader_type,
const char *pshader,
std::vector<unsigned int> &spv);
static bool m_use_spv;
@@ -96,7 +96,7 @@
void SetMessageOptions(EShMessages& messages);
void ProcessConfigFile();
EShLanguage FindLanguage(const std::string& name);
- EShLanguage FindLanguage(const VkPipelineShaderStage shader_type);
+ EShLanguage FindLanguage(const VkShaderStage shader_type);
std::string ConfigFile;
bool SetConfigFile(const std::string& name);
diff --git a/vk-generate.py b/vk-generate.py
index 1147620..39af406 100755
--- a/vk-generate.py
+++ b/vk-generate.py
@@ -233,7 +233,7 @@
% self.prefix)
func.append("%s PFN_vkGetProcAddr gpa,"
% (" " * len(self.prefix)))
- func.append("%s VkPhysicalGpu gpu)"
+ func.append("%s VkPhysicalDevice gpu)"
% (" " * len(self.prefix)))
func.append("{")
func.append(" %s" % "\n ".join(stmts))
@@ -392,7 +392,7 @@
library_exports = {
"all": [],
"icd": [
- "EnumerateGpus",
+ "EnumeratePhysicalDevices",
"CreateInstance",
"DestroyInstance",
"GetProcAddr",
@@ -488,7 +488,7 @@
special_lookups = []
# these functions require special trampoline code beyond just the normal create object trampoline code
- special_names = ["AllocDescriptorSets", "GetMultiGpuCompatibility"]
+ special_names = ["AllocDescriptorSets", "GetMultiDeviceCompatibility"]
for proto in self.protos:
if self._is_loader_special_case(proto) or self._does_function_create_object(proto) or proto.name in special_names:
special_lookups.append("if (!strcmp(name, \"%s\"))" % proto.name)
diff --git a/vk-layer-generate.py b/vk-layer-generate.py
index 44c0b36..749e6d5 100755
--- a/vk-layer-generate.py
+++ b/vk-layer-generate.py
@@ -151,7 +151,7 @@
r_body.append(' // This layer intercepts callbacks')
r_body.append(' VK_LAYER_DBG_FUNCTION_NODE *pNewDbgFuncNode = (VK_LAYER_DBG_FUNCTION_NODE*)malloc(sizeof(VK_LAYER_DBG_FUNCTION_NODE));')
r_body.append(' if (!pNewDbgFuncNode)')
- r_body.append(' return VK_ERROR_OUT_OF_MEMORY;')
+ r_body.append(' return VK_ERROR_OUT_OF_HOST_MEMORY;')
r_body.append(' pNewDbgFuncNode->pfnMsgCallback = pfnMsgCallback;')
r_body.append(' pNewDbgFuncNode->pUserData = pUserData;')
r_body.append(' pNewDbgFuncNode->pNext = g_pDbgFunctionHead;')
@@ -258,6 +258,28 @@
ggei_body.append('}')
return "\n".join(ggei_body)
+ def _gen_layer_get_extension_support(self, layer="Generic"):
+ ges_body = []
+ ges_body.append('VK_LAYER_EXPORT VkResult VKAPI xglGetExtensionSupport(VkPhysicalDevice gpu, const char* pExtName)')
+ ges_body.append('{')
+ ges_body.append(' VkResult result;')
+ ges_body.append(' VkBaseLayerObject* gpuw = (VkBaseLayerObject *) gpu;')
+ ges_body.append('')
+ ges_body.append(' /* This entrypoint is NOT going to init its own dispatch table since loader calls here early */')
+ ges_body.append(' if (!strncmp(pExtName, "%s", strlen("%s")))' % (layer, layer))
+ ges_body.append(' {')
+ ges_body.append(' result = VK_SUCCESS;')
+ ges_body.append(' } else if (nextTable.GetExtensionSupport != NULL)')
+ ges_body.append(' {')
+ ges_body.append(' result = nextTable.GetExtensionSupport((VkPhysicalDevice)gpuw->nextObject, pExtName);')
+ ges_body.append(' } else')
+ ges_body.append(' {')
+ ges_body.append(' result = VK_ERROR_INVALID_EXTENSION;')
+ ges_body.append(' }')
+ ges_body.append(' return result;')
+ ges_body.append('}')
+ return "\n".join(ges_body)
+
def _generate_dispatch_entrypoints(self, qual=""):
if qual:
qual += " "
@@ -354,7 +376,7 @@
def _generate_layer_gpa_function(self, extensions=[]):
func_body = []
- func_body.append("VK_LAYER_EXPORT void* VKAPI vkGetProcAddr(VkPhysicalGpu gpu, const char* funcName)\n"
+ func_body.append("VK_LAYER_EXPORT void* VKAPI vkGetProcAddr(VkPhysicalDevice gpu, const char* funcName)\n"
"{\n"
" VkBaseLayerObject* gpuw = (VkBaseLayerObject *) gpu;\n"
" void* addr;\n"
@@ -373,7 +395,7 @@
func_body.append(" else {\n"
" if (gpuw->pGPA == NULL)\n"
" return NULL;\n"
- " return gpuw->pGPA((VkPhysicalGpu)gpuw->nextObject, funcName);\n"
+ " return gpuw->pGPA((VkPhysicalDevice)gpuw->nextObject, funcName);\n"
" }\n"
"}\n")
return "\n".join(func_body)
@@ -403,7 +425,7 @@
' fpNextGPA = pCurObj->pGPA;\n'
' assert(fpNextGPA);\n')
- func_body.append(" layer_initialize_dispatch_table(&nextTable, fpNextGPA, (VkPhysicalGpu) pCurObj->nextObject);")
+ func_body.append(" layer_initialize_dispatch_table(&nextTable, fpNextGPA, (VkPhysicalDevice) pCurObj->nextObject);")
if lockname is not None:
func_body.append(" if (!%sLockInitialized)" % lockname)
func_body.append(" {")
@@ -422,7 +444,7 @@
' fpNextGPA = pCurObj->pGPA;\n'
' assert(fpNextGPA);\n' % self.layer_name)
- func_body.append(" layer_initialize_dispatch_table(&nextTable, fpNextGPA, (VkPhysicalGpu) pCurObj->nextObject);\n")
+ func_body.append(" layer_initialize_dispatch_table(&nextTable, fpNextGPA, (VkPhysicalDevice) pCurObj->nextObject);\n")
func_body.append(" if (!printLockInitialized)")
func_body.append(" {")
func_body.append(" // TODO/TBD: Need to delete this mutex sometime. How???")
@@ -624,7 +646,7 @@
func_body.append(' PFN_vkGetProcAddr fpNextGPA;')
func_body.append(' fpNextGPA = pCurObj->pGPA;')
func_body.append(' assert(fpNextGPA);')
- func_body.append(' layer_initialize_dispatch_table(&nextTable, fpNextGPA, (VkPhysicalGpu) pCurObj->nextObject);')
+ func_body.append(' layer_initialize_dispatch_table(&nextTable, fpNextGPA, (VkPhysicalDevice) pCurObj->nextObject);')
func_body.append('')
func_body.append(' if (!printLockInitialized)')
func_body.append(' {')
@@ -732,7 +754,7 @@
if 'WsiX11AssociateConnection' == proto.name:
funcs.append("#if defined(__linux__) || defined(XCB_NVIDIA)")
if proto.name == "EnumerateLayers":
- c_call = proto.c_call().replace("(" + proto.params[0].name, "((VkPhysicalGpu)gpuw->nextObject", 1)
+ c_call = proto.c_call().replace("(" + proto.params[0].name, "((VkPhysicalDevice)gpuw->nextObject", 1)
funcs.append('%s%s\n'
'{\n'
' using namespace StreamControl;\n'
@@ -751,6 +773,36 @@
' return VK_SUCCESS;\n'
' }\n'
'}' % (qual, decl, self.layer_name, ret_val, proto.c_call(),f_open, log_func, f_close, stmt, self.layer_name))
+ elif 'GetExtensionSupport' == proto.name:
+ funcs.append('%s%s\n'
+ '{\n'
+ ' VkResult result;\n'
+ ' /* This entrypoint is NOT going to init its own dispatch table since loader calls here early */\n'
+ ' if (!strncmp(pExtName, "%s", strlen("%s")))\n'
+ ' {\n'
+ ' result = VK_SUCCESS;\n'
+ ' } else if (nextTable.GetExtensionSupport != NULL)\n'
+ ' {\n'
+ ' result = nextTable.%s;\n'
+ ' %s %s %s\n'
+ ' } else\n'
+ ' {\n'
+ ' result = VK_ERROR_INVALID_EXTENSION;\n'
+ ' }\n'
+ '%s'
+ '}' % (qual, decl, self.layer_name, self.layer_name, proto.c_call(), f_open, log_func, f_close, stmt))
+# elif 'vkphysicalgpu' == proto.params[0].ty.lower():
+# c_call = proto.c_call().replace("(" + proto.params[0].name, "((VkPhysicalDevice)gpuw->nextObject", 1)
+# funcs.append('%s%s\n'
+# '{\n'
+# ' using namespace StreamControl;\n'
+# ' VkBaseLayerObject* gpuw = (VkBaseLayerObject *) %s;\n'
+# ' pCurObj = gpuw;\n'
+# ' loader_platform_thread_once(&tabOnce, init%s);\n'
+# ' %snextTable.%s;\n'
+# ' %s%s%s\n'
+# '%s'
+# '}' % (qual, decl, proto.params[0].name, self.layer_name, ret_val, c_call, f_open, log_func, f_close, stmt))
else:
funcs.append('%s%s\n'
'{\n'
@@ -801,7 +853,7 @@
header_txt.append('')
header_txt.append('// For each Queue\'s doubly linked-list of mem refs')
header_txt.append('typedef struct _OT_MEM_INFO {')
- header_txt.append(' VkGpuMemory mem;')
+ header_txt.append(' VkDeviceMemory mem;')
header_txt.append(' struct _OT_MEM_INFO *pNextMI;')
header_txt.append(' struct _OT_MEM_INFO *pPrevMI;')
header_txt.append('')
@@ -831,7 +883,7 @@
header_txt.append(' }')
header_txt.append(' else {')
header_txt.append(' char str[1024];')
- header_txt.append(' sprintf(str, "ERROR: VK_ERROR_OUT_OF_MEMORY -- could not allocate memory for Queue Information");')
+ header_txt.append(' sprintf(str, "ERROR: VK_ERROR_OUT_OF_HOST_MEMORY -- could not allocate memory for Queue Information");')
header_txt.append(' layerCbMsg(VK_DBG_MSG_ERROR, VK_VALIDATION_LEVEL_0, queue, 0, OBJTRACK_INTERNAL_ERROR, "OBJTRACK", str);')
header_txt.append(' }')
header_txt.append('}')
@@ -1017,13 +1069,13 @@
header_txt.append('')
header_txt.append(' while (pTrav) {')
header_txt.append(' if (pTrav->obj.pObj == pObj) {')
- header_txt.append(' if (stateBindPoint == VK_STATE_BIND_VIEWPORT) {')
+ header_txt.append(' if (stateBindPoint == VK_STATE_BIND_POINT_VIEWPORT) {')
header_txt.append(' pTrav->obj.status |= OBJSTATUS_VIEWPORT_BOUND;')
- header_txt.append(' } else if (stateBindPoint == VK_STATE_BIND_RASTER) {')
+ header_txt.append(' } else if (stateBindPoint == VK_STATE_BIND_POINT_RASTER) {')
header_txt.append(' pTrav->obj.status |= OBJSTATUS_RASTER_BOUND;')
- header_txt.append(' } else if (stateBindPoint == VK_STATE_BIND_COLOR_BLEND) {')
+ header_txt.append(' } else if (stateBindPoint == VK_STATE_BIND_POINT_COLOR_BLEND) {')
header_txt.append(' pTrav->obj.status |= OBJSTATUS_COLOR_BLEND_BOUND;')
- header_txt.append(' } else if (stateBindPoint == VK_STATE_BIND_DEPTH_STENCIL) {')
+ header_txt.append(' } else if (stateBindPoint == VK_STATE_BIND_POINT_DEPTH_STENCIL) {')
header_txt.append(' pTrav->obj.status |= OBJSTATUS_DEPTH_STENCIL_BOUND;')
header_txt.append(' }')
header_txt.append(' return;')
@@ -1084,7 +1136,7 @@
header_txt.append('}')
header_txt.append('')
header_txt.append('static void setGpuQueueInfoState(void *pData) {')
- header_txt.append(' maxMemReferences = ((VkPhysicalGpuQueueProperties *)pData)->maxMemReferences;')
+ header_txt.append(' maxMemReferences = ((VkPhysicalDeviceQueueProperties *)pData)->maxMemReferences;')
header_txt.append('}')
return "\n".join(header_txt)
@@ -1128,9 +1180,9 @@
elif 'CmdDraw' in proto.name:
using_line += ' validate_draw_state_flags((void *)cmdBuffer);\n'
elif 'MapMemory' in proto.name:
- using_line += ' set_status((void*)mem, VkObjectTypeGpuMemory, OBJSTATUS_GPU_MEM_MAPPED);\n'
+ using_line += ' set_status((void*)mem, VkObjectTypeDeviceMemory, OBJSTATUS_GPU_MEM_MAPPED);\n'
elif 'UnmapMemory' in proto.name:
- using_line += ' reset_status((void*)mem, VkObjectTypeGpuMemory, OBJSTATUS_GPU_MEM_MAPPED);\n'
+ using_line += ' reset_status((void*)mem, VkObjectTypeDeviceMemory, OBJSTATUS_GPU_MEM_MAPPED);\n'
if 'AllocDescriptor' in proto.name: # Allocates array of DSs
create_line = ' for (uint32_t i = 0; i < *pCount; i++) {\n'
create_line += ' loader_platform_thread_lock_mutex(&objLock);\n'
@@ -1201,8 +1253,28 @@
' return VK_SUCCESS;\n'
' }\n'
'}' % (qual, decl, using_line, self.layer_name, ret_val, proto.c_call(), create_line, destroy_line, stmt, self.layer_name))
- elif 'GetGpuInfo' in proto.name:
- gpu_state = ' if (infoType == VK_INFO_TYPE_PHYSICAL_GPU_QUEUE_PROPERTIES) {\n'
+ elif 'GetExtensionSupport' == proto.name:
+ funcs.append('%s%s\n'
+ '{\n'
+ ' VkResult result;\n'
+ ' /* This entrypoint is NOT going to init its own dispatch table since loader calls this early */\n'
+ ' if (!strncmp(pExtName, "%s", strlen("%s")) ||\n'
+ ' !strncmp(pExtName, "objTrackGetObjectCount", strlen("objTrackGetObjectCount")) ||\n'
+ ' !strncmp(pExtName, "objTrackGetObjects", strlen("objTrackGetObjects")))\n'
+ ' {\n'
+ ' result = VK_SUCCESS;\n'
+ ' } else if (nextTable.GetExtensionSupport != NULL)\n'
+ ' {\n'
+ ' %s'
+ ' result = nextTable.%s;\n'
+ ' } else\n'
+ ' {\n'
+ ' result = VK_ERROR_INVALID_EXTENSION;\n'
+ ' }\n'
+ '%s'
+ '}' % (qual, decl, self.layer_name, self.layer_name, using_line, proto.c_call(), stmt))
+ elif 'GetPhysicalDeviceInfo' in proto.name:
+ gpu_state = ' if (infoType == VK_PHYSICAL_DEVICE_INFO_TYPE_QUEUE_PROPERTIES) {\n'
gpu_state += ' if (pData != NULL) {\n'
gpu_state += ' setGpuQueueInfoState(pData);\n'
gpu_state += ' }\n'
@@ -1335,7 +1407,7 @@
# Only watch core objects passed as first parameter
elif proto.params[0].ty not in vulkan.core.objects:
return None
- elif proto.params[0].ty != "VkPhysicalGpu":
+ elif proto.params[0].ty != "VkPhysicalDevice":
funcs.append('%s%s\n'
'{\n'
' useObject((VkObject) %s, "%s");\n'
diff --git a/vulkan.py b/vulkan.py
index 3e49704..18b6793 100755
--- a/vulkan.py
+++ b/vulkan.py
@@ -185,11 +185,11 @@
objects=[
"VkInstance",
- "VkPhysicalGpu",
+ "VkPhysicalDevice",
"VkBaseObject",
"VkDevice",
"VkQueue",
- "VkGpuMemory",
+ "VkDeviceMemory",
"VkObject",
"VkBuffer",
"VkBufferView",
@@ -228,20 +228,20 @@
Proto("VkResult", "EnumeratePhysicalDevices",
[Param("VkInstance", "instance"),
Param("uint32_t*", "pPhysicalDeviceCount"),
- Param("VkPhysicalGpu*", "pPhysicalDevices")]),
+ Param("VkPhysicalDevice*", "pPhysicalDevices")]),
- Proto("VkResult", "GetGpuInfo",
- [Param("VkPhysicalGpu", "gpu"),
- Param("VkPhysicalGpuInfoType", "infoType"),
+ Proto("VkResult", "GetPhysicalDeviceInfo",
+ [Param("VkPhysicalDevice", "gpu"),
+ Param("VkPhysicalDeviceInfoType", "infoType"),
Param("size_t*", "pDataSize"),
Param("void*", "pData")]),
Proto("void*", "GetProcAddr",
- [Param("VkPhysicalGpu", "gpu"),
+ [Param("VkPhysicalDevice", "gpu"),
Param("const char*", "pName")]),
Proto("VkResult", "CreateDevice",
- [Param("VkPhysicalGpu", "gpu"),
+ [Param("VkPhysicalDevice", "gpu"),
Param("const VkDeviceCreateInfo*", "pCreateInfo"),
Param("VkDevice*", "pDevice")]),
@@ -255,14 +255,14 @@
Param("void*", "pData")]),
Proto("VkResult", "GetPhysicalDeviceExtensionInfo",
- [Param("VkPhysicalGpu", "gpu"),
+ [Param("VkPhysicalDevice", "gpu"),
Param("VkExtensionInfoType", "infoType"),
Param("uint32_t", "extensionIndex"),
Param("size_t*", "pDataSize"),
Param("void*", "pData")]),
Proto("VkResult", "EnumerateLayers",
- [Param("VkPhysicalGpu", "gpu"),
+ [Param("VkPhysicalDevice", "gpu"),
Param("size_t", "maxLayerCount"),
Param("size_t", "maxStringSize"),
Param("size_t*", "pOutLayerCount"),
@@ -284,12 +284,12 @@
Proto("VkResult", "QueueAddMemReferences",
[Param("VkQueue", "queue"),
Param("uint32_t", "count"),
- Param("const VkGpuMemory*", "pMems")]),
+ Param("const VkDeviceMemory*", "pMems")]),
Proto("VkResult", "QueueRemoveMemReferences",
[Param("VkQueue", "queue"),
Param("uint32_t", "count"),
- Param("const VkGpuMemory*", "pMems")]),
+ Param("const VkDeviceMemory*", "pMems")]),
Proto("VkResult", "QueueWaitIdle",
[Param("VkQueue", "queue")]),
@@ -300,38 +300,38 @@
Proto("VkResult", "AllocMemory",
[Param("VkDevice", "device"),
Param("const VkMemoryAllocInfo*", "pAllocInfo"),
- Param("VkGpuMemory*", "pMem")]),
+ Param("VkDeviceMemory*", "pMem")]),
Proto("VkResult", "FreeMemory",
- [Param("VkGpuMemory", "mem")]),
+ [Param("VkDeviceMemory", "mem")]),
Proto("VkResult", "SetMemoryPriority",
- [Param("VkGpuMemory", "mem"),
+ [Param("VkDeviceMemory", "mem"),
Param("VkMemoryPriority", "priority")]),
Proto("VkResult", "MapMemory",
- [Param("VkGpuMemory", "mem"),
+ [Param("VkDeviceMemory", "mem"),
Param("VkFlags", "flags"),
Param("void**", "ppData")]),
Proto("VkResult", "UnmapMemory",
- [Param("VkGpuMemory", "mem")]),
+ [Param("VkDeviceMemory", "mem")]),
Proto("VkResult", "PinSystemMemory",
[Param("VkDevice", "device"),
Param("const void*", "pSysMem"),
Param("size_t", "memSize"),
- Param("VkGpuMemory*", "pMem")]),
+ Param("VkDeviceMemory*", "pMem")]),
- Proto("VkResult", "GetMultiGpuCompatibility",
- [Param("VkPhysicalGpu", "gpu0"),
- Param("VkPhysicalGpu", "gpu1"),
- Param("VkGpuCompatibilityInfo*", "pInfo")]),
+ Proto("VkResult", "GetMultiDeviceCompatibility",
+ [Param("VkPhysicalDevice", "gpu0"),
+ Param("VkPhysicalDevice", "gpu1"),
+ Param("VkPhysicalDeviceCompatibilityInfo*", "pInfo")]),
Proto("VkResult", "OpenSharedMemory",
[Param("VkDevice", "device"),
Param("const VkMemoryOpenInfo*", "pOpenInfo"),
- Param("VkGpuMemory*", "pMem")]),
+ Param("VkDeviceMemory*", "pMem")]),
Proto("VkResult", "OpenSharedSemaphore",
[Param("VkDevice", "device"),
@@ -341,13 +341,13 @@
Proto("VkResult", "OpenPeerMemory",
[Param("VkDevice", "device"),
Param("const VkPeerMemoryOpenInfo*", "pOpenInfo"),
- Param("VkGpuMemory*", "pMem")]),
+ Param("VkDeviceMemory*", "pMem")]),
Proto("VkResult", "OpenPeerImage",
[Param("VkDevice", "device"),
Param("const VkPeerImageOpenInfo*", "pOpenInfo"),
Param("VkImage*", "pImage"),
- Param("VkGpuMemory*", "pMem")]),
+ Param("VkDeviceMemory*", "pMem")]),
Proto("VkResult", "DestroyObject",
[Param("VkObject", "object")]),
@@ -362,25 +362,25 @@
[Param("VkQueue", "queue"),
Param("VkObject", "object"),
Param("uint32_t", "allocationIdx"),
- Param("VkGpuMemory", "mem"),
- Param("VkGpuSize", "offset")]),
+ Param("VkDeviceMemory", "mem"),
+ Param("VkDeviceSize", "offset")]),
Proto("VkResult", "QueueBindObjectMemoryRange",
[Param("VkQueue", "queue"),
Param("VkObject", "object"),
Param("uint32_t", "allocationIdx"),
- Param("VkGpuSize", "rangeOffset"),
- Param("VkGpuSize", "rangeSize"),
- Param("VkGpuMemory", "mem"),
- Param("VkGpuSize", "memOffset")]),
+ Param("VkDeviceSize", "rangeOffset"),
+ Param("VkDeviceSize", "rangeSize"),
+ Param("VkDeviceMemory", "mem"),
+ Param("VkDeviceSize", "memOffset")]),
Proto("VkResult", "QueueBindImageMemoryRange",
[Param("VkQueue", "queue"),
Param("VkImage", "image"),
Param("uint32_t", "allocationIdx"),
Param("const VkImageMemoryBindInfo*", "pBindInfo"),
- Param("VkGpuMemory", "mem"),
- Param("VkGpuSize", "memOffset")]),
+ Param("VkDeviceMemory", "mem"),
+ Param("VkDeviceSize", "memOffset")]),
Proto("VkResult", "CreateFence",
[Param("VkDevice", "device"),
@@ -439,7 +439,8 @@
Param("uint32_t", "startQuery"),
Param("uint32_t", "queryCount"),
Param("size_t*", "pDataSize"),
- Param("void*", "pData")]),
+ Param("void*", "pData"),
+ Param("VkQueryResultFlags", "flags")]),
Proto("VkResult", "GetFormatInfo",
[Param("VkDevice", "device"),
@@ -635,12 +636,13 @@
Param("uint32_t", "startBinding"),
Param("uint32_t", "bindingCount"),
Param("const VkBuffer*", "pBuffers"),
- Param("const VkGpuSize*", "pOffsets")]),
+ Param("const VkDeviceSize*", "pOffsets")]),
+
Proto("void", "CmdBindIndexBuffer",
[Param("VkCmdBuffer", "cmdBuffer"),
Param("VkBuffer", "buffer"),
- Param("VkGpuSize", "offset"),
+ Param("VkDeviceSize", "offset"),
Param("VkIndexType", "indexType")]),
Proto("void", "CmdDraw",
@@ -661,14 +663,14 @@
Proto("void", "CmdDrawIndirect",
[Param("VkCmdBuffer", "cmdBuffer"),
Param("VkBuffer", "buffer"),
- Param("VkGpuSize", "offset"),
+ Param("VkDeviceSize", "offset"),
Param("uint32_t", "count"),
Param("uint32_t", "stride")]),
Proto("void", "CmdDrawIndexedIndirect",
[Param("VkCmdBuffer", "cmdBuffer"),
Param("VkBuffer", "buffer"),
- Param("VkGpuSize", "offset"),
+ Param("VkDeviceSize", "offset"),
Param("uint32_t", "count"),
Param("uint32_t", "stride")]),
@@ -681,7 +683,7 @@
Proto("void", "CmdDispatchIndirect",
[Param("VkCmdBuffer", "cmdBuffer"),
Param("VkBuffer", "buffer"),
- Param("VkGpuSize", "offset")]),
+ Param("VkDeviceSize", "offset")]),
Proto("void", "CmdCopyBuffer",
[Param("VkCmdBuffer", "cmdBuffer"),
@@ -734,15 +736,15 @@
Proto("void", "CmdUpdateBuffer",
[Param("VkCmdBuffer", "cmdBuffer"),
Param("VkBuffer", "destBuffer"),
- Param("VkGpuSize", "destOffset"),
- Param("VkGpuSize", "dataSize"),
+ Param("VkDeviceSize", "destOffset"),
+ Param("VkDeviceSize", "dataSize"),
Param("const uint32_t*", "pData")]),
Proto("void", "CmdFillBuffer",
[Param("VkCmdBuffer", "cmdBuffer"),
Param("VkBuffer", "destBuffer"),
- Param("VkGpuSize", "destOffset"),
- Param("VkGpuSize", "fillSize"),
+ Param("VkDeviceSize", "destOffset"),
+ Param("VkDeviceSize", "fillSize"),
Param("uint32_t", "data")]),
Proto("void", "CmdClearColorImage",
@@ -783,11 +785,19 @@
Proto("void", "CmdWaitEvents",
[Param("VkCmdBuffer", "cmdBuffer"),
- Param("const VkEventWaitInfo*", "pWaitInfo")]),
+ Param("VkWaitEvent", "waitEvent"),
+ Param("uint32_t", "eventCount"),
+ Param("const VkEvent*", "pEvents"),
+ Param("uint32_t", "memBarrierCount"),
+ Param("const void**", "ppMemBarriers")]),
Proto("void", "CmdPipelineBarrier",
[Param("VkCmdBuffer", "cmdBuffer"),
- Param("const VkPipelineBarrier*", "pBarrier")]),
+ Param("VkWaitEvent", "waitEvent"),
+ Param("uint32_t", "pipeEventCount"),
+ Param("const VkPipeEvent*", "pPipeEvents"),
+ Param("uint32_t", "memBarrierCount"),
+ Param("const void**", "ppMemBarriers")]),
Proto("void", "CmdBeginQuery",
[Param("VkCmdBuffer", "cmdBuffer"),
@@ -810,7 +820,7 @@
[Param("VkCmdBuffer", "cmdBuffer"),
Param("VkTimestampType", "timestampType"),
Param("VkBuffer", "destBuffer"),
- Param("VkGpuSize", "destOffset")]),
+ Param("VkDeviceSize", "destOffset")]),
Proto("void", "CmdCopyQueryPoolResults",
[Param("VkCmdBuffer", "cmdBuffer"),
@@ -818,8 +828,8 @@
Param("uint32_t", "startQuery"),
Param("uint32_t", "queryCount"),
Param("VkBuffer", "destBuffer"),
- Param("VkGpuSize", "destOffset"),
- Param("VkGpuSize", "destStride"),
+ Param("VkDeviceSize", "destOffset"),
+ Param("VkDeviceSize", "destStride"),
Param("VkFlags", "flags")]),
Proto("void", "CmdInitAtomicCounters",
@@ -835,7 +845,7 @@
Param("uint32_t", "startCounter"),
Param("uint32_t", "counterCount"),
Param("VkBuffer", "srcBuffer"),
- Param("VkGpuSize", "srcOffset")]),
+ Param("VkDeviceSize", "srcOffset")]),
Proto("void", "CmdSaveAtomicCounters",
[Param("VkCmdBuffer", "cmdBuffer"),
@@ -843,7 +853,7 @@
Param("uint32_t", "startCounter"),
Param("uint32_t", "counterCount"),
Param("VkBuffer", "destBuffer"),
- Param("VkGpuSize", "destOffset")]),
+ Param("VkDeviceSize", "destOffset")]),
Proto("VkResult", "CreateFramebuffer",
[Param("VkDevice", "device"),
@@ -913,7 +923,7 @@
objects=[],
protos=[
Proto("VkResult", "WsiX11AssociateConnection",
- [Param("VkPhysicalGpu", "gpu"),
+ [Param("VkPhysicalDevice", "gpu"),
Param("const VK_WSI_X11_CONNECTION_INFO*", "pConnectionInfo")]),
Proto("VkResult", "WsiX11GetMSC",
@@ -926,7 +936,7 @@
[Param("VkDevice", "device"),
Param("const VK_WSI_X11_PRESENTABLE_IMAGE_CREATE_INFO*", "pCreateInfo"),
Param("VkImage*", "pImage"),
- Param("VkGpuMemory*", "pMem")]),
+ Param("VkDeviceMemory*", "pMem")]),
Proto("VkResult", "WsiX11QueuePresent",
[Param("VkQueue", "queue"),
@@ -939,14 +949,14 @@
object_root_list = [
"VkInstance",
- "VkPhysicalGpu",
+ "VkPhysicalDevice",
"VkBaseObject"
]
object_base_list = [
"VkDevice",
"VkQueue",
- "VkGpuMemory",
+ "VkDeviceMemory",
"VkObject"
]