/* * * Copyright (c) 2015-2016 The Khronos Group Inc. * Copyright (c) 2015-2016 Valve Corporation * Copyright (c) 2015-2016 LunarG, Inc. * Copyright (C) 2015 Google Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. * * Author: Courtney Goeltzenleuchter * Author: Jon Ashburn * Author: Tony Barbour * Author: Chia-I Wu */ #ifndef _GNU_SOURCE #define _GNU_SOURCE #endif #include #include #include "vk_loader_platform.h" #include "loader.h" #include "debug_utils.h" #include "wsi.h" #include "vk_loader_extensions.h" #include "gpa_helper.h" // Trampoline entrypoints are in this file for core Vulkan commands // Get an instance level or global level entry point address. // @param instance // @param pName // @return // If instance == NULL returns a global level functions only // If instance is valid returns a trampoline entry point for all dispatchable Vulkan // functions both core and extensions. LOADER_EXPORT VKAPI_ATTR PFN_vkVoidFunction VKAPI_CALL vkGetInstanceProcAddr(VkInstance instance, const char *pName) { void *addr; addr = globalGetProcAddr(pName); if (instance == VK_NULL_HANDLE || addr != NULL) { return addr; } struct loader_instance *ptr_instance = loader_get_instance(instance); if (ptr_instance == NULL) return NULL; // Return trampoline code for non-global entrypoints including any extensions. // Device extensions are returned if a layer or ICD supports the extension. // Instance extensions are returned if the extension is enabled and the // loader or someone else supports the extension return trampolineGetProcAddr(ptr_instance, pName); } // Get a device level or global level entry point address. // @param device // @param pName // @return // If device is valid, returns a device relative entry point for device level // entry points both core and extensions. // Device relative means call down the device chain. LOADER_EXPORT VKAPI_ATTR PFN_vkVoidFunction VKAPI_CALL vkGetDeviceProcAddr(VkDevice device, const char *pName) { void *addr; // For entrypoints that loader must handle (ie non-dispatchable or create object) // make sure the loader entrypoint is returned addr = loader_non_passthrough_gdpa(pName); if (addr) { return addr; } // Although CreateDevice is on device chain it's dispatchable object isn't // a VkDevice or child of VkDevice so return NULL. if (!strcmp(pName, "CreateDevice")) return NULL; // Return the dispatch table entrypoint for the fastest case const VkLayerDispatchTable *disp_table = *(VkLayerDispatchTable **)device; if (disp_table == NULL) return NULL; addr = loader_lookup_device_dispatch_table(disp_table, pName); if (addr) return addr; if (disp_table->GetDeviceProcAddr == NULL) return NULL; return disp_table->GetDeviceProcAddr(device, pName); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkEnumerateInstanceExtensionProperties(const char *pLayerName, uint32_t *pPropertyCount, VkExtensionProperties *pProperties) { tls_instance = NULL; LOADER_PLATFORM_THREAD_ONCE(&once_init, loader_initialize); // We know we need to call at least the terminator VkResult res = VK_SUCCESS; VkEnumerateInstanceExtensionPropertiesChain chain_tail = { .header = { .type = VK_CHAIN_TYPE_ENUMERATE_INSTANCE_EXTENSION_PROPERTIES, .version = VK_CURRENT_CHAIN_VERSION, .size = sizeof(chain_tail), }, .pfnNextLayer = &terminator_EnumerateInstanceExtensionProperties, .pNextLink = NULL, }; VkEnumerateInstanceExtensionPropertiesChain *chain_head = &chain_tail; // Get the implicit layers struct loader_layer_list layers; memset(&layers, 0, sizeof(layers)); loaderScanForImplicitLayers(NULL, &layers); // We'll need to save the dl handles so we can close them later loader_platform_dl_handle *libs = malloc(sizeof(loader_platform_dl_handle) * layers.count); if (libs == NULL) { return VK_ERROR_OUT_OF_HOST_MEMORY; } size_t lib_count = 0; // Prepend layers onto the chain if they implment this entry point for (uint32_t i = 0; i < layers.count; ++i) { if (!loaderImplicitLayerIsEnabled(NULL, layers.list + i) || layers.list[i].pre_instance_functions.enumerate_instance_extension_properties[0] == '\0') { continue; } loader_platform_dl_handle layer_lib = loader_platform_open_library(layers.list[i].lib_name); if (layer_lib == NULL) { loader_log(NULL, VK_DEBUG_REPORT_WARNING_BIT_EXT, 0, "%s: Unable to load implicit layer library \"%s\"", __FUNCTION__, layers.list[i].lib_name); continue; } libs[lib_count++] = layer_lib; void *pfn = loader_platform_get_proc_address(layer_lib, layers.list[i].pre_instance_functions.enumerate_instance_extension_properties); if (pfn == NULL) { loader_log(NULL, VK_DEBUG_REPORT_WARNING_BIT_EXT, 0, "%s: Unable to resolve symbol \"%s\" in implicit layer library \"%s\"", __FUNCTION__, layers.list[i].pre_instance_functions.enumerate_instance_extension_properties, layers.list[i].lib_name); continue; } VkEnumerateInstanceExtensionPropertiesChain *chain_link = malloc(sizeof(VkEnumerateInstanceExtensionPropertiesChain)); if (chain_link == NULL) { res = VK_ERROR_OUT_OF_HOST_MEMORY; break; } chain_link->header.type = VK_CHAIN_TYPE_ENUMERATE_INSTANCE_EXTENSION_PROPERTIES; chain_link->header.version = VK_CURRENT_CHAIN_VERSION; chain_link->header.size = sizeof(*chain_link); chain_link->pfnNextLayer = pfn; chain_link->pNextLink = chain_head; chain_head = chain_link; } // Call down the chain if (res == VK_SUCCESS) { res = chain_head->pfnNextLayer(chain_head->pNextLink, pLayerName, pPropertyCount, pProperties); } // Free up the layers loaderDeleteLayerListAndProperties(NULL, &layers); // Tear down the chain while (chain_head != &chain_tail) { VkEnumerateInstanceExtensionPropertiesChain *holder = chain_head; chain_head = (VkEnumerateInstanceExtensionPropertiesChain *)chain_head->pNextLink; free(holder); } // Close the dl handles for (size_t i = 0; i < lib_count; ++i) { loader_platform_close_library(libs[i]); } free(libs); return res; } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkEnumerateInstanceLayerProperties(uint32_t *pPropertyCount, VkLayerProperties *pProperties) { tls_instance = NULL; LOADER_PLATFORM_THREAD_ONCE(&once_init, loader_initialize); // We know we need to call at least the terminator VkResult res = VK_SUCCESS; VkEnumerateInstanceLayerPropertiesChain chain_tail = { .header = { .type = VK_CHAIN_TYPE_ENUMERATE_INSTANCE_LAYER_PROPERTIES, .version = VK_CURRENT_CHAIN_VERSION, .size = sizeof(chain_tail), }, .pfnNextLayer = &terminator_EnumerateInstanceLayerProperties, .pNextLink = NULL, }; VkEnumerateInstanceLayerPropertiesChain *chain_head = &chain_tail; // Get the implicit layers struct loader_layer_list layers; memset(&layers, 0, sizeof(layers)); loaderScanForImplicitLayers(NULL, &layers); // We'll need to save the dl handles so we can close them later loader_platform_dl_handle *libs = malloc(sizeof(loader_platform_dl_handle) * layers.count); if (libs == NULL) { return VK_ERROR_OUT_OF_HOST_MEMORY; } size_t lib_count = 0; // Prepend layers onto the chain if they implment this entry point for (uint32_t i = 0; i < layers.count; ++i) { if (!loaderImplicitLayerIsEnabled(NULL, layers.list + i) || layers.list[i].pre_instance_functions.enumerate_instance_layer_properties[0] == '\0') { continue; } loader_platform_dl_handle layer_lib = loader_platform_open_library(layers.list[i].lib_name); if (layer_lib == NULL) { loader_log(NULL, VK_DEBUG_REPORT_WARNING_BIT_EXT, 0, "%s: Unable to load implicit layer library \"%s\"", __FUNCTION__, layers.list[i].lib_name); continue; } libs[lib_count++] = layer_lib; void *pfn = loader_platform_get_proc_address(layer_lib, layers.list[i].pre_instance_functions.enumerate_instance_layer_properties); if (pfn == NULL) { loader_log(NULL, VK_DEBUG_REPORT_WARNING_BIT_EXT, 0, "%s: Unable to resolve symbol \"%s\" in implicit layer library \"%s\"", __FUNCTION__, layers.list[i].pre_instance_functions.enumerate_instance_layer_properties, layers.list[i].lib_name); continue; } VkEnumerateInstanceLayerPropertiesChain *chain_link = malloc(sizeof(VkEnumerateInstanceLayerPropertiesChain)); if (chain_link == NULL) { res = VK_ERROR_OUT_OF_HOST_MEMORY; break; } chain_link->header.type = VK_CHAIN_TYPE_ENUMERATE_INSTANCE_LAYER_PROPERTIES; chain_link->header.version = VK_CURRENT_CHAIN_VERSION; chain_link->header.size = sizeof(*chain_link); chain_link->pfnNextLayer = pfn; chain_link->pNextLink = chain_head; chain_head = chain_link; } // Call down the chain if (res == VK_SUCCESS) { res = chain_head->pfnNextLayer(chain_head->pNextLink, pPropertyCount, pProperties); } // Free up the layers loaderDeleteLayerListAndProperties(NULL, &layers); // Tear down the chain while (chain_head != &chain_tail) { VkEnumerateInstanceLayerPropertiesChain *holder = chain_head; chain_head = (VkEnumerateInstanceLayerPropertiesChain *)chain_head->pNextLink; free(holder); } // Close the dl handles for (size_t i = 0; i < lib_count; ++i) { loader_platform_close_library(libs[i]); } free(libs); return res; } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkEnumerateInstanceVersion(uint32_t* pApiVersion) { tls_instance = NULL; LOADER_PLATFORM_THREAD_ONCE(&once_init, loader_initialize); // We know we need to call at least the terminator VkResult res = VK_SUCCESS; VkEnumerateInstanceVersionChain chain_tail = { .header = { .type = VK_CHAIN_TYPE_ENUMERATE_INSTANCE_VERSION, .version = VK_CURRENT_CHAIN_VERSION, .size = sizeof(chain_tail), }, .pfnNextLayer = &terminator_EnumerateInstanceVersion, .pNextLink = NULL, }; VkEnumerateInstanceVersionChain *chain_head = &chain_tail; // Get the implicit layers struct loader_layer_list layers; memset(&layers, 0, sizeof(layers)); loaderScanForImplicitLayers(NULL, &layers); // We'll need to save the dl handles so we can close them later loader_platform_dl_handle *libs = malloc(sizeof(loader_platform_dl_handle) * layers.count); if (libs == NULL) { return VK_ERROR_OUT_OF_HOST_MEMORY; } size_t lib_count = 0; // Prepend layers onto the chain if they implment this entry point for (uint32_t i = 0; i < layers.count; ++i) { if (!loaderImplicitLayerIsEnabled(NULL, layers.list + i) || layers.list[i].pre_instance_functions.enumerate_instance_version[0] == '\0') { continue; } loader_platform_dl_handle layer_lib = loader_platform_open_library(layers.list[i].lib_name); if (layer_lib == NULL) { loader_log(NULL, VK_DEBUG_REPORT_WARNING_BIT_EXT, 0, "%s: Unable to load implicit layer library \"%s\"", __FUNCTION__, layers.list[i].lib_name); continue; } libs[lib_count++] = layer_lib; void *pfn = loader_platform_get_proc_address(layer_lib, layers.list[i].pre_instance_functions.enumerate_instance_version); if (pfn == NULL) { loader_log(NULL, VK_DEBUG_REPORT_WARNING_BIT_EXT, 0, "%s: Unable to resolve symbol \"%s\" in implicit layer library \"%s\"", __FUNCTION__, layers.list[i].pre_instance_functions.enumerate_instance_version, layers.list[i].lib_name); continue; } VkEnumerateInstanceVersionChain *chain_link = malloc(sizeof(VkEnumerateInstanceVersionChain)); if (chain_link == NULL) { res = VK_ERROR_OUT_OF_HOST_MEMORY; break; } chain_link->header.type = VK_CHAIN_TYPE_ENUMERATE_INSTANCE_VERSION; chain_link->header.version = VK_CURRENT_CHAIN_VERSION; chain_link->header.size = sizeof(*chain_link); chain_link->pfnNextLayer = pfn; chain_link->pNextLink = chain_head; chain_head = chain_link; } // Call down the chain if (res == VK_SUCCESS) { res = chain_head->pfnNextLayer(chain_head->pNextLink, pApiVersion); } // Free up the layers loaderDeleteLayerListAndProperties(NULL, &layers); // Tear down the chain while (chain_head != &chain_tail) { VkEnumerateInstanceVersionChain *holder = chain_head; chain_head = (VkEnumerateInstanceVersionChain *)chain_head->pNextLink; free(holder); } // Close the dl handles for (size_t i = 0; i < lib_count; ++i) { loader_platform_close_library(libs[i]); } free(libs); return res; } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkCreateInstance(const VkInstanceCreateInfo *pCreateInfo, const VkAllocationCallbacks *pAllocator, VkInstance *pInstance) { struct loader_instance *ptr_instance = NULL; VkInstance created_instance = VK_NULL_HANDLE; bool loaderLocked = false; VkResult res = VK_ERROR_INITIALIZATION_FAILED; LOADER_PLATFORM_THREAD_ONCE(&once_init, loader_initialize); #if (DEBUG_DISABLE_APP_ALLOCATORS == 1) { #else if (pAllocator) { ptr_instance = (struct loader_instance *)pAllocator->pfnAllocation(pAllocator->pUserData, sizeof(struct loader_instance), sizeof(int *), VK_SYSTEM_ALLOCATION_SCOPE_INSTANCE); } else { #endif ptr_instance = (struct loader_instance *)malloc(sizeof(struct loader_instance)); } VkInstanceCreateInfo ici = *pCreateInfo; if (ptr_instance == NULL) { res = VK_ERROR_OUT_OF_HOST_MEMORY; goto out; } tls_instance = ptr_instance; loader_platform_thread_lock_mutex(&loader_lock); loaderLocked = true; memset(ptr_instance, 0, sizeof(struct loader_instance)); if (pAllocator) { ptr_instance->alloc_callbacks = *pAllocator; } // Save the application version if (NULL == pCreateInfo || NULL == pCreateInfo->pApplicationInfo || 0 == pCreateInfo->pApplicationInfo->apiVersion) { ptr_instance->app_api_major_version = 1; ptr_instance->app_api_minor_version = 0; } else { ptr_instance->app_api_major_version = VK_VERSION_MAJOR(pCreateInfo->pApplicationInfo->apiVersion); ptr_instance->app_api_minor_version = VK_VERSION_MINOR(pCreateInfo->pApplicationInfo->apiVersion); } // Look for one or more VK_EXT_debug_report or VK_EXT_debug_utils create info structures // and setup a callback(s) for each one found. ptr_instance->num_tmp_report_callbacks = 0; ptr_instance->tmp_report_create_infos = NULL; ptr_instance->tmp_report_callbacks = NULL; ptr_instance->num_tmp_messengers = 0; ptr_instance->tmp_messenger_create_infos = NULL; ptr_instance->tmp_messengers = NULL; // Handle cases of VK_EXT_debug_utils if (util_CopyDebugUtilsMessengerCreateInfos(pCreateInfo->pNext, pAllocator, &ptr_instance->num_tmp_messengers, &ptr_instance->tmp_messenger_create_infos, &ptr_instance->tmp_messengers)) { // One or more were found, but allocation failed. Therefore, clean up and fail this function: res = VK_ERROR_OUT_OF_HOST_MEMORY; goto out; } else if (ptr_instance->num_tmp_messengers > 0) { // Setup the temporary messenger(s) here to catch early issues: if (util_CreateDebugUtilsMessengers(ptr_instance, pAllocator, ptr_instance->num_tmp_messengers, ptr_instance->tmp_messenger_create_infos, ptr_instance->tmp_messengers)) { // Failure of setting up one or more of the messenger. Therefore, clean up and fail this function: res = VK_ERROR_OUT_OF_HOST_MEMORY; goto out; } } // Handle cases of VK_EXT_debug_report if (util_CopyDebugReportCreateInfos(pCreateInfo->pNext, pAllocator, &ptr_instance->num_tmp_report_callbacks, &ptr_instance->tmp_report_create_infos, &ptr_instance->tmp_report_callbacks)) { // One or more were found, but allocation failed. Therefore, clean up and fail this function: res = VK_ERROR_OUT_OF_HOST_MEMORY; goto out; } else if (ptr_instance->num_tmp_report_callbacks > 0) { // Setup the temporary callback(s) here to catch early issues: if (util_CreateDebugReportCallbacks(ptr_instance, pAllocator, ptr_instance->num_tmp_report_callbacks, ptr_instance->tmp_report_create_infos, ptr_instance->tmp_report_callbacks)) { // Failure of setting up one or more of the callback. Therefore, clean up and fail this function: res = VK_ERROR_OUT_OF_HOST_MEMORY; goto out; } } // Due to implicit layers need to get layer list even if // enabledLayerCount == 0 and VK_INSTANCE_LAYERS is unset. For now always // get layer list via loaderScanForLayers(). memset(&ptr_instance->instance_layer_list, 0, sizeof(ptr_instance->instance_layer_list)); loaderScanForLayers(ptr_instance, &ptr_instance->instance_layer_list); // Validate the app requested layers to be enabled if (pCreateInfo->enabledLayerCount > 0) { res = loaderValidateLayers(ptr_instance, pCreateInfo->enabledLayerCount, pCreateInfo->ppEnabledLayerNames, &ptr_instance->instance_layer_list); if (res != VK_SUCCESS) { goto out; } } // Scan/discover all ICD libraries memset(&ptr_instance->icd_tramp_list, 0, sizeof(ptr_instance->icd_tramp_list)); res = loader_icd_scan(ptr_instance, &ptr_instance->icd_tramp_list); if (res != VK_SUCCESS) { goto out; } // Get extensions from all ICD's, merge so no duplicates, then validate res = loader_get_icd_loader_instance_extensions(ptr_instance, &ptr_instance->icd_tramp_list, &ptr_instance->ext_list); if (res != VK_SUCCESS) { goto out; } res = loader_validate_instance_extensions(ptr_instance, &ptr_instance->ext_list, &ptr_instance->instance_layer_list, &ici); if (res != VK_SUCCESS) { goto out; } ptr_instance->disp = loader_instance_heap_alloc(ptr_instance, sizeof(struct loader_instance_dispatch_table), VK_SYSTEM_ALLOCATION_SCOPE_INSTANCE); if (ptr_instance->disp == NULL) { loader_log(ptr_instance, VK_DEBUG_REPORT_ERROR_BIT_EXT, 0, "vkCreateInstance: Failed to allocate Loader's full Instance dispatch table."); res = VK_ERROR_OUT_OF_HOST_MEMORY; goto out; } memcpy(&ptr_instance->disp->layer_inst_disp, &instance_disp, sizeof(instance_disp)); ptr_instance->next = loader.instances; loader.instances = ptr_instance; // Activate any layers on instance chain res = loaderEnableInstanceLayers(ptr_instance, &ici, &ptr_instance->instance_layer_list); if (res != VK_SUCCESS) { goto out; } created_instance = (VkInstance)ptr_instance; res = loader_create_instance_chain(&ici, pAllocator, ptr_instance, &created_instance); if (res == VK_SUCCESS) { memset(ptr_instance->enabled_known_extensions.padding, 0, sizeof(uint64_t) * 4); wsi_create_instance(ptr_instance, &ici); debug_utils_CreateInstance(ptr_instance, &ici); extensions_create_instance(ptr_instance, &ici); *pInstance = created_instance; // Finally have the layers in place and everyone has seen // the CreateInstance command go by. This allows the layer's // GetInstanceProcAddr functions to return valid extension functions // if enabled. loaderActivateInstanceLayerExtensions(ptr_instance, *pInstance); } out: if (NULL != ptr_instance) { if (res != VK_SUCCESS) { if (NULL != ptr_instance->next) { loader.instances = ptr_instance->next; } if (NULL != ptr_instance->disp) { loader_instance_heap_free(ptr_instance, ptr_instance->disp); } if (ptr_instance->num_tmp_report_callbacks > 0) { // Remove temporary VK_EXT_debug_report items util_DestroyDebugReportCallbacks(ptr_instance, pAllocator, ptr_instance->num_tmp_report_callbacks, ptr_instance->tmp_report_callbacks); util_FreeDebugReportCreateInfos(pAllocator, ptr_instance->tmp_report_create_infos, ptr_instance->tmp_report_callbacks); } if (ptr_instance->num_tmp_messengers > 0) { // Remove temporary VK_EXT_debug_utils items util_DestroyDebugUtilsMessengers(ptr_instance, pAllocator, ptr_instance->num_tmp_messengers, ptr_instance->tmp_messengers); util_FreeDebugUtilsMessengerCreateInfos(pAllocator, ptr_instance->tmp_messenger_create_infos, ptr_instance->tmp_messengers); } if (NULL != ptr_instance->expanded_activated_layer_list.list) { loaderDeactivateLayers(ptr_instance, NULL, &ptr_instance->expanded_activated_layer_list); } if (NULL != ptr_instance->app_activated_layer_list.list) { loaderDestroyLayerList(ptr_instance, NULL, &ptr_instance->app_activated_layer_list); } loaderDeleteLayerListAndProperties(ptr_instance, &ptr_instance->instance_layer_list); loader_scanned_icd_clear(ptr_instance, &ptr_instance->icd_tramp_list); loader_destroy_generic_list(ptr_instance, (struct loader_generic_list *)&ptr_instance->ext_list); loader_instance_heap_free(ptr_instance, ptr_instance); } else { // Remove temporary VK_EXT_debug_report or VK_EXT_debug_utils items util_DestroyDebugUtilsMessengers(ptr_instance, pAllocator, ptr_instance->num_tmp_messengers, ptr_instance->tmp_messengers); util_DestroyDebugReportCallbacks(ptr_instance, pAllocator, ptr_instance->num_tmp_report_callbacks, ptr_instance->tmp_report_callbacks); } if (loaderLocked) { loader_platform_thread_unlock_mutex(&loader_lock); } } return res; } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkDestroyInstance(VkInstance instance, const VkAllocationCallbacks *pAllocator) { const VkLayerInstanceDispatchTable *disp; struct loader_instance *ptr_instance = NULL; bool callback_setup = false; bool messenger_setup = false; if (instance == VK_NULL_HANDLE) { return; } disp = loader_get_instance_layer_dispatch(instance); loader_platform_thread_lock_mutex(&loader_lock); ptr_instance = loader_get_instance(instance); if (pAllocator) { ptr_instance->alloc_callbacks = *pAllocator; } if (ptr_instance->num_tmp_messengers > 0) { // Setup the temporary VK_EXT_debug_utils messenger(s) here to catch cleanup issues: if (!util_CreateDebugUtilsMessengers(ptr_instance, pAllocator, ptr_instance->num_tmp_messengers, ptr_instance->tmp_messenger_create_infos, ptr_instance->tmp_messengers)) { messenger_setup = true; } } if (ptr_instance->num_tmp_report_callbacks > 0) { // Setup the temporary VK_EXT_debug_report callback(s) here to catch cleanup issues: if (!util_CreateDebugReportCallbacks(ptr_instance, pAllocator, ptr_instance->num_tmp_report_callbacks, ptr_instance->tmp_report_create_infos, ptr_instance->tmp_report_callbacks)) { callback_setup = true; } } disp->DestroyInstance(instance, pAllocator); if (NULL != ptr_instance->expanded_activated_layer_list.list) { loaderDeactivateLayers(ptr_instance, NULL, &ptr_instance->expanded_activated_layer_list); } if (NULL != ptr_instance->app_activated_layer_list.list) { loaderDestroyLayerList(ptr_instance, NULL, &ptr_instance->app_activated_layer_list); } if (ptr_instance->phys_devs_tramp) { for (uint32_t i = 0; i < ptr_instance->phys_dev_count_tramp; i++) { loader_instance_heap_free(ptr_instance, ptr_instance->phys_devs_tramp[i]); } loader_instance_heap_free(ptr_instance, ptr_instance->phys_devs_tramp); } if (ptr_instance->phys_dev_groups_tramp) { for (uint32_t i = 0; i < ptr_instance->phys_dev_group_count_tramp; i++) { loader_instance_heap_free(ptr_instance, ptr_instance->phys_dev_groups_tramp[i]); } loader_instance_heap_free(ptr_instance, ptr_instance->phys_dev_groups_tramp); } if (messenger_setup) { util_DestroyDebugUtilsMessengers(ptr_instance, pAllocator, ptr_instance->num_tmp_messengers, ptr_instance->tmp_messengers); util_FreeDebugUtilsMessengerCreateInfos(pAllocator, ptr_instance->tmp_messenger_create_infos, ptr_instance->tmp_messengers); } if (callback_setup) { util_DestroyDebugReportCallbacks(ptr_instance, pAllocator, ptr_instance->num_tmp_report_callbacks, ptr_instance->tmp_report_callbacks); util_FreeDebugReportCreateInfos(pAllocator, ptr_instance->tmp_report_create_infos, ptr_instance->tmp_report_callbacks); } loader_instance_heap_free(ptr_instance, ptr_instance->disp); loader_instance_heap_free(ptr_instance, ptr_instance); loader_platform_thread_unlock_mutex(&loader_lock); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkEnumeratePhysicalDevices(VkInstance instance, uint32_t *pPhysicalDeviceCount, VkPhysicalDevice *pPhysicalDevices) { VkResult res = VK_SUCCESS; uint32_t count; uint32_t i; struct loader_instance *inst; loader_platform_thread_lock_mutex(&loader_lock); inst = loader_get_instance(instance); if (NULL == inst) { res = VK_ERROR_INITIALIZATION_FAILED; goto out; } if (NULL == pPhysicalDeviceCount) { loader_log(inst, VK_DEBUG_REPORT_ERROR_BIT_EXT, 0, "vkEnumeratePhysicalDevices: Received NULL pointer for physical device count return value."); res = VK_ERROR_INITIALIZATION_FAILED; goto out; } // Setup the trampoline loader physical devices. This will actually // call down and setup the terminator loader physical devices during the // process. VkResult setup_res = setupLoaderTrampPhysDevs(instance); if (setup_res != VK_SUCCESS && setup_res != VK_INCOMPLETE) { res = setup_res; goto out; } count = inst->phys_dev_count_tramp; // Wrap the PhysDev object for loader usage, return wrapped objects if (NULL != pPhysicalDevices) { if (inst->phys_dev_count_tramp > *pPhysicalDeviceCount) { loader_log(inst, VK_DEBUG_REPORT_INFORMATION_BIT_EXT, 0, "vkEnumeratePhysicalDevices: Trimming device count down" " by application request from %d to %d physical devices", inst->phys_dev_count_tramp, *pPhysicalDeviceCount); count = *pPhysicalDeviceCount; res = VK_INCOMPLETE; } for (i = 0; i < count; i++) { pPhysicalDevices[i] = (VkPhysicalDevice)inst->phys_devs_tramp[i]; } } *pPhysicalDeviceCount = count; out: loader_platform_thread_unlock_mutex(&loader_lock); return res; } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkGetPhysicalDeviceFeatures(VkPhysicalDevice physicalDevice, VkPhysicalDeviceFeatures *pFeatures) { const VkLayerInstanceDispatchTable *disp; VkPhysicalDevice unwrapped_phys_dev = loader_unwrap_physical_device(physicalDevice); disp = loader_get_instance_layer_dispatch(physicalDevice); disp->GetPhysicalDeviceFeatures(unwrapped_phys_dev, pFeatures); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkGetPhysicalDeviceFormatProperties(VkPhysicalDevice physicalDevice, VkFormat format, VkFormatProperties *pFormatInfo) { const VkLayerInstanceDispatchTable *disp; VkPhysicalDevice unwrapped_pd = loader_unwrap_physical_device(physicalDevice); disp = loader_get_instance_layer_dispatch(physicalDevice); disp->GetPhysicalDeviceFormatProperties(unwrapped_pd, format, pFormatInfo); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkGetPhysicalDeviceImageFormatProperties( VkPhysicalDevice physicalDevice, VkFormat format, VkImageType type, VkImageTiling tiling, VkImageUsageFlags usage, VkImageCreateFlags flags, VkImageFormatProperties *pImageFormatProperties) { const VkLayerInstanceDispatchTable *disp; VkPhysicalDevice unwrapped_phys_dev = loader_unwrap_physical_device(physicalDevice); disp = loader_get_instance_layer_dispatch(physicalDevice); return disp->GetPhysicalDeviceImageFormatProperties(unwrapped_phys_dev, format, type, tiling, usage, flags, pImageFormatProperties); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkGetPhysicalDeviceProperties(VkPhysicalDevice physicalDevice, VkPhysicalDeviceProperties *pProperties) { const VkLayerInstanceDispatchTable *disp; VkPhysicalDevice unwrapped_phys_dev = loader_unwrap_physical_device(physicalDevice); disp = loader_get_instance_layer_dispatch(physicalDevice); disp->GetPhysicalDeviceProperties(unwrapped_phys_dev, pProperties); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkGetPhysicalDeviceQueueFamilyProperties(VkPhysicalDevice physicalDevice, uint32_t *pQueueFamilyPropertyCount, VkQueueFamilyProperties *pQueueProperties) { const VkLayerInstanceDispatchTable *disp; VkPhysicalDevice unwrapped_phys_dev = loader_unwrap_physical_device(physicalDevice); disp = loader_get_instance_layer_dispatch(physicalDevice); disp->GetPhysicalDeviceQueueFamilyProperties(unwrapped_phys_dev, pQueueFamilyPropertyCount, pQueueProperties); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkGetPhysicalDeviceMemoryProperties(VkPhysicalDevice physicalDevice, VkPhysicalDeviceMemoryProperties *pMemoryProperties) { const VkLayerInstanceDispatchTable *disp; VkPhysicalDevice unwrapped_phys_dev = loader_unwrap_physical_device(physicalDevice); disp = loader_get_instance_layer_dispatch(physicalDevice); disp->GetPhysicalDeviceMemoryProperties(unwrapped_phys_dev, pMemoryProperties); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkCreateDevice(VkPhysicalDevice physicalDevice, const VkDeviceCreateInfo *pCreateInfo, const VkAllocationCallbacks *pAllocator, VkDevice *pDevice) { loader_platform_thread_lock_mutex(&loader_lock); VkResult res = loader_layer_create_device(NULL, physicalDevice, pCreateInfo, pAllocator, pDevice, NULL, NULL); loader_platform_thread_unlock_mutex(&loader_lock); return res; } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkDestroyDevice(VkDevice device, const VkAllocationCallbacks *pAllocator) { const VkLayerDispatchTable *disp; if (device == VK_NULL_HANDLE) { return; } disp = loader_get_dispatch(device); loader_platform_thread_lock_mutex(&loader_lock); loader_layer_destroy_device(device, pAllocator, disp->DestroyDevice); loader_platform_thread_unlock_mutex(&loader_lock); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkEnumerateDeviceExtensionProperties(VkPhysicalDevice physicalDevice, const char *pLayerName, uint32_t *pPropertyCount, VkExtensionProperties *pProperties) { VkResult res = VK_SUCCESS; struct loader_physical_device_tramp *phys_dev; const VkLayerInstanceDispatchTable *disp; phys_dev = (struct loader_physical_device_tramp *)physicalDevice; loader_platform_thread_lock_mutex(&loader_lock); // always pass this call down the instance chain which will terminate // in the ICD. This allows layers to filter the extensions coming back // up the chain. In the terminator we look up layer extensions from the // manifest file if it wasn't provided by the layer itself. disp = loader_get_instance_layer_dispatch(physicalDevice); res = disp->EnumerateDeviceExtensionProperties(phys_dev->phys_dev, pLayerName, pPropertyCount, pProperties); loader_platform_thread_unlock_mutex(&loader_lock); return res; } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkEnumerateDeviceLayerProperties(VkPhysicalDevice physicalDevice, uint32_t *pPropertyCount, VkLayerProperties *pProperties) { uint32_t copy_size; struct loader_physical_device_tramp *phys_dev; struct loader_layer_list *enabled_layers, layers_list; memset(&layers_list, 0, sizeof(layers_list)); loader_platform_thread_lock_mutex(&loader_lock); // Don't dispatch this call down the instance chain, want all device layers // enumerated and instance chain may not contain all device layers // TODO re-evaluate the above statement we maybe able to start calling // down the chain phys_dev = (struct loader_physical_device_tramp *)physicalDevice; const struct loader_instance *inst = phys_dev->this_instance; uint32_t count = inst->app_activated_layer_list.count; if (count == 0 || pProperties == NULL) { *pPropertyCount = count; loader_platform_thread_unlock_mutex(&loader_lock); return VK_SUCCESS; } enabled_layers = (struct loader_layer_list *)&inst->app_activated_layer_list; copy_size = (*pPropertyCount < count) ? *pPropertyCount : count; for (uint32_t i = 0; i < copy_size; i++) { memcpy(&pProperties[i], &(enabled_layers->list[i].info), sizeof(VkLayerProperties)); } *pPropertyCount = copy_size; if (copy_size < count) { loader_platform_thread_unlock_mutex(&loader_lock); return VK_INCOMPLETE; } loader_platform_thread_unlock_mutex(&loader_lock); return VK_SUCCESS; } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkGetDeviceQueue(VkDevice device, uint32_t queueNodeIndex, uint32_t queueIndex, VkQueue *pQueue) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); disp->GetDeviceQueue(device, queueNodeIndex, queueIndex, pQueue); loader_set_dispatch(*pQueue, disp); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkQueueSubmit(VkQueue queue, uint32_t submitCount, const VkSubmitInfo *pSubmits, VkFence fence) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(queue); return disp->QueueSubmit(queue, submitCount, pSubmits, fence); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkQueueWaitIdle(VkQueue queue) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(queue); return disp->QueueWaitIdle(queue); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkDeviceWaitIdle(VkDevice device) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->DeviceWaitIdle(device); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkAllocateMemory(VkDevice device, const VkMemoryAllocateInfo *pAllocateInfo, const VkAllocationCallbacks *pAllocator, VkDeviceMemory *pMemory) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->AllocateMemory(device, pAllocateInfo, pAllocator, pMemory); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkFreeMemory(VkDevice device, VkDeviceMemory mem, const VkAllocationCallbacks *pAllocator) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); disp->FreeMemory(device, mem, pAllocator); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkMapMemory(VkDevice device, VkDeviceMemory mem, VkDeviceSize offset, VkDeviceSize size, VkFlags flags, void **ppData) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->MapMemory(device, mem, offset, size, flags, ppData); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkUnmapMemory(VkDevice device, VkDeviceMemory mem) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); disp->UnmapMemory(device, mem); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkFlushMappedMemoryRanges(VkDevice device, uint32_t memoryRangeCount, const VkMappedMemoryRange *pMemoryRanges) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->FlushMappedMemoryRanges(device, memoryRangeCount, pMemoryRanges); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkInvalidateMappedMemoryRanges(VkDevice device, uint32_t memoryRangeCount, const VkMappedMemoryRange *pMemoryRanges) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->InvalidateMappedMemoryRanges(device, memoryRangeCount, pMemoryRanges); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkGetDeviceMemoryCommitment(VkDevice device, VkDeviceMemory memory, VkDeviceSize *pCommittedMemoryInBytes) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); disp->GetDeviceMemoryCommitment(device, memory, pCommittedMemoryInBytes); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkBindBufferMemory(VkDevice device, VkBuffer buffer, VkDeviceMemory mem, VkDeviceSize offset) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->BindBufferMemory(device, buffer, mem, offset); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkBindImageMemory(VkDevice device, VkImage image, VkDeviceMemory mem, VkDeviceSize offset) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->BindImageMemory(device, image, mem, offset); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkGetBufferMemoryRequirements(VkDevice device, VkBuffer buffer, VkMemoryRequirements *pMemoryRequirements) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); disp->GetBufferMemoryRequirements(device, buffer, pMemoryRequirements); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkGetImageMemoryRequirements(VkDevice device, VkImage image, VkMemoryRequirements *pMemoryRequirements) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); disp->GetImageMemoryRequirements(device, image, pMemoryRequirements); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkGetImageSparseMemoryRequirements(VkDevice device, VkImage image, uint32_t *pSparseMemoryRequirementCount, VkSparseImageMemoryRequirements *pSparseMemoryRequirements) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); disp->GetImageSparseMemoryRequirements(device, image, pSparseMemoryRequirementCount, pSparseMemoryRequirements); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkGetPhysicalDeviceSparseImageFormatProperties( VkPhysicalDevice physicalDevice, VkFormat format, VkImageType type, VkSampleCountFlagBits samples, VkImageUsageFlags usage, VkImageTiling tiling, uint32_t *pPropertyCount, VkSparseImageFormatProperties *pProperties) { const VkLayerInstanceDispatchTable *disp; VkPhysicalDevice unwrapped_phys_dev = loader_unwrap_physical_device(physicalDevice); disp = loader_get_instance_layer_dispatch(physicalDevice); disp->GetPhysicalDeviceSparseImageFormatProperties(unwrapped_phys_dev, format, type, samples, usage, tiling, pPropertyCount, pProperties); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkQueueBindSparse(VkQueue queue, uint32_t bindInfoCount, const VkBindSparseInfo *pBindInfo, VkFence fence) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(queue); return disp->QueueBindSparse(queue, bindInfoCount, pBindInfo, fence); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkCreateFence(VkDevice device, const VkFenceCreateInfo *pCreateInfo, const VkAllocationCallbacks *pAllocator, VkFence *pFence) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->CreateFence(device, pCreateInfo, pAllocator, pFence); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkDestroyFence(VkDevice device, VkFence fence, const VkAllocationCallbacks *pAllocator) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); disp->DestroyFence(device, fence, pAllocator); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkResetFences(VkDevice device, uint32_t fenceCount, const VkFence *pFences) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->ResetFences(device, fenceCount, pFences); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkGetFenceStatus(VkDevice device, VkFence fence) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->GetFenceStatus(device, fence); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkWaitForFences(VkDevice device, uint32_t fenceCount, const VkFence *pFences, VkBool32 waitAll, uint64_t timeout) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->WaitForFences(device, fenceCount, pFences, waitAll, timeout); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkCreateSemaphore(VkDevice device, const VkSemaphoreCreateInfo *pCreateInfo, const VkAllocationCallbacks *pAllocator, VkSemaphore *pSemaphore) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->CreateSemaphore(device, pCreateInfo, pAllocator, pSemaphore); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkDestroySemaphore(VkDevice device, VkSemaphore semaphore, const VkAllocationCallbacks *pAllocator) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); disp->DestroySemaphore(device, semaphore, pAllocator); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkCreateEvent(VkDevice device, const VkEventCreateInfo *pCreateInfo, const VkAllocationCallbacks *pAllocator, VkEvent *pEvent) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->CreateEvent(device, pCreateInfo, pAllocator, pEvent); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkDestroyEvent(VkDevice device, VkEvent event, const VkAllocationCallbacks *pAllocator) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); disp->DestroyEvent(device, event, pAllocator); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkGetEventStatus(VkDevice device, VkEvent event) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->GetEventStatus(device, event); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkSetEvent(VkDevice device, VkEvent event) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->SetEvent(device, event); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkResetEvent(VkDevice device, VkEvent event) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->ResetEvent(device, event); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkCreateQueryPool(VkDevice device, const VkQueryPoolCreateInfo *pCreateInfo, const VkAllocationCallbacks *pAllocator, VkQueryPool *pQueryPool) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->CreateQueryPool(device, pCreateInfo, pAllocator, pQueryPool); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkDestroyQueryPool(VkDevice device, VkQueryPool queryPool, const VkAllocationCallbacks *pAllocator) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); disp->DestroyQueryPool(device, queryPool, pAllocator); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkGetQueryPoolResults(VkDevice device, VkQueryPool queryPool, uint32_t firstQuery, uint32_t queryCount, size_t dataSize, void *pData, VkDeviceSize stride, VkQueryResultFlags flags) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->GetQueryPoolResults(device, queryPool, firstQuery, queryCount, dataSize, pData, stride, flags); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkCreateBuffer(VkDevice device, const VkBufferCreateInfo *pCreateInfo, const VkAllocationCallbacks *pAllocator, VkBuffer *pBuffer) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->CreateBuffer(device, pCreateInfo, pAllocator, pBuffer); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkDestroyBuffer(VkDevice device, VkBuffer buffer, const VkAllocationCallbacks *pAllocator) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); disp->DestroyBuffer(device, buffer, pAllocator); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkCreateBufferView(VkDevice device, const VkBufferViewCreateInfo *pCreateInfo, const VkAllocationCallbacks *pAllocator, VkBufferView *pView) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->CreateBufferView(device, pCreateInfo, pAllocator, pView); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkDestroyBufferView(VkDevice device, VkBufferView bufferView, const VkAllocationCallbacks *pAllocator) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); disp->DestroyBufferView(device, bufferView, pAllocator); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkCreateImage(VkDevice device, const VkImageCreateInfo *pCreateInfo, const VkAllocationCallbacks *pAllocator, VkImage *pImage) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->CreateImage(device, pCreateInfo, pAllocator, pImage); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkDestroyImage(VkDevice device, VkImage image, const VkAllocationCallbacks *pAllocator) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); disp->DestroyImage(device, image, pAllocator); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkGetImageSubresourceLayout(VkDevice device, VkImage image, const VkImageSubresource *pSubresource, VkSubresourceLayout *pLayout) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); disp->GetImageSubresourceLayout(device, image, pSubresource, pLayout); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkCreateImageView(VkDevice device, const VkImageViewCreateInfo *pCreateInfo, const VkAllocationCallbacks *pAllocator, VkImageView *pView) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->CreateImageView(device, pCreateInfo, pAllocator, pView); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkDestroyImageView(VkDevice device, VkImageView imageView, const VkAllocationCallbacks *pAllocator) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); disp->DestroyImageView(device, imageView, pAllocator); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkCreateShaderModule(VkDevice device, const VkShaderModuleCreateInfo *pCreateInfo, const VkAllocationCallbacks *pAllocator, VkShaderModule *pShader) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->CreateShaderModule(device, pCreateInfo, pAllocator, pShader); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkDestroyShaderModule(VkDevice device, VkShaderModule shaderModule, const VkAllocationCallbacks *pAllocator) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); disp->DestroyShaderModule(device, shaderModule, pAllocator); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkCreatePipelineCache(VkDevice device, const VkPipelineCacheCreateInfo *pCreateInfo, const VkAllocationCallbacks *pAllocator, VkPipelineCache *pPipelineCache) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->CreatePipelineCache(device, pCreateInfo, pAllocator, pPipelineCache); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkDestroyPipelineCache(VkDevice device, VkPipelineCache pipelineCache, const VkAllocationCallbacks *pAllocator) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); disp->DestroyPipelineCache(device, pipelineCache, pAllocator); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkGetPipelineCacheData(VkDevice device, VkPipelineCache pipelineCache, size_t *pDataSize, void *pData) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->GetPipelineCacheData(device, pipelineCache, pDataSize, pData); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkMergePipelineCaches(VkDevice device, VkPipelineCache dstCache, uint32_t srcCacheCount, const VkPipelineCache *pSrcCaches) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->MergePipelineCaches(device, dstCache, srcCacheCount, pSrcCaches); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkCreateGraphicsPipelines(VkDevice device, VkPipelineCache pipelineCache, uint32_t createInfoCount, const VkGraphicsPipelineCreateInfo *pCreateInfos, const VkAllocationCallbacks *pAllocator, VkPipeline *pPipelines) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->CreateGraphicsPipelines(device, pipelineCache, createInfoCount, pCreateInfos, pAllocator, pPipelines); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkCreateComputePipelines(VkDevice device, VkPipelineCache pipelineCache, uint32_t createInfoCount, const VkComputePipelineCreateInfo *pCreateInfos, const VkAllocationCallbacks *pAllocator, VkPipeline *pPipelines) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->CreateComputePipelines(device, pipelineCache, createInfoCount, pCreateInfos, pAllocator, pPipelines); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkDestroyPipeline(VkDevice device, VkPipeline pipeline, const VkAllocationCallbacks *pAllocator) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); disp->DestroyPipeline(device, pipeline, pAllocator); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkCreatePipelineLayout(VkDevice device, const VkPipelineLayoutCreateInfo *pCreateInfo, const VkAllocationCallbacks *pAllocator, VkPipelineLayout *pPipelineLayout) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->CreatePipelineLayout(device, pCreateInfo, pAllocator, pPipelineLayout); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkDestroyPipelineLayout(VkDevice device, VkPipelineLayout pipelineLayout, const VkAllocationCallbacks *pAllocator) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); disp->DestroyPipelineLayout(device, pipelineLayout, pAllocator); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkCreateSampler(VkDevice device, const VkSamplerCreateInfo *pCreateInfo, const VkAllocationCallbacks *pAllocator, VkSampler *pSampler) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->CreateSampler(device, pCreateInfo, pAllocator, pSampler); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkDestroySampler(VkDevice device, VkSampler sampler, const VkAllocationCallbacks *pAllocator) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); disp->DestroySampler(device, sampler, pAllocator); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkCreateDescriptorSetLayout(VkDevice device, const VkDescriptorSetLayoutCreateInfo *pCreateInfo, const VkAllocationCallbacks *pAllocator, VkDescriptorSetLayout *pSetLayout) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->CreateDescriptorSetLayout(device, pCreateInfo, pAllocator, pSetLayout); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkDestroyDescriptorSetLayout(VkDevice device, VkDescriptorSetLayout descriptorSetLayout, const VkAllocationCallbacks *pAllocator) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); disp->DestroyDescriptorSetLayout(device, descriptorSetLayout, pAllocator); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkCreateDescriptorPool(VkDevice device, const VkDescriptorPoolCreateInfo *pCreateInfo, const VkAllocationCallbacks *pAllocator, VkDescriptorPool *pDescriptorPool) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->CreateDescriptorPool(device, pCreateInfo, pAllocator, pDescriptorPool); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkDestroyDescriptorPool(VkDevice device, VkDescriptorPool descriptorPool, const VkAllocationCallbacks *pAllocator) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); disp->DestroyDescriptorPool(device, descriptorPool, pAllocator); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkResetDescriptorPool(VkDevice device, VkDescriptorPool descriptorPool, VkDescriptorPoolResetFlags flags) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->ResetDescriptorPool(device, descriptorPool, flags); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkAllocateDescriptorSets(VkDevice device, const VkDescriptorSetAllocateInfo *pAllocateInfo, VkDescriptorSet *pDescriptorSets) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->AllocateDescriptorSets(device, pAllocateInfo, pDescriptorSets); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkFreeDescriptorSets(VkDevice device, VkDescriptorPool descriptorPool, uint32_t descriptorSetCount, const VkDescriptorSet *pDescriptorSets) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->FreeDescriptorSets(device, descriptorPool, descriptorSetCount, pDescriptorSets); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkUpdateDescriptorSets(VkDevice device, uint32_t descriptorWriteCount, const VkWriteDescriptorSet *pDescriptorWrites, uint32_t descriptorCopyCount, const VkCopyDescriptorSet *pDescriptorCopies) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); disp->UpdateDescriptorSets(device, descriptorWriteCount, pDescriptorWrites, descriptorCopyCount, pDescriptorCopies); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkCreateFramebuffer(VkDevice device, const VkFramebufferCreateInfo *pCreateInfo, const VkAllocationCallbacks *pAllocator, VkFramebuffer *pFramebuffer) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->CreateFramebuffer(device, pCreateInfo, pAllocator, pFramebuffer); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkDestroyFramebuffer(VkDevice device, VkFramebuffer framebuffer, const VkAllocationCallbacks *pAllocator) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); disp->DestroyFramebuffer(device, framebuffer, pAllocator); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkCreateRenderPass(VkDevice device, const VkRenderPassCreateInfo *pCreateInfo, const VkAllocationCallbacks *pAllocator, VkRenderPass *pRenderPass) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->CreateRenderPass(device, pCreateInfo, pAllocator, pRenderPass); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkDestroyRenderPass(VkDevice device, VkRenderPass renderPass, const VkAllocationCallbacks *pAllocator) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); disp->DestroyRenderPass(device, renderPass, pAllocator); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkGetRenderAreaGranularity(VkDevice device, VkRenderPass renderPass, VkExtent2D *pGranularity) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); disp->GetRenderAreaGranularity(device, renderPass, pGranularity); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkCreateCommandPool(VkDevice device, const VkCommandPoolCreateInfo *pCreateInfo, const VkAllocationCallbacks *pAllocator, VkCommandPool *pCommandPool) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->CreateCommandPool(device, pCreateInfo, pAllocator, pCommandPool); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkDestroyCommandPool(VkDevice device, VkCommandPool commandPool, const VkAllocationCallbacks *pAllocator) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); disp->DestroyCommandPool(device, commandPool, pAllocator); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkResetCommandPool(VkDevice device, VkCommandPool commandPool, VkCommandPoolResetFlags flags) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); return disp->ResetCommandPool(device, commandPool, flags); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkAllocateCommandBuffers(VkDevice device, const VkCommandBufferAllocateInfo *pAllocateInfo, VkCommandBuffer *pCommandBuffers) { const VkLayerDispatchTable *disp; VkResult res; disp = loader_get_dispatch(device); res = disp->AllocateCommandBuffers(device, pAllocateInfo, pCommandBuffers); if (res == VK_SUCCESS) { for (uint32_t i = 0; i < pAllocateInfo->commandBufferCount; i++) { if (pCommandBuffers[i]) { loader_init_dispatch(pCommandBuffers[i], disp); } } } return res; } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkFreeCommandBuffers(VkDevice device, VkCommandPool commandPool, uint32_t commandBufferCount, const VkCommandBuffer *pCommandBuffers) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(device); disp->FreeCommandBuffers(device, commandPool, commandBufferCount, pCommandBuffers); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkBeginCommandBuffer(VkCommandBuffer commandBuffer, const VkCommandBufferBeginInfo *pBeginInfo) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); return disp->BeginCommandBuffer(commandBuffer, pBeginInfo); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkEndCommandBuffer(VkCommandBuffer commandBuffer) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); return disp->EndCommandBuffer(commandBuffer); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkResetCommandBuffer(VkCommandBuffer commandBuffer, VkCommandBufferResetFlags flags) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); return disp->ResetCommandBuffer(commandBuffer, flags); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdBindPipeline(VkCommandBuffer commandBuffer, VkPipelineBindPoint pipelineBindPoint, VkPipeline pipeline) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdBindPipeline(commandBuffer, pipelineBindPoint, pipeline); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdSetViewport(VkCommandBuffer commandBuffer, uint32_t firstViewport, uint32_t viewportCount, const VkViewport *pViewports) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdSetViewport(commandBuffer, firstViewport, viewportCount, pViewports); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdSetScissor(VkCommandBuffer commandBuffer, uint32_t firstScissor, uint32_t scissorCount, const VkRect2D *pScissors) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdSetScissor(commandBuffer, firstScissor, scissorCount, pScissors); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdSetLineWidth(VkCommandBuffer commandBuffer, float lineWidth) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdSetLineWidth(commandBuffer, lineWidth); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdSetDepthBias(VkCommandBuffer commandBuffer, float depthBiasConstantFactor, float depthBiasClamp, float depthBiasSlopeFactor) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdSetDepthBias(commandBuffer, depthBiasConstantFactor, depthBiasClamp, depthBiasSlopeFactor); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdSetBlendConstants(VkCommandBuffer commandBuffer, const float blendConstants[4]) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdSetBlendConstants(commandBuffer, blendConstants); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdSetDepthBounds(VkCommandBuffer commandBuffer, float minDepthBounds, float maxDepthBounds) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdSetDepthBounds(commandBuffer, minDepthBounds, maxDepthBounds); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdSetStencilCompareMask(VkCommandBuffer commandBuffer, VkStencilFaceFlags faceMask, uint32_t compareMask) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdSetStencilCompareMask(commandBuffer, faceMask, compareMask); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdSetStencilWriteMask(VkCommandBuffer commandBuffer, VkStencilFaceFlags faceMask, uint32_t writeMask) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdSetStencilWriteMask(commandBuffer, faceMask, writeMask); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdSetStencilReference(VkCommandBuffer commandBuffer, VkStencilFaceFlags faceMask, uint32_t reference) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdSetStencilReference(commandBuffer, faceMask, reference); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdBindDescriptorSets(VkCommandBuffer commandBuffer, VkPipelineBindPoint pipelineBindPoint, VkPipelineLayout layout, uint32_t firstSet, uint32_t descriptorSetCount, const VkDescriptorSet *pDescriptorSets, uint32_t dynamicOffsetCount, const uint32_t *pDynamicOffsets) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdBindDescriptorSets(commandBuffer, pipelineBindPoint, layout, firstSet, descriptorSetCount, pDescriptorSets, dynamicOffsetCount, pDynamicOffsets); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdBindIndexBuffer(VkCommandBuffer commandBuffer, VkBuffer buffer, VkDeviceSize offset, VkIndexType indexType) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdBindIndexBuffer(commandBuffer, buffer, offset, indexType); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdBindVertexBuffers(VkCommandBuffer commandBuffer, uint32_t firstBinding, uint32_t bindingCount, const VkBuffer *pBuffers, const VkDeviceSize *pOffsets) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdBindVertexBuffers(commandBuffer, firstBinding, bindingCount, pBuffers, pOffsets); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdDraw(VkCommandBuffer commandBuffer, uint32_t vertexCount, uint32_t instanceCount, uint32_t firstVertex, uint32_t firstInstance) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdDraw(commandBuffer, vertexCount, instanceCount, firstVertex, firstInstance); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdDrawIndexed(VkCommandBuffer commandBuffer, uint32_t indexCount, uint32_t instanceCount, uint32_t firstIndex, int32_t vertexOffset, uint32_t firstInstance) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdDrawIndexed(commandBuffer, indexCount, instanceCount, firstIndex, vertexOffset, firstInstance); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdDrawIndirect(VkCommandBuffer commandBuffer, VkBuffer buffer, VkDeviceSize offset, uint32_t drawCount, uint32_t stride) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdDrawIndirect(commandBuffer, buffer, offset, drawCount, stride); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdDrawIndexedIndirect(VkCommandBuffer commandBuffer, VkBuffer buffer, VkDeviceSize offset, uint32_t drawCount, uint32_t stride) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdDrawIndexedIndirect(commandBuffer, buffer, offset, drawCount, stride); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdDispatch(VkCommandBuffer commandBuffer, uint32_t x, uint32_t y, uint32_t z) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdDispatch(commandBuffer, x, y, z); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdDispatchIndirect(VkCommandBuffer commandBuffer, VkBuffer buffer, VkDeviceSize offset) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdDispatchIndirect(commandBuffer, buffer, offset); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdCopyBuffer(VkCommandBuffer commandBuffer, VkBuffer srcBuffer, VkBuffer dstBuffer, uint32_t regionCount, const VkBufferCopy *pRegions) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdCopyBuffer(commandBuffer, srcBuffer, dstBuffer, regionCount, pRegions); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdCopyImage(VkCommandBuffer commandBuffer, VkImage srcImage, VkImageLayout srcImageLayout, VkImage dstImage, VkImageLayout dstImageLayout, uint32_t regionCount, const VkImageCopy *pRegions) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdCopyImage(commandBuffer, srcImage, srcImageLayout, dstImage, dstImageLayout, regionCount, pRegions); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdBlitImage(VkCommandBuffer commandBuffer, VkImage srcImage, VkImageLayout srcImageLayout, VkImage dstImage, VkImageLayout dstImageLayout, uint32_t regionCount, const VkImageBlit *pRegions, VkFilter filter) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdBlitImage(commandBuffer, srcImage, srcImageLayout, dstImage, dstImageLayout, regionCount, pRegions, filter); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdCopyBufferToImage(VkCommandBuffer commandBuffer, VkBuffer srcBuffer, VkImage dstImage, VkImageLayout dstImageLayout, uint32_t regionCount, const VkBufferImageCopy *pRegions) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdCopyBufferToImage(commandBuffer, srcBuffer, dstImage, dstImageLayout, regionCount, pRegions); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdCopyImageToBuffer(VkCommandBuffer commandBuffer, VkImage srcImage, VkImageLayout srcImageLayout, VkBuffer dstBuffer, uint32_t regionCount, const VkBufferImageCopy *pRegions) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdCopyImageToBuffer(commandBuffer, srcImage, srcImageLayout, dstBuffer, regionCount, pRegions); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdUpdateBuffer(VkCommandBuffer commandBuffer, VkBuffer dstBuffer, VkDeviceSize dstOffset, VkDeviceSize dataSize, const void *pData) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdUpdateBuffer(commandBuffer, dstBuffer, dstOffset, dataSize, pData); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdFillBuffer(VkCommandBuffer commandBuffer, VkBuffer dstBuffer, VkDeviceSize dstOffset, VkDeviceSize size, uint32_t data) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdFillBuffer(commandBuffer, dstBuffer, dstOffset, size, data); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdClearColorImage(VkCommandBuffer commandBuffer, VkImage image, VkImageLayout imageLayout, const VkClearColorValue *pColor, uint32_t rangeCount, const VkImageSubresourceRange *pRanges) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdClearColorImage(commandBuffer, image, imageLayout, pColor, rangeCount, pRanges); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdClearDepthStencilImage(VkCommandBuffer commandBuffer, VkImage image, VkImageLayout imageLayout, const VkClearDepthStencilValue *pDepthStencil, uint32_t rangeCount, const VkImageSubresourceRange *pRanges) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdClearDepthStencilImage(commandBuffer, image, imageLayout, pDepthStencil, rangeCount, pRanges); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdClearAttachments(VkCommandBuffer commandBuffer, uint32_t attachmentCount, const VkClearAttachment *pAttachments, uint32_t rectCount, const VkClearRect *pRects) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdClearAttachments(commandBuffer, attachmentCount, pAttachments, rectCount, pRects); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdResolveImage(VkCommandBuffer commandBuffer, VkImage srcImage, VkImageLayout srcImageLayout, VkImage dstImage, VkImageLayout dstImageLayout, uint32_t regionCount, const VkImageResolve *pRegions) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdResolveImage(commandBuffer, srcImage, srcImageLayout, dstImage, dstImageLayout, regionCount, pRegions); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdSetEvent(VkCommandBuffer commandBuffer, VkEvent event, VkPipelineStageFlags stageMask) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdSetEvent(commandBuffer, event, stageMask); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdResetEvent(VkCommandBuffer commandBuffer, VkEvent event, VkPipelineStageFlags stageMask) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdResetEvent(commandBuffer, event, stageMask); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdWaitEvents(VkCommandBuffer commandBuffer, uint32_t eventCount, const VkEvent *pEvents, VkPipelineStageFlags sourceStageMask, VkPipelineStageFlags dstStageMask, uint32_t memoryBarrierCount, const VkMemoryBarrier *pMemoryBarriers, uint32_t bufferMemoryBarrierCount, const VkBufferMemoryBarrier *pBufferMemoryBarriers, uint32_t imageMemoryBarrierCount, const VkImageMemoryBarrier *pImageMemoryBarriers) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdWaitEvents(commandBuffer, eventCount, pEvents, sourceStageMask, dstStageMask, memoryBarrierCount, pMemoryBarriers, bufferMemoryBarrierCount, pBufferMemoryBarriers, imageMemoryBarrierCount, pImageMemoryBarriers); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdPipelineBarrier(VkCommandBuffer commandBuffer, VkPipelineStageFlags srcStageMask, VkPipelineStageFlags dstStageMask, VkDependencyFlags dependencyFlags, uint32_t memoryBarrierCount, const VkMemoryBarrier *pMemoryBarriers, uint32_t bufferMemoryBarrierCount, const VkBufferMemoryBarrier *pBufferMemoryBarriers, uint32_t imageMemoryBarrierCount, const VkImageMemoryBarrier *pImageMemoryBarriers) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdPipelineBarrier(commandBuffer, srcStageMask, dstStageMask, dependencyFlags, memoryBarrierCount, pMemoryBarriers, bufferMemoryBarrierCount, pBufferMemoryBarriers, imageMemoryBarrierCount, pImageMemoryBarriers); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdBeginQuery(VkCommandBuffer commandBuffer, VkQueryPool queryPool, uint32_t slot, VkFlags flags) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdBeginQuery(commandBuffer, queryPool, slot, flags); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdEndQuery(VkCommandBuffer commandBuffer, VkQueryPool queryPool, uint32_t slot) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdEndQuery(commandBuffer, queryPool, slot); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdResetQueryPool(VkCommandBuffer commandBuffer, VkQueryPool queryPool, uint32_t firstQuery, uint32_t queryCount) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdResetQueryPool(commandBuffer, queryPool, firstQuery, queryCount); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdWriteTimestamp(VkCommandBuffer commandBuffer, VkPipelineStageFlagBits pipelineStage, VkQueryPool queryPool, uint32_t slot) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdWriteTimestamp(commandBuffer, pipelineStage, queryPool, slot); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdCopyQueryPoolResults(VkCommandBuffer commandBuffer, VkQueryPool queryPool, uint32_t firstQuery, uint32_t queryCount, VkBuffer dstBuffer, VkDeviceSize dstOffset, VkDeviceSize stride, VkFlags flags) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdCopyQueryPoolResults(commandBuffer, queryPool, firstQuery, queryCount, dstBuffer, dstOffset, stride, flags); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdPushConstants(VkCommandBuffer commandBuffer, VkPipelineLayout layout, VkShaderStageFlags stageFlags, uint32_t offset, uint32_t size, const void *pValues) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdPushConstants(commandBuffer, layout, stageFlags, offset, size, pValues); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdBeginRenderPass(VkCommandBuffer commandBuffer, const VkRenderPassBeginInfo *pRenderPassBegin, VkSubpassContents contents) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdBeginRenderPass(commandBuffer, pRenderPassBegin, contents); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdNextSubpass(VkCommandBuffer commandBuffer, VkSubpassContents contents) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdNextSubpass(commandBuffer, contents); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdEndRenderPass(VkCommandBuffer commandBuffer) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdEndRenderPass(commandBuffer); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdExecuteCommands(VkCommandBuffer commandBuffer, uint32_t commandBuffersCount, const VkCommandBuffer *pCommandBuffers) { const VkLayerDispatchTable *disp; disp = loader_get_dispatch(commandBuffer); disp->CmdExecuteCommands(commandBuffer, commandBuffersCount, pCommandBuffers); } // ---- Vulkan core 1.1 trampolines VkResult setupLoaderTrampPhysDevGroups(VkInstance instance) { VkResult res = VK_SUCCESS; struct loader_instance *inst; uint32_t total_count = 0; VkPhysicalDeviceGroupPropertiesKHR **new_phys_dev_groups = NULL; VkPhysicalDeviceGroupPropertiesKHR *local_phys_dev_groups = NULL; PFN_vkEnumeratePhysicalDeviceGroups fpEnumeratePhysicalDeviceGroups = NULL; inst = loader_get_instance(instance); if (NULL == inst) { res = VK_ERROR_INITIALIZATION_FAILED; goto out; } // Get the function pointer to use to call into the ICD. This could be the core or KHR version if (inst->enabled_known_extensions.khr_device_group_creation) { fpEnumeratePhysicalDeviceGroups = inst->disp->layer_inst_disp.EnumeratePhysicalDeviceGroupsKHR; } else { fpEnumeratePhysicalDeviceGroups = inst->disp->layer_inst_disp.EnumeratePhysicalDeviceGroups; } // Setup the trampoline loader physical devices. This will actually // call down and setup the terminator loader physical devices during the // process. VkResult setup_res = setupLoaderTrampPhysDevs(instance); if (setup_res != VK_SUCCESS && setup_res != VK_INCOMPLETE) { res = setup_res; goto out; } // Query how many physical device groups there res = fpEnumeratePhysicalDeviceGroups(instance, &total_count, NULL); if (res != VK_SUCCESS) { loader_log(inst, VK_DEBUG_REPORT_ERROR_BIT_EXT, 0, "setupLoaderTrampPhysDevGroups: Failed during dispatch call of " "\'EnumeratePhysicalDeviceGroupsKHR\' to lower layers or " "loader to get count."); goto out; } // Create an array for the new physical device groups, which will be stored // in the instance for the trampoline code. new_phys_dev_groups = (VkPhysicalDeviceGroupPropertiesKHR **)loader_instance_heap_alloc( inst, total_count * sizeof(VkPhysicalDeviceGroupPropertiesKHR *), VK_SYSTEM_ALLOCATION_SCOPE_INSTANCE); if (NULL == new_phys_dev_groups) { loader_log(inst, VK_DEBUG_REPORT_ERROR_BIT_EXT, 0, "setupLoaderTrampPhysDevGroups: Failed to allocate new physical device" " group array of size %d", total_count); res = VK_ERROR_OUT_OF_HOST_MEMORY; goto out; } memset(new_phys_dev_groups, 0, total_count * sizeof(VkPhysicalDeviceGroupPropertiesKHR *)); // Create a temporary array (on the stack) to keep track of the // returned VkPhysicalDevice values. local_phys_dev_groups = loader_stack_alloc(sizeof(VkPhysicalDeviceGroupPropertiesKHR) * total_count); if (NULL == local_phys_dev_groups) { loader_log(inst, VK_DEBUG_REPORT_ERROR_BIT_EXT, 0, "setupLoaderTrampPhysDevGroups: Failed to allocate local " "physical device group array of size %d", total_count); res = VK_ERROR_OUT_OF_HOST_MEMORY; goto out; } // Initialize the memory to something valid memset(local_phys_dev_groups, 0, sizeof(VkPhysicalDeviceGroupPropertiesKHR) * total_count); for (uint32_t group = 0; group < total_count; group++) { local_phys_dev_groups[group].sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_GROUP_PROPERTIES_KHR; local_phys_dev_groups[group].pNext = NULL; local_phys_dev_groups[group].subsetAllocation = false; } // Call down and get the content fpEnumeratePhysicalDeviceGroups(instance, &total_count, local_phys_dev_groups); if (VK_SUCCESS != res) { loader_log(inst, VK_DEBUG_REPORT_ERROR_BIT_EXT, 0, "setupLoaderTrampPhysDevGroups: Failed during dispatch call of " "\'EnumeratePhysicalDeviceGroupsKHR\' to lower layers or " "loader to get content."); goto out; } // Replace all the physical device IDs with the proper loader values for (uint32_t group = 0; group < total_count; group++) { for (uint32_t group_gpu = 0; group_gpu < local_phys_dev_groups[group].physicalDeviceCount; group_gpu++) { bool found = false; for (uint32_t tramp_gpu = 0; tramp_gpu < inst->phys_dev_count_tramp; tramp_gpu++) { if (local_phys_dev_groups[group].physicalDevices[group_gpu] == inst->phys_devs_tramp[tramp_gpu]->phys_dev) { local_phys_dev_groups[group].physicalDevices[group_gpu] = (VkPhysicalDevice)inst->phys_devs_tramp[tramp_gpu]; found = true; break; } } if (!found) { loader_log(inst, VK_DEBUG_REPORT_ERROR_BIT_EXT, 0, "setupLoaderTrampPhysDevGroups: Failed to find GPU %d in group %d" " returned by \'EnumeratePhysicalDeviceGroupsKHR\' in list returned" " by \'EnumeratePhysicalDevices\'", group_gpu, group); res = VK_ERROR_INITIALIZATION_FAILED; goto out; } } } // Copy or create everything to fill the new array of physical device groups for (uint32_t new_idx = 0; new_idx < total_count; new_idx++) { // Check if this physical device group with the same contents is already in the old buffer for (uint32_t old_idx = 0; old_idx < inst->phys_dev_group_count_tramp; old_idx++) { if (local_phys_dev_groups[new_idx].physicalDeviceCount == inst->phys_dev_groups_tramp[old_idx]->physicalDeviceCount) { bool found_all_gpus = true; for (uint32_t old_gpu = 0; old_gpu < inst->phys_dev_groups_tramp[old_idx]->physicalDeviceCount; old_gpu++) { bool found_gpu = false; for (uint32_t new_gpu = 0; new_gpu < local_phys_dev_groups[new_idx].physicalDeviceCount; new_gpu++) { if (local_phys_dev_groups[new_idx].physicalDevices[new_gpu] == inst->phys_dev_groups_tramp[old_idx]->physicalDevices[old_gpu]) { found_gpu = true; break; } } if (!found_gpu) { found_all_gpus = false; break; } } if (!found_all_gpus) { continue; } else { new_phys_dev_groups[new_idx] = inst->phys_dev_groups_tramp[old_idx]; break; } } } // If this physical device group isn't in the old buffer, create it if (NULL == new_phys_dev_groups[new_idx]) { new_phys_dev_groups[new_idx] = (VkPhysicalDeviceGroupPropertiesKHR *)loader_instance_heap_alloc( inst, sizeof(VkPhysicalDeviceGroupPropertiesKHR), VK_SYSTEM_ALLOCATION_SCOPE_INSTANCE); if (NULL == new_phys_dev_groups[new_idx]) { loader_log(inst, VK_DEBUG_REPORT_ERROR_BIT_EXT, 0, "setupLoaderTrampPhysDevGroups: Failed to allocate " "physical device group trampoline object %d", new_idx); total_count = new_idx; res = VK_ERROR_OUT_OF_HOST_MEMORY; goto out; } memcpy(new_phys_dev_groups[new_idx], &local_phys_dev_groups[new_idx], sizeof(VkPhysicalDeviceGroupPropertiesKHR)); } } out: if (VK_SUCCESS != res) { if (NULL != new_phys_dev_groups) { for (uint32_t i = 0; i < total_count; i++) { loader_instance_heap_free(inst, new_phys_dev_groups[i]); } loader_instance_heap_free(inst, new_phys_dev_groups); } total_count = 0; } else { // Free everything that didn't carry over to the new array of // physical device groups if (NULL != inst->phys_dev_groups_tramp) { for (uint32_t i = 0; i < inst->phys_dev_group_count_tramp; i++) { bool found = false; for (uint32_t j = 0; j < total_count; j++) { if (inst->phys_dev_groups_tramp[i] == new_phys_dev_groups[j]) { found = true; break; } } if (!found) { loader_instance_heap_free(inst, inst->phys_dev_groups_tramp[i]); } } loader_instance_heap_free(inst, inst->phys_dev_groups_tramp); } // Swap in the new physical device group list inst->phys_dev_group_count_tramp = total_count; inst->phys_dev_groups_tramp = new_phys_dev_groups; } return res; } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkEnumeratePhysicalDeviceGroups( VkInstance instance, uint32_t *pPhysicalDeviceGroupCount, VkPhysicalDeviceGroupProperties *pPhysicalDeviceGroupProperties) { VkResult res = VK_SUCCESS; uint32_t count; uint32_t i; struct loader_instance *inst = NULL; loader_platform_thread_lock_mutex(&loader_lock); inst = loader_get_instance(instance); if (NULL == inst) { res = VK_ERROR_INITIALIZATION_FAILED; goto out; } if (NULL == pPhysicalDeviceGroupCount) { loader_log(inst, VK_DEBUG_REPORT_ERROR_BIT_EXT, 0, "vkEnumeratePhysicalDeviceGroupsKHR: Received NULL pointer for physical " "device group count return value."); res = VK_ERROR_INITIALIZATION_FAILED; goto out; } VkResult setup_res = setupLoaderTrampPhysDevGroups(instance); if (VK_SUCCESS != setup_res) { res = setup_res; goto out; } count = inst->phys_dev_group_count_tramp; // Wrap the PhysDev object for loader usage, return wrapped objects if (NULL != pPhysicalDeviceGroupProperties) { if (inst->phys_dev_group_count_tramp > *pPhysicalDeviceGroupCount) { loader_log(inst, VK_DEBUG_REPORT_INFORMATION_BIT_EXT, 0, "vkEnumeratePhysicalDeviceGroupsKHR: Trimming device group count down" " by application request from %d to %d physical device groups", inst->phys_dev_group_count_tramp, *pPhysicalDeviceGroupCount); count = *pPhysicalDeviceGroupCount; res = VK_INCOMPLETE; } for (i = 0; i < count; i++) { memcpy(&pPhysicalDeviceGroupProperties[i], inst->phys_dev_groups_tramp[i], sizeof(VkPhysicalDeviceGroupPropertiesKHR)); } } *pPhysicalDeviceGroupCount = count; out: loader_platform_thread_unlock_mutex(&loader_lock); return res; } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkGetPhysicalDeviceFeatures2(VkPhysicalDevice physicalDevice, VkPhysicalDeviceFeatures2 *pFeatures) { VkPhysicalDevice unwrapped_phys_dev = loader_unwrap_physical_device(physicalDevice); const VkLayerInstanceDispatchTable *disp = loader_get_instance_layer_dispatch(physicalDevice); const struct loader_instance *inst = ((struct loader_physical_device_tramp*) physicalDevice)->this_instance; if (inst != NULL && inst->enabled_known_extensions.khr_get_physical_device_properties2) { disp->GetPhysicalDeviceFeatures2KHR(unwrapped_phys_dev, pFeatures); } else { disp->GetPhysicalDeviceFeatures2(unwrapped_phys_dev, pFeatures); } } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkGetPhysicalDeviceProperties2(VkPhysicalDevice physicalDevice, VkPhysicalDeviceProperties2 *pProperties) { VkPhysicalDevice unwrapped_phys_dev = loader_unwrap_physical_device(physicalDevice); const VkLayerInstanceDispatchTable *disp = loader_get_instance_layer_dispatch(physicalDevice); const struct loader_instance *inst = ((struct loader_physical_device_tramp*) physicalDevice)->this_instance; if (inst != NULL && inst->enabled_known_extensions.khr_get_physical_device_properties2) { disp->GetPhysicalDeviceProperties2KHR(unwrapped_phys_dev, pProperties); } else { disp->GetPhysicalDeviceProperties2(unwrapped_phys_dev, pProperties); } } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkGetPhysicalDeviceFormatProperties2(VkPhysicalDevice physicalDevice, VkFormat format, VkFormatProperties2 *pFormatProperties) { VkPhysicalDevice unwrapped_phys_dev = loader_unwrap_physical_device(physicalDevice); const VkLayerInstanceDispatchTable *disp = loader_get_instance_layer_dispatch(physicalDevice); const struct loader_instance *inst = ((struct loader_physical_device_tramp*) physicalDevice)->this_instance; if (inst != NULL && inst->enabled_known_extensions.khr_get_physical_device_properties2) { disp->GetPhysicalDeviceFormatProperties2KHR(unwrapped_phys_dev, format, pFormatProperties); } else { disp->GetPhysicalDeviceFormatProperties2(unwrapped_phys_dev, format, pFormatProperties); } } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkGetPhysicalDeviceImageFormatProperties2( VkPhysicalDevice physicalDevice, const VkPhysicalDeviceImageFormatInfo2 *pImageFormatInfo, VkImageFormatProperties2 *pImageFormatProperties) { VkPhysicalDevice unwrapped_phys_dev = loader_unwrap_physical_device(physicalDevice); const VkLayerInstanceDispatchTable *disp = loader_get_instance_layer_dispatch(physicalDevice); const struct loader_instance *inst = ((struct loader_physical_device_tramp*) physicalDevice)->this_instance; if (inst != NULL && inst->enabled_known_extensions.khr_get_physical_device_properties2) { return disp->GetPhysicalDeviceImageFormatProperties2KHR(unwrapped_phys_dev, pImageFormatInfo, pImageFormatProperties); } else { return disp->GetPhysicalDeviceImageFormatProperties2(unwrapped_phys_dev, pImageFormatInfo, pImageFormatProperties); } } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkGetPhysicalDeviceQueueFamilyProperties2(VkPhysicalDevice physicalDevice, uint32_t *pQueueFamilyPropertyCount, VkQueueFamilyProperties2 *pQueueFamilyProperties) { VkPhysicalDevice unwrapped_phys_dev = loader_unwrap_physical_device(physicalDevice); const VkLayerInstanceDispatchTable *disp = loader_get_instance_layer_dispatch(physicalDevice); const struct loader_instance *inst = ((struct loader_physical_device_tramp*) physicalDevice)->this_instance; if (inst != NULL && inst->enabled_known_extensions.khr_get_physical_device_properties2) { disp->GetPhysicalDeviceQueueFamilyProperties2KHR(unwrapped_phys_dev, pQueueFamilyPropertyCount, pQueueFamilyProperties); } else { disp->GetPhysicalDeviceQueueFamilyProperties2(unwrapped_phys_dev, pQueueFamilyPropertyCount, pQueueFamilyProperties); } } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkGetPhysicalDeviceMemoryProperties2(VkPhysicalDevice physicalDevice, VkPhysicalDeviceMemoryProperties2 *pMemoryProperties) { VkPhysicalDevice unwrapped_phys_dev = loader_unwrap_physical_device(physicalDevice); const VkLayerInstanceDispatchTable *disp = loader_get_instance_layer_dispatch(physicalDevice); const struct loader_instance *inst = ((struct loader_physical_device_tramp*) physicalDevice)->this_instance; if (inst != NULL && inst->enabled_known_extensions.khr_get_physical_device_properties2) { disp->GetPhysicalDeviceMemoryProperties2KHR(unwrapped_phys_dev, pMemoryProperties); } else { disp->GetPhysicalDeviceMemoryProperties2(unwrapped_phys_dev, pMemoryProperties); } } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkGetPhysicalDeviceSparseImageFormatProperties2( VkPhysicalDevice physicalDevice, const VkPhysicalDeviceSparseImageFormatInfo2 *pFormatInfo, uint32_t *pPropertyCount, VkSparseImageFormatProperties2 *pProperties) { VkPhysicalDevice unwrapped_phys_dev = loader_unwrap_physical_device(physicalDevice); const VkLayerInstanceDispatchTable *disp = loader_get_instance_layer_dispatch(physicalDevice); const struct loader_instance *inst = ((struct loader_physical_device_tramp*) physicalDevice)->this_instance; if (inst != NULL && inst->enabled_known_extensions.khr_get_physical_device_properties2) { disp->GetPhysicalDeviceSparseImageFormatProperties2KHR(unwrapped_phys_dev, pFormatInfo, pPropertyCount, pProperties); } else { disp->GetPhysicalDeviceSparseImageFormatProperties2(unwrapped_phys_dev, pFormatInfo, pPropertyCount, pProperties); } } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkGetPhysicalDeviceExternalBufferProperties( VkPhysicalDevice physicalDevice, const VkPhysicalDeviceExternalBufferInfo *pExternalBufferInfo, VkExternalBufferProperties *pExternalBufferProperties) { VkPhysicalDevice unwrapped_phys_dev = loader_unwrap_physical_device(physicalDevice); const VkLayerInstanceDispatchTable *disp = loader_get_instance_layer_dispatch(physicalDevice); const struct loader_instance *inst = ((struct loader_physical_device_tramp*) physicalDevice)->this_instance; if (inst != NULL && inst->enabled_known_extensions.khr_external_memory_capabilities){ disp->GetPhysicalDeviceExternalBufferPropertiesKHR(unwrapped_phys_dev, pExternalBufferInfo, pExternalBufferProperties); } else { disp->GetPhysicalDeviceExternalBufferProperties(unwrapped_phys_dev, pExternalBufferInfo, pExternalBufferProperties); } } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkGetPhysicalDeviceExternalSemaphoreProperties( VkPhysicalDevice physicalDevice, const VkPhysicalDeviceExternalSemaphoreInfoKHR *pExternalSemaphoreInfo, VkExternalSemaphoreProperties *pExternalSemaphoreProperties) { VkPhysicalDevice unwrapped_phys_dev = loader_unwrap_physical_device(physicalDevice); const VkLayerInstanceDispatchTable *disp = loader_get_instance_layer_dispatch(physicalDevice); const struct loader_instance *inst = ((struct loader_physical_device_tramp*) physicalDevice)->this_instance; if (inst != NULL && inst->enabled_known_extensions.khr_external_semaphore_capabilities) { disp->GetPhysicalDeviceExternalSemaphorePropertiesKHR(unwrapped_phys_dev, pExternalSemaphoreInfo, pExternalSemaphoreProperties); } else { disp->GetPhysicalDeviceExternalSemaphoreProperties(unwrapped_phys_dev, pExternalSemaphoreInfo, pExternalSemaphoreProperties); } } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkGetPhysicalDeviceExternalFenceProperties( VkPhysicalDevice physicalDevice, const VkPhysicalDeviceExternalFenceInfo *pExternalFenceInfo, VkExternalFenceProperties *pExternalFenceProperties) { VkPhysicalDevice unwrapped_phys_dev = loader_unwrap_physical_device(physicalDevice); const VkLayerInstanceDispatchTable *disp = loader_get_instance_layer_dispatch(physicalDevice); const struct loader_instance *inst = ((struct loader_physical_device_tramp*) physicalDevice)->this_instance; if (inst != NULL && inst->enabled_known_extensions.khr_external_fence_capabilities) { disp->GetPhysicalDeviceExternalFencePropertiesKHR(unwrapped_phys_dev, pExternalFenceInfo, pExternalFenceProperties); } else { disp->GetPhysicalDeviceExternalFenceProperties(unwrapped_phys_dev, pExternalFenceInfo, pExternalFenceProperties); } } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkBindBufferMemory2( VkDevice device, uint32_t bindInfoCount, const VkBindBufferMemoryInfo* pBindInfos) { const VkLayerDispatchTable *disp = loader_get_dispatch(device); return disp->BindBufferMemory2(device, bindInfoCount, pBindInfos); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkBindImageMemory2( VkDevice device, uint32_t bindInfoCount, const VkBindImageMemoryInfo* pBindInfos) { const VkLayerDispatchTable *disp = loader_get_dispatch(device); return disp->BindImageMemory2(device, bindInfoCount, pBindInfos); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkGetDeviceGroupPeerMemoryFeatures( VkDevice device, uint32_t heapIndex, uint32_t localDeviceIndex, uint32_t remoteDeviceIndex, VkPeerMemoryFeatureFlags* pPeerMemoryFeatures) { const VkLayerDispatchTable *disp = loader_get_dispatch(device); disp->GetDeviceGroupPeerMemoryFeatures(device, heapIndex, localDeviceIndex, remoteDeviceIndex, pPeerMemoryFeatures); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdSetDeviceMask( VkCommandBuffer commandBuffer, uint32_t deviceMask) { const VkLayerDispatchTable *disp = loader_get_dispatch(commandBuffer); disp->CmdSetDeviceMask(commandBuffer, deviceMask); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdDispatchBase( VkCommandBuffer commandBuffer, uint32_t baseGroupX, uint32_t baseGroupY, uint32_t baseGroupZ, uint32_t groupCountX, uint32_t groupCountY, uint32_t groupCountZ) { const VkLayerDispatchTable *disp = loader_get_dispatch(commandBuffer); disp->CmdDispatchBase(commandBuffer, baseGroupX, baseGroupY, baseGroupZ, groupCountX, groupCountY, groupCountZ); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkGetImageMemoryRequirements2( VkDevice device, const VkImageMemoryRequirementsInfo2* pInfo, VkMemoryRequirements2* pMemoryRequirements) { const VkLayerDispatchTable *disp = loader_get_dispatch(device); disp->GetImageMemoryRequirements2(device, pInfo, pMemoryRequirements); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkGetBufferMemoryRequirements2( VkDevice device, const VkBufferMemoryRequirementsInfo2* pInfo, VkMemoryRequirements2* pMemoryRequirements) { const VkLayerDispatchTable *disp = loader_get_dispatch(device); disp->GetBufferMemoryRequirements2(device, pInfo, pMemoryRequirements); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkGetImageSparseMemoryRequirements2( VkDevice device, const VkImageSparseMemoryRequirementsInfo2* pInfo, uint32_t* pSparseMemoryRequirementCount, VkSparseImageMemoryRequirements2* pSparseMemoryRequirements) { const VkLayerDispatchTable *disp = loader_get_dispatch(device); disp->GetImageSparseMemoryRequirements2(device, pInfo, pSparseMemoryRequirementCount, pSparseMemoryRequirements); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkTrimCommandPool( VkDevice device, VkCommandPool commandPool, VkCommandPoolTrimFlags flags) { const VkLayerDispatchTable *disp = loader_get_dispatch(device); disp->TrimCommandPool(device, commandPool, flags); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkGetDeviceQueue2(VkDevice device, const VkDeviceQueueInfo2 *pQueueInfo, VkQueue *pQueue) { const VkLayerDispatchTable *disp = loader_get_dispatch(device); disp->GetDeviceQueue2(device, pQueueInfo, pQueue); if (*pQueue != VK_NULL_HANDLE) { loader_set_dispatch(*pQueue, disp); } } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkCreateSamplerYcbcrConversion( VkDevice device, const VkSamplerYcbcrConversionCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkSamplerYcbcrConversion* pYcbcrConversion) { const VkLayerDispatchTable *disp = loader_get_dispatch(device); return disp->CreateSamplerYcbcrConversion(device, pCreateInfo, pAllocator, pYcbcrConversion); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkDestroySamplerYcbcrConversion( VkDevice device, VkSamplerYcbcrConversion ycbcrConversion, const VkAllocationCallbacks* pAllocator) { const VkLayerDispatchTable *disp = loader_get_dispatch(device); disp->DestroySamplerYcbcrConversion(device, ycbcrConversion, pAllocator); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkGetDescriptorSetLayoutSupport( VkDevice device, const VkDescriptorSetLayoutCreateInfo* pCreateInfo, VkDescriptorSetLayoutSupport* pSupport) { const VkLayerDispatchTable *disp = loader_get_dispatch(device); disp->GetDescriptorSetLayoutSupport(device, pCreateInfo, pSupport); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkCreateDescriptorUpdateTemplate(VkDevice device, const VkDescriptorUpdateTemplateCreateInfo *pCreateInfo, const VkAllocationCallbacks *pAllocator, VkDescriptorUpdateTemplate *pDescriptorUpdateTemplate) { const VkLayerDispatchTable *disp = loader_get_dispatch(device); return disp->CreateDescriptorUpdateTemplate(device, pCreateInfo, pAllocator, pDescriptorUpdateTemplate); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkDestroyDescriptorUpdateTemplate(VkDevice device, VkDescriptorUpdateTemplate descriptorUpdateTemplate, const VkAllocationCallbacks *pAllocator) { const VkLayerDispatchTable *disp = loader_get_dispatch(device); disp->DestroyDescriptorUpdateTemplate(device, descriptorUpdateTemplate, pAllocator); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkUpdateDescriptorSetWithTemplate(VkDevice device, VkDescriptorSet descriptorSet, VkDescriptorUpdateTemplate descriptorUpdateTemplate, const void *pData) { const VkLayerDispatchTable *disp = loader_get_dispatch(device); disp->UpdateDescriptorSetWithTemplate(device, descriptorSet, descriptorUpdateTemplate, pData); } // ---- Vulkan core 1.2 trampolines LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkCreateRenderPass2(VkDevice device, const VkRenderPassCreateInfo2* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkRenderPass* pRenderPass) { const VkLayerDispatchTable *disp = loader_get_dispatch(device); return disp->CreateRenderPass2(device, pCreateInfo, pAllocator, pRenderPass); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdBeginRenderPass2(VkCommandBuffer commandBuffer, const VkRenderPassBeginInfo* pRenderPassBegin, const VkSubpassBeginInfo* pSubpassBeginInfo) { const VkLayerDispatchTable *disp = loader_get_dispatch(commandBuffer); disp->CmdBeginRenderPass2(commandBuffer, pRenderPassBegin, pSubpassBeginInfo); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdNextSubpass2(VkCommandBuffer commandBuffer, const VkSubpassBeginInfo* pSubpassBeginInfo, const VkSubpassEndInfo* pSubpassEndInfo) { const VkLayerDispatchTable *disp = loader_get_dispatch(commandBuffer); disp->CmdNextSubpass2(commandBuffer, pSubpassBeginInfo, pSubpassEndInfo); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdEndRenderPass2(VkCommandBuffer commandBuffer, const VkSubpassEndInfo* pSubpassEndInfo) { const VkLayerDispatchTable *disp = loader_get_dispatch(commandBuffer); disp->CmdEndRenderPass2(commandBuffer, pSubpassEndInfo); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdDrawIndirectCount(VkCommandBuffer commandBuffer, VkBuffer buffer, VkDeviceSize offset, VkBuffer countBuffer, VkDeviceSize countBufferOffset, uint32_t maxDrawCount, uint32_t stride) { const VkLayerDispatchTable *disp = loader_get_dispatch(commandBuffer); disp->CmdDrawIndirectCount(commandBuffer, buffer, offset, countBuffer, countBufferOffset, maxDrawCount, stride); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkCmdDrawIndexedIndirectCount(VkCommandBuffer commandBuffer, VkBuffer buffer, VkDeviceSize offset, VkBuffer countBuffer, VkDeviceSize countBufferOffset, uint32_t maxDrawCount, uint32_t stride) { const VkLayerDispatchTable *disp = loader_get_dispatch(commandBuffer); disp->CmdDrawIndexedIndirectCount(commandBuffer, buffer, offset, countBuffer, countBufferOffset, maxDrawCount, stride); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkGetSemaphoreCounterValue(VkDevice device, VkSemaphore semaphore, uint64_t* pValue) { const VkLayerDispatchTable *disp = loader_get_dispatch(device); return disp->GetSemaphoreCounterValue(device, semaphore, pValue); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkWaitSemaphores(VkDevice device, const VkSemaphoreWaitInfo* pWaitInfo, uint64_t timeout) { const VkLayerDispatchTable *disp = loader_get_dispatch(device); return disp->WaitSemaphores(device, pWaitInfo, timeout); } LOADER_EXPORT VKAPI_ATTR VkResult VKAPI_CALL vkSignalSemaphore(VkDevice device, const VkSemaphoreSignalInfo* pSignalInfo) { const VkLayerDispatchTable *disp = loader_get_dispatch(device); return disp->SignalSemaphore(device, pSignalInfo); } LOADER_EXPORT VKAPI_ATTR VkDeviceAddress VKAPI_CALL vkGetBufferDeviceAddress(VkDevice device, const VkBufferDeviceAddressInfo* pInfo) { const VkLayerDispatchTable *disp = loader_get_dispatch(device); return disp->GetBufferDeviceAddress(device, pInfo); } LOADER_EXPORT VKAPI_ATTR uint64_t VKAPI_CALL vkGetBufferOpaqueCaptureAddress(VkDevice device, const VkBufferDeviceAddressInfo* pInfo) { const VkLayerDispatchTable *disp = loader_get_dispatch(device); return disp->GetBufferOpaqueCaptureAddress(device, pInfo); } LOADER_EXPORT VKAPI_ATTR uint64_t VKAPI_CALL vkGetDeviceMemoryOpaqueCaptureAddress(VkDevice device, const VkDeviceMemoryOpaqueCaptureAddressInfo* pInfo) { const VkLayerDispatchTable *disp = loader_get_dispatch(device); return disp->GetDeviceMemoryOpaqueCaptureAddress(device, pInfo); } LOADER_EXPORT VKAPI_ATTR void VKAPI_CALL vkResetQueryPool(VkDevice device, VkQueryPool queryPool, uint32_t firstQuery, uint32_t queryCount) { const VkLayerDispatchTable *disp = loader_get_dispatch(device); disp->ResetQueryPool(device, queryPool, firstQuery, queryCount); }