/* Copyright (c) 2015-2019 The Khronos Group Inc. * Copyright (c) 2015-2019 Valve Corporation * Copyright (c) 2015-2019 LunarG, Inc. * Copyright (C) 2015-2019 Google Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. * * Author: Mark Lobodzinski * Author: Jon Ashburn * Author: Tobin Ehlis */ // shared_mutex support added in MSVC 2015 update 2 #if defined(_MSC_FULL_VER) && _MSC_FULL_VER >= 190023918 #include typedef std::shared_mutex object_lifetime_mutex_t; typedef std::shared_lock read_object_lifetime_mutex_t; typedef std::unique_lock write_object_lifetime_mutex_t; #else typedef std::mutex object_lifetime_mutex_t; typedef std::unique_lock read_object_lifetime_mutex_t; typedef std::unique_lock write_object_lifetime_mutex_t; #endif // Suppress unused warning on Linux #if defined(__GNUC__) #define DECORATE_UNUSED __attribute__((unused)) #else #define DECORATE_UNUSED #endif // clang-format off static const char DECORATE_UNUSED *kVUID_ObjectTracker_Info = "UNASSIGNED-ObjectTracker-Info"; static const char DECORATE_UNUSED *kVUID_ObjectTracker_InternalError = "UNASSIGNED-ObjectTracker-InternalError"; static const char DECORATE_UNUSED *kVUID_ObjectTracker_ObjectLeak = "UNASSIGNED-ObjectTracker-ObjectLeak"; static const char DECORATE_UNUSED *kVUID_ObjectTracker_UnknownObject = "UNASSIGNED-ObjectTracker-UnknownObject"; // clang-format on #undef DECORATE_UNUSED extern uint64_t object_track_index; // Object Status -- used to track state of individual objects typedef VkFlags ObjectStatusFlags; enum ObjectStatusFlagBits { OBJSTATUS_NONE = 0x00000000, // No status is set OBJSTATUS_FENCE_IS_SUBMITTED = 0x00000001, // Fence has been submitted OBJSTATUS_VIEWPORT_BOUND = 0x00000002, // Viewport state object has been bound OBJSTATUS_RASTER_BOUND = 0x00000004, // Viewport state object has been bound OBJSTATUS_COLOR_BLEND_BOUND = 0x00000008, // Viewport state object has been bound OBJSTATUS_DEPTH_STENCIL_BOUND = 0x00000010, // Viewport state object has been bound OBJSTATUS_GPU_MEM_MAPPED = 0x00000020, // Memory object is currently mapped OBJSTATUS_COMMAND_BUFFER_SECONDARY = 0x00000040, // Command Buffer is of type SECONDARY OBJSTATUS_CUSTOM_ALLOCATOR = 0x00000080, // Allocated with custom allocator }; // Object and state information structure struct ObjTrackState { uint64_t handle; // Object handle (new) VulkanObjectType object_type; // Object type identifier ObjectStatusFlags status; // Object state uint64_t parent_object; // Parent object std::unique_ptr > child_objects; // Child objects (used for VkDescriptorPool only) }; // Track Queue information struct ObjTrackQueueInfo { uint32_t queue_node_index; VkQueue queue; }; typedef std::unordered_map object_map_type; class ObjectLifetimes : public ValidationObject { public: // Override chassis read/write locks for this validation object // This override takes a deferred lock. i.e. it is not acquired. // This class does its own locking with a shared mutex. virtual std::unique_lock write_lock() { return std::unique_lock(validation_object_mutex, std::defer_lock); } object_lifetime_mutex_t object_lifetime_mutex; write_object_lifetime_mutex_t write_shared_lock() { return write_object_lifetime_mutex_t(object_lifetime_mutex); } read_object_lifetime_mutex_t read_shared_lock() { return read_object_lifetime_mutex_t(object_lifetime_mutex); } uint64_t num_objects[kVulkanObjectTypeMax + 1]; uint64_t num_total_objects; // Vector of unordered_maps per object type to hold ObjTrackState info std::vector object_map; // Special-case map for swapchain images std::unordered_map swapchainImageMap; // Map of queue information structures, one per queue std::unordered_map queue_info_map; std::vector queue_family_properties; // Constructor for object lifetime tracking ObjectLifetimes() : num_objects{}, num_total_objects(0), object_map{} { object_map.resize(kVulkanObjectTypeMax + 1); } bool DeviceReportUndestroyedObjects(VkDevice device, VulkanObjectType object_type, const std::string &error_code); void DeviceDestroyUndestroyedObjects(VkDevice device, VulkanObjectType object_type); void CreateQueue(VkDevice device, VkQueue vkObj); void AddQueueInfo(VkDevice device, uint32_t queue_node_index, VkQueue queue); void ValidateQueueFlags(VkQueue queue, const char *function); void AllocateCommandBuffer(VkDevice device, const VkCommandPool command_pool, const VkCommandBuffer command_buffer, VkCommandBufferLevel level); void AllocateDescriptorSet(VkDevice device, VkDescriptorPool descriptor_pool, VkDescriptorSet descriptor_set); void CreateSwapchainImageObject(VkDevice dispatchable_object, VkImage swapchain_image, VkSwapchainKHR swapchain); bool ReportUndestroyedObjects(VkDevice device, const std::string &error_code); void DestroyUndestroyedObjects(VkDevice device); bool ValidateDeviceObject(const VulkanTypedHandle &device_typed, const char *invalid_handle_code, const char *wrong_device_code); void DestroyQueueDataStructures(VkDevice device); bool ValidateCommandBuffer(VkDevice device, VkCommandPool command_pool, VkCommandBuffer command_buffer); bool ValidateDescriptorSet(VkDevice device, VkDescriptorPool descriptor_pool, VkDescriptorSet descriptor_set); bool ValidateSamplerObjects(VkDevice device, const VkDescriptorSetLayoutCreateInfo *pCreateInfo); template bool ValidateDescriptorWrite(DispObj disp, VkWriteDescriptorSet const *desc, bool isPush); ObjectLifetimes *GetObjectLifetimeData(std::vector &object_dispatch) { for (auto layer_object : object_dispatch) { if (layer_object->container_type == LayerObjectTypeObjectTracker) { return (reinterpret_cast(layer_object)); } } return nullptr; }; template bool ValidateObject(T1 dispatchable_object, T2 object, VulkanObjectType object_type, bool null_allowed, const char *invalid_handle_code, const char *wrong_device_code) { if (null_allowed && (object == VK_NULL_HANDLE)) { return false; } auto object_handle = HandleToUint64(object); if (object_type == kVulkanObjectTypeDevice) { return ValidateDeviceObject(VulkanTypedHandle(object, object_type), invalid_handle_code, wrong_device_code); } VkDebugReportObjectTypeEXT debug_object_type = get_debug_report_enum[object_type]; // Look for object in object map if (object_map[object_type].find(object_handle) == object_map[object_type].end()) { // If object is an image, also look for it in the swapchain image map if ((object_type != kVulkanObjectTypeImage) || (swapchainImageMap.find(object_handle) == swapchainImageMap.end())) { // Object not found, look for it in other device object maps for (auto other_device_data : layer_data_map) { for (auto layer_object_data : other_device_data.second->object_dispatch) { if (layer_object_data->container_type == LayerObjectTypeObjectTracker) { auto object_lifetime_data = reinterpret_cast(layer_object_data); if (object_lifetime_data && (object_lifetime_data != this)) { if (object_lifetime_data->object_map[object_type].find(object_handle) != object_lifetime_data->object_map[object_type].end() || (object_type == kVulkanObjectTypeImage && object_lifetime_data->swapchainImageMap.find(object_handle) != object_lifetime_data->swapchainImageMap.end())) { // Object found on other device, report an error if object has a device parent error code if ((wrong_device_code != kVUIDUndefined) && (object_type != kVulkanObjectTypeSurfaceKHR)) { return log_msg(report_data, VK_DEBUG_REPORT_ERROR_BIT_EXT, debug_object_type, object_handle, wrong_device_code, "Object 0x%" PRIxLEAST64 " was not created, allocated or retrieved from the correct device.", object_handle); } else { return false; } } } } } } // Report an error if object was not found anywhere return log_msg(report_data, VK_DEBUG_REPORT_ERROR_BIT_EXT, debug_object_type, object_handle, invalid_handle_code, "Invalid %s Object 0x%" PRIxLEAST64 ".", object_string[object_type], object_handle); } } return false; } template void CreateObject(T1 dispatchable_object, T2 object, VulkanObjectType object_type, const VkAllocationCallbacks *pAllocator) { uint64_t object_handle = HandleToUint64(object); bool custom_allocator = (pAllocator != nullptr); if (!object_map[object_type].count(object_handle)) { ObjTrackState *pNewObjNode = new ObjTrackState; pNewObjNode->object_type = object_type; pNewObjNode->status = custom_allocator ? OBJSTATUS_CUSTOM_ALLOCATOR : OBJSTATUS_NONE; pNewObjNode->handle = object_handle; object_map[object_type][object_handle] = pNewObjNode; num_objects[object_type]++; num_total_objects++; if (object_type == kVulkanObjectTypeDescriptorPool) { pNewObjNode->child_objects.reset(new std::unordered_set); } } } template void DestroyObjectSilently(T1 object, VulkanObjectType object_type) { auto object_handle = HandleToUint64(object); assert(object_handle != VK_NULL_HANDLE); auto item = object_map[object_type].find(object_handle); assert(item != object_map[object_type].end()); ObjTrackState *pNode = item->second; assert(num_total_objects > 0); num_total_objects--; assert(num_objects[pNode->object_type] > 0); num_objects[pNode->object_type]--; delete pNode; object_map[object_type].erase(item); } template void RecordDestroyObject(T1 dispatchable_object, T2 object, VulkanObjectType object_type) { auto object_handle = HandleToUint64(object); if (object_handle != VK_NULL_HANDLE) { auto item = object_map[object_type].find(object_handle); if (item != object_map[object_type].end()) { DestroyObjectSilently(object, object_type); } } } template bool ValidateDestroyObject(T1 dispatchable_object, T2 object, VulkanObjectType object_type, const VkAllocationCallbacks *pAllocator, const char *expected_custom_allocator_code, const char *expected_default_allocator_code) { auto object_handle = HandleToUint64(object); bool custom_allocator = pAllocator != nullptr; VkDebugReportObjectTypeEXT debug_object_type = get_debug_report_enum[object_type]; bool skip = false; if (object_handle != VK_NULL_HANDLE) { auto item = object_map[object_type].find(object_handle); if (item != object_map[object_type].end()) { ObjTrackState *pNode = item->second; auto allocated_with_custom = (pNode->status & OBJSTATUS_CUSTOM_ALLOCATOR) ? true : false; if (allocated_with_custom && !custom_allocator && expected_custom_allocator_code != kVUIDUndefined) { // This check only verifies that custom allocation callbacks were provided to both Create and Destroy calls, // it cannot verify that these allocation callbacks are compatible with each other. skip |= log_msg(report_data, VK_DEBUG_REPORT_ERROR_BIT_EXT, debug_object_type, object_handle, expected_custom_allocator_code, "Custom allocator not specified while destroying %s obj 0x%" PRIxLEAST64 " but specified at creation.", object_string[object_type], object_handle); } else if (!allocated_with_custom && custom_allocator && expected_default_allocator_code != kVUIDUndefined) { skip |= log_msg(report_data, VK_DEBUG_REPORT_ERROR_BIT_EXT, debug_object_type, object_handle, expected_default_allocator_code, "Custom allocator specified while destroying %s obj 0x%" PRIxLEAST64 " but not specified at creation.", object_string[object_type], object_handle); } } } return skip; } #include "object_tracker.h" };