1/* Copyright (c) 2015-2016 The Khronos Group Inc. 2 * Copyright (c) 2015-2016 Valve Corporation 3 * Copyright (c) 2015-2016 LunarG, Inc. 4 * Copyright (C) 2015-2016 Google Inc. 5 * 6 * Licensed under the Apache License, Version 2.0 (the "License"); 7 * you may not use this file except in compliance with the License. 8 * You may obtain a copy of the License at 9 * 10 * http://www.apache.org/licenses/LICENSE-2.0 11 * 12 * Unless required by applicable law or agreed to in writing, software 13 * distributed under the License is distributed on an "AS IS" BASIS, 14 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 15 * See the License for the specific language governing permissions and 16 * limitations under the License. 17 * 18 * Author: Tobin Ehlis <tobine@google.com> 19 * John Zulauf <jzulauf@lunarg.com> 20 */ 21 22// Allow use of STL min and max functions in Windows 23#define NOMINMAX 24 25#include "descriptor_sets.h" 26#include "vk_enum_string_helper.h" 27#include "vk_safe_struct.h" 28#include "buffer_validation.h" 29#include <sstream> 30#include <algorithm> 31 32struct BindingNumCmp { 33 bool operator()(const VkDescriptorSetLayoutBinding *a, const VkDescriptorSetLayoutBinding *b) const { 34 return a->binding < b->binding; 35 } 36}; 37 38// Construct DescriptorSetLayout instance from given create info 39// Proactively reserve and resize as possible, as the reallocation was visible in profiling 40cvdescriptorset::DescriptorSetLayout::DescriptorSetLayout(const VkDescriptorSetLayoutCreateInfo *p_create_info, 41 const VkDescriptorSetLayout layout) 42 : layout_(layout), 43 layout_destroyed_(false), 44 flags_(p_create_info->flags), 45 binding_count_(0), 46 descriptor_count_(0), 47 dynamic_descriptor_count_(0) { 48 binding_type_stats_ = {0, 0, 0}; 49 std::set<const VkDescriptorSetLayoutBinding *, BindingNumCmp> sorted_bindings; 50 const uint32_t input_bindings_count = p_create_info->bindingCount; 51 // Sort the input bindings in binding number order, eliminating duplicates 52 for (uint32_t i = 0; i < input_bindings_count; i++) { 53 sorted_bindings.insert(p_create_info->pBindings + i); 54 } 55 56 // Store the create info in the sorted order from above 57 std::map<uint32_t, uint32_t> binding_to_dyn_count; 58 uint32_t index = 0; 59 binding_count_ = static_cast<uint32_t>(sorted_bindings.size()); 60 bindings_.reserve(binding_count_); 61 binding_to_index_map_.reserve(binding_count_); 62 for (auto input_binding : sorted_bindings) { 63 // Add to binding and map, s.t. it is robust to invalid duplication of binding_num 64 const auto binding_num = input_binding->binding; 65 binding_to_index_map_[binding_num] = index++; 66 bindings_.emplace_back(input_binding); 67 auto &binding_info = bindings_.back(); 68 69 descriptor_count_ += binding_info.descriptorCount; 70 if (binding_info.descriptorCount > 0) { 71 non_empty_bindings_.insert(binding_num); 72 } 73 74 if (binding_info.descriptorType == VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER_DYNAMIC || 75 binding_info.descriptorType == VK_DESCRIPTOR_TYPE_STORAGE_BUFFER_DYNAMIC) { 76 binding_to_dyn_count[binding_num] = binding_info.descriptorCount; 77 dynamic_descriptor_count_ += binding_info.descriptorCount; 78 binding_type_stats_.dynamic_buffer_count++; 79 } else if ((binding_info.descriptorType == VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER) || 80 (binding_info.descriptorType == VK_DESCRIPTOR_TYPE_STORAGE_BUFFER)) { 81 binding_type_stats_.non_dynamic_buffer_count++; 82 } else { 83 binding_type_stats_.image_sampler_count++; 84 } 85 } 86 assert(bindings_.size() == binding_count_); 87 uint32_t global_index = 0; 88 binding_to_global_index_range_map_.reserve(binding_count_); 89 // Vector order is finalized so create maps of bindings to descriptors and descriptors to indices 90 for (uint32_t i = 0; i < binding_count_; ++i) { 91 auto binding_num = bindings_[i].binding; 92 auto final_index = global_index + bindings_[i].descriptorCount; 93 binding_to_global_index_range_map_[binding_num] = IndexRange(global_index, final_index); 94 if (final_index != global_index) { 95 global_start_to_index_map_[global_index] = i; 96 } 97 global_index = final_index; 98 } 99 100 // Now create dyn offset array mapping for any dynamic descriptors 101 uint32_t dyn_array_idx = 0; 102 binding_to_dynamic_array_idx_map_.reserve(binding_to_dyn_count.size()); 103 for (const auto &bc_pair : binding_to_dyn_count) { 104 binding_to_dynamic_array_idx_map_[bc_pair.first] = dyn_array_idx; 105 dyn_array_idx += bc_pair.second; 106 } 107} 108 109// Validate descriptor set layout create info 110bool cvdescriptorset::DescriptorSetLayout::ValidateCreateInfo(const debug_report_data *report_data, 111 const VkDescriptorSetLayoutCreateInfo *create_info, 112 const bool push_descriptor_ext, const uint32_t max_push_descriptors) { 113 bool skip = false; 114 std::unordered_set<uint32_t> bindings; 115 uint64_t total_descriptors = 0; 116 117 const bool push_descriptor_set = create_info->flags & VK_DESCRIPTOR_SET_LAYOUT_CREATE_PUSH_DESCRIPTOR_BIT_KHR; 118 if (push_descriptor_set && !push_descriptor_ext) { 119 skip |= log_msg(report_data, VK_DEBUG_REPORT_ERROR_BIT_EXT, VK_DEBUG_REPORT_OBJECT_TYPE_UNKNOWN_EXT, 0, __LINE__, 120 DRAWSTATE_EXTENSION_NOT_ENABLED, "DS", 121 "Attemped to use %s in %s but its required extension %s has not been enabled.\n", 122 "VK_DESCRIPTOR_SET_LAYOUT_CREATE_PUSH_DESCRIPTOR_BIT_KHR", "VkDescriptorSetLayoutCreateInfo::flags", 123 VK_KHR_PUSH_DESCRIPTOR_EXTENSION_NAME); 124 } 125 126 auto valid_type = [push_descriptor_set](const VkDescriptorType type) { 127 return !push_descriptor_set || 128 ((type != VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER_DYNAMIC) && (type != VK_DESCRIPTOR_TYPE_STORAGE_BUFFER_DYNAMIC)); 129 }; 130 131 for (uint32_t i = 0; i < create_info->bindingCount; ++i) { 132 const auto &binding_info = create_info->pBindings[i]; 133 if (!bindings.insert(binding_info.binding).second) { 134 skip |= log_msg(report_data, VK_DEBUG_REPORT_ERROR_BIT_EXT, VK_DEBUG_REPORT_OBJECT_TYPE_UNKNOWN_EXT, 0, __LINE__, 135 VALIDATION_ERROR_0500022e, "DS", "duplicated binding number in VkDescriptorSetLayoutBinding. %s", 136 validation_error_map[VALIDATION_ERROR_0500022e]); 137 } 138 if (!valid_type(binding_info.descriptorType)) { 139 skip |= 140 log_msg(report_data, VK_DEBUG_REPORT_ERROR_BIT_EXT, VK_DEBUG_REPORT_OBJECT_TYPE_UNKNOWN_EXT, 0, __LINE__, 141 VALIDATION_ERROR_05000230, "DS", 142 "invalid type %s ,for push descriptors in VkDescriptorSetLayoutBinding entry %" PRIu32 ". %s", 143 string_VkDescriptorType(binding_info.descriptorType), i, validation_error_map[VALIDATION_ERROR_05000230]); 144 } 145 total_descriptors += binding_info.descriptorCount; 146 } 147 148 if ((push_descriptor_set) && (total_descriptors > max_push_descriptors)) { 149 const char *undefined = push_descriptor_ext ? "" : " -- undefined"; 150 skip |= log_msg(report_data, VK_DEBUG_REPORT_ERROR_BIT_EXT, VK_DEBUG_REPORT_OBJECT_TYPE_UNKNOWN_EXT, 0, __LINE__, 151 VALIDATION_ERROR_05000232, "DS", 152 "for push descriptor, total descriptor count in layout (%" PRIu64 153 ") must not be greater than VkPhysicalDevicePushDescriptorPropertiesKHR::maxPushDescriptors (%" PRIu32 154 "%s). %s", 155 total_descriptors, max_push_descriptors, undefined, validation_error_map[VALIDATION_ERROR_05000232]); 156 } 157 158 return skip; 159} 160 161// Return valid index or "end" i.e. binding_count_; 162// The asserts in "Get" are reduced to the set where no valid answer(like null or 0) could be given 163// Common code for all binding lookups. 164uint32_t cvdescriptorset::DescriptorSetLayout::GetIndexFromBinding(uint32_t binding) const { 165 const auto &bi_itr = binding_to_index_map_.find(binding); 166 if (bi_itr != binding_to_index_map_.cend()) return bi_itr->second; 167 return GetBindingCount(); 168} 169VkDescriptorSetLayoutBinding const *cvdescriptorset::DescriptorSetLayout::GetDescriptorSetLayoutBindingPtrFromIndex( 170 const uint32_t index) const { 171 if (index >= bindings_.size()) return nullptr; 172 return bindings_[index].ptr(); 173} 174// Return descriptorCount for given index, 0 if index is unavailable 175uint32_t cvdescriptorset::DescriptorSetLayout::GetDescriptorCountFromIndex(const uint32_t index) const { 176 if (index >= bindings_.size()) return 0; 177 return bindings_[index].descriptorCount; 178} 179// For the given index, return descriptorType 180VkDescriptorType cvdescriptorset::DescriptorSetLayout::GetTypeFromIndex(const uint32_t index) const { 181 assert(index < bindings_.size()); 182 if (index < bindings_.size()) return bindings_[index].descriptorType; 183 return VK_DESCRIPTOR_TYPE_MAX_ENUM; 184} 185// For the given index, return stageFlags 186VkShaderStageFlags cvdescriptorset::DescriptorSetLayout::GetStageFlagsFromIndex(const uint32_t index) const { 187 assert(index < bindings_.size()); 188 if (index < bindings_.size()) return bindings_[index].stageFlags; 189 return VkShaderStageFlags(0); 190} 191 192// For the given global index, return index 193uint32_t cvdescriptorset::DescriptorSetLayout::GetIndexFromGlobalIndex(const uint32_t global_index) const { 194 auto start_it = global_start_to_index_map_.upper_bound(global_index); 195 uint32_t index = binding_count_; 196 assert(start_it != global_start_to_index_map_.cbegin()); 197 if (start_it != global_start_to_index_map_.cbegin()) { 198 --start_it; 199 index = start_it->second; 200#ifndef NDEBUG 201 const auto &range = GetGlobalIndexRangeFromBinding(bindings_[index].binding); 202 assert(range.start <= global_index && global_index < range.end); 203#endif 204 } 205 return index; 206} 207 208// For the given binding, return the global index range 209// As start and end are often needed in pairs, get both with a single hash lookup. 210const cvdescriptorset::IndexRange &cvdescriptorset::DescriptorSetLayout::GetGlobalIndexRangeFromBinding( 211 const uint32_t binding) const { 212 assert(binding_to_global_index_range_map_.count(binding)); 213 // In error case max uint32_t so index is out of bounds to break ASAP 214 const static IndexRange kInvalidRange = {0xFFFFFFFF, 0xFFFFFFFF}; 215 const auto &range_it = binding_to_global_index_range_map_.find(binding); 216 if (range_it != binding_to_global_index_range_map_.end()) { 217 return range_it->second; 218 } 219 return kInvalidRange; 220} 221 222// For given binding, return ptr to ImmutableSampler array 223VkSampler const *cvdescriptorset::DescriptorSetLayout::GetImmutableSamplerPtrFromBinding(const uint32_t binding) const { 224 const auto &bi_itr = binding_to_index_map_.find(binding); 225 if (bi_itr != binding_to_index_map_.end()) { 226 return bindings_[bi_itr->second].pImmutableSamplers; 227 } 228 return nullptr; 229} 230// Move to next valid binding having a non-zero binding count 231uint32_t cvdescriptorset::DescriptorSetLayout::GetNextValidBinding(const uint32_t binding) const { 232 auto it = non_empty_bindings_.upper_bound(binding); 233 assert(it != non_empty_bindings_.cend()); 234 if (it != non_empty_bindings_.cend()) return *it; 235 return GetMaxBinding() + 1; 236} 237// For given index, return ptr to ImmutableSampler array 238VkSampler const *cvdescriptorset::DescriptorSetLayout::GetImmutableSamplerPtrFromIndex(const uint32_t index) const { 239 if (index < bindings_.size()) { 240 return bindings_[index].pImmutableSamplers; 241 } 242 return nullptr; 243} 244// If our layout is compatible with rh_ds_layout, return true, 245// else return false and fill in error_msg will description of what causes incompatibility 246bool cvdescriptorset::DescriptorSetLayout::IsCompatible(DescriptorSetLayout const *const rh_ds_layout, 247 std::string *error_msg) const { 248 // Trivial case 249 if (layout_ == rh_ds_layout->GetDescriptorSetLayout()) return true; 250 if (descriptor_count_ != rh_ds_layout->descriptor_count_) { 251 std::stringstream error_str; 252 error_str << "DescriptorSetLayout " << layout_ << " has " << descriptor_count_ << " descriptors, but DescriptorSetLayout " 253 << rh_ds_layout->GetDescriptorSetLayout() << ", which comes from pipelineLayout, has " 254 << rh_ds_layout->descriptor_count_ << " descriptors."; 255 *error_msg = error_str.str(); 256 return false; // trivial fail case 257 } 258 // Descriptor counts match so need to go through bindings one-by-one 259 // and verify that type and stageFlags match 260 for (auto binding : bindings_) { 261 // TODO : Do we also need to check immutable samplers? 262 // VkDescriptorSetLayoutBinding *rh_binding; 263 if (binding.descriptorCount != rh_ds_layout->GetDescriptorCountFromBinding(binding.binding)) { 264 std::stringstream error_str; 265 error_str << "Binding " << binding.binding << " for DescriptorSetLayout " << layout_ << " has a descriptorCount of " 266 << binding.descriptorCount << " but binding " << binding.binding << " for DescriptorSetLayout " 267 << rh_ds_layout->GetDescriptorSetLayout() << ", which comes from pipelineLayout, has a descriptorCount of " 268 << rh_ds_layout->GetDescriptorCountFromBinding(binding.binding); 269 *error_msg = error_str.str(); 270 return false; 271 } else if (binding.descriptorType != rh_ds_layout->GetTypeFromBinding(binding.binding)) { 272 std::stringstream error_str; 273 error_str << "Binding " << binding.binding << " for DescriptorSetLayout " << layout_ << " is type '" 274 << string_VkDescriptorType(binding.descriptorType) << "' but binding " << binding.binding 275 << " for DescriptorSetLayout " << rh_ds_layout->GetDescriptorSetLayout() 276 << ", which comes from pipelineLayout, is type '" 277 << string_VkDescriptorType(rh_ds_layout->GetTypeFromBinding(binding.binding)) << "'"; 278 *error_msg = error_str.str(); 279 return false; 280 } else if (binding.stageFlags != rh_ds_layout->GetStageFlagsFromBinding(binding.binding)) { 281 std::stringstream error_str; 282 error_str << "Binding " << binding.binding << " for DescriptorSetLayout " << layout_ << " has stageFlags " 283 << binding.stageFlags << " but binding " << binding.binding << " for DescriptorSetLayout " 284 << rh_ds_layout->GetDescriptorSetLayout() << ", which comes from pipelineLayout, has stageFlags " 285 << rh_ds_layout->GetStageFlagsFromBinding(binding.binding); 286 *error_msg = error_str.str(); 287 return false; 288 } 289 } 290 return true; 291} 292 293bool cvdescriptorset::DescriptorSetLayout::IsNextBindingConsistent(const uint32_t binding) const { 294 if (!binding_to_index_map_.count(binding + 1)) return false; 295 auto const &bi_itr = binding_to_index_map_.find(binding); 296 if (bi_itr != binding_to_index_map_.end()) { 297 const auto &next_bi_itr = binding_to_index_map_.find(binding + 1); 298 if (next_bi_itr != binding_to_index_map_.end()) { 299 auto type = bindings_[bi_itr->second].descriptorType; 300 auto stage_flags = bindings_[bi_itr->second].stageFlags; 301 auto immut_samp = bindings_[bi_itr->second].pImmutableSamplers ? true : false; 302 if ((type != bindings_[next_bi_itr->second].descriptorType) || 303 (stage_flags != bindings_[next_bi_itr->second].stageFlags) || 304 (immut_samp != (bindings_[next_bi_itr->second].pImmutableSamplers ? true : false))) { 305 return false; 306 } 307 return true; 308 } 309 } 310 return false; 311} 312// Starting at offset descriptor of given binding, parse over update_count 313// descriptor updates and verify that for any binding boundaries that are crossed, the next binding(s) are all consistent 314// Consistency means that their type, stage flags, and whether or not they use immutable samplers matches 315// If so, return true. If not, fill in error_msg and return false 316bool cvdescriptorset::DescriptorSetLayout::VerifyUpdateConsistency(uint32_t current_binding, uint32_t offset, uint32_t update_count, 317 const char *type, const VkDescriptorSet set, 318 std::string *error_msg) const { 319 // Verify consecutive bindings match (if needed) 320 auto orig_binding = current_binding; 321 // Track count of descriptors in the current_bindings that are remaining to be updated 322 auto binding_remaining = GetDescriptorCountFromBinding(current_binding); 323 // First, it's legal to offset beyond your own binding so handle that case 324 // Really this is just searching for the binding in which the update begins and adjusting offset accordingly 325 while (offset >= binding_remaining) { 326 // Advance to next binding, decrement offset by binding size 327 offset -= binding_remaining; 328 binding_remaining = GetDescriptorCountFromBinding(++current_binding); 329 } 330 binding_remaining -= offset; 331 while (update_count > binding_remaining) { // While our updates overstep current binding 332 // Verify next consecutive binding matches type, stage flags & immutable sampler use 333 if (!IsNextBindingConsistent(current_binding++)) { 334 std::stringstream error_str; 335 error_str << "Attempting " << type << " descriptor set " << set << " binding #" << orig_binding << " with #" 336 << update_count 337 << " descriptors being updated but this update oversteps the bounds of this binding and the next binding is " 338 "not consistent with current binding so this update is invalid."; 339 *error_msg = error_str.str(); 340 return false; 341 } 342 // For sake of this check consider the bindings updated and grab count for next binding 343 update_count -= binding_remaining; 344 binding_remaining = GetDescriptorCountFromBinding(current_binding); 345 } 346 return true; 347} 348 349cvdescriptorset::AllocateDescriptorSetsData::AllocateDescriptorSetsData(uint32_t count) 350 : required_descriptors_by_type{}, layout_nodes(count, nullptr) {} 351 352cvdescriptorset::DescriptorSet::DescriptorSet(const VkDescriptorSet set, const VkDescriptorPool pool, 353 const std::shared_ptr<DescriptorSetLayout const> &layout, layer_data *dev_data) 354 : some_update_(false), 355 set_(set), 356 pool_state_(nullptr), 357 p_layout_(layout), 358 device_data_(dev_data), 359 limits_(GetPhysDevProperties(dev_data)->properties.limits) { 360 pool_state_ = GetDescriptorPoolState(dev_data, pool); 361 // Foreach binding, create default descriptors of given type 362 descriptors_.reserve(p_layout_->GetTotalDescriptorCount()); 363 for (uint32_t i = 0; i < p_layout_->GetBindingCount(); ++i) { 364 auto type = p_layout_->GetTypeFromIndex(i); 365 switch (type) { 366 case VK_DESCRIPTOR_TYPE_SAMPLER: { 367 auto immut_sampler = p_layout_->GetImmutableSamplerPtrFromIndex(i); 368 for (uint32_t di = 0; di < p_layout_->GetDescriptorCountFromIndex(i); ++di) { 369 if (immut_sampler) { 370 descriptors_.emplace_back(new SamplerDescriptor(immut_sampler + di)); 371 some_update_ = true; // Immutable samplers are updated at creation 372 } else 373 descriptors_.emplace_back(new SamplerDescriptor(nullptr)); 374 } 375 break; 376 } 377 case VK_DESCRIPTOR_TYPE_COMBINED_IMAGE_SAMPLER: { 378 auto immut = p_layout_->GetImmutableSamplerPtrFromIndex(i); 379 for (uint32_t di = 0; di < p_layout_->GetDescriptorCountFromIndex(i); ++di) { 380 if (immut) { 381 descriptors_.emplace_back(new ImageSamplerDescriptor(immut + di)); 382 some_update_ = true; // Immutable samplers are updated at creation 383 } else 384 descriptors_.emplace_back(new ImageSamplerDescriptor(nullptr)); 385 } 386 break; 387 } 388 // ImageDescriptors 389 case VK_DESCRIPTOR_TYPE_SAMPLED_IMAGE: 390 case VK_DESCRIPTOR_TYPE_INPUT_ATTACHMENT: 391 case VK_DESCRIPTOR_TYPE_STORAGE_IMAGE: 392 for (uint32_t di = 0; di < p_layout_->GetDescriptorCountFromIndex(i); ++di) 393 descriptors_.emplace_back(new ImageDescriptor(type)); 394 break; 395 case VK_DESCRIPTOR_TYPE_UNIFORM_TEXEL_BUFFER: 396 case VK_DESCRIPTOR_TYPE_STORAGE_TEXEL_BUFFER: 397 for (uint32_t di = 0; di < p_layout_->GetDescriptorCountFromIndex(i); ++di) 398 descriptors_.emplace_back(new TexelDescriptor(type)); 399 break; 400 case VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER: 401 case VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER_DYNAMIC: 402 case VK_DESCRIPTOR_TYPE_STORAGE_BUFFER: 403 case VK_DESCRIPTOR_TYPE_STORAGE_BUFFER_DYNAMIC: 404 for (uint32_t di = 0; di < p_layout_->GetDescriptorCountFromIndex(i); ++di) 405 descriptors_.emplace_back(new BufferDescriptor(type)); 406 break; 407 default: 408 assert(0); // Bad descriptor type specified 409 break; 410 } 411 } 412} 413 414cvdescriptorset::DescriptorSet::~DescriptorSet() { InvalidateBoundCmdBuffers(); } 415 416static std::string string_descriptor_req_view_type(descriptor_req req) { 417 std::string result(""); 418 for (unsigned i = 0; i <= VK_IMAGE_VIEW_TYPE_END_RANGE; i++) { 419 if (req & (1 << i)) { 420 if (result.size()) result += ", "; 421 result += string_VkImageViewType(VkImageViewType(i)); 422 } 423 } 424 425 if (!result.size()) result = "(none)"; 426 427 return result; 428} 429 430// Is this sets underlying layout compatible with passed in layout according to "Pipeline Layout Compatibility" in spec? 431bool cvdescriptorset::DescriptorSet::IsCompatible(DescriptorSetLayout const *const layout, std::string *error) const { 432 return layout->IsCompatible(p_layout_.get(), error); 433} 434 435// Validate that the state of this set is appropriate for the given bindings and dynamic_offsets at Draw time 436// This includes validating that all descriptors in the given bindings are updated, 437// that any update buffers are valid, and that any dynamic offsets are within the bounds of their buffers. 438// Return true if state is acceptable, or false and write an error message into error string 439bool cvdescriptorset::DescriptorSet::ValidateDrawState(const std::map<uint32_t, descriptor_req> &bindings, 440 const std::vector<uint32_t> &dynamic_offsets, GLOBAL_CB_NODE *cb_node, 441 const char *caller, std::string *error) const { 442 for (auto binding_pair : bindings) { 443 auto binding = binding_pair.first; 444 if (!p_layout_->HasBinding(binding)) { 445 std::stringstream error_str; 446 error_str << "Attempting to validate DrawState for binding #" << binding 447 << " which is an invalid binding for this descriptor set."; 448 *error = error_str.str(); 449 return false; 450 } 451 IndexRange index_range = p_layout_->GetGlobalIndexRangeFromBinding(binding); 452 auto array_idx = 0; // Track array idx if we're dealing with array descriptors 453 for (uint32_t i = index_range.start; i < index_range.end; ++i, ++array_idx) { 454 if (!descriptors_[i]->updated) { 455 std::stringstream error_str; 456 error_str << "Descriptor in binding #" << binding << " at global descriptor index " << i 457 << " is being used in draw but has not been updated."; 458 *error = error_str.str(); 459 return false; 460 } else { 461 auto descriptor_class = descriptors_[i]->GetClass(); 462 if (descriptor_class == GeneralBuffer) { 463 // Verify that buffers are valid 464 auto buffer = static_cast<BufferDescriptor *>(descriptors_[i].get())->GetBuffer(); 465 auto buffer_node = GetBufferState(device_data_, buffer); 466 if (!buffer_node) { 467 std::stringstream error_str; 468 error_str << "Descriptor in binding #" << binding << " at global descriptor index " << i 469 << " references invalid buffer " << buffer << "."; 470 *error = error_str.str(); 471 return false; 472 } else if (!buffer_node->sparse) { 473 for (auto mem_binding : buffer_node->GetBoundMemory()) { 474 if (!GetMemObjInfo(device_data_, mem_binding)) { 475 std::stringstream error_str; 476 error_str << "Descriptor in binding #" << binding << " at global descriptor index " << i 477 << " uses buffer " << buffer << " that references invalid memory " << mem_binding << "."; 478 *error = error_str.str(); 479 return false; 480 } 481 } 482 } else { 483 // Enqueue sparse resource validation, as these can only be validated at submit time 484 auto device_data_copy = device_data_; // Cannot capture members by value, so make capturable copy. 485 std::function<bool(void)> function = [device_data_copy, caller, buffer_node]() { 486 return core_validation::ValidateBufferMemoryIsValid(device_data_copy, buffer_node, caller); 487 }; 488 cb_node->queue_submit_functions.push_back(function); 489 } 490 if (descriptors_[i]->IsDynamic()) { 491 // Validate that dynamic offsets are within the buffer 492 auto buffer_size = buffer_node->createInfo.size; 493 auto range = static_cast<BufferDescriptor *>(descriptors_[i].get())->GetRange(); 494 auto desc_offset = static_cast<BufferDescriptor *>(descriptors_[i].get())->GetOffset(); 495 auto dyn_offset = dynamic_offsets[GetDynamicOffsetIndexFromBinding(binding) + array_idx]; 496 if (VK_WHOLE_SIZE == range) { 497 if ((dyn_offset + desc_offset) > buffer_size) { 498 std::stringstream error_str; 499 error_str << "Dynamic descriptor in binding #" << binding << " at global descriptor index " << i 500 << " uses buffer " << buffer << " with update range of VK_WHOLE_SIZE has dynamic offset " 501 << dyn_offset << " combined with offset " << desc_offset 502 << " that oversteps the buffer size of " << buffer_size << "."; 503 *error = error_str.str(); 504 return false; 505 } 506 } else { 507 if ((dyn_offset + desc_offset + range) > buffer_size) { 508 std::stringstream error_str; 509 error_str << "Dynamic descriptor in binding #" << binding << " at global descriptor index " << i 510 << " uses buffer " << buffer << " with dynamic offset " << dyn_offset 511 << " combined with offset " << desc_offset << " and range " << range 512 << " that oversteps the buffer size of " << buffer_size << "."; 513 *error = error_str.str(); 514 return false; 515 } 516 } 517 } 518 } else if (descriptor_class == ImageSampler || descriptor_class == Image) { 519 VkImageView image_view; 520 VkImageLayout image_layout; 521 if (descriptor_class == ImageSampler) { 522 image_view = static_cast<ImageSamplerDescriptor *>(descriptors_[i].get())->GetImageView(); 523 image_layout = static_cast<ImageSamplerDescriptor *>(descriptors_[i].get())->GetImageLayout(); 524 } else { 525 image_view = static_cast<ImageDescriptor *>(descriptors_[i].get())->GetImageView(); 526 image_layout = static_cast<ImageDescriptor *>(descriptors_[i].get())->GetImageLayout(); 527 } 528 auto reqs = binding_pair.second; 529 530 auto image_view_state = GetImageViewState(device_data_, image_view); 531 if (nullptr == image_view_state) { 532 // Image view must have been destroyed since initial update. Could potentially flag the descriptor 533 // as "invalid" (updated = false) at DestroyImageView() time and detect this error at bind time 534 std::stringstream error_str; 535 error_str << "Descriptor in binding #" << binding << " at global descriptor index " << i 536 << " is using imageView " << image_view << " that has been destroyed."; 537 *error = error_str.str(); 538 return false; 539 } 540 auto image_view_ci = image_view_state->create_info; 541 542 if ((reqs & DESCRIPTOR_REQ_ALL_VIEW_TYPE_BITS) && (~reqs & (1 << image_view_ci.viewType))) { 543 // bad view type 544 std::stringstream error_str; 545 error_str << "Descriptor in binding #" << binding << " at global descriptor index " << i 546 << " requires an image view of type " << string_descriptor_req_view_type(reqs) << " but got " 547 << string_VkImageViewType(image_view_ci.viewType) << "."; 548 *error = error_str.str(); 549 return false; 550 } 551 552 auto image_node = GetImageState(device_data_, image_view_ci.image); 553 assert(image_node); 554 // Verify Image Layout 555 // Copy first mip level into sub_layers and loop over each mip level to verify layout 556 VkImageSubresourceLayers sub_layers; 557 sub_layers.aspectMask = image_view_ci.subresourceRange.aspectMask; 558 sub_layers.baseArrayLayer = image_view_ci.subresourceRange.baseArrayLayer; 559 sub_layers.layerCount = image_view_ci.subresourceRange.layerCount; 560 bool hit_error = false; 561 for (auto cur_level = image_view_ci.subresourceRange.baseMipLevel; 562 cur_level < image_view_ci.subresourceRange.levelCount; ++cur_level) { 563 sub_layers.mipLevel = cur_level; 564 VerifyImageLayout(device_data_, cb_node, image_node, sub_layers, image_layout, VK_IMAGE_LAYOUT_UNDEFINED, 565 caller, VALIDATION_ERROR_046002b0, &hit_error); 566 if (hit_error) { 567 *error = 568 "Image layout specified at vkUpdateDescriptorSets() time doesn't match actual image layout at time " 569 "descriptor is used. See previous error callback for specific details."; 570 return false; 571 } 572 } 573 // Verify Sample counts 574 if ((reqs & DESCRIPTOR_REQ_SINGLE_SAMPLE) && image_node->createInfo.samples != VK_SAMPLE_COUNT_1_BIT) { 575 std::stringstream error_str; 576 error_str << "Descriptor in binding #" << binding << " at global descriptor index " << i 577 << " requires bound image to have VK_SAMPLE_COUNT_1_BIT but got " 578 << string_VkSampleCountFlagBits(image_node->createInfo.samples) << "."; 579 *error = error_str.str(); 580 return false; 581 } 582 if ((reqs & DESCRIPTOR_REQ_MULTI_SAMPLE) && image_node->createInfo.samples == VK_SAMPLE_COUNT_1_BIT) { 583 std::stringstream error_str; 584 error_str << "Descriptor in binding #" << binding << " at global descriptor index " << i 585 << " requires bound image to have multiple samples, but got VK_SAMPLE_COUNT_1_BIT."; 586 *error = error_str.str(); 587 return false; 588 } 589 } 590 } 591 } 592 } 593 return true; 594} 595 596// For given bindings, place any update buffers or images into the passed-in unordered_sets 597uint32_t cvdescriptorset::DescriptorSet::GetStorageUpdates(const std::map<uint32_t, descriptor_req> &bindings, 598 std::unordered_set<VkBuffer> *buffer_set, 599 std::unordered_set<VkImageView> *image_set) const { 600 auto num_updates = 0; 601 for (auto binding_pair : bindings) { 602 auto binding = binding_pair.first; 603 // If a binding doesn't exist, skip it 604 if (!p_layout_->HasBinding(binding)) { 605 continue; 606 } 607 uint32_t start_idx = p_layout_->GetGlobalIndexRangeFromBinding(binding).start; 608 if (descriptors_[start_idx]->IsStorage()) { 609 if (Image == descriptors_[start_idx]->descriptor_class) { 610 for (uint32_t i = 0; i < p_layout_->GetDescriptorCountFromBinding(binding); ++i) { 611 if (descriptors_[start_idx + i]->updated) { 612 image_set->insert(static_cast<ImageDescriptor *>(descriptors_[start_idx + i].get())->GetImageView()); 613 num_updates++; 614 } 615 } 616 } else if (TexelBuffer == descriptors_[start_idx]->descriptor_class) { 617 for (uint32_t i = 0; i < p_layout_->GetDescriptorCountFromBinding(binding); ++i) { 618 if (descriptors_[start_idx + i]->updated) { 619 auto bufferview = static_cast<TexelDescriptor *>(descriptors_[start_idx + i].get())->GetBufferView(); 620 auto bv_state = GetBufferViewState(device_data_, bufferview); 621 if (bv_state) { 622 buffer_set->insert(bv_state->create_info.buffer); 623 num_updates++; 624 } 625 } 626 } 627 } else if (GeneralBuffer == descriptors_[start_idx]->descriptor_class) { 628 for (uint32_t i = 0; i < p_layout_->GetDescriptorCountFromBinding(binding); ++i) { 629 if (descriptors_[start_idx + i]->updated) { 630 buffer_set->insert(static_cast<BufferDescriptor *>(descriptors_[start_idx + i].get())->GetBuffer()); 631 num_updates++; 632 } 633 } 634 } 635 } 636 } 637 return num_updates; 638} 639// Set is being deleted or updates so invalidate all bound cmd buffers 640void cvdescriptorset::DescriptorSet::InvalidateBoundCmdBuffers() { 641 core_validation::invalidateCommandBuffers(device_data_, cb_bindings, {HandleToUint64(set_), kVulkanObjectTypeDescriptorSet}); 642} 643// Perform write update in given update struct 644void cvdescriptorset::DescriptorSet::PerformWriteUpdate(const VkWriteDescriptorSet *update) { 645 // Perform update on a per-binding basis as consecutive updates roll over to next binding 646 auto descriptors_remaining = update->descriptorCount; 647 auto binding_being_updated = update->dstBinding; 648 auto offset = update->dstArrayElement; 649 uint32_t update_index = 0; 650 while (descriptors_remaining) { 651 uint32_t update_count = std::min(descriptors_remaining, GetDescriptorCountFromBinding(binding_being_updated)); 652 auto global_idx = p_layout_->GetGlobalIndexRangeFromBinding(binding_being_updated).start + offset; 653 // Loop over the updates for a single binding at a time 654 for (uint32_t di = 0; di < update_count; ++di, ++update_index) { 655 descriptors_[global_idx + di]->WriteUpdate(update, update_index); 656 } 657 // Roll over to next binding in case of consecutive update 658 descriptors_remaining -= update_count; 659 offset = 0; 660 binding_being_updated++; 661 } 662 if (update->descriptorCount) some_update_ = true; 663 664 InvalidateBoundCmdBuffers(); 665} 666// Validate Copy update 667bool cvdescriptorset::DescriptorSet::ValidateCopyUpdate(const debug_report_data *report_data, const VkCopyDescriptorSet *update, 668 const DescriptorSet *src_set, UNIQUE_VALIDATION_ERROR_CODE *error_code, 669 std::string *error_msg) { 670 // Verify dst layout still valid 671 if (p_layout_->IsDestroyed()) { 672 *error_code = VALIDATION_ERROR_03207601; 673 string_sprintf(error_msg, 674 "Cannot call vkUpdateDescriptorSets() to perform copy update on descriptor set dstSet 0x%" PRIxLEAST64 675 " created with destroyed VkDescriptorSetLayout 0x%" PRIxLEAST64, 676 HandleToUint64(set_), HandleToUint64(p_layout_->GetDescriptorSetLayout())); 677 return false; 678 } 679 680 // Verify src layout still valid 681 if (src_set->p_layout_->IsDestroyed()) { 682 *error_code = VALIDATION_ERROR_0322d201; 683 string_sprintf( 684 error_msg, 685 "Cannot call vkUpdateDescriptorSets() to perform copy update of dstSet 0x%" PRIxLEAST64 686 " from descriptor set srcSet 0x%" PRIxLEAST64 " created with destroyed VkDescriptorSetLayout 0x%" PRIxLEAST64, 687 HandleToUint64(set_), HandleToUint64(src_set->set_), HandleToUint64(src_set->p_layout_->GetDescriptorSetLayout())); 688 return false; 689 } 690 691 // Verify idle ds 692 if (in_use.load()) { 693 // TODO : Re-using Free Idle error code, need copy update idle error code 694 *error_code = VALIDATION_ERROR_2860026a; 695 std::stringstream error_str; 696 error_str << "Cannot call vkUpdateDescriptorSets() to perform copy update on descriptor set " << set_ 697 << " that is in use by a command buffer"; 698 *error_msg = error_str.str(); 699 return false; 700 } 701 if (!p_layout_->HasBinding(update->dstBinding)) { 702 *error_code = VALIDATION_ERROR_032002b6; 703 std::stringstream error_str; 704 error_str << "DescriptorSet " << set_ << " does not have copy update dest binding of " << update->dstBinding; 705 *error_msg = error_str.str(); 706 return false; 707 } 708 if (!src_set->HasBinding(update->srcBinding)) { 709 *error_code = VALIDATION_ERROR_032002b2; 710 std::stringstream error_str; 711 error_str << "DescriptorSet " << set_ << " does not have copy update src binding of " << update->srcBinding; 712 *error_msg = error_str.str(); 713 return false; 714 } 715 // src & dst set bindings are valid 716 // Check bounds of src & dst 717 auto src_start_idx = src_set->GetGlobalIndexRangeFromBinding(update->srcBinding).start + update->srcArrayElement; 718 if ((src_start_idx + update->descriptorCount) > src_set->GetTotalDescriptorCount()) { 719 // SRC update out of bounds 720 *error_code = VALIDATION_ERROR_032002b4; 721 std::stringstream error_str; 722 error_str << "Attempting copy update from descriptorSet " << update->srcSet << " binding#" << update->srcBinding 723 << " with offset index of " << src_set->GetGlobalIndexRangeFromBinding(update->srcBinding).start 724 << " plus update array offset of " << update->srcArrayElement << " and update of " << update->descriptorCount 725 << " descriptors oversteps total number of descriptors in set: " << src_set->GetTotalDescriptorCount(); 726 *error_msg = error_str.str(); 727 return false; 728 } 729 auto dst_start_idx = p_layout_->GetGlobalIndexRangeFromBinding(update->dstBinding).start + update->dstArrayElement; 730 if ((dst_start_idx + update->descriptorCount) > p_layout_->GetTotalDescriptorCount()) { 731 // DST update out of bounds 732 *error_code = VALIDATION_ERROR_032002b8; 733 std::stringstream error_str; 734 error_str << "Attempting copy update to descriptorSet " << set_ << " binding#" << update->dstBinding 735 << " with offset index of " << p_layout_->GetGlobalIndexRangeFromBinding(update->dstBinding).start 736 << " plus update array offset of " << update->dstArrayElement << " and update of " << update->descriptorCount 737 << " descriptors oversteps total number of descriptors in set: " << p_layout_->GetTotalDescriptorCount(); 738 *error_msg = error_str.str(); 739 return false; 740 } 741 // Check that types match 742 // TODO : Base default error case going from here is VALIDATION_ERROR_0002b8012ba which covers all consistency issues, need more 743 // fine-grained error codes 744 *error_code = VALIDATION_ERROR_032002ba; 745 auto src_type = src_set->GetTypeFromBinding(update->srcBinding); 746 auto dst_type = p_layout_->GetTypeFromBinding(update->dstBinding); 747 if (src_type != dst_type) { 748 std::stringstream error_str; 749 error_str << "Attempting copy update to descriptorSet " << set_ << " binding #" << update->dstBinding << " with type " 750 << string_VkDescriptorType(dst_type) << " from descriptorSet " << src_set->GetSet() << " binding #" 751 << update->srcBinding << " with type " << string_VkDescriptorType(src_type) << ". Types do not match"; 752 *error_msg = error_str.str(); 753 return false; 754 } 755 // Verify consistency of src & dst bindings if update crosses binding boundaries 756 if ((!src_set->GetLayout()->VerifyUpdateConsistency(update->srcBinding, update->srcArrayElement, update->descriptorCount, 757 "copy update from", src_set->GetSet(), error_msg)) || 758 (!p_layout_->VerifyUpdateConsistency(update->dstBinding, update->dstArrayElement, update->descriptorCount, "copy update to", 759 set_, error_msg))) { 760 return false; 761 } 762 // Update parameters all look good and descriptor updated so verify update contents 763 if (!VerifyCopyUpdateContents(update, src_set, src_type, src_start_idx, error_code, error_msg)) return false; 764 765 // All checks passed so update is good 766 return true; 767} 768// Perform Copy update 769void cvdescriptorset::DescriptorSet::PerformCopyUpdate(const VkCopyDescriptorSet *update, const DescriptorSet *src_set) { 770 auto src_start_idx = src_set->GetGlobalIndexRangeFromBinding(update->srcBinding).start + update->srcArrayElement; 771 auto dst_start_idx = p_layout_->GetGlobalIndexRangeFromBinding(update->dstBinding).start + update->dstArrayElement; 772 // Update parameters all look good so perform update 773 for (uint32_t di = 0; di < update->descriptorCount; ++di) { 774 auto src = src_set->descriptors_[src_start_idx + di].get(); 775 auto dst = descriptors_[dst_start_idx + di].get(); 776 if (src->updated) { 777 dst->CopyUpdate(src); 778 some_update_ = true; 779 } else { 780 dst->updated = false; 781 } 782 } 783 784 InvalidateBoundCmdBuffers(); 785} 786 787// Bind cb_node to this set and this set to cb_node. 788// Prereq: This should be called for a set that has been confirmed to be active for the given cb_node, meaning it's going 789// to be used in a draw by the given cb_node 790void cvdescriptorset::DescriptorSet::BindCommandBuffer(GLOBAL_CB_NODE *cb_node, 791 const std::map<uint32_t, descriptor_req> &binding_req_map) { 792 // bind cb to this descriptor set 793 cb_bindings.insert(cb_node); 794 // Add bindings for descriptor set, the set's pool, and individual objects in the set 795 cb_node->object_bindings.insert({HandleToUint64(set_), kVulkanObjectTypeDescriptorSet}); 796 pool_state_->cb_bindings.insert(cb_node); 797 cb_node->object_bindings.insert({HandleToUint64(pool_state_->pool), kVulkanObjectTypeDescriptorPool}); 798 // For the active slots, use set# to look up descriptorSet from boundDescriptorSets, and bind all of that descriptor set's 799 // resources 800 for (auto binding_req_pair : binding_req_map) { 801 auto binding = binding_req_pair.first; 802 auto range = p_layout_->GetGlobalIndexRangeFromBinding(binding); 803 for (uint32_t i = range.start; i < range.end; ++i) { 804 descriptors_[i]->BindCommandBuffer(device_data_, cb_node); 805 } 806 } 807} 808void cvdescriptorset::DescriptorSet::FilterAndTrackOneBindingReq(const BindingReqMap::value_type &binding_req_pair, 809 const BindingReqMap &in_req, BindingReqMap *out_req, 810 TrackedBindings *bindings) { 811 assert(out_req); 812 assert(bindings); 813 const auto binding = binding_req_pair.first; 814 // Use insert and look at the boolean ("was inserted") in the returned pair to see if this is a new set member. 815 // Saves one hash lookup vs. find ... compare w/ end ... insert. 816 const auto it_bool_pair = bindings->insert(binding); 817 if (it_bool_pair.second) { 818 out_req->emplace(binding_req_pair); 819 } 820} 821void cvdescriptorset::DescriptorSet::FilterAndTrackOneBindingReq(const BindingReqMap::value_type &binding_req_pair, 822 const BindingReqMap &in_req, BindingReqMap *out_req, 823 TrackedBindings *bindings, uint32_t limit) { 824 if (bindings->size() < limit) FilterAndTrackOneBindingReq(binding_req_pair, in_req, out_req, bindings); 825} 826 827void cvdescriptorset::DescriptorSet::FilterAndTrackBindingReqs(GLOBAL_CB_NODE *cb_state, const BindingReqMap &in_req, 828 BindingReqMap *out_req) { 829 TrackedBindings &bound = cached_validation_[cb_state].command_binding_and_usage; 830 if (bound.size() == GetBindingCount()) { 831 return; // All bindings are bound, out req is empty 832 } 833 for (const auto &binding_req_pair : in_req) { 834 const auto binding = binding_req_pair.first; 835 // If a binding doesn't exist, or has already been bound, skip it 836 if (p_layout_->HasBinding(binding)) { 837 FilterAndTrackOneBindingReq(binding_req_pair, in_req, out_req, &bound); 838 } 839 } 840} 841 842void cvdescriptorset::DescriptorSet::FilterAndTrackBindingReqs(GLOBAL_CB_NODE *cb_state, PIPELINE_STATE *pipeline, 843 const BindingReqMap &in_req, BindingReqMap *out_req) { 844 auto &validated = cached_validation_[cb_state]; 845 auto &image_sample_val = validated.image_samplers[pipeline]; 846 auto *const dynamic_buffers = &validated.dynamic_buffers; 847 auto *const non_dynamic_buffers = &validated.non_dynamic_buffers; 848 const auto &stats = p_layout_->GetBindingTypeStats(); 849 for (const auto &binding_req_pair : in_req) { 850 auto binding = binding_req_pair.first; 851 VkDescriptorSetLayoutBinding const *layout_binding = p_layout_->GetDescriptorSetLayoutBindingPtrFromBinding(binding); 852 if (!layout_binding) { 853 continue; 854 } 855 // Caching criteria differs per type. 856 // If image_layout have changed , the image descriptors need to be validated against them. 857 if ((layout_binding->descriptorType == VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER_DYNAMIC) || 858 (layout_binding->descriptorType == VK_DESCRIPTOR_TYPE_STORAGE_BUFFER_DYNAMIC)) { 859 FilterAndTrackOneBindingReq(binding_req_pair, in_req, out_req, dynamic_buffers, stats.dynamic_buffer_count); 860 } else if ((layout_binding->descriptorType == VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER) || 861 (layout_binding->descriptorType == VK_DESCRIPTOR_TYPE_STORAGE_BUFFER)) { 862 FilterAndTrackOneBindingReq(binding_req_pair, in_req, out_req, non_dynamic_buffers, stats.non_dynamic_buffer_count); 863 } else { 864 // This is rather crude, as the changed layouts may not impact the bound descriptors, 865 // but the simple "versioning" is a simple "dirt" test. 866 auto &version = image_sample_val[binding]; // Take advantage of default construtor zero initialzing new entries 867 if (version != cb_state->image_layout_change_count) { 868 version = cb_state->image_layout_change_count; 869 out_req->emplace(binding_req_pair); 870 } 871 } 872 } 873} 874 875cvdescriptorset::SamplerDescriptor::SamplerDescriptor(const VkSampler *immut) : sampler_(VK_NULL_HANDLE), immutable_(false) { 876 updated = false; 877 descriptor_class = PlainSampler; 878 if (immut) { 879 sampler_ = *immut; 880 immutable_ = true; 881 updated = true; 882 } 883} 884// Validate given sampler. Currently this only checks to make sure it exists in the samplerMap 885bool cvdescriptorset::ValidateSampler(const VkSampler sampler, const layer_data *dev_data) { 886 return (GetSamplerState(dev_data, sampler) != nullptr); 887} 888 889bool cvdescriptorset::ValidateImageUpdate(VkImageView image_view, VkImageLayout image_layout, VkDescriptorType type, 890 const layer_data *dev_data, UNIQUE_VALIDATION_ERROR_CODE *error_code, 891 std::string *error_msg) { 892 // TODO : Defaulting to 00943 for all cases here. Need to create new error codes for various cases. 893 *error_code = VALIDATION_ERROR_15c0028c; 894 auto iv_state = GetImageViewState(dev_data, image_view); 895 if (!iv_state) { 896 std::stringstream error_str; 897 error_str << "Invalid VkImageView: " << image_view; 898 *error_msg = error_str.str(); 899 return false; 900 } 901 // Note that when an imageview is created, we validated that memory is bound so no need to re-check here 902 // Validate that imageLayout is compatible with aspect_mask and image format 903 // and validate that image usage bits are correct for given usage 904 VkImageAspectFlags aspect_mask = iv_state->create_info.subresourceRange.aspectMask; 905 VkImage image = iv_state->create_info.image; 906 VkFormat format = VK_FORMAT_MAX_ENUM; 907 VkImageUsageFlags usage = 0; 908 auto image_node = GetImageState(dev_data, image); 909 if (image_node) { 910 format = image_node->createInfo.format; 911 usage = image_node->createInfo.usage; 912 // Validate that memory is bound to image 913 // TODO: This should have its own valid usage id apart from 2524 which is from CreateImageView case. The only 914 // the error here occurs is if memory bound to a created imageView has been freed. 915 if (ValidateMemoryIsBoundToImage(dev_data, image_node, "vkUpdateDescriptorSets()", VALIDATION_ERROR_0ac007f8)) { 916 *error_code = VALIDATION_ERROR_0ac007f8; 917 *error_msg = "No memory bound to image."; 918 return false; 919 } 920 921 // KHR_maintenance1 allows rendering into 2D or 2DArray views which slice a 3D image, 922 // but not binding them to descriptor sets. 923 if (image_node->createInfo.imageType == VK_IMAGE_TYPE_3D && 924 (iv_state->create_info.viewType == VK_IMAGE_VIEW_TYPE_2D || 925 iv_state->create_info.viewType == VK_IMAGE_VIEW_TYPE_2D_ARRAY)) { 926 *error_code = VALIDATION_ERROR_046002ae; 927 *error_msg = "ImageView must not be a 2D or 2DArray view of a 3D image"; 928 return false; 929 } 930 } 931 // First validate that format and layout are compatible 932 if (format == VK_FORMAT_MAX_ENUM) { 933 std::stringstream error_str; 934 error_str << "Invalid image (" << image << ") in imageView (" << image_view << ")."; 935 *error_msg = error_str.str(); 936 return false; 937 } 938 // TODO : The various image aspect and format checks here are based on general spec language in 11.5 Image Views section under 939 // vkCreateImageView(). What's the best way to create unique id for these cases? 940 bool ds = FormatIsDepthOrStencil(format); 941 switch (image_layout) { 942 case VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL: 943 // Only Color bit must be set 944 if ((aspect_mask & VK_IMAGE_ASPECT_COLOR_BIT) != VK_IMAGE_ASPECT_COLOR_BIT) { 945 std::stringstream error_str; 946 error_str 947 << "ImageView (" << image_view 948 << ") uses layout VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL but does not have VK_IMAGE_ASPECT_COLOR_BIT set."; 949 *error_msg = error_str.str(); 950 return false; 951 } 952 // format must NOT be DS 953 if (ds) { 954 std::stringstream error_str; 955 error_str << "ImageView (" << image_view 956 << ") uses layout VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL but the image format is " 957 << string_VkFormat(format) << " which is not a color format."; 958 *error_msg = error_str.str(); 959 return false; 960 } 961 break; 962 case VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL: 963 case VK_IMAGE_LAYOUT_DEPTH_STENCIL_READ_ONLY_OPTIMAL: 964 // Depth or stencil bit must be set, but both must NOT be set 965 if (aspect_mask & VK_IMAGE_ASPECT_DEPTH_BIT) { 966 if (aspect_mask & VK_IMAGE_ASPECT_STENCIL_BIT) { 967 // both must NOT be set 968 std::stringstream error_str; 969 error_str << "ImageView (" << image_view << ") has both STENCIL and DEPTH aspects set"; 970 *error_msg = error_str.str(); 971 return false; 972 } 973 } else if (!(aspect_mask & VK_IMAGE_ASPECT_STENCIL_BIT)) { 974 // Neither were set 975 std::stringstream error_str; 976 error_str << "ImageView (" << image_view << ") has layout " << string_VkImageLayout(image_layout) 977 << " but does not have STENCIL or DEPTH aspects set"; 978 *error_msg = error_str.str(); 979 return false; 980 } 981 // format must be DS 982 if (!ds) { 983 std::stringstream error_str; 984 error_str << "ImageView (" << image_view << ") has layout " << string_VkImageLayout(image_layout) 985 << " but the image format is " << string_VkFormat(format) << " which is not a depth/stencil format."; 986 *error_msg = error_str.str(); 987 return false; 988 } 989 break; 990 default: 991 // For other layouts if the source is depth/stencil image, both aspect bits must not be set 992 if (ds) { 993 if (aspect_mask & VK_IMAGE_ASPECT_DEPTH_BIT) { 994 if (aspect_mask & VK_IMAGE_ASPECT_STENCIL_BIT) { 995 // both must NOT be set 996 std::stringstream error_str; 997 error_str << "ImageView (" << image_view << ") has layout " << string_VkImageLayout(image_layout) 998 << " and is using depth/stencil image of format " << string_VkFormat(format) 999 << " but it has both STENCIL and DEPTH aspects set, which is illegal. When using a depth/stencil " 1000 "image in a descriptor set, please only set either VK_IMAGE_ASPECT_DEPTH_BIT or " 1001 "VK_IMAGE_ASPECT_STENCIL_BIT depending on whether it will be used for depth reads or stencil " 1002 "reads respectively."; 1003 *error_msg = error_str.str(); 1004 return false; 1005 } 1006 } 1007 } 1008 break; 1009 } 1010 // Now validate that usage flags are correctly set for given type of update 1011 // As we're switching per-type, if any type has specific layout requirements, check those here as well 1012 // TODO : The various image usage bit requirements are in general spec language for VkImageUsageFlags bit block in 11.3 Images 1013 // under vkCreateImage() 1014 // TODO : Need to also validate case VALIDATION_ERROR_15c002a0 where STORAGE_IMAGE & INPUT_ATTACH types must have been created 1015 // with identify swizzle 1016 std::string error_usage_bit; 1017 switch (type) { 1018 case VK_DESCRIPTOR_TYPE_SAMPLED_IMAGE: 1019 case VK_DESCRIPTOR_TYPE_COMBINED_IMAGE_SAMPLER: { 1020 if (!(usage & VK_IMAGE_USAGE_SAMPLED_BIT)) { 1021 error_usage_bit = "VK_IMAGE_USAGE_SAMPLED_BIT"; 1022 } 1023 break; 1024 } 1025 case VK_DESCRIPTOR_TYPE_STORAGE_IMAGE: { 1026 if (!(usage & VK_IMAGE_USAGE_STORAGE_BIT)) { 1027 error_usage_bit = "VK_IMAGE_USAGE_STORAGE_BIT"; 1028 } else if (VK_IMAGE_LAYOUT_GENERAL != image_layout) { 1029 std::stringstream error_str; 1030 // TODO : Need to create custom enum error codes for these cases 1031 if (image_node->shared_presentable) { 1032 if (VK_IMAGE_LAYOUT_SHARED_PRESENT_KHR != image_layout) { 1033 error_str << "ImageView (" << image_view 1034 << ") of VK_DESCRIPTOR_TYPE_STORAGE_IMAGE type with a front-buffered image is being updated with " 1035 "layout " 1036 << string_VkImageLayout(image_layout) 1037 << " but according to spec section 13.1 Descriptor Types, 'Front-buffered images that report " 1038 "support for VK_FORMAT_FEATURE_STORAGE_IMAGE_BIT must be in the " 1039 "VK_IMAGE_LAYOUT_SHARED_PRESENT_KHR layout.'"; 1040 *error_msg = error_str.str(); 1041 return false; 1042 } 1043 } else if (VK_IMAGE_LAYOUT_GENERAL != image_layout) { 1044 error_str << "ImageView (" << image_view 1045 << ") of VK_DESCRIPTOR_TYPE_STORAGE_IMAGE type is being updated with layout " 1046 << string_VkImageLayout(image_layout) 1047 << " but according to spec section 13.1 Descriptor Types, 'Load and store operations on storage " 1048 "images can only be done on images in VK_IMAGE_LAYOUT_GENERAL layout.'"; 1049 *error_msg = error_str.str(); 1050 return false; 1051 } 1052 } 1053 break; 1054 } 1055 case VK_DESCRIPTOR_TYPE_INPUT_ATTACHMENT: { 1056 if (!(usage & VK_IMAGE_USAGE_INPUT_ATTACHMENT_BIT)) { 1057 error_usage_bit = "VK_IMAGE_USAGE_INPUT_ATTACHMENT_BIT"; 1058 } 1059 break; 1060 } 1061 default: 1062 break; 1063 } 1064 if (!error_usage_bit.empty()) { 1065 std::stringstream error_str; 1066 error_str << "ImageView (" << image_view << ") with usage mask 0x" << usage 1067 << " being used for a descriptor update of type " << string_VkDescriptorType(type) << " does not have " 1068 << error_usage_bit << " set."; 1069 *error_msg = error_str.str(); 1070 return false; 1071 } 1072 return true; 1073} 1074 1075void cvdescriptorset::SamplerDescriptor::WriteUpdate(const VkWriteDescriptorSet *update, const uint32_t index) { 1076 sampler_ = update->pImageInfo[index].sampler; 1077 updated = true; 1078} 1079 1080void cvdescriptorset::SamplerDescriptor::CopyUpdate(const Descriptor *src) { 1081 if (!immutable_) { 1082 auto update_sampler = static_cast<const SamplerDescriptor *>(src)->sampler_; 1083 sampler_ = update_sampler; 1084 } 1085 updated = true; 1086} 1087 1088void cvdescriptorset::SamplerDescriptor::BindCommandBuffer(const layer_data *dev_data, GLOBAL_CB_NODE *cb_node) { 1089 if (!immutable_) { 1090 auto sampler_state = GetSamplerState(dev_data, sampler_); 1091 if (sampler_state) core_validation::AddCommandBufferBindingSampler(cb_node, sampler_state); 1092 } 1093} 1094 1095cvdescriptorset::ImageSamplerDescriptor::ImageSamplerDescriptor(const VkSampler *immut) 1096 : sampler_(VK_NULL_HANDLE), immutable_(false), image_view_(VK_NULL_HANDLE), image_layout_(VK_IMAGE_LAYOUT_UNDEFINED) { 1097 updated = false; 1098 descriptor_class = ImageSampler; 1099 if (immut) { 1100 sampler_ = *immut; 1101 immutable_ = true; 1102 } 1103} 1104 1105void cvdescriptorset::ImageSamplerDescriptor::WriteUpdate(const VkWriteDescriptorSet *update, const uint32_t index) { 1106 updated = true; 1107 const auto &image_info = update->pImageInfo[index]; 1108 sampler_ = image_info.sampler; 1109 image_view_ = image_info.imageView; 1110 image_layout_ = image_info.imageLayout; 1111} 1112 1113void cvdescriptorset::ImageSamplerDescriptor::CopyUpdate(const Descriptor *src) { 1114 if (!immutable_) { 1115 auto update_sampler = static_cast<const ImageSamplerDescriptor *>(src)->sampler_; 1116 sampler_ = update_sampler; 1117 } 1118 auto image_view = static_cast<const ImageSamplerDescriptor *>(src)->image_view_; 1119 auto image_layout = static_cast<const ImageSamplerDescriptor *>(src)->image_layout_; 1120 updated = true; 1121 image_view_ = image_view; 1122 image_layout_ = image_layout; 1123} 1124 1125void cvdescriptorset::ImageSamplerDescriptor::BindCommandBuffer(const layer_data *dev_data, GLOBAL_CB_NODE *cb_node) { 1126 // First add binding for any non-immutable sampler 1127 if (!immutable_) { 1128 auto sampler_state = GetSamplerState(dev_data, sampler_); 1129 if (sampler_state) core_validation::AddCommandBufferBindingSampler(cb_node, sampler_state); 1130 } 1131 // Add binding for image 1132 auto iv_state = GetImageViewState(dev_data, image_view_); 1133 if (iv_state) { 1134 core_validation::AddCommandBufferBindingImageView(dev_data, cb_node, iv_state); 1135 } 1136} 1137 1138cvdescriptorset::ImageDescriptor::ImageDescriptor(const VkDescriptorType type) 1139 : storage_(false), image_view_(VK_NULL_HANDLE), image_layout_(VK_IMAGE_LAYOUT_UNDEFINED) { 1140 updated = false; 1141 descriptor_class = Image; 1142 if (VK_DESCRIPTOR_TYPE_STORAGE_IMAGE == type) storage_ = true; 1143} 1144 1145void cvdescriptorset::ImageDescriptor::WriteUpdate(const VkWriteDescriptorSet *update, const uint32_t index) { 1146 updated = true; 1147 const auto &image_info = update->pImageInfo[index]; 1148 image_view_ = image_info.imageView; 1149 image_layout_ = image_info.imageLayout; 1150} 1151 1152void cvdescriptorset::ImageDescriptor::CopyUpdate(const Descriptor *src) { 1153 auto image_view = static_cast<const ImageDescriptor *>(src)->image_view_; 1154 auto image_layout = static_cast<const ImageDescriptor *>(src)->image_layout_; 1155 updated = true; 1156 image_view_ = image_view; 1157 image_layout_ = image_layout; 1158} 1159 1160void cvdescriptorset::ImageDescriptor::BindCommandBuffer(const layer_data *dev_data, GLOBAL_CB_NODE *cb_node) { 1161 // Add binding for image 1162 auto iv_state = GetImageViewState(dev_data, image_view_); 1163 if (iv_state) { 1164 core_validation::AddCommandBufferBindingImageView(dev_data, cb_node, iv_state); 1165 } 1166} 1167 1168cvdescriptorset::BufferDescriptor::BufferDescriptor(const VkDescriptorType type) 1169 : storage_(false), dynamic_(false), buffer_(VK_NULL_HANDLE), offset_(0), range_(0) { 1170 updated = false; 1171 descriptor_class = GeneralBuffer; 1172 if (VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER_DYNAMIC == type) { 1173 dynamic_ = true; 1174 } else if (VK_DESCRIPTOR_TYPE_STORAGE_BUFFER == type) { 1175 storage_ = true; 1176 } else if (VK_DESCRIPTOR_TYPE_STORAGE_BUFFER_DYNAMIC == type) { 1177 dynamic_ = true; 1178 storage_ = true; 1179 } 1180} 1181void cvdescriptorset::BufferDescriptor::WriteUpdate(const VkWriteDescriptorSet *update, const uint32_t index) { 1182 updated = true; 1183 const auto &buffer_info = update->pBufferInfo[index]; 1184 buffer_ = buffer_info.buffer; 1185 offset_ = buffer_info.offset; 1186 range_ = buffer_info.range; 1187} 1188 1189void cvdescriptorset::BufferDescriptor::CopyUpdate(const Descriptor *src) { 1190 auto buff_desc = static_cast<const BufferDescriptor *>(src); 1191 updated = true; 1192 buffer_ = buff_desc->buffer_; 1193 offset_ = buff_desc->offset_; 1194 range_ = buff_desc->range_; 1195} 1196 1197void cvdescriptorset::BufferDescriptor::BindCommandBuffer(const layer_data *dev_data, GLOBAL_CB_NODE *cb_node) { 1198 auto buffer_node = GetBufferState(dev_data, buffer_); 1199 if (buffer_node) core_validation::AddCommandBufferBindingBuffer(dev_data, cb_node, buffer_node); 1200} 1201 1202cvdescriptorset::TexelDescriptor::TexelDescriptor(const VkDescriptorType type) : buffer_view_(VK_NULL_HANDLE), storage_(false) { 1203 updated = false; 1204 descriptor_class = TexelBuffer; 1205 if (VK_DESCRIPTOR_TYPE_STORAGE_TEXEL_BUFFER == type) storage_ = true; 1206} 1207 1208void cvdescriptorset::TexelDescriptor::WriteUpdate(const VkWriteDescriptorSet *update, const uint32_t index) { 1209 updated = true; 1210 buffer_view_ = update->pTexelBufferView[index]; 1211} 1212 1213void cvdescriptorset::TexelDescriptor::CopyUpdate(const Descriptor *src) { 1214 updated = true; 1215 buffer_view_ = static_cast<const TexelDescriptor *>(src)->buffer_view_; 1216} 1217 1218void cvdescriptorset::TexelDescriptor::BindCommandBuffer(const layer_data *dev_data, GLOBAL_CB_NODE *cb_node) { 1219 auto bv_state = GetBufferViewState(dev_data, buffer_view_); 1220 if (bv_state) { 1221 core_validation::AddCommandBufferBindingBufferView(dev_data, cb_node, bv_state); 1222 } 1223} 1224 1225// This is a helper function that iterates over a set of Write and Copy updates, pulls the DescriptorSet* for updated 1226// sets, and then calls their respective Validate[Write|Copy]Update functions. 1227// If the update hits an issue for which the callback returns "true", meaning that the call down the chain should 1228// be skipped, then true is returned. 1229// If there is no issue with the update, then false is returned. 1230bool cvdescriptorset::ValidateUpdateDescriptorSets(const debug_report_data *report_data, const layer_data *dev_data, 1231 uint32_t write_count, const VkWriteDescriptorSet *p_wds, uint32_t copy_count, 1232 const VkCopyDescriptorSet *p_cds) { 1233 bool skip = false; 1234 // Validate Write updates 1235 for (uint32_t i = 0; i < write_count; i++) { 1236 auto dest_set = p_wds[i].dstSet; 1237 auto set_node = core_validation::GetSetNode(dev_data, dest_set); 1238 if (!set_node) { 1239 skip |= 1240 log_msg(report_data, VK_DEBUG_REPORT_ERROR_BIT_EXT, VK_DEBUG_REPORT_OBJECT_TYPE_DESCRIPTOR_SET_EXT, 1241 HandleToUint64(dest_set), __LINE__, DRAWSTATE_INVALID_DESCRIPTOR_SET, "DS", 1242 "Cannot call vkUpdateDescriptorSets() on descriptor set 0x%" PRIxLEAST64 " that has not been allocated.", 1243 HandleToUint64(dest_set)); 1244 } else { 1245 UNIQUE_VALIDATION_ERROR_CODE error_code; 1246 std::string error_str; 1247 if (!set_node->ValidateWriteUpdate(report_data, &p_wds[i], &error_code, &error_str)) { 1248 skip |= log_msg(report_data, VK_DEBUG_REPORT_ERROR_BIT_EXT, VK_DEBUG_REPORT_OBJECT_TYPE_DESCRIPTOR_SET_EXT, 1249 HandleToUint64(dest_set), __LINE__, error_code, "DS", 1250 "vkUpdateDescriptorSets() failed write update validation for Descriptor Set 0x%" PRIx64 1251 " with error: %s. %s", 1252 HandleToUint64(dest_set), error_str.c_str(), validation_error_map[error_code]); 1253 } 1254 } 1255 } 1256 // Now validate copy updates 1257 for (uint32_t i = 0; i < copy_count; ++i) { 1258 auto dst_set = p_cds[i].dstSet; 1259 auto src_set = p_cds[i].srcSet; 1260 auto src_node = core_validation::GetSetNode(dev_data, src_set); 1261 auto dst_node = core_validation::GetSetNode(dev_data, dst_set); 1262 // Object_tracker verifies that src & dest descriptor set are valid 1263 assert(src_node); 1264 assert(dst_node); 1265 UNIQUE_VALIDATION_ERROR_CODE error_code; 1266 std::string error_str; 1267 if (!dst_node->ValidateCopyUpdate(report_data, &p_cds[i], src_node, &error_code, &error_str)) { 1268 skip |= log_msg(report_data, VK_DEBUG_REPORT_ERROR_BIT_EXT, VK_DEBUG_REPORT_OBJECT_TYPE_DESCRIPTOR_SET_EXT, 1269 HandleToUint64(dst_set), __LINE__, error_code, "DS", 1270 "vkUpdateDescriptorSets() failed copy update from Descriptor Set 0x%" PRIx64 1271 " to Descriptor Set 0x%" PRIx64 " with error: %s. %s", 1272 HandleToUint64(src_set), HandleToUint64(dst_set), error_str.c_str(), validation_error_map[error_code]); 1273 } 1274 } 1275 return skip; 1276} 1277// This is a helper function that iterates over a set of Write and Copy updates, pulls the DescriptorSet* for updated 1278// sets, and then calls their respective Perform[Write|Copy]Update functions. 1279// Prerequisite : ValidateUpdateDescriptorSets() should be called and return "false" prior to calling PerformUpdateDescriptorSets() 1280// with the same set of updates. 1281// This is split from the validate code to allow validation prior to calling down the chain, and then update after 1282// calling down the chain. 1283void cvdescriptorset::PerformUpdateDescriptorSets(const layer_data *dev_data, uint32_t write_count, 1284 const VkWriteDescriptorSet *p_wds, uint32_t copy_count, 1285 const VkCopyDescriptorSet *p_cds) { 1286 // Write updates first 1287 uint32_t i = 0; 1288 for (i = 0; i < write_count; ++i) { 1289 auto dest_set = p_wds[i].dstSet; 1290 auto set_node = core_validation::GetSetNode(dev_data, dest_set); 1291 if (set_node) { 1292 set_node->PerformWriteUpdate(&p_wds[i]); 1293 } 1294 } 1295 // Now copy updates 1296 for (i = 0; i < copy_count; ++i) { 1297 auto dst_set = p_cds[i].dstSet; 1298 auto src_set = p_cds[i].srcSet; 1299 auto src_node = core_validation::GetSetNode(dev_data, src_set); 1300 auto dst_node = core_validation::GetSetNode(dev_data, dst_set); 1301 if (src_node && dst_node) { 1302 dst_node->PerformCopyUpdate(&p_cds[i], src_node); 1303 } 1304 } 1305} 1306// This helper function carries out the state updates for descriptor updates peformed via update templates. It basically collects 1307// data and leverages the PerformUpdateDescriptor helper functions to do this. 1308void cvdescriptorset::PerformUpdateDescriptorSetsWithTemplateKHR(layer_data *device_data, VkDescriptorSet descriptorSet, 1309 std::unique_ptr<TEMPLATE_STATE> const &template_state, 1310 const void *pData) { 1311 auto const &create_info = template_state->create_info; 1312 1313 // Create a vector of write structs 1314 std::vector<VkWriteDescriptorSet> desc_writes; 1315 auto layout_obj = GetDescriptorSetLayout(device_data, create_info.descriptorSetLayout); 1316 1317 // Create a WriteDescriptorSet struct for each template update entry 1318 for (uint32_t i = 0; i < create_info.descriptorUpdateEntryCount; i++) { 1319 auto binding_count = layout_obj->GetDescriptorCountFromBinding(create_info.pDescriptorUpdateEntries[i].dstBinding); 1320 auto binding_being_updated = create_info.pDescriptorUpdateEntries[i].dstBinding; 1321 auto dst_array_element = create_info.pDescriptorUpdateEntries[i].dstArrayElement; 1322 1323 desc_writes.reserve(desc_writes.size() + create_info.pDescriptorUpdateEntries[i].descriptorCount); 1324 for (uint32_t j = 0; j < create_info.pDescriptorUpdateEntries[i].descriptorCount; j++) { 1325 desc_writes.emplace_back(); 1326 auto &write_entry = desc_writes.back(); 1327 1328 size_t offset = create_info.pDescriptorUpdateEntries[i].offset + j * create_info.pDescriptorUpdateEntries[i].stride; 1329 char *update_entry = (char *)(pData) + offset; 1330 1331 if (dst_array_element >= binding_count) { 1332 dst_array_element = 0; 1333 binding_being_updated = layout_obj->GetNextValidBinding(binding_being_updated); 1334 } 1335 1336 write_entry.sType = VK_STRUCTURE_TYPE_WRITE_DESCRIPTOR_SET; 1337 write_entry.pNext = NULL; 1338 write_entry.dstSet = descriptorSet; 1339 write_entry.dstBinding = binding_being_updated; 1340 write_entry.dstArrayElement = dst_array_element; 1341 write_entry.descriptorCount = 1; 1342 write_entry.descriptorType = create_info.pDescriptorUpdateEntries[i].descriptorType; 1343 1344 switch (create_info.pDescriptorUpdateEntries[i].descriptorType) { 1345 case VK_DESCRIPTOR_TYPE_SAMPLER: 1346 case VK_DESCRIPTOR_TYPE_COMBINED_IMAGE_SAMPLER: 1347 case VK_DESCRIPTOR_TYPE_SAMPLED_IMAGE: 1348 case VK_DESCRIPTOR_TYPE_STORAGE_IMAGE: 1349 case VK_DESCRIPTOR_TYPE_INPUT_ATTACHMENT: 1350 write_entry.pImageInfo = reinterpret_cast<VkDescriptorImageInfo *>(update_entry); 1351 break; 1352 1353 case VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER: 1354 case VK_DESCRIPTOR_TYPE_STORAGE_BUFFER: 1355 case VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER_DYNAMIC: 1356 case VK_DESCRIPTOR_TYPE_STORAGE_BUFFER_DYNAMIC: 1357 write_entry.pBufferInfo = reinterpret_cast<VkDescriptorBufferInfo *>(update_entry); 1358 break; 1359 1360 case VK_DESCRIPTOR_TYPE_UNIFORM_TEXEL_BUFFER: 1361 case VK_DESCRIPTOR_TYPE_STORAGE_TEXEL_BUFFER: 1362 write_entry.pTexelBufferView = reinterpret_cast<VkBufferView *>(update_entry); 1363 break; 1364 default: 1365 assert(0); 1366 break; 1367 } 1368 dst_array_element++; 1369 } 1370 } 1371 PerformUpdateDescriptorSets(device_data, static_cast<uint32_t>(desc_writes.size()), desc_writes.data(), 0, NULL); 1372} 1373// Validate the state for a given write update but don't actually perform the update 1374// If an error would occur for this update, return false and fill in details in error_msg string 1375bool cvdescriptorset::DescriptorSet::ValidateWriteUpdate(const debug_report_data *report_data, const VkWriteDescriptorSet *update, 1376 UNIQUE_VALIDATION_ERROR_CODE *error_code, std::string *error_msg) { 1377 // Verify dst layout still valid 1378 if (p_layout_->IsDestroyed()) { 1379 *error_code = VALIDATION_ERROR_15c00280; 1380 string_sprintf(error_msg, 1381 "Cannot call vkUpdateDescriptorSets() to perform write update on descriptor set 0x%" PRIxLEAST64 1382 " created with destroyed VkDescriptorSetLayout 0x%" PRIxLEAST64, 1383 HandleToUint64(set_), HandleToUint64(p_layout_->GetDescriptorSetLayout())); 1384 return false; 1385 } 1386 // Verify idle ds 1387 if (in_use.load()) { 1388 // TODO : Re-using Free Idle error code, need write update idle error code 1389 *error_code = VALIDATION_ERROR_2860026a; 1390 std::stringstream error_str; 1391 error_str << "Cannot call vkUpdateDescriptorSets() to perform write update on descriptor set " << set_ 1392 << " that is in use by a command buffer"; 1393 *error_msg = error_str.str(); 1394 return false; 1395 } 1396 // Verify dst binding exists 1397 if (!p_layout_->HasBinding(update->dstBinding)) { 1398 *error_code = VALIDATION_ERROR_15c00276; 1399 std::stringstream error_str; 1400 error_str << "DescriptorSet " << set_ << " does not have binding " << update->dstBinding; 1401 *error_msg = error_str.str(); 1402 return false; 1403 } else { 1404 // Make sure binding isn't empty 1405 if (0 == p_layout_->GetDescriptorCountFromBinding(update->dstBinding)) { 1406 *error_code = VALIDATION_ERROR_15c00278; 1407 std::stringstream error_str; 1408 error_str << "DescriptorSet " << set_ << " cannot updated binding " << update->dstBinding << " that has 0 descriptors"; 1409 *error_msg = error_str.str(); 1410 return false; 1411 } 1412 } 1413 // We know that binding is valid, verify update and do update on each descriptor 1414 auto start_idx = p_layout_->GetGlobalIndexRangeFromBinding(update->dstBinding).start + update->dstArrayElement; 1415 auto type = p_layout_->GetTypeFromBinding(update->dstBinding); 1416 if (type != update->descriptorType) { 1417 *error_code = VALIDATION_ERROR_15c0027e; 1418 std::stringstream error_str; 1419 error_str << "Attempting write update to descriptor set " << set_ << " binding #" << update->dstBinding << " with type " 1420 << string_VkDescriptorType(type) << " but update type is " << string_VkDescriptorType(update->descriptorType); 1421 *error_msg = error_str.str(); 1422 return false; 1423 } 1424 if (update->descriptorCount > (descriptors_.size() - start_idx)) { 1425 *error_code = VALIDATION_ERROR_15c00282; 1426 std::stringstream error_str; 1427 error_str << "Attempting write update to descriptor set " << set_ << " binding #" << update->dstBinding << " with " 1428 << descriptors_.size() - start_idx 1429 << " descriptors in that binding and all successive bindings of the set, but update of " 1430 << update->descriptorCount << " descriptors combined with update array element offset of " 1431 << update->dstArrayElement << " oversteps the available number of consecutive descriptors"; 1432 *error_msg = error_str.str(); 1433 return false; 1434 } 1435 // Verify consecutive bindings match (if needed) 1436 if (!p_layout_->VerifyUpdateConsistency(update->dstBinding, update->dstArrayElement, update->descriptorCount, "write update to", 1437 set_, error_msg)) { 1438 // TODO : Should break out "consecutive binding updates" language into valid usage statements 1439 *error_code = VALIDATION_ERROR_15c00282; 1440 return false; 1441 } 1442 // Update is within bounds and consistent so last step is to validate update contents 1443 if (!VerifyWriteUpdateContents(update, start_idx, error_code, error_msg)) { 1444 std::stringstream error_str; 1445 error_str << "Write update to descriptor in set " << set_ << " binding #" << update->dstBinding 1446 << " failed with error message: " << error_msg->c_str(); 1447 *error_msg = error_str.str(); 1448 return false; 1449 } 1450 // All checks passed, update is clean 1451 return true; 1452} 1453// For the given buffer, verify that its creation parameters are appropriate for the given type 1454// If there's an error, update the error_msg string with details and return false, else return true 1455bool cvdescriptorset::DescriptorSet::ValidateBufferUsage(BUFFER_STATE const *buffer_node, VkDescriptorType type, 1456 UNIQUE_VALIDATION_ERROR_CODE *error_code, std::string *error_msg) const { 1457 // Verify that usage bits set correctly for given type 1458 auto usage = buffer_node->createInfo.usage; 1459 std::string error_usage_bit; 1460 switch (type) { 1461 case VK_DESCRIPTOR_TYPE_UNIFORM_TEXEL_BUFFER: 1462 if (!(usage & VK_BUFFER_USAGE_UNIFORM_TEXEL_BUFFER_BIT)) { 1463 *error_code = VALIDATION_ERROR_15c0029c; 1464 error_usage_bit = "VK_BUFFER_USAGE_UNIFORM_TEXEL_BUFFER_BIT"; 1465 } 1466 break; 1467 case VK_DESCRIPTOR_TYPE_STORAGE_TEXEL_BUFFER: 1468 if (!(usage & VK_BUFFER_USAGE_STORAGE_TEXEL_BUFFER_BIT)) { 1469 *error_code = VALIDATION_ERROR_15c0029e; 1470 error_usage_bit = "VK_BUFFER_USAGE_STORAGE_TEXEL_BUFFER_BIT"; 1471 } 1472 break; 1473 case VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER: 1474 case VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER_DYNAMIC: 1475 if (!(usage & VK_BUFFER_USAGE_UNIFORM_BUFFER_BIT)) { 1476 *error_code = VALIDATION_ERROR_15c00292; 1477 error_usage_bit = "VK_BUFFER_USAGE_UNIFORM_BUFFER_BIT"; 1478 } 1479 break; 1480 case VK_DESCRIPTOR_TYPE_STORAGE_BUFFER: 1481 case VK_DESCRIPTOR_TYPE_STORAGE_BUFFER_DYNAMIC: 1482 if (!(usage & VK_BUFFER_USAGE_STORAGE_BUFFER_BIT)) { 1483 *error_code = VALIDATION_ERROR_15c00296; 1484 error_usage_bit = "VK_BUFFER_USAGE_STORAGE_BUFFER_BIT"; 1485 } 1486 break; 1487 default: 1488 break; 1489 } 1490 if (!error_usage_bit.empty()) { 1491 std::stringstream error_str; 1492 error_str << "Buffer (" << buffer_node->buffer << ") with usage mask 0x" << usage 1493 << " being used for a descriptor update of type " << string_VkDescriptorType(type) << " does not have " 1494 << error_usage_bit << " set."; 1495 *error_msg = error_str.str(); 1496 return false; 1497 } 1498 return true; 1499} 1500// For buffer descriptor updates, verify the buffer usage and VkDescriptorBufferInfo struct which includes: 1501// 1. buffer is valid 1502// 2. buffer was created with correct usage flags 1503// 3. offset is less than buffer size 1504// 4. range is either VK_WHOLE_SIZE or falls in (0, (buffer size - offset)] 1505// 5. range and offset are within the device's limits 1506// If there's an error, update the error_msg string with details and return false, else return true 1507bool cvdescriptorset::DescriptorSet::ValidateBufferUpdate(VkDescriptorBufferInfo const *buffer_info, VkDescriptorType type, 1508 UNIQUE_VALIDATION_ERROR_CODE *error_code, std::string *error_msg) const { 1509 // First make sure that buffer is valid 1510 auto buffer_node = GetBufferState(device_data_, buffer_info->buffer); 1511 // Any invalid buffer should already be caught by object_tracker 1512 assert(buffer_node); 1513 if (ValidateMemoryIsBoundToBuffer(device_data_, buffer_node, "vkUpdateDescriptorSets()", VALIDATION_ERROR_15c00294)) { 1514 *error_code = VALIDATION_ERROR_15c00294; 1515 *error_msg = "No memory bound to buffer."; 1516 return false; 1517 } 1518 // Verify usage bits 1519 if (!ValidateBufferUsage(buffer_node, type, error_code, error_msg)) { 1520 // error_msg will have been updated by ValidateBufferUsage() 1521 return false; 1522 } 1523 // offset must be less than buffer size 1524 if (buffer_info->offset >= buffer_node->createInfo.size) { 1525 *error_code = VALIDATION_ERROR_044002a8; 1526 std::stringstream error_str; 1527 error_str << "VkDescriptorBufferInfo offset of " << buffer_info->offset << " is greater than or equal to buffer " 1528 << buffer_node->buffer << " size of " << buffer_node->createInfo.size; 1529 *error_msg = error_str.str(); 1530 return false; 1531 } 1532 if (buffer_info->range != VK_WHOLE_SIZE) { 1533 // Range must be VK_WHOLE_SIZE or > 0 1534 if (!buffer_info->range) { 1535 *error_code = VALIDATION_ERROR_044002aa; 1536 std::stringstream error_str; 1537 error_str << "VkDescriptorBufferInfo range is not VK_WHOLE_SIZE and is zero, which is not allowed."; 1538 *error_msg = error_str.str(); 1539 return false; 1540 } 1541 // Range must be VK_WHOLE_SIZE or <= (buffer size - offset) 1542 if (buffer_info->range > (buffer_node->createInfo.size - buffer_info->offset)) { 1543 *error_code = VALIDATION_ERROR_044002ac; 1544 std::stringstream error_str; 1545 error_str << "VkDescriptorBufferInfo range is " << buffer_info->range << " which is greater than buffer size (" 1546 << buffer_node->createInfo.size << ") minus requested offset of " << buffer_info->offset; 1547 *error_msg = error_str.str(); 1548 return false; 1549 } 1550 } 1551 // Check buffer update sizes against device limits 1552 if (VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER == type || VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER_DYNAMIC == type) { 1553 auto max_ub_range = limits_.maxUniformBufferRange; 1554 // TODO : If range is WHOLE_SIZE, need to make sure underlying buffer size doesn't exceed device max 1555 if (buffer_info->range != VK_WHOLE_SIZE && buffer_info->range > max_ub_range) { 1556 *error_code = VALIDATION_ERROR_15c00298; 1557 std::stringstream error_str; 1558 error_str << "VkDescriptorBufferInfo range is " << buffer_info->range 1559 << " which is greater than this device's maxUniformBufferRange (" << max_ub_range << ")"; 1560 *error_msg = error_str.str(); 1561 return false; 1562 } 1563 } else if (VK_DESCRIPTOR_TYPE_STORAGE_BUFFER == type || VK_DESCRIPTOR_TYPE_STORAGE_BUFFER_DYNAMIC == type) { 1564 auto max_sb_range = limits_.maxStorageBufferRange; 1565 // TODO : If range is WHOLE_SIZE, need to make sure underlying buffer size doesn't exceed device max 1566 if (buffer_info->range != VK_WHOLE_SIZE && buffer_info->range > max_sb_range) { 1567 *error_code = VALIDATION_ERROR_15c0029a; 1568 std::stringstream error_str; 1569 error_str << "VkDescriptorBufferInfo range is " << buffer_info->range 1570 << " which is greater than this device's maxStorageBufferRange (" << max_sb_range << ")"; 1571 *error_msg = error_str.str(); 1572 return false; 1573 } 1574 } 1575 return true; 1576} 1577 1578// Verify that the contents of the update are ok, but don't perform actual update 1579bool cvdescriptorset::DescriptorSet::VerifyWriteUpdateContents(const VkWriteDescriptorSet *update, const uint32_t index, 1580 UNIQUE_VALIDATION_ERROR_CODE *error_code, 1581 std::string *error_msg) const { 1582 switch (update->descriptorType) { 1583 case VK_DESCRIPTOR_TYPE_COMBINED_IMAGE_SAMPLER: { 1584 for (uint32_t di = 0; di < update->descriptorCount; ++di) { 1585 // Validate image 1586 auto image_view = update->pImageInfo[di].imageView; 1587 auto image_layout = update->pImageInfo[di].imageLayout; 1588 if (!ValidateImageUpdate(image_view, image_layout, update->descriptorType, device_data_, error_code, error_msg)) { 1589 std::stringstream error_str; 1590 error_str << "Attempted write update to combined image sampler descriptor failed due to: " 1591 << error_msg->c_str(); 1592 *error_msg = error_str.str(); 1593 return false; 1594 } 1595 } 1596 // Intentional fall-through to validate sampler 1597 } 1598 case VK_DESCRIPTOR_TYPE_SAMPLER: { 1599 for (uint32_t di = 0; di < update->descriptorCount; ++di) { 1600 if (!descriptors_[index + di].get()->IsImmutableSampler()) { 1601 if (!ValidateSampler(update->pImageInfo[di].sampler, device_data_)) { 1602 *error_code = VALIDATION_ERROR_15c0028a; 1603 std::stringstream error_str; 1604 error_str << "Attempted write update to sampler descriptor with invalid sampler: " 1605 << update->pImageInfo[di].sampler << "."; 1606 *error_msg = error_str.str(); 1607 return false; 1608 } 1609 } else { 1610 // TODO : Warn here 1611 } 1612 } 1613 break; 1614 } 1615 case VK_DESCRIPTOR_TYPE_SAMPLED_IMAGE: 1616 case VK_DESCRIPTOR_TYPE_INPUT_ATTACHMENT: 1617 case VK_DESCRIPTOR_TYPE_STORAGE_IMAGE: { 1618 for (uint32_t di = 0; di < update->descriptorCount; ++di) { 1619 auto image_view = update->pImageInfo[di].imageView; 1620 auto image_layout = update->pImageInfo[di].imageLayout; 1621 if (!ValidateImageUpdate(image_view, image_layout, update->descriptorType, device_data_, error_code, error_msg)) { 1622 std::stringstream error_str; 1623 error_str << "Attempted write update to image descriptor failed due to: " << error_msg->c_str(); 1624 *error_msg = error_str.str(); 1625 return false; 1626 } 1627 } 1628 break; 1629 } 1630 case VK_DESCRIPTOR_TYPE_UNIFORM_TEXEL_BUFFER: 1631 case VK_DESCRIPTOR_TYPE_STORAGE_TEXEL_BUFFER: { 1632 for (uint32_t di = 0; di < update->descriptorCount; ++di) { 1633 auto buffer_view = update->pTexelBufferView[di]; 1634 auto bv_state = GetBufferViewState(device_data_, buffer_view); 1635 if (!bv_state) { 1636 *error_code = VALIDATION_ERROR_15c00286; 1637 std::stringstream error_str; 1638 error_str << "Attempted write update to texel buffer descriptor with invalid buffer view: " << buffer_view; 1639 *error_msg = error_str.str(); 1640 return false; 1641 } 1642 auto buffer = bv_state->create_info.buffer; 1643 auto buffer_state = GetBufferState(device_data_, buffer); 1644 // Verify that buffer underlying the view hasn't been destroyed prematurely 1645 if (!buffer_state) { 1646 *error_code = VALIDATION_ERROR_15c00286; 1647 std::stringstream error_str; 1648 error_str << "Attempted write update to texel buffer descriptor failed because underlying buffer (" << buffer 1649 << ") has been destroyed: " << error_msg->c_str(); 1650 *error_msg = error_str.str(); 1651 return false; 1652 } else if (!ValidateBufferUsage(buffer_state, update->descriptorType, error_code, error_msg)) { 1653 std::stringstream error_str; 1654 error_str << "Attempted write update to texel buffer descriptor failed due to: " << error_msg->c_str(); 1655 *error_msg = error_str.str(); 1656 return false; 1657 } 1658 } 1659 break; 1660 } 1661 case VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER: 1662 case VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER_DYNAMIC: 1663 case VK_DESCRIPTOR_TYPE_STORAGE_BUFFER: 1664 case VK_DESCRIPTOR_TYPE_STORAGE_BUFFER_DYNAMIC: { 1665 for (uint32_t di = 0; di < update->descriptorCount; ++di) { 1666 if (!ValidateBufferUpdate(update->pBufferInfo + di, update->descriptorType, error_code, error_msg)) { 1667 std::stringstream error_str; 1668 error_str << "Attempted write update to buffer descriptor failed due to: " << error_msg->c_str(); 1669 *error_msg = error_str.str(); 1670 return false; 1671 } 1672 } 1673 break; 1674 } 1675 default: 1676 assert(0); // We've already verified update type so should never get here 1677 break; 1678 } 1679 // All checks passed so update contents are good 1680 return true; 1681} 1682// Verify that the contents of the update are ok, but don't perform actual update 1683bool cvdescriptorset::DescriptorSet::VerifyCopyUpdateContents(const VkCopyDescriptorSet *update, const DescriptorSet *src_set, 1684 VkDescriptorType type, uint32_t index, 1685 UNIQUE_VALIDATION_ERROR_CODE *error_code, 1686 std::string *error_msg) const { 1687 // Note : Repurposing some Write update error codes here as specific details aren't called out for copy updates like they are 1688 // for write updates 1689 switch (src_set->descriptors_[index]->descriptor_class) { 1690 case PlainSampler: { 1691 for (uint32_t di = 0; di < update->descriptorCount; ++di) { 1692 const auto src_desc = src_set->descriptors_[index + di].get(); 1693 if (!src_desc->updated) continue; 1694 if (!src_desc->IsImmutableSampler()) { 1695 auto update_sampler = static_cast<SamplerDescriptor *>(src_desc)->GetSampler(); 1696 if (!ValidateSampler(update_sampler, device_data_)) { 1697 *error_code = VALIDATION_ERROR_15c0028a; 1698 std::stringstream error_str; 1699 error_str << "Attempted copy update to sampler descriptor with invalid sampler: " << update_sampler << "."; 1700 *error_msg = error_str.str(); 1701 return false; 1702 } 1703 } else { 1704 // TODO : Warn here 1705 } 1706 } 1707 break; 1708 } 1709 case ImageSampler: { 1710 for (uint32_t di = 0; di < update->descriptorCount; ++di) { 1711 const auto src_desc = src_set->descriptors_[index + di].get(); 1712 if (!src_desc->updated) continue; 1713 auto img_samp_desc = static_cast<const ImageSamplerDescriptor *>(src_desc); 1714 // First validate sampler 1715 if (!img_samp_desc->IsImmutableSampler()) { 1716 auto update_sampler = img_samp_desc->GetSampler(); 1717 if (!ValidateSampler(update_sampler, device_data_)) { 1718 *error_code = VALIDATION_ERROR_15c0028a; 1719 std::stringstream error_str; 1720 error_str << "Attempted copy update to sampler descriptor with invalid sampler: " << update_sampler << "."; 1721 *error_msg = error_str.str(); 1722 return false; 1723 } 1724 } else { 1725 // TODO : Warn here 1726 } 1727 // Validate image 1728 auto image_view = img_samp_desc->GetImageView(); 1729 auto image_layout = img_samp_desc->GetImageLayout(); 1730 if (!ValidateImageUpdate(image_view, image_layout, type, device_data_, error_code, error_msg)) { 1731 std::stringstream error_str; 1732 error_str << "Attempted copy update to combined image sampler descriptor failed due to: " << error_msg->c_str(); 1733 *error_msg = error_str.str(); 1734 return false; 1735 } 1736 } 1737 break; 1738 } 1739 case Image: { 1740 for (uint32_t di = 0; di < update->descriptorCount; ++di) { 1741 const auto src_desc = src_set->descriptors_[index + di].get(); 1742 if (!src_desc->updated) continue; 1743 auto img_desc = static_cast<const ImageDescriptor *>(src_desc); 1744 auto image_view = img_desc->GetImageView(); 1745 auto image_layout = img_desc->GetImageLayout(); 1746 if (!ValidateImageUpdate(image_view, image_layout, type, device_data_, error_code, error_msg)) { 1747 std::stringstream error_str; 1748 error_str << "Attempted copy update to image descriptor failed due to: " << error_msg->c_str(); 1749 *error_msg = error_str.str(); 1750 return false; 1751 } 1752 } 1753 break; 1754 } 1755 case TexelBuffer: { 1756 for (uint32_t di = 0; di < update->descriptorCount; ++di) { 1757 const auto src_desc = src_set->descriptors_[index + di].get(); 1758 if (!src_desc->updated) continue; 1759 auto buffer_view = static_cast<TexelDescriptor *>(src_desc)->GetBufferView(); 1760 auto bv_state = GetBufferViewState(device_data_, buffer_view); 1761 if (!bv_state) { 1762 *error_code = VALIDATION_ERROR_15c00286; 1763 std::stringstream error_str; 1764 error_str << "Attempted copy update to texel buffer descriptor with invalid buffer view: " << buffer_view; 1765 *error_msg = error_str.str(); 1766 return false; 1767 } 1768 auto buffer = bv_state->create_info.buffer; 1769 if (!ValidateBufferUsage(GetBufferState(device_data_, buffer), type, error_code, error_msg)) { 1770 std::stringstream error_str; 1771 error_str << "Attempted copy update to texel buffer descriptor failed due to: " << error_msg->c_str(); 1772 *error_msg = error_str.str(); 1773 return false; 1774 } 1775 } 1776 break; 1777 } 1778 case GeneralBuffer: { 1779 for (uint32_t di = 0; di < update->descriptorCount; ++di) { 1780 const auto src_desc = src_set->descriptors_[index + di].get(); 1781 if (!src_desc->updated) continue; 1782 auto buffer = static_cast<BufferDescriptor *>(src_desc)->GetBuffer(); 1783 if (!ValidateBufferUsage(GetBufferState(device_data_, buffer), type, error_code, error_msg)) { 1784 std::stringstream error_str; 1785 error_str << "Attempted copy update to buffer descriptor failed due to: " << error_msg->c_str(); 1786 *error_msg = error_str.str(); 1787 return false; 1788 } 1789 } 1790 break; 1791 } 1792 default: 1793 assert(0); // We've already verified update type so should never get here 1794 break; 1795 } 1796 // All checks passed so update contents are good 1797 return true; 1798} 1799// Update the common AllocateDescriptorSetsData 1800void cvdescriptorset::UpdateAllocateDescriptorSetsData(const layer_data *dev_data, const VkDescriptorSetAllocateInfo *p_alloc_info, 1801 AllocateDescriptorSetsData *ds_data) { 1802 for (uint32_t i = 0; i < p_alloc_info->descriptorSetCount; i++) { 1803 auto layout = GetDescriptorSetLayout(dev_data, p_alloc_info->pSetLayouts[i]); 1804 if (layout) { 1805 ds_data->layout_nodes[i] = layout; 1806 // Count total descriptors required per type 1807 for (uint32_t j = 0; j < layout->GetBindingCount(); ++j) { 1808 const auto &binding_layout = layout->GetDescriptorSetLayoutBindingPtrFromIndex(j); 1809 uint32_t typeIndex = static_cast<uint32_t>(binding_layout->descriptorType); 1810 ds_data->required_descriptors_by_type[typeIndex] += binding_layout->descriptorCount; 1811 } 1812 } 1813 // Any unknown layouts will be flagged as errors during ValidateAllocateDescriptorSets() call 1814 } 1815} 1816// Verify that the state at allocate time is correct, but don't actually allocate the sets yet 1817bool cvdescriptorset::ValidateAllocateDescriptorSets(const core_validation::layer_data *dev_data, 1818 const VkDescriptorSetAllocateInfo *p_alloc_info, 1819 const AllocateDescriptorSetsData *ds_data) { 1820 bool skip = false; 1821 auto report_data = core_validation::GetReportData(dev_data); 1822 1823 for (uint32_t i = 0; i < p_alloc_info->descriptorSetCount; i++) { 1824 auto layout = GetDescriptorSetLayout(dev_data, p_alloc_info->pSetLayouts[i]); 1825 if (layout) { // nullptr layout indicates no valid layout handle for this device, validated/logged in object_tracker 1826 if (layout->GetCreateFlags() & VK_DESCRIPTOR_SET_LAYOUT_CREATE_PUSH_DESCRIPTOR_BIT_KHR) { 1827 skip |= log_msg(report_data, VK_DEBUG_REPORT_ERROR_BIT_EXT, VK_DEBUG_REPORT_OBJECT_TYPE_DESCRIPTOR_SET_LAYOUT_EXT, 1828 HandleToUint64(p_alloc_info->pSetLayouts[i]), __LINE__, VALIDATION_ERROR_04c00268, "DS", 1829 "Layout 0x%" PRIxLEAST64 " specified at pSetLayouts[%" PRIu32 1830 "] in vkAllocateDescriptorSets() was created with invalid flag %s set. %s", 1831 HandleToUint64(p_alloc_info->pSetLayouts[i]), i, 1832 "VK_DESCRIPTOR_SET_LAYOUT_CREATE_PUSH_DESCRIPTOR_BIT_KHR", 1833 validation_error_map[VALIDATION_ERROR_04c00268]); 1834 } 1835 } 1836 } 1837 if (!GetDeviceExtensions(dev_data)->vk_khr_maintenance1) { 1838 auto pool_state = GetDescriptorPoolState(dev_data, p_alloc_info->descriptorPool); 1839 // Track number of descriptorSets allowable in this pool 1840 if (pool_state->availableSets < p_alloc_info->descriptorSetCount) { 1841 skip |= log_msg(report_data, VK_DEBUG_REPORT_ERROR_BIT_EXT, VK_DEBUG_REPORT_OBJECT_TYPE_DESCRIPTOR_POOL_EXT, 1842 HandleToUint64(pool_state->pool), __LINE__, VALIDATION_ERROR_04c00264, "DS", 1843 "Unable to allocate %u descriptorSets from pool 0x%" PRIxLEAST64 1844 ". This pool only has %d descriptorSets remaining. %s", 1845 p_alloc_info->descriptorSetCount, HandleToUint64(pool_state->pool), pool_state->availableSets, 1846 validation_error_map[VALIDATION_ERROR_04c00264]); 1847 } 1848 // Determine whether descriptor counts are satisfiable 1849 for (uint32_t i = 0; i < VK_DESCRIPTOR_TYPE_RANGE_SIZE; i++) { 1850 if (ds_data->required_descriptors_by_type[i] > pool_state->availableDescriptorTypeCount[i]) { 1851 skip |= log_msg(report_data, VK_DEBUG_REPORT_ERROR_BIT_EXT, VK_DEBUG_REPORT_OBJECT_TYPE_DESCRIPTOR_POOL_EXT, 1852 HandleToUint64(pool_state->pool), __LINE__, VALIDATION_ERROR_04c00266, "DS", 1853 "Unable to allocate %u descriptors of type %s from pool 0x%" PRIxLEAST64 1854 ". This pool only has %d descriptors of this type remaining. %s", 1855 ds_data->required_descriptors_by_type[i], string_VkDescriptorType(VkDescriptorType(i)), 1856 HandleToUint64(pool_state->pool), pool_state->availableDescriptorTypeCount[i], 1857 validation_error_map[VALIDATION_ERROR_04c00266]); 1858 } 1859 } 1860 } 1861 1862 return skip; 1863} 1864// Decrement allocated sets from the pool and insert new sets into set_map 1865void cvdescriptorset::PerformAllocateDescriptorSets(const VkDescriptorSetAllocateInfo *p_alloc_info, 1866 const VkDescriptorSet *descriptor_sets, 1867 const AllocateDescriptorSetsData *ds_data, 1868 std::unordered_map<VkDescriptorPool, DESCRIPTOR_POOL_STATE *> *pool_map, 1869 std::unordered_map<VkDescriptorSet, cvdescriptorset::DescriptorSet *> *set_map, 1870 layer_data *dev_data) { 1871 auto pool_state = (*pool_map)[p_alloc_info->descriptorPool]; 1872 // Account for sets and individual descriptors allocated from pool 1873 pool_state->availableSets -= p_alloc_info->descriptorSetCount; 1874 for (uint32_t i = 0; i < VK_DESCRIPTOR_TYPE_RANGE_SIZE; i++) { 1875 pool_state->availableDescriptorTypeCount[i] -= ds_data->required_descriptors_by_type[i]; 1876 } 1877 // Create tracking object for each descriptor set; insert into global map and the pool's set. 1878 for (uint32_t i = 0; i < p_alloc_info->descriptorSetCount; i++) { 1879 auto new_ds = new cvdescriptorset::DescriptorSet(descriptor_sets[i], p_alloc_info->descriptorPool, ds_data->layout_nodes[i], 1880 dev_data); 1881 1882 pool_state->sets.insert(new_ds); 1883 new_ds->in_use.store(0); 1884 (*set_map)[descriptor_sets[i]] = new_ds; 1885 } 1886} 1887 1888cvdescriptorset::PrefilterBindRequestMap::PrefilterBindRequestMap(cvdescriptorset::DescriptorSet &ds, const BindingReqMap &in_map, 1889 GLOBAL_CB_NODE *cb_state) 1890 : filtered_map_(), orig_map_(in_map) { 1891 if (ds.GetTotalDescriptorCount() > kManyDescriptors_) { 1892 filtered_map_.reset(new std::map<uint32_t, descriptor_req>()); 1893 ds.FilterAndTrackBindingReqs(cb_state, orig_map_, filtered_map_.get()); 1894 } 1895} 1896cvdescriptorset::PrefilterBindRequestMap::PrefilterBindRequestMap(cvdescriptorset::DescriptorSet &ds, const BindingReqMap &in_map, 1897 GLOBAL_CB_NODE *cb_state, PIPELINE_STATE *pipeline) 1898 : filtered_map_(), orig_map_(in_map) { 1899 if (ds.GetTotalDescriptorCount() > kManyDescriptors_) { 1900 filtered_map_.reset(new std::map<uint32_t, descriptor_req>()); 1901 ds.FilterAndTrackBindingReqs(cb_state, pipeline, orig_map_, filtered_map_.get()); 1902 } 1903} 1904