/external/deqp/external/vulkancts/modules/vulkan/draw/ |
H A D | vktDrawBufferObjectUtil.cpp | 42 void Buffer::bindMemory (de::MovePtr<vk::Allocation> allocation) argument 44 DE_ASSERT(allocation); 45 VK_CHECK(m_vk.bindBufferMemory(m_device, *m_object, allocation->getMemory(), allocation->getOffset())); 48 m_allocation = allocation;
|
H A D | vktDrawImageObjectUtil.cpp | 847 void Image::bindMemory (de::MovePtr<vk::Allocation> allocation) argument 849 DE_ASSERT(allocation); 850 VK_CHECK(m_vk.bindImageMemory(m_device, *m_object, allocation->getMemory(), allocation->getOffset())); 853 m_allocation = allocation;
|
/external/deqp/external/vulkancts/modules/vulkan/dynamic_state/ |
H A D | vktDynamicStateBufferObjectUtil.cpp | 42 void Buffer::bindMemory (de::MovePtr<vk::Allocation> allocation) argument 44 DE_ASSERT(allocation); 45 VK_CHECK(m_vk.bindBufferMemory(m_device, *m_object, allocation->getMemory(), allocation->getOffset())); 48 m_allocation = allocation;
|
H A D | vktDynamicStateImageObjectUtil.cpp | 847 void Image::bindMemory (de::MovePtr<vk::Allocation> allocation) argument 849 DE_ASSERT(allocation); 850 VK_CHECK(m_vk.bindImageMemory(m_device, *m_object, allocation->getMemory(), allocation->getOffset())); 853 m_allocation = allocation;
|
/external/deqp/external/vulkancts/modules/vulkan/query_pool/ |
H A D | vktQueryPoolBufferObjectUtil.cpp | 42 void Buffer::bindMemory (de::MovePtr<vk::Allocation> allocation) argument 44 DE_ASSERT(allocation); 45 VK_CHECK(m_vk.bindBufferMemory(m_device, *m_object, allocation->getMemory(), allocation->getOffset())); 48 m_allocation = allocation;
|
H A D | vktQueryPoolImageObjectUtil.cpp | 826 void Image::bindMemory (de::MovePtr<vk::Allocation> allocation) argument 828 DE_ASSERT(allocation); 829 VK_CHECK(m_vk.bindImageMemory(m_device, *m_object, allocation->getMemory(), allocation->getOffset())); 832 m_allocation = allocation;
|
/external/deqp/external/vulkancts/modules/vulkan/api/ |
H A D | vktApiComputeInstanceResultBuffer.cpp | 71 de::MovePtr<Allocation> allocation = allocator.allocate(requirements, MemoryRequirement::HostVisible); local 73 VK_CHECK(vki.bindBufferMemory(device, *buffer, allocation->getMemory(), allocation->getOffset())); 76 void* mapPtr = allocation->getHostPtr(); 81 flushMappedMemoryRange(vki, device, allocation->getMemory(), allocation->getOffset(), (VkDeviceSize) DATA_SIZE); 83 *outAllocation = allocation;
|
H A D | vktApiBufferComputeInstance.cpp | 64 de::MovePtr<Allocation> allocation = allocator.allocate(requirements, MemoryRequirement::HostVisible); local 66 VK_CHECK(vki.bindBufferMemory(device, *buffer, allocation->getMemory(), allocation->getOffset())); 69 void* mapPtr = allocation->getHostPtr(); 79 flushMappedMemoryRange(vki, device, allocation->getMemory(), allocation->getOffset(), bufferSize); 81 *outAllocation = allocation;
|
/external/clang/test/Sema/ |
H A D | warn-unused-function.c | 50 static void cleanupMalloc(char * const * const allocation) { } argument
|
/external/webrtc/webrtc/call/ |
H A D | bitrate_allocator.cc | 43 ObserverBitrateMap allocation = AllocateBitrates(); local 44 for (const auto& kv : allocation) { 88 ObserverBitrateMap allocation = AllocateBitrates(); local 90 for (auto& kv : allocation) { 150 ObserverBitrateMap allocation; local 163 allocation[max_it->second.observer] = max_it->first; 165 allocation[max_it->second.observer] = observer_allowance; 171 return allocation; 176 ObserverBitrateMap allocation; local 180 allocation[observe [all...] |
/external/deqp/external/vulkancts/framework/vulkan/ |
H A D | vkAllocationCallbackUtil.hpp | 23 * \brief Memory allocation callback utilities. 82 } allocation; member in union:vk::AllocationCallbackRecord::__anon3736 109 static AllocationCallbackRecord allocation (size_t size, size_t alignment, VkSystemAllocationScope scope, void* returnedPtr);
|
H A D | vkAllocationCallbackUtil.cpp | 21 * \brief Memory allocation callback utilities. 130 AllocationCallbackRecord AllocationCallbackRecord::allocation (size_t size, size_t alignment, VkSystemAllocationScope scope, void* returnedPtr) function in class:vk::AllocationCallbackRecord 135 record.data.allocation.size = size; 136 record.data.allocation.alignment = alignment; 137 record.data.allocation.scope = scope; 138 record.data.allocation.returnedPtr = returnedPtr; 243 m_records.append(AllocationCallbackRecord::allocation(size, alignment, allocationScope, ptr)); 350 return record.data.allocation.alignment; 377 const VkSystemAllocationScope* const scopePtr = record.type == AllocationCallbackRecord::TYPE_ALLOCATION ? &record.data.allocation.scope 395 // Validate actual allocation behavio [all...] |
/external/deqp/external/vulkancts/modules/vulkan/sparse_resources/ |
H A D | vktSparseResourcesBufferSparseBinding.cpp | 304 const Allocation& allocation = outputBuffer->getAllocation(); local 306 invalidateMappedMemoryRange(deviceInterface, *m_logicalDevice, allocation.getMemory(), allocation.getOffset(), m_bufferSize); 308 const deUint8* outputData = static_cast<const deUint8*>(allocation.getHostPtr());
|
H A D | vktSparseResourcesBufferSparseResidency.cpp | 375 const Allocation& allocation = outputBuffer->getAllocation(); local 377 invalidateMappedMemoryRange(deviceInterface, *m_logicalDevice, allocation.getMemory(), allocation.getOffset(), m_bufferSize); 379 const deUint8* outputData = static_cast<const deUint8*>(allocation.getHostPtr());
|
H A D | vktSparseResourcesImageSparseBinding.cpp | 374 const Allocation& allocation = outputBuffer->getAllocation(); local 376 invalidateMappedMemoryRange(deviceInterface, *m_logicalDevice, allocation.getMemory(), allocation.getOffset(), imageSizeInBytes); 378 const deUint8* outputData = static_cast<const deUint8*>(allocation.getHostPtr());
|
H A D | vktSparseResourcesBufferMemoryAliasing.cpp | 385 const Allocation& allocation = outputBuffer->getAllocation(); local 387 invalidateMappedMemoryRange(deviceInterface, *m_logicalDevice, allocation.getMemory(), allocation.getOffset(), m_bufferSizeInBytes); 389 const deUint8* outputData = static_cast<const deUint8*>(allocation.getHostPtr());
|
H A D | vktSparseResourcesMipmapSparseResidency.cpp | 524 const Allocation& allocation = outputBuffer->getAllocation(); local 526 invalidateMappedMemoryRange(deviceInterface, *m_logicalDevice, allocation.getMemory(), allocation.getOffset(), imageSizeInBytes); 528 const deUint8* outputData = static_cast<const deUint8*>(allocation.getHostPtr());
|
H A D | vktSparseResourcesImageSparseResidency.cpp | 639 const Allocation& allocation = outputBuffer->getAllocation(); local 641 invalidateMappedMemoryRange(deviceInterface, *m_logicalDevice, allocation.getMemory(), allocation.getOffset(), imageSizeInBytes); 643 const deUint8* outputData = static_cast<const deUint8*>(allocation.getHostPtr());
|
/external/v8/src/heap/ |
H A D | scavenger.cc | 117 // over-allocation. 153 AllocationResult allocation = local 157 if (allocation.To(&target)) { 181 AllocationResult allocation = local 185 if (allocation.To(&target)) {
|
/external/v8/test/cctest/heap/ |
H A D | test-mark-compact.cc | 137 AllocationResult allocation; local 139 allocation = heap->AllocateFixedArray(arraysize); 140 } while (!allocation.IsRetry()); 146 allocation = heap->AllocateMap(JS_OBJECT_TYPE, JSObject::kHeaderSize); 147 } while (!allocation.IsRetry());
|
H A D | test-spaces.cc | 468 { AllocationResult allocation = lo->AllocateRaw(lo_size, NOT_EXECUTABLE); local 469 if (allocation.IsRetry()) break; 521 AllocationResult allocation = space->AllocateRawUnaligned(size); local 522 CHECK(!allocation.IsRetry()); 524 CHECK(allocation.To(&filler)); 531 AllocationResult allocation = space->AllocateRaw(size, kDoubleUnaligned); local 532 CHECK(!allocation.IsRetry()); 534 CHECK(allocation.To(&filler)); 541 AllocationResult allocation = space->AllocateRaw(size, EXECUTABLE); local 542 CHECK(!allocation [all...] |
/external/v8/test/unittests/compiler/ |
H A D | escape-analysis-unittest.cc | 77 Node* Store(const FieldAccess& access, Node* allocation, Node* value, argument 86 allocation, value, effect, control); 89 Node* StoreElement(const ElementAccess& access, Node* allocation, Node* index, argument 99 graph()->NewNode(simplified()->StoreElement(access), allocation, 213 Node* allocation = Allocate(Constant(kPointerSize)); local 214 Store(FieldAccessAtIndex(0), allocation, object1); 215 Node* finish = FinishRegion(allocation); 222 ExpectVirtual(allocation); 235 Node* allocation = Allocate(Constant(kPointerSize)); local 236 Store(FieldAccessAtIndex(0), allocation, object 260 Node* allocation = Allocate(Constant(kPointerSize)); local 311 Node* allocation = Allocate(Constant(kPointerSize)); local 345 Node* allocation = Allocate(Constant(kPointerSize)); local 376 Node* allocation = Allocate(Constant(kPointerSize)); local 410 Node* allocation = Allocate(Constant(kPointerSize)); local 435 Node* allocation = Allocate(Constant(kPointerSize)); local 474 Node* allocation = Allocate(Constant(kPointerSize * 2)); local 476 Store(FieldAccessAtIndex(kPointerSize), allocation, allocation); local [all...] |
/external/ImageMagick/MagickCore/ |
H A D | memory.c | 40 % Our custom memory allocation manager implements a best-fit allocation policy 106 *allocation, 153 allocation; 433 memory_pool.allocation+=size; 482 memory_pool.allocation=SegmentSize; 752 memory_pool.segments[i]->allocation); 754 (void) UnmapBlob(memory_pool.segments[i]->allocation, 820 segment_info->allocation=segment; 823 for ( ; (i >= 0) && (memory_pool.segments[i]->allocation > segmen 103 *allocation, member in struct:_DataSegmentInfo 150 allocation; member in struct:_MemoryPool [all...] |
/external/skia/src/gpu/text/ |
H A D | GrAtlasTextBlob.cpp | 29 void* allocation = pool->allocate(size); local 31 sk_bzero(allocation, size); 34 GrAtlasTextBlob* cacheBlob = new (allocation) GrAtlasTextBlob;
|
/external/v8/src/ast/ |
H A D | scopeinfo.cc | 341 VariableAllocationInfo allocation = ReceiverVariableField::decode(Flags()); local 342 return allocation == STACK || allocation == CONTEXT;
|