/external/compiler-rt/lib/lsan/ |
H A D | lsan_allocator.cc | 31 uptr requested_size : 54; member in struct:__lsan::ChunkMetadata 77 m->requested_size = size; 134 return m->requested_size; 163 if (addr < chunk + m->requested_size) 165 if (IsSpecialCaseOfOperatorNew0(chunk, m->requested_size, addr)) 191 uptr LsanMetadata::requested_size() const { function in class:__lsan::LsanMetadata 192 return reinterpret_cast<ChunkMetadata *>(metadata_)->requested_size; 208 if (m->allocated && (uptr)p < (uptr)chunk + m->requested_size) {
|
H A D | lsan_common.cc | 170 pp, p, chunk, chunk + m.requested_size(), m.requested_size()); 176 chunk, chunk + m.requested_size(), m.requested_size()); 311 ScanRangeForPointers(next_chunk, next_chunk + m.requested_size(), frontier, 322 ScanRangeForPointers(chunk, chunk + m.requested_size(), 335 chunk, chunk + m.requested_size(), m.requested_size()); 397 leak_report->AddLeakedChunk(chunk, stack_trace_id, m.requested_size(),
|
H A D | lsan_common.h | 181 uptr requested_size() const;
|
/external/compiler-rt/lib/msan/ |
H A D | msan_allocator.cc | 26 uptr requested_size; member in struct:__msan::Metadata 129 meta->requested_size = size; 148 uptr size = meta->requested_size; 149 meta->requested_size = 0; 186 uptr old_size = meta->requested_size; 190 meta->requested_size = new_size; 217 return b->requested_size;
|
/external/v8/src/zone/ |
H A D | accounting-allocator.cc | 126 Segment* AccountingAllocator::GetSegmentFromPool(size_t requested_size) { argument 127 if (requested_size > (1 << kMaxSegmentSizePower)) { 132 while (requested_size > (static_cast<size_t>(1) << power)) power++; 154 DCHECK_GE(segment->size(), requested_size);
|
H A D | zone.cc | 115 Segment* Zone::NewSegment(size_t requested_size) { argument 116 Segment* result = allocator_->GetSegment(requested_size); 118 DCHECK_GE(result->size(), requested_size);
|
H A D | accounting-allocator.h | 68 Segment* GetSegmentFromPool(size_t requested_size);
|
H A D | zone.h | 100 inline Segment* NewSegment(size_t requested_size);
|
/external/valgrind/coregrind/m_ume/ |
H A D | macho.c | 404 handle_lcmain ( vki_size_t requested_size, argument 407 if (requested_size == 0) { 408 requested_size = default_stack_size(); 410 requested_size = VG_PGROUNDUP(requested_size); 413 requested_size += HACK; 415 SysRes res = VG_(am_mmap_anon_float_client)(requested_size, 417 check_mmap_float(res, requested_size, "handle_lcmain"); 420 out_info->stack_end = out_info->stack_start + requested_size - 1;
|
/external/tensorflow/tensorflow/core/common_runtime/gpu/ |
H A D | gpu_bfc_allocator_test.cc | 389 const size_t requested_size = 2 * (256 << i); local 390 EXPECT_EQ(requested_size, a.RequestedSize(initial_ptrs[2 * i]) + 396 EXPECT_EQ(bin_info.total_requested_bytes_in_use, requested_size); 426 size_t requested_size = 256 << i; local 427 EXPECT_EQ(requested_size, a.RequestedSize(initial_ptrs[2 * i])); 432 EXPECT_EQ(bin_info.total_requested_bytes_in_use, requested_size);
|
/external/tensorflow/tensorflow/core/framework/ |
H A D | tracking_allocator.h | 124 size_t requested_size; member in struct:tensorflow::TrackingAllocator::Chunk
|
H A D | tracking_allocator.cc | 121 return (*it).second.requested_size;
|
/external/tensorflow/tensorflow/core/common_runtime/ |
H A D | bfc_allocator.cc | 314 chunk->requested_size = num_bytes; 530 return c->requested_size; 604 size_t wasted = c->size - c->requested_size; 607 region_offset + c->requested_size, region.ptr(), c->ptr, 612 region.ptr(), c->ptr, c->requested_size, '*'); 708 bin_info.total_requested_bytes_in_use += c->requested_size;
|
H A D | bfc_allocator.h | 100 // fragmentation. requested_size keeps track of what the client 103 size_t requested_size = 0; member in struct:tensorflow::BFCAllocator::Chunk 132 " | Requested Size: ", strings::HumanReadableNumBytes(requested_size),
|
/external/libchrome/base/memory/ |
H A D | shared_memory.h | 184 size_t requested_size() const { return requested_size_; } function in class:base::SharedMemory
|
H A D | shared_memory_unittest.cc | 159 EXPECT_EQ(memory1.requested_size(), kDataSize); 183 EXPECT_EQ(memory2.requested_size(), 0U); 495 EXPECT_TRUE(shared_memory.Map(shared_memory.requested_size())); 497 EXPECT_EQ(0, mprotect(shared_memory.memory(), shared_memory.requested_size(),
|
/external/pdfium/third_party/base/allocator/partition_allocator/ |
H A D | partition_alloc.h | 711 size_t requested_size = size; 719 PartitionAllocHooks::AllocationHookIfEnabled(result, requested_size, 795 size_t requested_size = size; 803 PartitionAllocHooks::AllocationHookIfEnabled(ret, requested_size, type_name);
|
/external/python/cpython3/Objects/ |
H A D | bytearrayobject.c | 168 PyByteArray_Resize(PyObject *self, Py_ssize_t requested_size) argument 176 size_t size = (size_t) requested_size; 181 assert(requested_size >= 0); 183 if (requested_size == Py_SIZE(self)) { 227 Py_MIN(requested_size, Py_SIZE(self)));
|
/external/mesa3d/src/gallium/winsys/amdgpu/drm/ |
H A D | amdgpu_cs.c | 791 unsigned requested_size = rcs->prev_dw + rcs->current.cdw + dw; local 797 if (requested_size > amdgpu_ib_max_submit_dwords(ib->ib_type)) 800 ib->max_ib_size = MAX2(ib->max_ib_size, requested_size);
|
/external/v8/src/heap/ |
H A D | spaces.cc | 204 Address CodeRange::AllocateRawMemory(const size_t requested_size, argument 210 requested_size - 2 * MemoryAllocator::CodePageGuardSize()); 212 if (!ReserveBlock(requested_size, ¤t)) { 257 bool CodeRange::ReserveBlock(const size_t requested_size, FreeBlock* block) { argument 262 requested_size > allocation_list_[current_allocation_block_index_].size) { 264 if (!GetNextAllocationBlock(requested_size)) return false; 267 size_t aligned_requested = RoundUp(requested_size, MemoryChunk::kAlignment);
|
H A D | spaces.h | 959 bool SetUp(size_t requested_size); 979 MUST_USE_RESULT Address AllocateRawMemory(const size_t requested_size, 1017 bool ReserveBlock(const size_t requested_size, FreeBlock* block);
|
/external/compiler-rt/lib/asan/ |
H A D | asan_allocator.cc | 856 uptr LsanMetadata::requested_size() const { function in class:__lsan::LsanMetadata
|