/art/runtime/base/ |
H A D | stringprintf_test.cc | 24 size_t size = 0x00107e59; local 25 EXPECT_STREQ("00107e59", StringPrintf("%08zx", size).c_str()); 26 EXPECT_STREQ("0x00107e59", StringPrintf("0x%08zx", size).c_str());
|
H A D | bounded_fifo.h | 22 // A bounded fifo is a fifo which has a bounded size. The power of two version uses a bit mask to 39 return size() == 0; 42 size_t size() const { function in class:art::BoundedFifoPowerOfTwo
|
H A D | allocator.cc | 37 virtual void* Alloc(size_t size) { argument 38 return calloc(sizeof(uint8_t), size); 56 virtual void* Alloc(size_t size) { argument
|
/art/compiler/utils/ |
H A D | stack_checks.h | 37 static inline bool FrameNeedsStackCheck(size_t size, InstructionSet isa) { argument 38 return size >= kLargeFrameSize;
|
H A D | allocation.h | 27 // Allocate a new ArenaObject of 'size' bytes in the Arena. 28 void* operator new(size_t size, ArenaAllocator* allocator) { argument 29 return allocator->Alloc(size, kArenaAllocMisc); 39 void* operator new(size_t size) { argument
|
H A D | arena_bit_vector.cc | 28 virtual void* Alloc(size_t size) { argument 29 return arena_->Alloc(size, kArenaAllocGrowableBitMap); 34 static void* operator new(size_t size, ArenaAlloc* arena) { argument
|
H A D | arena_bit_vector.h | 61 static void* operator new(size_t size, ArenaAllocator* arena) { argument 64 static void* operator new(size_t size, ScopedArenaAllocator* arena) { argument
|
H A D | array_ref.h | 64 template <size_t size> 65 constexpr ArrayRef(T (&array)[size]) 66 : array_(array), size_(size) { 69 template <typename U, size_t size> 70 constexpr ArrayRef(U (&array)[size], 72 : array_(array), size_(size) { 75 constexpr ArrayRef(T* array, size_t size) argument 76 : array_(array), size_(size) { 80 constexpr ArrayRef(U* array, size_t size, argument 82 : array_(array), size_(size) { 127 size_type size() const { return size_; } function in class:art::ArrayRef [all...] |
H A D | managed_register.h | 86 // ManagedRegisterSpill contains information about data type size and location in caller frame 88 ManagedRegisterSpill(const ManagedRegister& other, uint32_t size, uint32_t spill_offset) argument 89 : ManagedRegister(other), size_(size), spill_offset_(spill_offset) { } 94 explicit ManagedRegisterSpill(const ManagedRegister& other, int32_t size) argument 95 : ManagedRegister(other), size_(size), spill_offset_(-1) { } 112 // The ManagedRegister does not have information about size and offset. 113 // In this case it's size and offset determined by BuildFrame (assembler)
|
/art/runtime/gc/allocator/ |
H A D | rosalloc-inl.h | 27 inline ALWAYS_INLINE void* RosAlloc::Alloc(Thread* self, size_t size, size_t* bytes_allocated) { argument 28 if (UNLIKELY(size > kLargeSizeThreshold)) { 29 return AllocLargeObject(self, size, bytes_allocated); 33 m = AllocFromRun(self, size, bytes_allocated); 35 m = AllocFromRunThreadUnsafe(self, size, bytes_allocated); 40 for (size_t i = 0; i < size; ++i) {
|
/art/compiler/dex/ |
H A D | mir_field_info.cc | 84 size_t size = OFFSETOF_MEMBER(MirSFieldLoweringInfo, storage_index_) + local 86 DCHECK_EQ(memcmp(&unresolved, &*it, size), 0);
|
/art/compiler/optimizing/ |
H A D | stack_map_test.cc | 44 size_t size = stream.ComputeNeededSize(); local 45 void* memory = arena.Alloc(size, kArenaAllocMisc); 46 MemoryRegion region(memory, size); 91 size_t size = stream.ComputeNeededSize(); local 92 void* memory = arena.Alloc(size, kArenaAllocMisc); 93 MemoryRegion region(memory, size);
|
H A D | codegen_test.cc | 34 virtual uint8_t* Allocate(size_t size) { argument 35 size_ = size; 36 memory_.reset(new uint8_t[size]);
|
H A D | optimizing_compiler.cc | 41 virtual uint8_t* Allocate(size_t size) { argument 42 size_ = size; 43 memory_.resize(size);
|
/art/runtime/gc/space/ |
H A D | dlmalloc_space-inl.h | 45 size_t size = mspace_usable_size(obj_ptr); local 47 *usable_size = size; 49 return size + kChunkOverhead;
|
H A D | rosalloc_space-inl.h | 30 // obj is a valid object. Use its class in the header to get the size. 32 size_t size = obj->SizeOf<kVerifyNone>(); local 33 size_t size_by_size = rosalloc_->UsableSize(size); 37 LOG(INFO) << "Found a bad sized obj of size " << size
|
H A D | large_object_space_test.cc | 49 while (requests.size() < num_allocations) { 65 for (size_t j = 0; j < requests.size(); ++j) { 66 std::swap(requests[j], requests[test_rand(&rand_seed) % requests.size()]); 71 size_t limit = !phase ? requests.size() / 2 : 0; 72 while (requests.size() > limit) { 100 AllocRaceTask(size_t id, size_t iterations, size_t size, LargeObjectSpace* los) : argument 101 id_(id), iterations_(iterations), size_(size), los_(los) {}
|
/art/runtime/ |
H A D | dex_file_verifier.h | 29 static bool Verify(const DexFile* dex_file, const byte* begin, size_t size, 37 DexFileVerifier(const DexFile* dex_file, const byte* begin, size_t size, const char* location) argument 38 : dex_file_(dex_file), begin_(begin), size_(size), location_(location), 46 // Check a list. The head is assumed to be at *ptr, and elements to be of size element_size. If 49 // Checks whether the offset is zero (when size is zero) or that the offset falls within the area 51 bool CheckValidOffsetAndSize(uint32_t offset, uint32_t size, const char* label); 57 uint32_t ReadUnsignedLittleEndian(uint32_t size);
|
H A D | vmap_table.h | 38 size_t size = DecodeUnsignedLeb128(&table); local 39 CHECK_LT(n, size); 102 DecodeUnsignedLeb128(&table); // Skip size.
|
H A D | leb128.h | 146 void Reserve(uint32_t size) { argument 147 data_.reserve(size);
|
H A D | mapping_table.h | 119 uint32_t size = DexToPcSize(); local 120 return DexToPcIterator(this, size); 138 DecodeUnsignedLeb128(&table); // PC to Dex size, unused. 196 uint32_t size = PcToDexSize(); local 197 return PcToDexIterator(this, size);
|
H A D | memory_region.h | 34 MemoryRegion(void* pointer, uword size) : pointer_(pointer), size_(size) {} argument 37 size_t size() const { return size_; } function in class:art::MemoryRegion 80 MemoryRegion Subregion(uintptr_t offset, uintptr_t size) const { 81 CHECK_GE(this->size(), size); 82 CHECK_LE(offset, this->size() - size); 83 return MemoryRegion(reinterpret_cast<void*>(start() + offset), size); 89 size_ = (region.size() [all...] |
/art/runtime/base/unix_file/ |
H A D | mapped_file.cc | 53 PLOG(WARNING) << "Failed to mmap file '" << GetPath() << "' of size " 71 << "' to size " << file_size; 80 PLOG(WARNING) << "Failed to mmap file '" << GetPath() << "' of size " 92 PLOG(WARNING) << "Failed unmap file '" << GetPath() << "' of size " 155 int64_t MappedFile::size() const { function in class:unix_file::MappedFile
|
/art/compiler/jni/quick/x86_64/ |
H A D | calling_convention_x86_64.cc | 106 if (entry_spills_.size() == 0) { 111 int32_t size = IsParamALongOrDouble(itr_args_)? 8 : 4; local 113 ManagedRegisterSpill spill(in_reg, size, spill_offset); 149 // Method*, return address and callee save area size, local reference segment state 151 (2 + CalleeSaveRegisters().size()) * kFramePointerSize; 154 // Plus return value spill area size
|
/art/runtime/jdwp/ |
H A D | jdwp_event.h | 77 int size; /* JdwpStepSize */ member in struct:art::JDWP::JdwpEventMod::__anon107
|