/art/compiler/utils/ |
H A D | jni_macro_assembler.h | 75 virtual void Store(FrameOffset offs, ManagedRegister src, size_t size) = 0; 93 virtual void Load(ManagedRegister dest, FrameOffset src, size_t size) = 0; 97 size_t size) = 0; 111 virtual void Move(ManagedRegister dest, ManagedRegister src, size_t size) = 0; 123 virtual void Copy(FrameOffset dest, FrameOffset src, ManagedRegister scratch, size_t size) = 0; 129 size_t size) = 0; 135 size_t size) = 0; 141 size_t size) = 0; 148 size_t size) = 0; 155 size_t size) [all...] |
H A D | swap_space.h | 38 void* Alloc(size_t size) REQUIRES(!lock_); 39 void Free(void* ptr, size_t size) REQUIRES(!lock_); 51 mutable size_t size; member in struct:art::SwapSpace::SpaceChunk 57 return reinterpret_cast<uintptr_t>(ptr) + size; 70 // Map size to an iterator to free_by_start_'s entry. 73 : size(sz), free_by_start_entry(entry) { } 77 mutable size_t size; member in struct:art::SwapSpace::FreeBySizeEntry 82 if (lhs.size != rhs.size) { 83 return lhs.size < rh [all...] |
H A D | test_dex_file_builder.h | 109 header->string_ids_size_ = strings_.size(); 112 uint32_t type_ids_offset = string_ids_offset + strings_.size() * sizeof(DexFile::StringId); 118 header->type_ids_size_ = types_.size(); 121 uint32_t proto_ids_offset = type_ids_offset + types_.size() * sizeof(DexFile::TypeId); 126 size_t num_args = entry.first.args.size(); 134 header->proto_ids_size_ = protos_.size(); 137 uint32_t field_ids_offset = proto_ids_offset + protos_.size() * sizeof(DexFile::ProtoId); 143 header->field_ids_size_ = fields_.size(); 146 uint32_t method_ids_offset = field_ids_offset + fields_.size() * sizeof(DexFile::FieldId); 152 header->method_ids_size_ = methods_.size(); [all...] |
/art/compiler/utils/x86/ |
H A D | jni_macro_assembler_x86.h | 58 void Store(FrameOffset offs, ManagedRegister src, size_t size) OVERRIDE; 74 void Load(ManagedRegister dest, FrameOffset src, size_t size) OVERRIDE; 76 void LoadFromThread(ManagedRegister dest, ThreadOffset32 src, size_t size) OVERRIDE; 88 void Move(ManagedRegister dest, ManagedRegister src, size_t size) OVERRIDE; 99 void Copy(FrameOffset dest, FrameOffset src, ManagedRegister scratch, size_t size) OVERRIDE; 102 size_t size) OVERRIDE; 105 size_t size) OVERRIDE; 108 size_t size) OVERRIDE; 111 ManagedRegister scratch, size_t size) OVERRIDE; 114 ManagedRegister scratch, size_t size) OVERRID [all...] |
H A D | jni_macro_assembler_x86.cc | 52 for (int i = spill_regs.size() - 1; i >= 0; --i) { 70 for (size_t i = 0; i < entry_spills.size(); ++i) { 92 int adjust = frame_size - spill_regs.size() * kFramePointerSize - kFramePointerSize; 95 for (size_t i = 0; i < spill_regs.size(); ++i) { 123 void X86JNIMacroAssembler::Store(FrameOffset offs, ManagedRegister msrc, size_t size) { argument 126 CHECK_EQ(0u, size); 128 CHECK_EQ(4u, size); 131 CHECK_EQ(8u, size); 135 if (size == 4) { 142 if (size 186 Load(ManagedRegister mdest, FrameOffset src, size_t size) argument 213 LoadFromThread(ManagedRegister mdest, ThreadOffset32 src, size_t size) argument 274 SignExtend(ManagedRegister mreg, size_t size) argument 285 ZeroExtend(ManagedRegister mreg, size_t size) argument 296 Move(ManagedRegister mdest, ManagedRegister msrc, size_t size) argument 347 Copy(FrameOffset dest, FrameOffset src, ManagedRegister mscratch, size_t size) argument 370 Copy(ManagedRegister dest_base, Offset dest_offset, FrameOffset src, ManagedRegister scratch, size_t size) argument 381 Copy(FrameOffset dest, FrameOffset src_base, Offset src_offset, ManagedRegister mscratch, size_t size) argument 393 Copy(ManagedRegister dest, Offset dest_offset, ManagedRegister src, Offset src_offset, ManagedRegister scratch, size_t size) argument 405 Copy(FrameOffset dest, Offset dest_offset, FrameOffset src, Offset src_offset, ManagedRegister mscratch, size_t size) argument [all...] |
/art/compiler/utils/x86_64/ |
H A D | jni_macro_assembler_x86_64.h | 59 void Store(FrameOffset offs, ManagedRegister src, size_t size) OVERRIDE; 77 void Load(ManagedRegister dest, FrameOffset src, size_t size) OVERRIDE; 79 void LoadFromThread(ManagedRegister dest, ThreadOffset64 src, size_t size) OVERRIDE; 93 void Move(ManagedRegister dest, ManagedRegister src, size_t size); 104 void Copy(FrameOffset dest, FrameOffset src, ManagedRegister scratch, size_t size) OVERRIDE; 110 size_t size) OVERRIDE; 116 size_t size) OVERRIDE; 122 size_t size) OVERRIDE; 129 size_t size) OVERRIDE; 136 size_t size) OVERRID [all...] |
/art/runtime/base/ |
H A D | allocator.cc | 32 void* Alloc(size_t size) { argument 33 return calloc(sizeof(uint8_t), size); 51 void* Alloc(size_t size ATTRIBUTE_UNUSED) {
|
H A D | iteration_range.h | 63 template <typename T, size_t size> 64 inline auto ReverseRange(T (&array)[size]) { 65 return ReverseRange(MakeIterationRange<T*>(array, array+size));
|
H A D | arena_allocator.cc | 142 << num_allocations << ", avg size: " << bytes_allocated / num_allocations << "\n"; 166 void ArenaAllocatorMemoryTool::DoMakeDefined(void* ptr, size_t size) { argument 167 MEMORY_TOOL_MAKE_DEFINED(ptr, size); 170 void ArenaAllocatorMemoryTool::DoMakeUndefined(void* ptr, size_t size) { argument 171 MEMORY_TOOL_MAKE_UNDEFINED(ptr, size); 174 void ArenaAllocatorMemoryTool::DoMakeInaccessible(void* ptr, size_t size) { argument 175 MEMORY_TOOL_MAKE_NOACCESS(ptr, size); 183 explicit MallocArena(size_t size = Arena::kDefaultSize); 195 MallocArena::MallocArena(size_t size) { argument 199 unaligned_memory_ = reinterpret_cast<uint8_t*>(calloc(1, size 238 MemMapArena(size_t size, bool low_4gb, const char* name) argument 302 AllocArena(size_t size) argument [all...] |
/art/runtime/gc/space/ |
H A D | dlmalloc_space-inl.h | 47 size_t size = mspace_usable_size(obj_ptr); local 49 *usable_size = size; 51 return size + kChunkOverhead;
|
/art/tools/dexfuzz/src/dexfuzz/rawdex/ |
H A D | MapItem.java | 42 public int size; field in class:MapItem 49 size = file.readUInt(); 61 file.writeUInt(size);
|
/art/compiler/linker/x86_64/ |
H A D | relative_patcher_x86_64_test.cc | 66 LinkerPatch::RelativeCodePatch(kCallCode.size() - 4u, nullptr, 1u), 79 LinkerPatch::RelativeCodePatch(kCallCode.size() - 4u, nullptr, 2u), 83 LinkerPatch::RelativeCodePatch(kCallCode.size() - 4u, nullptr, 1u), 90 uint32_t diff_after = method2_offset - (method1_offset + kCallCode.size() /* PC adjustment */); 99 uint32_t diff_before = method1_offset - (method2_offset + kCallCode.size() /* PC adjustment */); 112 LinkerPatch::RelativeCodePatch(kCallCode.size() - 4u, nullptr, 2u), 119 uint32_t diff = kTrampolineOffset - (result.second + kCallCode.size()); 134 LinkerPatch::DexCacheArrayPatch(kDexCacheLoadCode.size() - 4u, nullptr, 0u, kElementOffset), 142 dex_cache_arrays_begin_ + kElementOffset - (result.second + kDexCacheLoadCode.size()); 159 kStringReferenceCode.size() [all...] |
/art/profman/ |
H A D | profile_assistant.cc | 46 for (size_t i = 0; i < profile_files.size(); i++) { 91 explicit ScopedCollectionFlock(size_t size) : flocks_(size) {} argument 95 for (size_t i = 0; i < filenames.size(); i++) { 106 for (size_t i = 0; i < fds.size(); i++) { 127 ScopedCollectionFlock profile_files_flocks(profile_files_fd.size()); 146 ScopedCollectionFlock profile_files_flocks(profile_files.size());
|
/art/compiler/ |
H A D | common_compiler_test.cc | 55 // If the code size is 0 it means the method was skipped due to profile guided compilation. 56 if (compiled_method != nullptr && compiled_method->GetQuickCode().size() != 0u) { 58 const uint32_t code_size = code.size(); 61 : sizeof(OatQuickMethodHeader) + vmap_table.size(); 65 : vmap_table_offset + method_info.size(); 77 const size_t size = method_info.size() + vmap_table.size() + sizeof(method_header) + code_size; local 78 chunk->reserve(size + max_padding); 84 CHECK_EQ(chunk->size(), siz [all...] |
/art/runtime/jit/ |
H A D | profile_compilation_info.cc | 57 "InlineCache::kIndividualCacheSize does not have the expect type size"); 102 if (classes.size() + 1 >= InlineCache::kIndividualCacheSize) { 122 DCHECK(last_sep_index < dex_location.size()); 204 int64_t size = GetFileSizeBytes(filename); 205 if (size != -1) { 208 << size; 210 *bytes_written = static_cast<uint64_t>(size); 246 2 * sizeof(uint16_t) + // class_set.size + dex_location.size 247 2 * sizeof(uint32_t); // method_map.size [all...] |
/art/runtime/ |
H A D | memory_region.h | 40 return lhs.size() == rhs.size() && memcmp(lhs.begin(), rhs.begin(), lhs.size()) == 0; 48 size_t size() const { return size_; } function in class:art::FINAL 177 CopyFrom(offset, MemoryRegion(vector.data(), vector.size())); 183 CHECK_GE(this->size(), size_in); 184 CHECK_LE(offset, this->size() - size_in); 191 size_ = (region.size() + extra); 197 CHECK_GE(size(), sizeof(T)); 198 CHECK_LE(offset, size() [all...] |
H A D | cha_test.cc | 47 ASSERT_EQ(dependents->size(), 1u); 55 ASSERT_EQ(dependents->size(), 2u); 67 ASSERT_EQ(dependents->size(), 1u); 74 ASSERT_EQ(dependents->size(), 1u); 76 ASSERT_EQ(dependents->size(), 1u); 83 ASSERT_EQ(dependents->size(), 1u);
|
H A D | dex_method_iterator.h | 36 CHECK_NE(0U, dex_files_.size()); 45 if (dex_file_index_ == dex_files_.size()) { 129 CHECK_LT(dex_file_index_, dex_files_.size());
|
/art/compiler/utils/arm64/ |
H A D | jni_macro_assembler_arm64.h | 66 void Store(FrameOffset offs, ManagedRegister src, size_t size) OVERRIDE; 80 void Load(ManagedRegister dest, FrameOffset src, size_t size) OVERRIDE; 81 void LoadFromThread(ManagedRegister dest, ThreadOffset64 src, size_t size) OVERRIDE; 91 void Move(ManagedRegister dest, ManagedRegister src, size_t size) OVERRIDE; 98 void Copy(FrameOffset dest, FrameOffset src, ManagedRegister scratch, size_t size) OVERRIDE; 103 size_t size) OVERRIDE; 108 size_t size) OVERRIDE; 113 size_t size) OVERRIDE; 119 size_t size) OVERRIDE; 125 size_t size) OVERRID [all...] |
H A D | jni_macro_assembler_arm64.cc | 127 void Arm64JNIMacroAssembler::Store(FrameOffset offs, ManagedRegister m_src, size_t size) { argument 130 CHECK_EQ(0u, size); 132 CHECK_EQ(4u, size); 135 CHECK_EQ(8u, size); 257 size_t size) { 259 CHECK_EQ(0u, size) << dest; 261 CHECK_EQ(4u, size) << dest; 266 if (size == 1u) { 268 } else if (size == 4u) { 271 CHECK_EQ(8u, size) << des 254 Load(Arm64ManagedRegister dest, XRegister base, int32_t offset, size_t size) argument 282 Load(ManagedRegister m_dst, FrameOffset src, size_t size) argument 286 LoadFromThread(ManagedRegister m_dst, ThreadOffset64 src, size_t size) argument 332 Move(ManagedRegister m_dst, ManagedRegister m_src, size_t size) argument 388 Copy(FrameOffset dest, FrameOffset src, ManagedRegister m_scratch, size_t size) argument 406 Copy(FrameOffset dest, ManagedRegister src_base, Offset src_offset, ManagedRegister m_scratch, size_t size) argument 428 Copy(ManagedRegister m_dest_base, Offset dest_offs, FrameOffset src, ManagedRegister m_scratch, size_t size) argument 458 Copy(ManagedRegister m_dest, Offset dest_offset, ManagedRegister m_src, Offset src_offset, ManagedRegister m_scratch, size_t size) argument 505 SignExtend(ManagedRegister mreg, size_t size) argument 516 ZeroExtend(ManagedRegister mreg, size_t size) argument [all...] |
/art/runtime/verifier/ |
H A D | reg_type_cache.cc | 68 DCHECK_EQ(entries_.size(), small_precise_constants_[i]->GetId()); 71 DCHECK_EQ(entries_.size(), primitive_count_); 178 for (size_t i = primitive_count_; i < entries_.size(); i++) { 200 entry = new (&arena_) PreciseReferenceType(klass, AddString(sp_descriptor), entries_.size()); 202 entry = new (&arena_) ReferenceType(klass, AddString(sp_descriptor), entries_.size()); 216 new (&arena_) UnresolvedReferenceType(AddString(sp_descriptor), entries_.size())); 251 new (&arena_) PreciseReferenceType(klass, descriptor, entries_.size())) 252 : new (&arena_) ReferenceType(klass, descriptor, entries_.size()); 283 DCHECK_LE(primitive_count_, entries_.size()); 411 for (size_t i = primitive_count_; i < entries_.size(); [all...] |
/art/compiler/debug/dwarf/ |
H A D | debug_abbrev_writer.h | 54 has_children_offset_ = current_abbrev_.size(); 75 this->PushData(abbrev.data(), abbrev.size()); 86 return dchecked_integral_cast<uint32_t>(1 + abbrev_codes_.size());
|
/art/compiler/debug/ |
H A D | elf_debug_frame_writer.h | 181 // compressed it reduces the size since similar opcodes sequences are grouped. 183 sorted_method_infos.reserve(method_infos.size()); 184 for (size_t i = 0; i < method_infos.size(); i++) { 204 binary_search_table.reserve(2 * sorted_method_infos.size()); 206 patch_locations.reserve(sorted_method_infos.size()); 221 cfi_section->WriteFully(buffer.data(), buffer.size()); 222 buffer_address += buffer.size(); 237 cfi_section->WriteFully(buffer.data(), buffer.size()); 238 buffer_address += buffer.size(); 255 // Encoding of binary search table size [all...] |
/art/compiler/linker/ |
H A D | relative_patcher_test.h | 119 offset += code.size(); 183 out_.WriteFully(&code[0], code.size()); 184 offset += code.size(); 188 CHECK_EQ(output_.size(), output_size); 192 // Sanity check: original code size must match linked_code.size(). 201 CHECK_NE(idx, compiled_method_refs_.size()); 202 CHECK_EQ(compiled_methods_[idx]->GetQuickCode().size(), expected_code.size()); 207 CHECK_LT(offset, output_.size()); [all...] |
/art/compiler/utils/arm/ |
H A D | assembler_thumb2.h | 255 void vpaddld(DRegister dd, DRegister dm, int32_t size, bool is_unsigned) OVERRIDE; 313 Literal* NewLiteral(size_t size, const uint8_t* data) OVERRIDE; 398 // size before hand, so we assume it is the smallest available size and determine the final 402 // The actual instruction encoding for these is delayed until we know the final size of 432 // Calculated size of branch instruction based on type and offset. 478 static Fixup Branch(uint32_t location, Type type, Size size = kBranch16Bit, 482 DCHECK(size == kBranch16Bit || size == kBranch32Bit); 483 DCHECK(size 496 LoadNarrowLiteral(uint32_t location, Register rt, Size size) argument 532 LoadLiteralAddress(uint32_t location, Register rt, Size size) argument 629 Fixup(Register rn, Register rt2, SRegister sd, DRegister dd, Condition cond, Type type, Size size, uint32_t location) argument [all...] |