Searched refs:size (Results 51 - 75 of 479) sorted by relevance

1234567891011>>

/art/compiler/utils/
H A Djni_macro_assembler.h75 virtual void Store(FrameOffset offs, ManagedRegister src, size_t size) = 0;
93 virtual void Load(ManagedRegister dest, FrameOffset src, size_t size) = 0;
97 size_t size) = 0;
111 virtual void Move(ManagedRegister dest, ManagedRegister src, size_t size) = 0;
123 virtual void Copy(FrameOffset dest, FrameOffset src, ManagedRegister scratch, size_t size) = 0;
129 size_t size) = 0;
135 size_t size) = 0;
141 size_t size) = 0;
148 size_t size) = 0;
155 size_t size)
[all...]
H A Dswap_space.h38 void* Alloc(size_t size) REQUIRES(!lock_);
39 void Free(void* ptr, size_t size) REQUIRES(!lock_);
51 mutable size_t size; member in struct:art::SwapSpace::SpaceChunk
57 return reinterpret_cast<uintptr_t>(ptr) + size;
70 // Map size to an iterator to free_by_start_'s entry.
73 : size(sz), free_by_start_entry(entry) { }
77 mutable size_t size; member in struct:art::SwapSpace::FreeBySizeEntry
82 if (lhs.size != rhs.size) {
83 return lhs.size < rh
[all...]
H A Dtest_dex_file_builder.h109 header->string_ids_size_ = strings_.size();
112 uint32_t type_ids_offset = string_ids_offset + strings_.size() * sizeof(DexFile::StringId);
118 header->type_ids_size_ = types_.size();
121 uint32_t proto_ids_offset = type_ids_offset + types_.size() * sizeof(DexFile::TypeId);
126 size_t num_args = entry.first.args.size();
134 header->proto_ids_size_ = protos_.size();
137 uint32_t field_ids_offset = proto_ids_offset + protos_.size() * sizeof(DexFile::ProtoId);
143 header->field_ids_size_ = fields_.size();
146 uint32_t method_ids_offset = field_ids_offset + fields_.size() * sizeof(DexFile::FieldId);
152 header->method_ids_size_ = methods_.size();
[all...]
/art/compiler/utils/x86/
H A Djni_macro_assembler_x86.h58 void Store(FrameOffset offs, ManagedRegister src, size_t size) OVERRIDE;
74 void Load(ManagedRegister dest, FrameOffset src, size_t size) OVERRIDE;
76 void LoadFromThread(ManagedRegister dest, ThreadOffset32 src, size_t size) OVERRIDE;
88 void Move(ManagedRegister dest, ManagedRegister src, size_t size) OVERRIDE;
99 void Copy(FrameOffset dest, FrameOffset src, ManagedRegister scratch, size_t size) OVERRIDE;
102 size_t size) OVERRIDE;
105 size_t size) OVERRIDE;
108 size_t size) OVERRIDE;
111 ManagedRegister scratch, size_t size) OVERRIDE;
114 ManagedRegister scratch, size_t size) OVERRID
[all...]
H A Djni_macro_assembler_x86.cc52 for (int i = spill_regs.size() - 1; i >= 0; --i) {
70 for (size_t i = 0; i < entry_spills.size(); ++i) {
92 int adjust = frame_size - spill_regs.size() * kFramePointerSize - kFramePointerSize;
95 for (size_t i = 0; i < spill_regs.size(); ++i) {
123 void X86JNIMacroAssembler::Store(FrameOffset offs, ManagedRegister msrc, size_t size) { argument
126 CHECK_EQ(0u, size);
128 CHECK_EQ(4u, size);
131 CHECK_EQ(8u, size);
135 if (size == 4) {
142 if (size
186 Load(ManagedRegister mdest, FrameOffset src, size_t size) argument
213 LoadFromThread(ManagedRegister mdest, ThreadOffset32 src, size_t size) argument
274 SignExtend(ManagedRegister mreg, size_t size) argument
285 ZeroExtend(ManagedRegister mreg, size_t size) argument
296 Move(ManagedRegister mdest, ManagedRegister msrc, size_t size) argument
347 Copy(FrameOffset dest, FrameOffset src, ManagedRegister mscratch, size_t size) argument
370 Copy(ManagedRegister dest_base, Offset dest_offset, FrameOffset src, ManagedRegister scratch, size_t size) argument
381 Copy(FrameOffset dest, FrameOffset src_base, Offset src_offset, ManagedRegister mscratch, size_t size) argument
393 Copy(ManagedRegister dest, Offset dest_offset, ManagedRegister src, Offset src_offset, ManagedRegister scratch, size_t size) argument
405 Copy(FrameOffset dest, Offset dest_offset, FrameOffset src, Offset src_offset, ManagedRegister mscratch, size_t size) argument
[all...]
/art/compiler/utils/x86_64/
H A Djni_macro_assembler_x86_64.h59 void Store(FrameOffset offs, ManagedRegister src, size_t size) OVERRIDE;
77 void Load(ManagedRegister dest, FrameOffset src, size_t size) OVERRIDE;
79 void LoadFromThread(ManagedRegister dest, ThreadOffset64 src, size_t size) OVERRIDE;
93 void Move(ManagedRegister dest, ManagedRegister src, size_t size);
104 void Copy(FrameOffset dest, FrameOffset src, ManagedRegister scratch, size_t size) OVERRIDE;
110 size_t size) OVERRIDE;
116 size_t size) OVERRIDE;
122 size_t size) OVERRIDE;
129 size_t size) OVERRIDE;
136 size_t size) OVERRID
[all...]
/art/runtime/base/
H A Dallocator.cc32 void* Alloc(size_t size) { argument
33 return calloc(sizeof(uint8_t), size);
51 void* Alloc(size_t size ATTRIBUTE_UNUSED) {
H A Diteration_range.h63 template <typename T, size_t size>
64 inline auto ReverseRange(T (&array)[size]) {
65 return ReverseRange(MakeIterationRange<T*>(array, array+size));
H A Darena_allocator.cc142 << num_allocations << ", avg size: " << bytes_allocated / num_allocations << "\n";
166 void ArenaAllocatorMemoryTool::DoMakeDefined(void* ptr, size_t size) { argument
167 MEMORY_TOOL_MAKE_DEFINED(ptr, size);
170 void ArenaAllocatorMemoryTool::DoMakeUndefined(void* ptr, size_t size) { argument
171 MEMORY_TOOL_MAKE_UNDEFINED(ptr, size);
174 void ArenaAllocatorMemoryTool::DoMakeInaccessible(void* ptr, size_t size) { argument
175 MEMORY_TOOL_MAKE_NOACCESS(ptr, size);
183 explicit MallocArena(size_t size = Arena::kDefaultSize);
195 MallocArena::MallocArena(size_t size) { argument
199 unaligned_memory_ = reinterpret_cast<uint8_t*>(calloc(1, size
238 MemMapArena(size_t size, bool low_4gb, const char* name) argument
302 AllocArena(size_t size) argument
[all...]
/art/runtime/gc/space/
H A Ddlmalloc_space-inl.h47 size_t size = mspace_usable_size(obj_ptr); local
49 *usable_size = size;
51 return size + kChunkOverhead;
/art/tools/dexfuzz/src/dexfuzz/rawdex/
H A DMapItem.java42 public int size; field in class:MapItem
49 size = file.readUInt();
61 file.writeUInt(size);
/art/compiler/linker/x86_64/
H A Drelative_patcher_x86_64_test.cc66 LinkerPatch::RelativeCodePatch(kCallCode.size() - 4u, nullptr, 1u),
79 LinkerPatch::RelativeCodePatch(kCallCode.size() - 4u, nullptr, 2u),
83 LinkerPatch::RelativeCodePatch(kCallCode.size() - 4u, nullptr, 1u),
90 uint32_t diff_after = method2_offset - (method1_offset + kCallCode.size() /* PC adjustment */);
99 uint32_t diff_before = method1_offset - (method2_offset + kCallCode.size() /* PC adjustment */);
112 LinkerPatch::RelativeCodePatch(kCallCode.size() - 4u, nullptr, 2u),
119 uint32_t diff = kTrampolineOffset - (result.second + kCallCode.size());
134 LinkerPatch::DexCacheArrayPatch(kDexCacheLoadCode.size() - 4u, nullptr, 0u, kElementOffset),
142 dex_cache_arrays_begin_ + kElementOffset - (result.second + kDexCacheLoadCode.size());
159 kStringReferenceCode.size()
[all...]
/art/profman/
H A Dprofile_assistant.cc46 for (size_t i = 0; i < profile_files.size(); i++) {
91 explicit ScopedCollectionFlock(size_t size) : flocks_(size) {} argument
95 for (size_t i = 0; i < filenames.size(); i++) {
106 for (size_t i = 0; i < fds.size(); i++) {
127 ScopedCollectionFlock profile_files_flocks(profile_files_fd.size());
146 ScopedCollectionFlock profile_files_flocks(profile_files.size());
/art/compiler/
H A Dcommon_compiler_test.cc55 // If the code size is 0 it means the method was skipped due to profile guided compilation.
56 if (compiled_method != nullptr && compiled_method->GetQuickCode().size() != 0u) {
58 const uint32_t code_size = code.size();
61 : sizeof(OatQuickMethodHeader) + vmap_table.size();
65 : vmap_table_offset + method_info.size();
77 const size_t size = method_info.size() + vmap_table.size() + sizeof(method_header) + code_size; local
78 chunk->reserve(size + max_padding);
84 CHECK_EQ(chunk->size(), siz
[all...]
/art/runtime/jit/
H A Dprofile_compilation_info.cc57 "InlineCache::kIndividualCacheSize does not have the expect type size");
102 if (classes.size() + 1 >= InlineCache::kIndividualCacheSize) {
122 DCHECK(last_sep_index < dex_location.size());
204 int64_t size = GetFileSizeBytes(filename);
205 if (size != -1) {
208 << size;
210 *bytes_written = static_cast<uint64_t>(size);
246 2 * sizeof(uint16_t) + // class_set.size + dex_location.size
247 2 * sizeof(uint32_t); // method_map.size
[all...]
/art/runtime/
H A Dmemory_region.h40 return lhs.size() == rhs.size() && memcmp(lhs.begin(), rhs.begin(), lhs.size()) == 0;
48 size_t size() const { return size_; } function in class:art::FINAL
177 CopyFrom(offset, MemoryRegion(vector.data(), vector.size()));
183 CHECK_GE(this->size(), size_in);
184 CHECK_LE(offset, this->size() - size_in);
191 size_ = (region.size() + extra);
197 CHECK_GE(size(), sizeof(T));
198 CHECK_LE(offset, size()
[all...]
H A Dcha_test.cc47 ASSERT_EQ(dependents->size(), 1u);
55 ASSERT_EQ(dependents->size(), 2u);
67 ASSERT_EQ(dependents->size(), 1u);
74 ASSERT_EQ(dependents->size(), 1u);
76 ASSERT_EQ(dependents->size(), 1u);
83 ASSERT_EQ(dependents->size(), 1u);
H A Ddex_method_iterator.h36 CHECK_NE(0U, dex_files_.size());
45 if (dex_file_index_ == dex_files_.size()) {
129 CHECK_LT(dex_file_index_, dex_files_.size());
/art/compiler/utils/arm64/
H A Djni_macro_assembler_arm64.h66 void Store(FrameOffset offs, ManagedRegister src, size_t size) OVERRIDE;
80 void Load(ManagedRegister dest, FrameOffset src, size_t size) OVERRIDE;
81 void LoadFromThread(ManagedRegister dest, ThreadOffset64 src, size_t size) OVERRIDE;
91 void Move(ManagedRegister dest, ManagedRegister src, size_t size) OVERRIDE;
98 void Copy(FrameOffset dest, FrameOffset src, ManagedRegister scratch, size_t size) OVERRIDE;
103 size_t size) OVERRIDE;
108 size_t size) OVERRIDE;
113 size_t size) OVERRIDE;
119 size_t size) OVERRIDE;
125 size_t size) OVERRID
[all...]
H A Djni_macro_assembler_arm64.cc127 void Arm64JNIMacroAssembler::Store(FrameOffset offs, ManagedRegister m_src, size_t size) { argument
130 CHECK_EQ(0u, size);
132 CHECK_EQ(4u, size);
135 CHECK_EQ(8u, size);
257 size_t size) {
259 CHECK_EQ(0u, size) << dest;
261 CHECK_EQ(4u, size) << dest;
266 if (size == 1u) {
268 } else if (size == 4u) {
271 CHECK_EQ(8u, size) << des
254 Load(Arm64ManagedRegister dest, XRegister base, int32_t offset, size_t size) argument
282 Load(ManagedRegister m_dst, FrameOffset src, size_t size) argument
286 LoadFromThread(ManagedRegister m_dst, ThreadOffset64 src, size_t size) argument
332 Move(ManagedRegister m_dst, ManagedRegister m_src, size_t size) argument
388 Copy(FrameOffset dest, FrameOffset src, ManagedRegister m_scratch, size_t size) argument
406 Copy(FrameOffset dest, ManagedRegister src_base, Offset src_offset, ManagedRegister m_scratch, size_t size) argument
428 Copy(ManagedRegister m_dest_base, Offset dest_offs, FrameOffset src, ManagedRegister m_scratch, size_t size) argument
458 Copy(ManagedRegister m_dest, Offset dest_offset, ManagedRegister m_src, Offset src_offset, ManagedRegister m_scratch, size_t size) argument
505 SignExtend(ManagedRegister mreg, size_t size) argument
516 ZeroExtend(ManagedRegister mreg, size_t size) argument
[all...]
/art/runtime/verifier/
H A Dreg_type_cache.cc68 DCHECK_EQ(entries_.size(), small_precise_constants_[i]->GetId());
71 DCHECK_EQ(entries_.size(), primitive_count_);
178 for (size_t i = primitive_count_; i < entries_.size(); i++) {
200 entry = new (&arena_) PreciseReferenceType(klass, AddString(sp_descriptor), entries_.size());
202 entry = new (&arena_) ReferenceType(klass, AddString(sp_descriptor), entries_.size());
216 new (&arena_) UnresolvedReferenceType(AddString(sp_descriptor), entries_.size()));
251 new (&arena_) PreciseReferenceType(klass, descriptor, entries_.size()))
252 : new (&arena_) ReferenceType(klass, descriptor, entries_.size());
283 DCHECK_LE(primitive_count_, entries_.size());
411 for (size_t i = primitive_count_; i < entries_.size();
[all...]
/art/compiler/debug/dwarf/
H A Ddebug_abbrev_writer.h54 has_children_offset_ = current_abbrev_.size();
75 this->PushData(abbrev.data(), abbrev.size());
86 return dchecked_integral_cast<uint32_t>(1 + abbrev_codes_.size());
/art/compiler/debug/
H A Delf_debug_frame_writer.h181 // compressed it reduces the size since similar opcodes sequences are grouped.
183 sorted_method_infos.reserve(method_infos.size());
184 for (size_t i = 0; i < method_infos.size(); i++) {
204 binary_search_table.reserve(2 * sorted_method_infos.size());
206 patch_locations.reserve(sorted_method_infos.size());
221 cfi_section->WriteFully(buffer.data(), buffer.size());
222 buffer_address += buffer.size();
237 cfi_section->WriteFully(buffer.data(), buffer.size());
238 buffer_address += buffer.size();
255 // Encoding of binary search table size
[all...]
/art/compiler/linker/
H A Drelative_patcher_test.h119 offset += code.size();
183 out_.WriteFully(&code[0], code.size());
184 offset += code.size();
188 CHECK_EQ(output_.size(), output_size);
192 // Sanity check: original code size must match linked_code.size().
201 CHECK_NE(idx, compiled_method_refs_.size());
202 CHECK_EQ(compiled_methods_[idx]->GetQuickCode().size(), expected_code.size());
207 CHECK_LT(offset, output_.size());
[all...]
/art/compiler/utils/arm/
H A Dassembler_thumb2.h255 void vpaddld(DRegister dd, DRegister dm, int32_t size, bool is_unsigned) OVERRIDE;
313 Literal* NewLiteral(size_t size, const uint8_t* data) OVERRIDE;
398 // size before hand, so we assume it is the smallest available size and determine the final
402 // The actual instruction encoding for these is delayed until we know the final size of
432 // Calculated size of branch instruction based on type and offset.
478 static Fixup Branch(uint32_t location, Type type, Size size = kBranch16Bit,
482 DCHECK(size == kBranch16Bit || size == kBranch32Bit);
483 DCHECK(size
496 LoadNarrowLiteral(uint32_t location, Register rt, Size size) argument
532 LoadLiteralAddress(uint32_t location, Register rt, Size size) argument
629 Fixup(Register rn, Register rt2, SRegister sd, DRegister dd, Condition cond, Type type, Size size, uint32_t location) argument
[all...]

Completed in 318 milliseconds

1234567891011>>