/art/compiler/debug/dwarf/ |
H A D | headers.h | 49 size_t cie_header_start_ = writer.data()->size(); 57 writer.PushUleb128(1); // z: Augmentation data size. 75 writer.UpdateUint32(cie_header_start_, writer.data()->size() - cie_header_start_ - 4); 94 size_t fde_header_start = writer.data()->size(); 97 uint32_t cie_pointer = (buffer_address + buffer->size()) - cie_address; 106 code_address -= buffer_address + buffer->size(); 110 patch_locations->push_back(buffer_address + buffer->size() - section_address); 119 writer.PushUleb128(0); // Augmentation data size. 120 writer.PushData(opcodes.data(), opcodes.size()); 122 writer.UpdateUint32(fde_header_start, writer.data()->size() [all...] |
/art/compiler/linker/ |
H A D | vector_output_stream.cc | 24 : OutputStream(location), offset_(vector->size()), vector_(vector) {} 39 new_offset = vector_->size() + offset;
|
/art/runtime/ |
H A D | dex_method_iterator_test.cc | 33 CHECK_NE(boot_class_path_.size(), 0U); 34 for (size_t i = 0; i < boot_class_path_.size(); ++i) {
|
H A D | monitor_pool_test.cc | 55 // Note: for correct testing, make sure this is corresponding to monitor-pool's initial size. 64 if (monitors.size() == 0) { 66 } else if (monitors.size() == kMaxUsage) { 80 size_t index = r.next() % monitors.size(); 95 size_t target_size = monitors.size() + 2*kMaxUsage; 96 while (monitors.size() < target_size) { 98 static_cast<int32_t>(-monitors.size())); 112 size_t index = r.next() % monitors.size();
|
H A D | linear_alloc.cc | 31 void* LinearAlloc::Alloc(Thread* self, size_t size) { argument 33 return allocator_.Alloc(size);
|
/art/runtime/base/unix_file/ |
H A D | random_access_file_test.h | 76 ASSERT_EQ(content.size(), static_cast<uint64_t>(file->Write(content.data(), content.size(), 0))); 84 const int buf_size = content.size() + 10; 90 ASSERT_EQ(content.size(), static_cast<uint64_t>(file->Read(buf.get(), buf_size, 0))); 91 ASSERT_EQ(std::string(buf.get(), content.size()), content); 95 ASSERT_LT(short_request, content.size()); 116 ASSERT_EQ(content.size(), static_cast<uint64_t>(file->Write(content.data(), content.size(), 0))); 117 ASSERT_EQ(content.size(), static_cast<uint64_t>(file->GetLength())); 152 ASSERT_EQ(content.size(), static_cas [all...] |
H A D | random_access_file_utils.cc | 29 while ((n = src.Read(&buf[0], buf.size(), offset)) > 0) {
|
/art/compiler/optimizing/ |
H A D | stack_map_stream.cc | 34 current_entry_.dex_register_locations_start_index = dex_register_locations_.size(); 35 current_entry_.inline_infos_start_index = inline_infos_.size(); 79 size_t index = location_catalog_entries_.size(); 111 current_inline_info_.dex_register_locations_start_index = dex_register_locations_.size(); 142 inline_info_size_ = inline_infos_.size() * inline_info_encoding_.GetEntrySize(); 150 stack_maps_size_ = stack_maps_.size() * stack_map_size; 162 code_info_encoding.number_of_stack_maps = stack_maps_.size(); 166 code_info_encoding.number_of_location_catalog_entries = location_catalog_entries_.size(); 170 // to compute its size (note that we do not encode that size i 181 size_t size = DexRegisterLocationCatalog::kFixedSize; local 197 size_t size = DexRegisterMap::kFixedSize; local 212 size_t size = 0; local [all...] |
H A D | side_effects_analysis.h | 31 block_effects_(graph->GetBlocks().size(), 33 loop_effects_(graph->GetBlocks().size(),
|
/art/benchmark/scoped-primitive-array/ |
H A D | scoped_primitive_array.cc | 25 ret += sc[0] + sc[sc.size() - 1]; 35 ret += sc[0] + sc[sc.size() - 1]; 45 ret += sc[0] + sc[sc.size() - 1]; 55 ret += sc[0] + sc[sc.size() - 1];
|
/art/test/004-NativeAllocations/src/ |
H A D | Main.java | 65 int size = (int)(maxMem / 2 / count); 69 allocations[i % count] = new NativeAllocation(size, false); 76 allocations[i % count] = new NativeAllocation(size, true); 81 allocations[i % count] = new NativeAllocation(size, true);
|
/art/runtime/arch/x86/ |
H A D | asm_support_x86.S | 68 #define SIZE(name) .size name, .-name 75 #define CFI_ADJUST_CFA_OFFSET(size) .cfi_adjust_cfa_offset size 76 #define CFI_DEF_CFA(reg,size) .cfi_def_cfa reg,size 79 #define CFI_REL_OFFSET(reg,size) .cfi_rel_offset reg,size 86 #define CFI_ADJUST_CFA_OFFSET(size) 87 #define CFI_DEF_CFA(reg,size) 90 #define CFI_REL_OFFSET(reg,size) [all...] |
/art/runtime/arch/x86_64/ |
H A D | asm_support_x86_64.S | 65 #define SIZE(name) .size name, .-name 72 #define CFI_ADJUST_CFA_OFFSET(size) .cfi_adjust_cfa_offset size 73 #define CFI_DEF_CFA(reg,size) .cfi_def_cfa reg,size 76 #define CFI_REL_OFFSET(reg,size) .cfi_rel_offset reg,size 81 #define CFI_ADJUST_CFA_OFFSET(size) 82 #define CFI_DEF_CFA(reg,size) 85 #define CFI_REL_OFFSET(reg,size) [all...] |
/art/test/080-oom-throw/src/ |
H A D | Main.java | 90 int size = 32 * 1024 * 1024; 93 holder[i] = new char[size]; 96 size = size / 2; 97 if (size == 0) {
|
/art/runtime/base/ |
H A D | stringpiece.h | 54 : ptr_(str.data()), length_(str.size()) { } 62 size_type size() const { return length_; } function in class:art::StringPiece 107 return std::string(data(), size()); 115 return std::string(data(), size()); 152 const int r = memcmp(data(), rhs.data(), std::min(size(), rhs.size())); 156 if (size() < rhs.size()) { 158 } else if (size() > rhs.size()) { [all...] |
H A D | bounded_fifo.h | 25 // A bounded fifo is a fifo which has a bounded size. The power of two version uses a bit mask to 42 return size() == 0; 45 size_t size() const { function in class:art::BoundedFifoPowerOfTwo
|
H A D | allocator.cc | 33 void* Alloc(size_t size) { argument 34 return calloc(sizeof(uint8_t), size); 52 void* Alloc(size_t size ATTRIBUTE_UNUSED) {
|
H A D | arena_allocator.cc | 136 << num_allocations << ", avg size: " << bytes_allocated / num_allocations << "\n"; 148 void ArenaAllocatorMemoryTool::DoMakeDefined(void* ptr, size_t size) { argument 149 MEMORY_TOOL_MAKE_DEFINED(ptr, size); 152 void ArenaAllocatorMemoryTool::DoMakeUndefined(void* ptr, size_t size) { argument 153 MEMORY_TOOL_MAKE_UNDEFINED(ptr, size); 156 void ArenaAllocatorMemoryTool::DoMakeInaccessible(void* ptr, size_t size) { argument 157 MEMORY_TOOL_MAKE_NOACCESS(ptr, size); 163 MallocArena::MallocArena(size_t size) { argument 164 memory_ = reinterpret_cast<uint8_t*>(calloc(1, size)); 166 size_ = size; 173 MemMapArena(size_t size, bool low_4gb, const char* name) argument 231 AllocArena(size_t size) argument [all...] |
/art/runtime/gc/allocator/ |
H A D | rosalloc.h | 310 // Represents a run of memory slots of the same size. 353 uint8_t size_bracket_idx_; // The index of the size bracket of this run. 366 // Returns the byte size of the header. 424 // Add the given slot to the bulk free list. Returns the bracket size. 461 // size. 495 // The number of size brackets. 497 // The sizes (the slot sizes, in bytes) of the size brackets. 499 // The numbers of pages that are used for runs for each size bracket. 501 // The numbers of slots of the runs for each size bracket. 503 // The header sizes in bytes of the runs for each size bracke 516 BracketSizeToIndex(size_t size) argument 539 IsSizeForThreadLocal(size_t size) argument 546 RoundToBracketSize(size_t size) argument 560 SizeToIndex(size_t size) argument 575 SizeToIndexAndBracketSize(size_t size, size_t* bracket_size_out) argument [all...] |
/art/compiler/utils/ |
H A D | array_ref.h | 64 template <size_t size> 65 explicit constexpr ArrayRef(T (&array)[size]) 66 : array_(array), size_(size) { 69 template <typename U, size_t size> 70 explicit constexpr ArrayRef(U (&array)[size], 73 : array_(array), size_(size) { 84 : array_(v.data()), size_(v.size()) { 93 : array_(v.data()), size_(v.size()) { 130 size_type size() const { return size_; } function in class:art::ArrayRef 131 bool empty() const { return size() [all...] |
/art/compiler/dex/ |
H A D | verification_results.cc | 64 DCHECK_EQ(it->second->GetDevirtMap().size(), verified_method->GetDevirtMap().size()); 65 DCHECK_EQ(it->second->GetSafeCastSet().size(), verified_method->GetSafeCastSet().size());
|
/art/compiler/linker/x86_64/ |
H A D | relative_patcher_x86_64.cc | 28 DCHECK_LE(patch.LiteralOffset() + 4u, code->size());
|
H A D | relative_patcher_x86_64_test.cc | 66 LinkerPatch::RelativeCodePatch(kCallCode.size() - 4u, nullptr, 1u), 79 LinkerPatch::RelativeCodePatch(kCallCode.size() - 4u, nullptr, 2u), 83 LinkerPatch::RelativeCodePatch(kCallCode.size() - 4u, nullptr, 1u), 90 uint32_t diff_after = method2_offset - (method1_offset + kCallCode.size() /* PC adjustment */); 99 uint32_t diff_before = method1_offset - (method2_offset + kCallCode.size() /* PC adjustment */); 112 LinkerPatch::RelativeCodePatch(kCallCode.size() - 4u, nullptr, 2u), 119 uint32_t diff = kTrampolineOffset - (result.second + kCallCode.size()); 134 LinkerPatch::DexCacheArrayPatch(kDexCacheLoadCode.size() - 4u, nullptr, 0u, kElementOffset), 142 dex_cache_arrays_begin_ + kElementOffset - (result.second + kDexCacheLoadCode.size()); 159 kStringReferenceCode.size() [all...] |
/art/runtime/gc/space/ |
H A D | dlmalloc_space-inl.h | 47 size_t size = mspace_usable_size(obj_ptr); local 49 *usable_size = size; 51 return size + kChunkOverhead;
|
/art/tools/dexfuzz/src/dexfuzz/rawdex/ |
H A D | MapItem.java | 42 public int size; field in class:MapItem 49 size = file.readUInt(); 61 file.writeUInt(size);
|