/art/compiler/debug/ |
H A D | elf_debug_writer.cc | 67 compilation_units.reserve(class_to_compilation_unit.size()); 141 CHECK_GT(method_infos.size(), 0u); 159 if (method_infos.size() > 1) {
|
/art/compiler/jit/ |
H A D | jit_logger.cc | 66 bool res = perf_file_->WriteFully(str.c_str(), str.size()); 115 uint32_t size_; // Total size of header 126 // Each record starts with such basic information: event type, total size, and timestamp. 154 uint32_t size_; // Total size of this event record. 155 // For example, for kLoad event, size of the event record is: 156 // sizeof(PerfJitCodeLoad) + method_name.size() + compiled code size. 277 jit_code.size_ = sizeof(jit_code) + method_name.size() + 1 + code_size; 293 UNUSED(jit_dump_file_->WriteFully(method_name.c_str(), method_name.size() + 1));
|
/art/compiler/optimizing/ |
H A D | superblock_cloner.h | 73 if (from >= graph->GetBlocks().size() || to >= graph->GetBlocks().size()) { 277 DCHECK(block_id < graph_->GetBlocks().size());
|
H A D | gvn.cc | 169 // Hash table remains the same size. We copy the bucket pointers and leave 174 // Hash table size changes. We copy and rehash all entries, and set all 349 sets_(graph->GetBlocks().size(), nullptr, allocator_.Adapter(kArenaAllocGvn)), 351 &allocator_, graph->GetBlocks().size(), /* expandable */ false, kArenaAllocGvn) { 415 if (predecessors.size() == 0 || predecessors[0]->IsEntryBlock()) { 424 if (dominator->GetSuccessors().size() == 1) { 433 // into the recycled set, so we pass `dominator_set` as a reference for size. 462 } else if (predecessors.size() > 1) {
|
H A D | superblock_cloner.cc | 182 DCHECK(!first_phi_met || copy_succ->GetPredecessors().size() == phi_input_count); 226 ArenaBitVector visited(arena_, graph_->GetBlocks().size(), false, kArenaAllocSuperblockCloner); 231 ArenaBitVector visiting(arena_, graph_->GetBlocks().size(), false, kArenaAllocGraphBuilder); 233 ArenaVector<size_t> successors_visited(graph_->GetBlocks().size(), 248 if (successors_visited[current_id] == current->GetSuccessors().size()) { 357 arena_, graph_->GetBlocks().size(), false, kArenaAllocSuperblockCloner); 477 arena_, graph_->GetBlocks().size(), false, kArenaAllocSuperblockCloner);
|
H A D | nodes.cc | 50 block->SetBlockId(blocks_.size()); 62 &allocator, blocks_.size(), /* expandable */ false, kArenaAllocGraphBuilder); 65 ScopedArenaVector<size_t> successors_visited(blocks_.size(), 79 if (successors_visited[current_id] == current->GetSuccessors().size()) { 145 for (size_t i = 0; i < blocks_.size(); ++i) { 158 for (size_t i = 0; i < blocks_.size(); ++i) { 183 ArenaBitVector visited(&allocator, blocks_.size(), false, kArenaAllocGraphBuilder); 266 reverse_post_order_.reserve(blocks_.size()); 272 ScopedArenaVector<size_t> visits(blocks_.size(), 0u, allocator.Adapter(kArenaAllocGraphBuilder)); 274 ScopedArenaVector<size_t> successors_visited(blocks_.size(), 2006 size_t size = 0; local [all...] |
H A D | codegen_test_utils.h | 191 virtual uint8_t* Allocate(size_t size) { argument 192 size_ = size; 193 memory_.reset(new uint8_t[size]);
|
H A D | register_allocator_linear_scan.cc | 98 int_spill_slots_.size(), 99 long_spill_slots_.size(), 100 float_spill_slots_.size(), 101 double_spill_slots_.size(), 313 for (size_t safepoint_index = safepoints_.size(); safepoint_index > 0; --safepoint_index) { 318 DCHECK(safepoint_index == safepoints_.size() || 324 DCHECK_EQ(safepoint_index, safepoints_.size()); 519 // Remember the inactive_ size here since the ones moved to inactive_ from 521 size_t inactive_intervals_to_handle = inactive_.size(); 639 for (size_t i = 0; i < inputs.size(); [all...] |
/art/libdexfile/dex/ |
H A D | descriptors_names.cc | 86 CHECK_EQ(class_name[class_name.size() - 1], ';') << class_name; 88 class_name.erase(class_name.size() - 1, 1); 408 const char* end = p + s.size();
|
H A D | dex_file_tracking_registrar.cc | 63 // Poisons the size and offset information along with the first instruction. 128 inline void SetRegistrationRange(const void* begin, size_t size, bool should_poison) { argument 130 MEMORY_TOOL_MAKE_NOACCESS(begin, size); 134 MEMORY_TOOL_MAKE_DEFINED(begin, size);
|
H A D | utf_test.cc | 118 ASSERT_EQ(expected.size(), CountUtf8Bytes(&input[0], input.size())); 120 std::vector<uint8_t> output(expected.size()); 121 ConvertUtf16ToModifiedUtf8(reinterpret_cast<char*>(&output[0]), expected.size(), 122 &input[0], input.size());
|
/art/test/1940-ddms-ext/ |
H A D | ddm_ext.cc | 114 static void JNICALL PublishCB(jvmtiEnv* jvmti, JNIEnv* jnienv, jint type, jint size, jbyte* bytes) { argument 120 ScopedLocalRef<jbyteArray> res(jnienv, jnienv->NewByteArray(size)); 121 jnienv->SetByteArrayRegion(res.get(), 0, size, bytes);
|
/art/runtime/jdwp/ |
H A D | jdwp_handler.cc | 72 expandBufAdd4BE(reply, objects.size()); 73 for (size_t i = 0; i < objects.size(); ++i) { 160 expandBufAdd4BE(pReply, ids.size()); 162 for (size_t i = 0; i < ids.size(); ++i) { 190 expandBufAdd4BE(pReply, thread_ids.size()); 191 for (uint32_t i = 0; i < thread_ids.size(); ++i) { 290 expandBufAdd4BE(pReply, class_path.size()); 297 expandBufAdd4BE(pReply, boot_class_path.size()); 360 expandBufAdd4BE(pReply, classes.size()); 362 for (size_t i = 0; i < classes.size(); 1313 uint32_t size = request->ReadUnsigned32("step size"); local [all...] |
/art/runtime/ |
H A D | reference_table_test.cc | 268 } while (start < haystack.size()); 320 EXPECT_EQ(1u, strCounts.size()); 321 EXPECT_EQ(1u, b1Counts.size()); 322 EXPECT_EQ(1u, b2Counts.size()); 323 EXPECT_EQ(1u, cCounts.size());
|
H A D | elf_file.cc | 116 *error_msg = StringPrintf("File size of %zd bytes not large enough to contain ELF header of " 123 // first just map ELF header to get program header size information 140 *error_msg = StringPrintf("File size of %zd bytes not large enough to contain ELF program " 407 // It might not be mapped, but we can compare against the file size. 1033 bool ElfFileImpl<ElfTypes>::GetLoadedSize(size_t* size, std::string* error_msg) const { argument 1052 *size = static_cast<size_t>(-1); 1066 oss << "Loaded size is 0x" << std::hex << loaded_size << " but maximum size_t is 0x" 1069 *size = static_cast<size_t>(-1); 1072 *size = loaded_size; 1116 // of required size an 1861 GetLoadedSize(size_t* size, std::string* error_msg) const argument [all...] |
/art/test/ti-agent/ |
H A D | redefinition_helper.cc | 246 gTransformations[name_str].size() > 0) { 250 CHECK_EQ(JVMTI_ERROR_NONE, jvmti_env->Allocate(desired_array.size(), &new_data)); 251 memcpy(new_data, desired_array.data(), desired_array.size()); 253 *new_class_data_len = desired_array.size(); 273 gTransformations[name_str].size() > 0) {
|
/art/compiler/linker/arm/ |
H A D | relative_patcher_arm_base.cc | 44 return code_.size(); 103 return pending_offset_ != offsets_.size(); 136 return offsets_.size(); 193 pending_thunks_.reserve(thunks_.size()); 288 // Adjust code size for extra space required by the subclass. 289 uint32_t max_code_size = compiled_method->GetQuickCode().size() + max_extra_space; 439 size_t index = unreserved_thunks_.size(); 446 if (index + 1u != unreserved_thunks_.size()) {
|
/art/runtime/jit/ |
H A D | profile_compilation_info.h | 484 &bitmap_storage[0], bitmap_storage.size()), 0, ComputeBitmapBits(num_method_ids)); 503 DCHECK_EQ(bitmap_storage.size(), other.bitmap_storage.size()); 504 for (size_t i = 0; i < bitmap_storage.size(); ++i) { 595 // Inflate the input buffer (in_buffer) of size in_size. It returns a buffer of 596 // compressed data for the input buffer of "compressed_data_size" size. 601 // Inflate the input buffer(in_buffer) of size in_size. out_size is the expected output 602 // size of the buffer. It puts the output in out_buffer. It returns Z_STREAM_END on 676 explicit SafeBuffer(size_t size) : storage_(new uint8_t[size]) { argument [all...] |
/art/dexlayout/ |
H A D | dex_ir.h | 65 // Item size constants. 134 uint32_t Size() const { return collection_.size(); } 142 CHECK_EQ(map.size(), Size()); 190 uint32_t size() const { return collection_.size(); } function in class:art::dex_ir::CollectionMap 495 void SetSize(uint32_t size) { size_ = size; } argument 501 Item(uint32_t offset, uint32_t size) : offset_(offset), size_(size) { } argument 520 IndexedItem(uint32_t offset, uint32_t size, uint32_ argument 1317 DexFileSection(const std::string& name, uint16_t type, uint32_t size, uint32_t offset) argument 1323 uint32_t size; member in struct:art::dex_ir::DexFileSection [all...] |
H A D | compact_dex_writer.cc | 86 stream->Write(data.data(), data.size()); 97 for (size_t i = 0; i < debug_info_offsets.size(); ++i) { 102 << (end_time - start_time) / debug_info_offsets.size(); 280 // header size. 437 // Sort the debug infos by method index order, this reduces size by ~0.1% by reducing the size of 482 if (link_data.size() > 0) { 483 CHECK_EQ(header_->LinkSize(), static_cast<uint32_t>(link_data.size())); 489 data_stream->Write(&link_data[0], link_data.size()); 500 // Offset must be zero when the size i [all...] |
/art/runtime/gc/space/ |
H A D | large_object_space.cc | 69 *usable_size = num_bytes; // Since we have redzones, shrink the usable size. 207 CHECK(it != large_objects_.end()) << "Attempted to get size of a large object which is not live"; 248 // allocation. Each allocation has an AllocationInfo which contains the size of the previous free 259 // Returns the allocation size in bytes. 263 // Updates the allocation size and whether or not it is free. 264 void SetByteSize(size_t size, bool free) { argument 265 DCHECK_EQ(size & ~kFlagsMask, 0u); 266 DCHECK_ALIGNED(size, FreeListSpace::kAlignment); 267 alloc_size_ = (size / FreeListSpace::kAlignment) | (free ? kFlagFree : 0u); 308 // Update the size o 349 Create(const std::string& name, uint8_t* requested_begin, size_t size) argument 541 size_t size = cur_info->ByteSize(); local [all...] |
/art/compiler/utils/mips64/ |
H A D | assembler_mips64.h | 328 Literal(uint32_t size, const uint8_t* data) argument 329 : label_(), size_(size) { 330 DCHECK_LE(size, Literal::kMaxSize); 331 memcpy(data_, data, size); 374 return labels_.size() * sizeof(uint32_t); 480 void Dext(GpuRegister rs, GpuRegister rt, int pos, int size); // MIPS64 481 void Ins(GpuRegister rt, GpuRegister rs, int pos, int size); 482 void Dins(GpuRegister rt, GpuRegister rs, int pos, int size); // MIPS64 483 void Dinsm(GpuRegister rt, GpuRegister rs, int pos, int size); // MIPS64 484 void Dinsu(GpuRegister rt, GpuRegister rs, int pos, int size); // MIPS6 [all...] |
/art/compiler/utils/mips/ |
H A D | assembler_mips.cc | 30 "Unexpected Mips pointer size."); 31 static_assert(kMipsPointerSize == PointerSize::k32, "Unexpected Mips pointer size."); 67 // generator to CodePosition() to record the buffer size. 68 // Instructions after which the buffer size is recorded cannot 139 size_t size = buffer_.Size(); local 146 if (ds_fsm_target_pcs_.size() != 0 && ds_fsm_target_pcs_.back() + sizeof(uint32_t) == size) { 149 return size; 199 size_t size = advances.size(); local 678 Ext(Register rd, Register rt, int pos, int size) argument 681 CHECK(0 < pos + size && pos + size <= 32) << pos << " + " << size; local 686 Ins(Register rd, Register rt, int pos, int size) argument 689 CHECK(0 < pos + size && pos + size <= 32) << pos << " + " << size; local 953 size_t size = buffer_.Size(); local 3722 size_t size = buffer_.Size(); local 3779 NewLiteral(size_t size, const uint8_t* data) argument 3912 uint32_t size = end - branch.GetOldEndLocation(); local 4335 size_t size = buffer_.Size(); local 4709 EmitLoad(ManagedRegister m_dst, Register src_register, int32_t src_offset, size_t size) argument 4860 Store(FrameOffset dest, ManagedRegister msrc, size_t size) argument 4928 Load(ManagedRegister mdest, FrameOffset src, size_t size) argument 4932 LoadFromThread(ManagedRegister mdest, ThreadOffset32 src, size_t size) argument 4974 Move(ManagedRegister mdest, ManagedRegister msrc, size_t size) argument 5036 Copy(FrameOffset dest, FrameOffset src, ManagedRegister mscratch, size_t size) argument 5051 Copy(FrameOffset dest, ManagedRegister src_base, Offset src_offset, ManagedRegister mscratch, size_t size) argument 5059 Copy(ManagedRegister dest_base, Offset dest_offset, FrameOffset src, ManagedRegister mscratch, size_t size) argument 5075 Copy(ManagedRegister dest, Offset dest_offset, ManagedRegister src, Offset src_offset, ManagedRegister mscratch, size_t size) argument [all...] |
/art/compiler/utils/ |
H A D | assembler_test_base.h | 130 if (data.size() > res.code->size()) { 133 "is equal: this implies sub-optimal encoding! Our code size=" << data.size() << 134 ", gcc size=" << res.code->size(); 358 s_out.write(data, buffer.size()); 412 if (tokens.size() < OBJDUMP_SECTION_LINE_MIN_TOKENS) {
|
/art/runtime/entrypoints/ |
H A D | entrypoint_utils.cc | 69 if (args.size() > 0 || (target_sdk_version > 0 && target_sdk_version <= 21)) { 70 args_jobj = soa.Env()->NewObjectArray(args.size(), WellKnownClasses::java_lang_Object, nullptr); 75 for (size_t i = 0; i < args.size(); ++i) {
|