/art/compiler/optimizing/ |
H A D | nodes_mips.h | 73 int32_t num_entries, 79 num_entries_(num_entries) { 72 HMipsPackedSwitch(int32_t start_value, int32_t num_entries, HInstruction* input, HMipsComputeBaseMethodAddress* method_base, uint32_t dex_pc) argument
|
H A D | nodes_x86.h | 88 int32_t num_entries, 94 num_entries_(num_entries) { 87 HX86PackedSwitch(int32_t start_value, int32_t num_entries, HInstruction* input, HX86ComputeBaseMethodAddress* method_base, uint32_t dex_pc) argument
|
H A D | code_generator_arm_vixl.h | 126 uint32_t num_entries = switch_instr_->GetNumEntries(); local 127 for (uint32_t i = 0; i < num_entries; i++) {
|
H A D | code_generator.cc | 716 uint16_t num_entries = table.GetNumEntries(); local 720 for (size_t i = 0; i < num_entries; ++i) {
|
H A D | code_generator_arm64.cc | 81 // The compare/jump sequence will generate about (1.5 * num_entries + 3) instructions. While jump 82 // table version generates 7 instructions and num_entries literals. Compare/jump sequence will 83 // generates less code/data with a small num_entries. 646 uint32_t num_entries = switch_instr_->GetNumEntries(); local 647 DCHECK_GE(num_entries, kPackedSwitchCompareJumpThreshold); 652 num_entries * sizeof(int32_t), 657 for (uint32_t i = 0; i < num_entries; i++) { 5761 uint32_t num_entries = switch_instr->GetNumEntries(); 5773 if (num_entries <= kPackedSwitchCompareJumpThreshold || 5785 for (; num_entries [all...] |
H A D | code_generator_arm_vixl.cc | 2034 uint32_t num_entries = switch_instr_->GetNumEntries(); local 2035 DCHECK_GE(num_entries, kPackedSwitchCompareJumpThreshold); 2041 num_entries * sizeof(int32_t), 2045 for (uint32_t i = 0; i < num_entries; i++) { 2051 uint32_t num_entries = switch_instr_->GetNumEntries(); local 2052 DCHECK_GE(num_entries, kPackedSwitchCompareJumpThreshold); 2055 for (uint32_t i = 0; i < num_entries; i++) { 8698 uint32_t num_entries = switch_instr->GetNumEntries(); 8703 if (num_entries <= kPackedSwitchCompareJumpThreshold || 8718 for (; num_entries [all...] |
H A D | code_generator_x86.cc | 7413 uint32_t num_entries, 7419 DCHECK_GT(num_entries, 2u); 7437 for (; index + 1 < num_entries; index += 2) { 7446 if (index != num_entries) { 7448 DCHECK_EQ(index + 1, num_entries); 7461 uint32_t num_entries = switch_instr->GetNumEntries(); local 7467 num_entries, 7486 uint32_t num_entries = switch_instr->GetNumEntries(); local 7491 if (num_entries <= kPackedSwitchJumpTableThreshold) { 7494 num_entries, 7411 GenPackedSwitchWithCompares(Register value_reg, int32_t lower_bound, uint32_t num_entries, HBasicBlock* switch_block, HBasicBlock* default_block) argument 7670 int32_t num_entries = switch_instr_->GetNumEntries(); local [all...] |
H A D | code_generator_x86_64.cc | 44 // The compare/jump sequence will generate about (1.5 * num_entries) instructions. A jump 45 // table version generates 7 instructions and num_entries literals. Compare/jump sequence will 46 // generates less code/data with a small num_entries. 6782 uint32_t num_entries = switch_instr->GetNumEntries(); local 6790 if (num_entries <= kPackedSwitchJumpTableThreshold) { 6794 DCHECK_GT(num_entries, 2u); 6812 for (; index + 1 < num_entries; index += 2) { 6821 if (index != num_entries) { 6823 DCHECK_EQ(index + 1, num_entries); 6844 __ cmpl(value_reg, Immediate(num_entries 7023 int32_t num_entries = switch_instr_->GetNumEntries(); local [all...] |
H A D | nodes.h | 3152 uint32_t num_entries, 3157 num_entries_(num_entries) { 3151 HPackedSwitch(int32_t start_value, uint32_t num_entries, HInstruction* input, uint32_t dex_pc = kNoDexPc) argument
|
/art/runtime/ |
H A D | imt_conflict_table.h | 61 // num_entries excludes the header. 62 ImtConflictTable(size_t num_entries, PointerSize pointer_size) { argument 63 SetInterfaceMethod(num_entries, pointer_size, nullptr); 64 SetImplementationMethod(num_entries, pointer_size, nullptr); 171 static size_t ComputeSize(size_t num_entries, PointerSize pointer_size) { argument 172 return (num_entries + 1) * EntrySize(pointer_size); // Add one for null terminator.
|
H A D | stack_map.h | 1101 size_t num_entries; member in struct:art::DexRegisterLocation::BitEncodingTable 1108 EncodeUnsignedLeb128(dest, num_entries); 1113 num_entries = DecodeUnsignedLeb128(ptr); 1121 *offset += encoding.BitSize() * num_entries; 1127 DCHECK_LT(index, num_entries); 1138 size_t num_entries = 0; member in struct:art::DexRegisterLocation::ByteSizedTable 1148 EncodeUnsignedLeb128(dest, num_entries); 1153 num_entries = DecodeUnsignedLeb128(ptr); 1406 return encoding.location_catalog.num_entries; 1414 return encoding.stack_map.num_entries; [all...] |
H A D | class_linker.cc | 6506 const size_t num_entries = table->NumEntries(image_pointer_size_); local 6507 table->SetInterfaceMethod(num_entries, image_pointer_size_, interface_method); 6508 table->SetImplementationMethod(num_entries, image_pointer_size_, implementation_method); 6779 int32_t num_entries = vtable->GetLength(); local 6780 for (int32_t i = 0; i < num_entries; i++) { 6789 for (int32_t j = i + 1; j < num_entries; j++) {
|
/art/runtime/verifier/ |
H A D | verifier_deps.cc | 644 size_t num_entries = DecodeUint32WithOverflowCheck(in, end); local 645 for (size_t i = 0; i < num_entries; ++i) { 657 size_t num_entries = DecodeUint32WithOverflowCheck(in, end); local 658 vector->reserve(num_entries); 659 for (size_t i = 0; i < num_entries; ++i) {
|
/art/tools/dmtracedump/ |
H A D | tracedump.cc | 1224 int32_t num_entries = 0; local 1225 for (pTimed = list; pTimed; pTimed = pTimed->next) num_entries += 1; 1226 *num = num_entries; 1227 if (num_entries == 0) return nullptr; 1231 TimedMethod* sorted = new TimedMethod[num_entries]; 1234 qsort(sorted, num_entries, sizeof(TimedMethod), compareTimedMethod); 1237 for (ii = 0; ii < num_entries - 1; ++ii) sorted[ii].next = &sorted[ii + 1]; 1238 sorted[num_entries - 1].next = nullptr;
|