/art/runtime/base/ |
H A D | casts.h | 80 inline Dest bit_cast(const Source& source) { argument 85 memcpy(&dest, &source, sizeof(dest)); 92 inline Dest dchecked_integral_cast(const Source source) { argument 97 source >= static_cast<Source>(std::numeric_limits<Dest>::min())) && 101 source <= static_cast<Source>(std::numeric_limits<Dest>::max()))); 103 return static_cast<Dest>(source); 110 inline Dest reinterpret_cast64(Source source) { argument 116 DCHECK_EQ(static_cast<Source>(static_cast<uintptr_t>(source)), source); 117 return reinterpret_cast<Dest>(static_cast<uintptr_t>(source)); [all...] |
/art/runtime/interpreter/mterp/ |
H A D | gen_mterp.py | 19 # and assembly source files for the Dalvik interpreter. 105 source = tokens[1] 106 if source.endswith(".S"): 110 % source) 187 # Set location of a single alt opcode's source file. 207 # Set location of a single opcode's source file. 334 def loadAndEmitAltStub(source, opindex): 337 print " alt emit %s --> stub" % source 342 appendSourceFile(source, dict, asm_fp, None) 361 source [all...] |
/art/compiler/optimizing/ |
H A D | parallel_move_resolver.cc | 26 // moves to perform, ignoring any move that is redundant (the source is 104 // Update the source of `move`, knowing that `updated_location` has been swapped 108 Location source = move->GetSource(); local 109 if (LowOf(updated_location).Equals(source)) { 111 } else if (HighOf(updated_location).Equals(source)) { 114 DCHECK(updated_location.Equals(source)) << updated_location << " " << source; local 124 // which means that a call to PerformMove could change any source operand 143 // dependencies. Any unperformed, unpending move with a source the same 150 // Though PerformMove can change any source operan 220 Location source = move->GetSource(); local 330 Location source = move->GetSource(); local 448 Location source = move->GetSource(); local 523 AddPendingMove(Location source, Location destination, Primitive::Type type) argument [all...] |
H A D | register_allocator.h | 124 Location source, 128 Location source, 130 void InsertMoveAfter(HInstruction* instruction, Location source, Location destination) const; 133 Location source, 137 Location source, 141 Location source,
|
H A D | register_allocator.cc | 1408 Location source, 1415 && !source.IsConstant()) { 1416 move->AddMove(source.ToLow(), destination.ToLow(), Primitive::kPrimInt, instruction); 1417 move->AddMove(source.ToHigh(), destination.ToHigh(), Primitive::kPrimInt, nullptr); 1419 move->AddMove(source, destination, type, instruction); 1425 Location source, 1427 if (source.Equals(destination)) return; 1443 AddMove(move, source, destination, nullptr, input->GetType()); 1456 Location source, 1459 if (source 1407 AddMove(HParallelMove* move, Location source, Location destination, HInstruction* instruction, Primitive::Type type) const argument 1423 AddInputMoveFor(HInstruction* input, HInstruction* user, Location source, Location destination) const argument 1454 InsertParallelMoveAt(size_t position, HInstruction* instruction, Location source, Location destination) const argument 1526 InsertParallelMoveAtExitOf(HBasicBlock* block, HInstruction* instruction, Location source, Location destination) const argument 1556 InsertParallelMoveAtEntryOf(HBasicBlock* block, HInstruction* instruction, Location source, Location destination) const argument 1576 InsertMoveAfter(HInstruction* instruction, Location source, Location destination) const argument 1619 Location source = current->ToLocation(); local 1759 LiveInterval* source = interval->GetSiblingAt(source_position); local 1899 Location source = current->ToLocation(); local 1964 Location source = input->GetLiveInterval()->GetLocationAt( local [all...] |
H A D | code_generator_arm.cc | 1067 void CodeGeneratorARM::Move32(Location destination, Location source) { argument 1068 if (source.Equals(destination)) { 1072 if (source.IsRegister()) { 1073 __ Mov(destination.AsRegister<Register>(), source.AsRegister<Register>()); 1074 } else if (source.IsFpuRegister()) { 1075 __ vmovrs(destination.AsRegister<Register>(), source.AsFpuRegister<SRegister>()); 1077 __ LoadFromOffset(kLoadWord, destination.AsRegister<Register>(), SP, source.GetStackIndex()); 1080 if (source.IsRegister()) { 1081 __ vmovsr(destination.AsFpuRegister<SRegister>(), source.AsRegister<Register>()); 1082 } else if (source 1101 Move64(Location destination, Location source) argument 4522 Register source = value; local 4853 Location source = move->GetSource(); local 4997 Location source = move->GetSource(); local [all...] |
H A D | code_generator_x86.cc | 983 void CodeGeneratorX86::Move32(Location destination, Location source) { argument 984 if (source.Equals(destination)) { 988 if (source.IsRegister()) { 989 __ movl(destination.AsRegister<Register>(), source.AsRegister<Register>()); 990 } else if (source.IsFpuRegister()) { 991 __ movd(destination.AsRegister<Register>(), source.AsFpuRegister<XmmRegister>()); 993 DCHECK(source.IsStackSlot()); 994 __ movl(destination.AsRegister<Register>(), Address(ESP, source.GetStackIndex())); 997 if (source.IsRegister()) { 998 __ movd(destination.AsFpuRegister<XmmRegister>(), source 1023 Move64(Location destination, Location source) argument 3131 PushOntoFPStack(Location source, uint32_t temp_offset, uint32_t stack_adjustment, bool is_fp, bool is_wide) argument 5631 Location source = move->GetSource(); local 5761 LOG(FATAL) << "Unimplemented move: " << destination << " <- " << source; local 5805 Location source = move->GetSource(); local [all...] |
H A D | parallel_move_resolver.h | 152 // Called after a move which takes a scratch location as source. A subclass can defer the cleanup 180 void AddPendingMove(Location source, Location destination, Primitive::Type type);
|
H A D | code_generator_x86_64.cc | 1123 void CodeGeneratorX86_64::Move(Location destination, Location source) { argument 1124 if (source.Equals(destination)) { 1129 if (source.IsRegister()) { 1130 __ movq(dest, source.AsRegister<CpuRegister>()); 1131 } else if (source.IsFpuRegister()) { 1132 __ movd(dest, source.AsFpuRegister<XmmRegister>()); 1133 } else if (source.IsStackSlot()) { 1134 __ movl(dest, Address(CpuRegister(RSP), source.GetStackIndex())); 1135 } else if (source.IsConstant()) { 1136 HConstant* constant = source 3199 PushOntoFPStack(Location source, uint32_t temp_offset, uint32_t stack_adjustment, bool is_float) argument 5107 Location source = move->GetSource(); local 5261 Location source = move->GetSource(); local [all...] |
H A D | code_generator_mips.cc | 823 void CodeGeneratorMIPS::Move32(Location destination, Location source) { argument 824 if (source.Equals(destination)) { 829 if (source.IsRegister()) { 830 __ Move(destination.AsRegister<Register>(), source.AsRegister<Register>()); 831 } else if (source.IsFpuRegister()) { 832 __ Mfc1(destination.AsRegister<Register>(), source.AsFpuRegister<FRegister>()); 834 DCHECK(source.IsStackSlot()) << "Cannot move from " << source << " to " << destination; 835 __ LoadFromOffset(kLoadWord, destination.AsRegister<Register>(), SP, source.GetStackIndex()); 838 if (source 860 Move64(Location destination, Location source) argument [all...] |
H A D | code_generator_mips64.cc | 632 Location source, 634 if (source.Equals(destination)) { 638 // A valid move can always be inferred from the destination and source 646 HConstant* src_cst = source.IsConstant() ? source.GetConstant() : nullptr; 647 if (source.IsStackSlot() || 654 // If the source is a double stack slot or a 64bit constant, a 64bit 655 // type is appropriate. Else the source is a register, and since the 663 if (source.IsStackSlot() || source 631 MoveLocation(Location destination, Location source, Primitive::Type dst_type) argument [all...] |
H A D | code_generator_arm64.cc | 1184 Location source, 1186 if (source.Equals(destination)) { 1190 // A valid move can always be inferred from the destination and source 1198 HConstant* src_cst = source.IsConstant() ? source.GetConstant() : nullptr; 1199 if (source.IsStackSlot() || 1206 // If the source is a double stack slot or a 64bit constant, a 64bit 1207 // type is appropriate. Else the source is a register, and since the 1216 if (source.IsStackSlot() || source 1183 MoveLocation(Location destination, Location source, Primitive::Type dst_type) argument 1691 CPURegister source = value; local 2073 MemOperand source = HeapOperand(obj); local 2156 CPURegister source = value; local [all...] |
H A D | code_generator_x86.h | 271 void PushOntoFPStack(Location source, uint32_t temp_offset, 385 void Move32(Location destination, Location source); 387 void Move64(Location destination, Location source);
|
H A D | code_generator_mips.h | 270 void Move32(Location destination, Location source); 271 void Move64(Location destination, Location source);
|
H A D | code_generator_arm.h | 363 void Move32(Location destination, Location source); 365 void Move64(Location destination, Location source);
|
H A D | code_generator_x86_64.h | 264 void PushOntoFPStack(Location source, uint32_t temp_offset, 372 void Move(Location destination, Location source);
|
/art/test/021-string2/src/junit/framework/ |
H A D | ComparisonCompactor.java | 39 private String compactString(String source) { argument 40 String result= DELTA_START + source.substring(fPrefix, source.length() - fSuffix + 1) + DELTA_END;
|
/art/test/082-inline-execute/src/junit/framework/ |
H A D | ComparisonCompactor.java | 39 private String compactString(String source) { argument 40 String result= DELTA_START + source.substring(fPrefix, source.length() - fSuffix + 1) + DELTA_END;
|
/art/runtime/mirror/ |
H A D | dex_cache-inl.h | 148 mirror::String* source = src[i].Read<kReadBarrierOption>(); local 149 mirror::String* new_source = visitor(source); 158 mirror::Class* source = src[i].Read<kReadBarrierOption>(); local 159 mirror::Class* new_source = visitor(source);
|
/art/compiler/utils/arm64/ |
H A D | assembler_arm64.h | 257 void StoreWToOffset(StoreOperandType type, WRegister source, 259 void StoreToOffset(XRegister source, XRegister base, int32_t offset); 260 void StoreSToOffset(SRegister source, XRegister base, int32_t offset); 261 void StoreDToOffset(DRegister source, XRegister base, int32_t offset);
|
H A D | assembler_arm64.cc | 97 void Arm64Assembler::StoreWToOffset(StoreOperandType type, WRegister source, argument 101 ___ Strb(reg_w(source), MEM_OP(reg_x(base), offset)); 104 ___ Strh(reg_w(source), MEM_OP(reg_x(base), offset)); 107 ___ Str(reg_w(source), MEM_OP(reg_x(base), offset)); 114 void Arm64Assembler::StoreToOffset(XRegister source, XRegister base, int32_t offset) { argument 115 CHECK_NE(source, SP); 116 ___ Str(reg_x(source), MEM_OP(reg_x(base), offset)); 119 void Arm64Assembler::StoreSToOffset(SRegister source, XRegister base, int32_t offset) { argument 120 ___ Str(reg_s(source), MEM_OP(reg_x(base), offset)); 123 void Arm64Assembler::StoreDToOffset(DRegister source, XRegiste argument 193 Arm64ManagedRegister source = m_source.AsArm64(); local [all...] |
/art/runtime/jit/ |
H A D | offline_profiling_info.h | 135 const std::string& source,
|
H A D | offline_profiling_info.cc | 345 const std::string& source, 352 *error += "Profile EOF reached prematurely for " + source; 355 *error += "Profile IO error for " + source + strerror(errno); 343 FillFromFd( int fd, const std::string& source, std::string* error) argument
|
/art/runtime/ |
H A D | elf_file_impl.h | 169 bool CheckSectionsLinked(const uint8_t* source, const uint8_t* target) const;
|
/art/cmdline/ |
H A D | cmdline_types.h | 755 static std::string RemovePrefix(const std::string& source) { argument 756 size_t prefix_idx = source.find(":"); 762 return source.substr(prefix_idx + 1);
|