/art/runtime/base/ |
H A D | casts.h | 82 inline Dest bit_cast(const Source& source) { argument 87 memcpy(&dest, &source, sizeof(dest));
|
/art/compiler/optimizing/ |
H A D | register_allocator.h | 110 void InsertParallelMoveAtExitOf(HBasicBlock* block, Location source, Location destination) const; 111 void InsertParallelMoveAtEntryOf(HBasicBlock* block, Location source, Location destination) const; 112 void InsertMoveAfter(HInstruction* instruction, Location source, Location destination) const; 113 void AddInputMoveFor(HInstruction* instruction, Location source, Location destination) const; 114 void InsertParallelMoveAt(size_t position, Location source, Location destination) const;
|
H A D | register_allocator.cc | 676 Location source, 678 if (source.Equals(destination)) return; 694 move->AddMove(new (allocator_) MoveOperands(source, destination)); 698 Location source, 700 if (source.Equals(destination)) return; 737 move->AddMove(new (allocator_) MoveOperands(source, destination)); 741 Location source, 743 if (source.Equals(destination)) return; 759 move->AddMove(new (allocator_) MoveOperands(source, destination)); 763 Location source, 675 AddInputMoveFor(HInstruction* instruction, Location source, Location destination) const argument 697 InsertParallelMoveAt(size_t position, Location source, Location destination) const argument 740 InsertParallelMoveAtExitOf(HBasicBlock* block, Location source, Location destination) const argument 762 InsertParallelMoveAtEntryOf(HBasicBlock* block, Location source, Location destination) const argument 779 InsertMoveAfter(HInstruction* instruction, Location source, Location destination) const argument 817 Location source = ConvertToLocation(current); local 860 LiveInterval* source = nullptr; local 933 Location source = ConvertToLocation(current); local 973 Location source = FindLocationAt(input->GetLiveInterval(), local [all...] |
H A D | code_generator_x86_64.cc | 274 void CodeGeneratorX86_64::Move(Location destination, Location source) { argument 275 if (source.Equals(destination)) { 279 if (source.IsRegister()) { 280 __ movq(destination.AsX86_64().AsCpuRegister(), source.AsX86_64().AsCpuRegister()); 281 } else if (source.IsStackSlot()) { 282 __ movl(destination.AsX86_64().AsCpuRegister(), Address(CpuRegister(RSP), source.GetStackIndex())); 284 DCHECK(source.IsDoubleStackSlot()); 285 __ movq(destination.AsX86_64().AsCpuRegister(), Address(CpuRegister(RSP), source.GetStackIndex())); 288 if (source.IsRegister()) { 289 __ movl(Address(CpuRegister(RSP), destination.GetStackIndex()), source 1340 Location source = move->GetSource(); local 1437 Location source = move->GetSource(); local [all...] |
H A D | code_generator_arm.cc | 368 void CodeGeneratorARM::Move32(Location destination, Location source) { argument 369 if (source.Equals(destination)) { 373 if (source.IsRegister()) { 374 __ Mov(destination.AsArm().AsCoreRegister(), source.AsArm().AsCoreRegister()); 376 __ ldr(destination.AsArm().AsCoreRegister(), Address(SP, source.GetStackIndex())); 380 if (source.IsRegister()) { 381 __ str(source.AsArm().AsCoreRegister(), Address(SP, destination.GetStackIndex())); 383 __ ldr(IP, Address(SP, source.GetStackIndex())); 389 void CodeGeneratorARM::Move64(Location destination, Location source) { argument 390 if (source 1506 Location source = move->GetSource(); local 1559 Location source = move->GetSource(); local [all...] |
H A D | code_generator_x86.cc | 343 void CodeGeneratorX86::Move32(Location destination, Location source) { argument 344 if (source.Equals(destination)) { 348 if (source.IsRegister()) { 349 __ movl(destination.AsX86().AsCpuRegister(), source.AsX86().AsCpuRegister()); 351 DCHECK(source.IsStackSlot()); 352 __ movl(destination.AsX86().AsCpuRegister(), Address(ESP, source.GetStackIndex())); 355 if (source.IsRegister()) { 356 __ movl(Address(ESP, destination.GetStackIndex()), source.AsX86().AsCpuRegister()); 358 DCHECK(source.IsStackSlot()); 359 __ pushl(Address(ESP, source 365 Move64(Location destination, Location source) argument 1501 Location source = move->GetSource(); local 1561 Location source = move->GetSource(); local [all...] |
H A D | parallel_move_resolver.cc | 53 // moves to perform, ignoring any move that is redundant (the source is 70 // which means that a call to PerformMove could change any source operand 83 // dependencies. Any unperformed, unpending move with a source the same 89 // Though PerformMove can change any source operand in the move graph, 91 // not miss any). Assume there is a non-blocking move with source A 92 // and this move is blocked on source B and there is a swap of A and 107 // This move's source may have changed due to swaps to resolve cycles and 129 // Any unperformed (including pending) move with a source of either 130 // this move's source or destination needs to have their source 132 Location source = move->GetSource(); local [all...] |
H A D | code_generator_arm.h | 179 void Move32(Location destination, Location source); 181 void Move64(Location destination, Location source);
|
H A D | code_generator_x86.h | 181 void Move32(Location destination, Location source); 183 void Move64(Location destination, Location source);
|
H A D | code_generator_x86_64.h | 183 void Move(Location destination, Location source);
|
H A D | nodes.h | 1471 MoveOperands(Location source, Location destination) argument 1472 : source_(source), destination_(destination) {} 1481 // destination (but not the source). 1504 // A move is redundant if it's been eliminated, if its source and
|
/art/compiler/utils/arm64/ |
H A D | assembler_arm64.h | 222 void StoreWToOffset(StoreOperandType type, WRegister source, 224 void StoreToOffset(Register source, Register base, int32_t offset); 225 void StoreSToOffset(SRegister source, Register base, int32_t offset); 226 void StoreDToOffset(DRegister source, Register base, int32_t offset);
|
H A D | assembler_arm64.cc | 92 void Arm64Assembler::StoreWToOffset(StoreOperandType type, WRegister source, argument 96 ___ Strb(reg_w(source), MEM_OP(reg_x(base), offset)); 99 ___ Strh(reg_w(source), MEM_OP(reg_x(base), offset)); 102 ___ Str(reg_w(source), MEM_OP(reg_x(base), offset)); 109 void Arm64Assembler::StoreToOffset(Register source, Register base, int32_t offset) { argument 110 CHECK_NE(source, SP); 111 ___ Str(reg_x(source), MEM_OP(reg_x(base), offset)); 114 void Arm64Assembler::StoreSToOffset(SRegister source, Register base, int32_t offset) { argument 115 ___ Str(reg_s(source), MEM_OP(reg_x(base), offset)); 118 void Arm64Assembler::StoreDToOffset(DRegister source, Registe argument 188 Arm64ManagedRegister source = m_source.AsArm64(); local [all...] |
/art/runtime/ |
H A D | elf_file.h | 146 bool CheckSectionsLinked(const byte* source, const byte* target) const;
|
H A D | thread_pool.h | 142 virtual void StealFrom(Thread* self, WorkStealingTask* source) = 0;
|
H A D | elf_file.cc | 338 bool ElfFile::CheckSectionsLinked(const byte* source, const byte* target) const { argument 351 if (Begin() + section_header->sh_offset == source) { 352 // Found the source.
|
H A D | class_linker.cc | 3267 std::string source; local 3269 source += " from "; 3270 source += dex_cache->GetLocation()->ToModifiedUtf8(); 3272 LOG(INFO) << "Loaded class " << descriptor << source; local
|
/art/compiler/dex/ |
H A D | mir_graph.cc | 1790 bool MIRGraph::HasSuspendTestBetween(BasicBlock* source, BasicBlockId target_id) { argument 1793 if (source == nullptr || target == nullptr) 1799 if (bb == source) 1801 if (source->dominators->IsBitSet(bb->id) && bb->dominators->IsBitSet(target_id))
|
H A D | mir_graph.h | 1075 * source basic block of a backedge and being dominated by the target basic 1078 bool HasSuspendTestBetween(BasicBlock* source, BasicBlockId target_id);
|