Lines Matching refs:scratch

2520   X86_64ManagedRegister scratch = mscratch.AsX86_64();
2521 CHECK(scratch.IsCpuRegister());
2522 leaq(scratch.AsCpuRegister(), Address(CpuRegister(RSP), fr_offs));
2523 gs()->movq(Address::Absolute(thr_offs, true), scratch.AsCpuRegister());
2531 FrameOffset /*in_off*/, ManagedRegister /*scratch*/) {
2671 X86_64ManagedRegister scratch = mscratch.AsX86_64();
2672 CHECK(scratch.IsCpuRegister());
2673 movl(scratch.AsCpuRegister(), Address(CpuRegister(RSP), src));
2674 movl(Address(CpuRegister(RSP), dest), scratch.AsCpuRegister());
2680 X86_64ManagedRegister scratch = mscratch.AsX86_64();
2681 CHECK(scratch.IsCpuRegister());
2682 gs()->movq(scratch.AsCpuRegister(), Address::Absolute(thr_offs, true));
2683 Store(fr_offs, scratch, 8);
2689 X86_64ManagedRegister scratch = mscratch.AsX86_64();
2690 CHECK(scratch.IsCpuRegister());
2691 Load(scratch, fr_offs, 8);
2692 gs()->movq(Address::Absolute(thr_offs, true), scratch.AsCpuRegister());
2697 X86_64ManagedRegister scratch = mscratch.AsX86_64();
2698 if (scratch.IsCpuRegister() && size == 8) {
2699 Load(scratch, src, 4);
2700 Store(dest, scratch, 4);
2701 Load(scratch, FrameOffset(src.Int32Value() + 4), 4);
2702 Store(FrameOffset(dest.Int32Value() + 4), scratch, 4);
2704 Load(scratch, src, size);
2705 Store(dest, scratch, size);
2710 ManagedRegister /*scratch*/, size_t /*size*/) {
2715 ManagedRegister scratch, size_t size) {
2716 CHECK(scratch.IsNoRegister());
2724 CpuRegister scratch = mscratch.AsX86_64().AsCpuRegister();
2726 movq(scratch, Address(CpuRegister(RSP), src_base));
2727 movq(scratch, Address(scratch, src_offset));
2728 movq(Address(CpuRegister(RSP), dest), scratch);
2733 ManagedRegister scratch, size_t size) {
2735 CHECK(scratch.IsNoRegister());
2742 CpuRegister scratch = mscratch.AsX86_64().AsCpuRegister();
2745 movq(scratch, Address(CpuRegister(RSP), src));
2746 pushq(Address(scratch, src_offset));
2747 popq(Address(scratch, dest_offset));
2786 X86_64ManagedRegister scratch = mscratch.AsX86_64();
2787 CHECK(scratch.IsCpuRegister());
2790 movl(scratch.AsCpuRegister(), Address(CpuRegister(RSP), handle_scope_offset));
2791 testl(scratch.AsCpuRegister(), scratch.AsCpuRegister());
2793 leaq(scratch.AsCpuRegister(), Address(CpuRegister(RSP), handle_scope_offset));
2796 leaq(scratch.AsCpuRegister(), Address(CpuRegister(RSP), handle_scope_offset));
2798 Store(out_off, scratch, 8);
2834 CpuRegister scratch = mscratch.AsX86_64().AsCpuRegister();
2835 movq(scratch, Address(CpuRegister(RSP), base));
2836 call(Address(scratch, offset));
2848 X86_64ManagedRegister scratch = mscratch.AsX86_64();
2849 gs()->movq(scratch.AsCpuRegister(), Address::Absolute(Thread::SelfOffset<8>(), true));
2850 movq(Address(CpuRegister(RSP), offset), scratch.AsCpuRegister());
2862 void X86_64Assembler::ExceptionPoll(ManagedRegister /*scratch*/, size_t stack_adjust) {