/arch/x86/include/asm/ |
H A D | ftrace.h | 8 subq $0x38, %rsp 9 movq %rax, (%rsp) 10 movq %rcx, 8(%rsp) 11 movq %rdx, 16(%rsp) 12 movq %rsi, 24(%rsp) 13 movq %rdi, 32(%rsp) 14 movq %r8, 40(%rsp) 15 movq %r9, 48(%rsp) 19 movq 48(%rsp), %r9 20 movq 40(%rsp), [all...] |
H A D | calling.h | 10 ( rsp is obviously invariant across normal function calls. (gcc can 'merge' 89 subq $9*8+\addskip, %rsp 140 addq $ARG_SKIP+\addskip, %rsp 146 movq \offset(%rsp), %r11 147 movq \offset+8(%rsp), %r10 148 movq \offset+16(%rsp), %r9 149 movq \offset+24(%rsp), %r8 150 movq \offset+40(%rsp), %rcx 151 movq \offset+48(%rsp), %rdx 152 movq \offset+56(%rsp), [all...] |
H A D | dwarf2.h | 105 movq %\reg, \offset(%rsp) 110 movq \offset(%rsp), %\reg
|
H A D | ptrace.h | 92 unsigned long rsp; member in struct:pt_regs
|
H A D | sigcontext.h | 248 __u64 rsp; member in struct:sigcontext
|
/arch/x86/platform/efi/ |
H A D | efi_stub_64.S | 12 mov %rsp, %rax; \ 13 subq $0x70, %rsp; \ 14 and $~0xf, %rsp; \ 15 mov %rax, (%rsp); \ 18 mov %rax, 0x8(%rsp); \ 19 movaps %xmm0, 0x60(%rsp); \ 20 movaps %xmm1, 0x50(%rsp); \ 21 movaps %xmm2, 0x40(%rsp); \ 22 movaps %xmm3, 0x30(%rsp); \ 23 movaps %xmm4, 0x20(%rsp); \ [all...] |
/arch/x86/ia32/ |
H A D | ia32entry.S | 45 movq %rax,\offset+R11(%rsp) 46 movq %rax,\offset+R10(%rsp) 47 movq %\_r9,\offset+R9(%rsp) 48 movq %rax,\offset+R8(%rsp) 63 movl \offset+16(%rsp),%r9d 65 movl \offset+40(%rsp),%ecx 66 movl \offset+48(%rsp),%edx 67 movl \offset+56(%rsp),%esi 68 movl \offset+64(%rsp),%edi 119 CFI_DEF_CFA rsp, [all...] |
/arch/x86/xen/ |
H A D | xen-asm_64.S | 24 mov 8+0(%rsp), %rcx 25 mov 8+8(%rsp), %r11 33 * rsp 42 * rsp->rax } 71 movq %rsp, PER_CPU_VAR(old_rsp) 72 movq PER_CPU_VAR(kernel_stack), %rsp 90 movq %rsp, PER_CPU_VAR(old_rsp) 91 movq PER_CPU_VAR(kernel_stack), %rsp 108 * - kernel rsp 111 * rsp [all...] |
H A D | xen-head.S | 24 mov $init_thread_union+THREAD_SIZE,%rsp
|
/arch/x86/kernel/ |
H A D | entry_64.S | 81 movq 0x38(%rsp), %rdi 121 movq 0x38(%rsp), %rdi 142 movq 0x38(%rsp), %rsi 154 subq $24, %rsp 157 movq %rax, (%rsp) 158 movq %rdx, 8(%rsp) 164 movq 8(%rsp), %rdx 165 movq (%rsp), %rax 166 addq $24, %rsp 185 bt $9,EFLAGS-\offset(%rsp) /* interrupt [all...] |
H A D | relocate_kernel_64.S | 62 movq %rsp, RSP(%r11) 95 lea PAGE_SIZE(%r8), %rsp 149 * leave %rsp alone 174 leaq PAGE_SIZE(%r10), %rsp 178 movq 0(%rsp), %rbp 187 lea PAGE_SIZE(%r8), %rsp 194 movq RSP(%r8), %rsp
|
H A D | head_64.S | 200 movq stack_start(%rip),%rsp 202 /* zero EFLAGS after setting rsp */ 301 0: movq 0(%rsp),%rcx # get ip 302 movq 8(%rsp),%rdx # get cs 311 movq 0(%rsp),%rsi # get rip again
|
/arch/x86/um/ |
H A D | stub_64.S | 24 mov %rbx, %rsp 25 add $0x10, %rsp 28 mov 0x0(%rsp), %rax 37 mov %rsp, 8(%rbx) 40 add %rax, %rsp
|
H A D | setjmp_64.S | 10 # %rsp (post-return) 27 movq %rsp,8(%rdi) # Post-return %rsp! 46 movq 8(%rdi),%rsp
|
/arch/x86/crypto/ |
H A D | salsa20-x86_64-asm_64.S | 6 mov %rsp,%r11 9 sub %r11,%rsp 27 movq %r11,0(%rsp) 29 movq %r12,8(%rsp) 31 movq %r13,16(%rsp) 33 movq %r14,24(%rsp) 35 movq %r15,32(%rsp) 37 movq %rbx,40(%rsp) 39 movq %rbp,48(%rsp) 57 movq %rcx,56(%rsp) [all...] |
H A D | aesni-intel_asm.S | 362 movdqa \TMP1, HashKey_k(%rsp) 365 movdqa \TMP5, HashKey_2(%rsp) 369 movdqa \TMP1, HashKey_2_k(%rsp) 379 movdqa \TMP5, HashKey_3(%rsp) 382 movdqa \TMP1, HashKey_3_k(%rsp) 392 movdqa \TMP5, HashKey_4(%rsp) 395 movdqa \TMP1, HashKey_4_k(%rsp) 599 movdqa \TMP1, HashKey_k(%rsp) 602 movdqa \TMP5, HashKey_2(%rsp) 606 movdqa \TMP1, HashKey_2_k(%rsp) [all...] |
H A D | sha1_ssse3_asm.S | 64 #define WK(t) (((t) & 15) * 4)(%rsp) 80 mov %rsp, %r12 81 sub $64, %rsp # allocate workspace 82 and $~15, %rsp # align stack 98 mov %rsp, %rdi 102 mov %r12, %rsp # deallocate workspace
|
/arch/x86/lib/ |
H A D | csum-copy_64.S | 63 subq $7*8, %rsp 65 movq %rbx, 2*8(%rsp) 67 movq %r12, 3*8(%rsp) 69 movq %r14, 4*8(%rsp) 71 movq %r13, 5*8(%rsp) 73 movq %rbp, 6*8(%rsp) 76 movq %r8, (%rsp) 77 movq %r9, 1*8(%rsp) 219 movq 2*8(%rsp), %rbx 221 movq 3*8(%rsp), [all...] |
H A D | copy_page_64.S | 23 subq $2*8,%rsp 25 movq %rbx,(%rsp) 27 movq %r12,1*8(%rsp) 88 movq (%rsp),%rbx 90 movq 1*8(%rsp),%r12 92 addq $2*8,%rsp
|
/arch/arm/mach-msm/ |
H A D | scm.c | 146 * @rsp: response 150 static inline void *scm_get_response_buffer(const struct scm_response *rsp) argument 152 return (void *)rsp + rsp->buf_offset; 229 struct scm_response *rsp; local 245 rsp = scm_command_to_response(cmd); 247 u32 start = (u32)rsp; 248 u32 end = (u32)scm_get_response_buffer(rsp) + resp_len; 255 } while (!rsp->is_complete); 258 memcpy(resp_buf, scm_get_response_buffer(rsp), resp_le [all...] |
/arch/x86/kernel/acpi/ |
H A D | wakeup_64.S | 27 movq saved_rsp, %rsp 42 subq $8, %rsp 47 movq %rsp, pt_regs_sp(%rax) 67 movq %rsp, saved_rsp 73 addq $8, %rsp 92 movq pt_regs_sp(%rax), %rsp 109 addq $8, %rsp
|
/arch/x86/power/ |
H A D | hibernate_asm_64.S | 27 movq %rsp, pt_regs_sp(%rax) 124 movq pt_regs_sp(%rax), %rsp
|
/arch/x86/include/asm/xen/ |
H A D | interface_64.h | 65 * orb $3,1*8(%rsp) 82 /* Top of stack (%rsp at point of hypercall). */ 83 uint64_t rax, r11, rcx, flags, rip, cs, rflags, rsp, ss; member in struct:iret_context
|
/arch/powerpc/include/asm/ |
H A D | immap_qe.h | 396 struct rsp { struct 455 struct rsp rsp[0x2]; /* 0x104100 - 0x1042FF member in struct:qe_immap
|
/arch/mips/include/asm/octeon/ |
H A D | cvmx-pko.h | 198 * If set and rsp is set, word3 contains a pointer to 203 uint64_t rsp:1; member in struct:__anon2113::__anon2114
|