/arch/m68k/lib/ |
H A D | memset.c | 27 short *ss = s; local 28 *ss++ = c; 29 s = ss; 64 short *ss = s; local 65 *ss++ = c; 66 s = ss;
|
/arch/mips/boot/compressed/ |
H A D | string.c | 23 char *ss = s; local 26 ss[i] = c;
|
/arch/hexagon/kernel/ |
H A D | process.c | 77 struct hexagon_switch_stack *ss; local 92 ss = (struct hexagon_switch_stack *) ((unsigned long) childregs - 93 sizeof(*ss)); 94 ss->lr = (unsigned long)ret_from_fork; 95 p->thread.switch_sp = ss; 99 ss->r24 = usp; 100 ss->r25 = arg; 105 ss->r2524 = 0;
|
/arch/x86/boot/compressed/ |
H A D | string.c | 36 char *ss = s; local 39 ss[i] = c;
|
/arch/x86/include/asm/ |
H A D | suspend_32.h | 14 u16 es, fs, gs, ss; member in struct:saved_context
|
H A D | suspend_64.h | 22 u16 ds, es, fs, gs, ss; member in struct:saved_context
|
H A D | sigcontext.h | 26 unsigned short ss, __ssh; member in struct:sigcontext
|
H A D | user_32.h | 94 unsigned long ss; member in struct:user_regs_struct
|
H A D | a.out-core.h | 58 dump->regs.ss = (u16)regs->ss;
|
H A D | kexec.h | 72 * CPU does not save ss and sp on stack if execution is already 82 "movw %%ss, %%ax\n\t" 83 :"=a"(newregs->ss)); 89 * via panic otherwise just fix up the ss and sp if coming via kernel 108 asm volatile("movl %%ss, %%eax;" :"=a"(newregs->ss)); 130 asm volatile("movl %%ss, %%eax;" :"=a"(newregs->ss));
|
H A D | ptrace.h | 28 unsigned long ss; member in struct:pt_regs 57 unsigned long ss; member in struct:pt_regs 167 #define MAX_REG_OFFSET (offsetof(struct pt_regs, ss)) 185 * Traps from the kernel do not save sp and ss.
|
/arch/x86/power/ |
H A D | cpu.c | 85 savesegment(ss, ctxt->ss); 92 asm volatile ("movw %%ss, %0" : "=m" (ctxt->ss)); 207 loadsegment(ss, ctxt->ss); 220 asm volatile ("movw %0, %%ss" :: "r" (ctxt->ss));
|
/arch/x86/include/uapi/asm/ |
H A D | ptrace.h | 67 unsigned long ss; member in struct:pt_regs
|
H A D | sigcontext32.h | 71 unsigned short ss, __ssh; member in struct:sigcontext_ia32
|
/arch/x86/xen/ |
H A D | xen-asm_32.S | 105 movl %ss:xen_vcpu, %eax 118 setz %ss:XEN_vcpu_info_mask(%eax) 122 cmpw $0x0001, %ss:XEN_vcpu_info_pending(%eax) 130 movb $1, %ss:XEN_vcpu_info_mask(%eax) 166 * ss : (ss/esp may be present if we came from usermode) 175 * cs } (no ss/esp because we're nested
|
/arch/m32r/boot/compressed/ |
H A D | misc.c | 33 char *ss = s; local 36 *ss++ = c;
|
/arch/x86/kvm/ |
H A D | tss.h | 25 u32 ss; member in struct:tss_segment_32 54 u16 ss; member in struct:tss_segment_16
|
/arch/x86/purgatory/ |
H A D | setup-x86_64.S | 26 movl %eax, %ss
|
/arch/sh/boot/compressed/ |
H A D | misc.c | 81 char *ss = (char*)s; local 83 for (i=0;i<n;i++) ss[i] = c;
|
/arch/x86/kernel/ |
H A D | dumpstack.c | 256 unsigned short ss; local 280 ss = regs->ss & 0xffff; 283 savesegment(ss, ss); 287 printk(" SS:ESP %04x:%08lx\n", ss, sp);
|
H A D | process_32.c | 74 unsigned short ss, gs; local 78 ss = regs->ss & 0xffff; 82 savesegment(ss, ss); 96 (u16)regs->ds, (u16)regs->es, (u16)regs->fs, gs, ss); 204 regs->ss = __USER_DS;
|
/arch/ia64/mm/ |
H A D | tlb.c | 108 static inline void spinaphore_init(struct spinaphore *ss, int val) argument 110 ss->ticket = 0; 111 ss->serve = val; 114 static inline void down_spin(struct spinaphore *ss) argument 116 unsigned long t = ia64_fetchadd(1, &ss->ticket, acq), serve; 118 if (time_before(t, ss->serve)) 124 asm volatile ("ld8.c.nc %0=[%1]" : "=r"(serve) : "r"(&ss->serve) : "memory"); 131 static inline void up_spin(struct spinaphore *ss) argument 133 ia64_fetchadd(1, &ss->serve, rel);
|
/arch/x86/math-emu/ |
H A D | get_address.c | 50 offsetof(struct pt_regs, ss), 63 offsetof(struct pt_regs, ss), 73 u_char ss, index, base; local 81 ss = base >> 6; 92 /* A non-zero ss is illegal */ 93 if (ss) 96 offset += (REG_(index)) << ss;
|
/arch/x86/include/asm/xen/ |
H A D | interface_64.h | 83 uint64_t rax, r11, rcx, flags, rip, cs, rflags, rsp, ss; member in struct:iret_context 122 uint16_t ss, _pad2[3]; member in struct:cpu_user_regs
|
/arch/x86/realmode/rm/ |
H A D | reboot.S | 63 movl %ecx, %ss 115 movw %ax, %ss
|