/external/llvm/test/MC/X86/ |
H A D | x86_64-rand-encoding.s | 23 // CHECK: rdrandq %r11 25 rdrand %r11 47 // CHECK: rdseedq %r11 49 rdseed %r11
|
H A D | x86_64-bmi-encoding.s | 7 // CHECK: blsmskq %r11, %r10 9 blsmskq %r11, %r10 23 // CHECK: blsiq %r11, %r10 25 blsiq %r11, %r10 39 // CHECK: blsrq %r11, %r10 41 blsrq %r11, %r10 55 // CHECK: andnq (%rax), %r11, %r10 57 andnq (%rax), %r11, %r10 71 // CHECK: bextrq %r12, %r11, %r10 73 bextrq %r12, %r11, [all...] |
/external/compiler-rt/lib/builtins/arm/ |
H A D | aeabi_ldivmod.S | 22 push {r11, lr} 30 pop {r11, pc}
|
H A D | aeabi_uldivmod.S | 22 push {r11, lr} 30 pop {r11, pc}
|
/external/valgrind/coregrind/m_syswrap/ |
H A D | syscall-amd64-darwin.S | 117 movq -16(%rbp), %r11 /* r11 = VexGuestAMD64State * */ 118 movq OFFSET_amd64_RDI(%r11), %rdi 119 movq OFFSET_amd64_RSI(%r11), %rsi 120 movq OFFSET_amd64_RDX(%r11), %rdx 121 movq OFFSET_amd64_RCX(%r11), %r10 /* rcx is passed in r10 instead */ 122 movq OFFSET_amd64_R8(%r11), %r8 123 movq OFFSET_amd64_R9(%r11), %r9 125 movq OFFSET_amd64_RSP(%r11), %r11 /* r1 [all...] |
/external/libhevc/common/arm/ |
H A D | ihevc_intra_pred_chroma_ver.s | 122 lsl r11, r3, #2 124 add r11, r11, #0xfffffff0 132 vst2.8 {d22,d23}, [r2], r11 133 vst2.8 {d22,d23}, [r5], r11 134 vst2.8 {d22,d23}, [r8], r11 135 vst2.8 {d22,d23}, [r10], r11 145 vst2.8 {d22,d23}, [r2], r11 146 vst2.8 {d22,d23}, [r5], r11 147 vst2.8 {d22,d23}, [r8], r11 [all...] |
H A D | ihevc_inter_pred_filters_luma_horz.s | 130 mov r11,#1 191 vld1.u32 {d0},[r12],r11 @vector load pu1_src 192 vld1.u32 {d1},[r12],r11 193 vld1.u32 {d2},[r12],r11 194 vld1.u32 {d3},[r12],r11 215 vld1.u32 {d4},[r12],r11 217 vld1.u32 {d5},[r12],r11 219 vld1.u32 {d6},[r12],r11 221 vld1.u32 {d7},[r12],r11 223 vld1.u32 {d12},[r4],r11 [all...] |
/external/tremolo/Tremolo/ |
H A D | mdctLARM.s | 186 STMFD r13!,{r4,r6-r11,r14} 198 LDRB r11,[r9],#1 @ r11= *wL++ 202 MUL r11,r12,r11 @ r11 = *l * *wL++ 204 MLA r6, r7, r6, r11 @ r6 = *--r * *--wR 215 LDMFD r13!,{r4,r6-r11,PC} 226 STMFD r13!,{r4,r6-r11,r14} 237 LDRB r11,[r [all...] |
H A D | bitwiseARM.s | 45 STMFD r13!,{r10,r11,r14} 56 LDRLT r11,[r3,#4]! @ r11= ptr[1] 60 ORRLT r10,r10,r11,LSL r14 @ r10= Next 32 bits. 64 LDMFD r13!,{r10,r11,PC} 80 MOV r11,#1 83 RSB r11,r11,r11,LSL r5 @ r11 [all...] |
H A D | mdctARM.s | 188 STMFD r13!,{r4,r6-r11,r14} 198 LDR r11,[r9],#4 @ r11= *wL++ 203 SMULL r14,r11,r12,r11 @ (r14,r11) = *l * *wL++ 205 ADD r6, r6, r11 216 LDMFD r13!,{r4,r6-r11,PC} 227 STMFD r13!,{r4,r6-r11,r14} 237 LDR r11,[r [all...] |
/external/boringssl/linux-arm/crypto/sha/ |
H A D | sha1-armv4-large.S | 23 stmdb sp!,{r4,r5,r6,r7,r8,r9,r10,r11,r12,lr} 37 ldrb r11,[r1,#1] 42 orr r9,r9,r11,lsl#16 62 ldrb r11,[r1,#1] 67 orr r9,r9,r11,lsl#16 87 ldrb r11,[r1,#1] 92 orr r9,r9,r11,lsl#16 112 ldrb r11,[r1,#1] 117 orr r9,r9,r11,lsl#16 137 ldrb r11,[r [all...] |
/external/llvm/test/MC/ARM/ |
H A D | eh-directive-movsp-diagnostics.s | 22 .setfp r11, sp, #8 23 add r11, sp, #8 25 mov r7, r11 66 .movsp r11, 67 mov sp, r11 71 @ CHECK: .movsp r11, 80 .movsp r11, #constant 81 mov sp, r11 85 @ CHECK: .movsp r11, #constant
|
H A D | 2010-11-30-reloc-movt.s | 21 push {r11, lr} 25 pop {r11, pc}
|
H A D | eh-compact-pr0.s | 16 .save {r11, lr} 17 push {r11, lr} 18 .setfp r11, sp 19 mov r11, sp 20 pop {r11, lr} 30 .save {r11, lr} 31 push {r11, lr} 32 pop {r11, pc} 56 @ 0x9B = $sp can be found in $r11 57 @ 0x8480 = pop {r11, r1 [all...] |
H A D | eh-compact-pr1.s | 14 .save {r4, r5, r11, lr} 15 push {r4, r5, r11, lr} 17 .setfp r11, sp, #8 18 add r11, sp, #8 19 pop {r4, r5, r11, pc} 43 @ 0x9B = $sp can be found in $r11 45 @ 0x8483 = pop {r4, r5, r11, r14}
|
/external/strace/linux/avr32/ |
H A D | get_syscall_args.c | 2 tcp->u_arg[1] = avr32_regs.r11;
|
/external/boringssl/win-x86_64/crypto/bn/ |
H A D | x86_64-mont5.asm | 44 lea r11,[2+r9] 45 neg r11 46 lea rsp,[r11*8+rsp] 52 mov r11,r10 54 and r11,7 58 lea r12,[96+r11*8+r12] 102 mov r11,rdx 122 add r13,r11 123 mov r11,r10 130 add r11,ra [all...] |
/external/boringssl/linux-x86_64/crypto/bn/ |
H A D | x86_64-mont5.S | 27 leaq 2(%r9),%r11 28 negq %r11 29 leaq (%rsp,%r11,8),%rsp 35 movq %r10,%r11 37 andq $7,%r11 41 leaq 96(%r12,%r11,8),%r12 85 movq %rdx,%r11 105 addq %r11,%r13 106 movq %r10,%r11 113 addq %rax,%r11 [all...] |
H A D | x86_64-mont.S | 33 movq %rsp,%r11 38 movq %r11,8(%rsp,%r9,8) 54 movq %rdx,%r11 70 addq %r11,%r13 71 movq %r10,%r11 78 addq %rax,%r11 91 addq %r11,%r13 95 movq %r10,%r11 98 addq %r11,%r13 117 movq %rdx,%r11 [all...] |
/external/boringssl/mac-x86_64/crypto/bn/ |
H A D | x86_64-mont5.S | 26 leaq 2(%r9),%r11 27 negq %r11 28 leaq (%rsp,%r11,8),%rsp 34 movq %r10,%r11 36 andq $7,%r11 40 leaq 96(%r12,%r11,8),%r12 84 movq %rdx,%r11 104 addq %r11,%r13 105 movq %r10,%r11 112 addq %rax,%r11 [all...] |
H A D | x86_64-mont.S | 32 movq %rsp,%r11 37 movq %r11,8(%rsp,%r9,8) 53 movq %rdx,%r11 69 addq %r11,%r13 70 movq %r10,%r11 77 addq %rax,%r11 90 addq %r11,%r13 94 movq %r10,%r11 97 addq %r11,%r13 116 movq %rdx,%r11 [all...] |
/external/libvpx/libvpx/vp8/common/arm/armv6/ |
H A D | loopfilter_v6.asm | 64 stmdb sp!, {r4 - r11, lr} 75 ldr r11, [src], pstep ; p1 92 uqsub8 r8, r10, r11 ; p2 - p1 93 uqsub8 r10, r11, r10 ; p1 - p2 99 uqsub8 r6, r11, r12 ; p1 - p0 101 uqsub8 r7, r12, r11 ; p0 - p1 109 uqsub8 r6, r11, r10 ; p1 - q1 110 uqsub8 r7, r10, r11 ; q1 - p1 111 uqsub8 r11, r12, r9 ; p0 - q0 115 orr r12, r11, r1 [all...] |
H A D | idct_v6.asm | 26 stmdb sp!, {r4-r11, lr} 47 smulbt r11, r5, r12 ; (ip[13] * cospi8sqrt2minus1) >> 16 58 pkhtb r9, r11, r9, asr #16 ; 13c | 12c 59 ldr r11, [r0] ; i1 | i0 65 uadd16 r10, r11, r14 ; a 66 usub16 r8, r11, r14 ; b 96 pkhbt r11, r6, r0, lsl #16 ; i0 | i4 103 uadd16 r10, r11, r9 ; a 104 usub16 r9, r11, r9 ; b 110 smulwt r11, r [all...] |
H A D | dequant_idct_v6.asm | 21 stmdb sp!, {r4-r11, lr} 70 smulwt r11, r3, r12 77 pkhbt r9, r9, r11, lsl #16 78 ldr r11, [r0], #4 83 uadd16 r10, r11, r14 84 usub16 r8, r11, r14 107 pkhbt r11, r8, r6, lsl #16 113 uadd16 r10, r11, lr 114 usub16 lr, r11, lr 119 smulwt r11, r [all...] |
/external/webrtc/src/modules/audio_coding/codecs/isac/fix/source/ |
H A D | lattice_armv7.S | 26 @ r6, r7, r8, r10, r11: scratch 37 .save {r4-r11} 38 push {r4-r11} 59 smlabb r11, r7, r5, r12 @ sth_Q15[k] * tmpAR + 16384 62 smlabb r11, r6, r8, r11 @ cth_Q15[k]*ar_g_Q0[k]+(sth_Q15[k]*tmpAR+16384) 65 ssat r11, #16, r11, asr #15 67 strh r11, [r0], #-2 @ Output: ar_g_Q0[k+1] 78 pop {r4-r11} [all...] |