/external/llvm/test/MC/X86/ |
H A D | x86_64-bmi-encoding.s | 7 // CHECK: blsmskq %r11, %r10 9 blsmskq %r11, %r10 15 // CHECK: blsmskq (%rax), %r10 17 blsmskq (%rax), %r10 23 // CHECK: blsiq %r11, %r10 25 blsiq %r11, %r10 31 // CHECK: blsiq (%rax), %r10 33 blsiq (%rax), %r10 39 // CHECK: blsrq %r11, %r10 41 blsrq %r11, %r10 [all...] |
/external/openssl/crypto/sha/asm/ |
H A D | sha1-armv4-large.s | 22 ldrb r10,[r1,#2] 27 orr r9,r9,r10,lsl#8 28 eor r10,r5,r6 @ F_xx_xx 35 eor r10,r5,r6 @ F_xx_xx 41 and r10,r4,r10,ror#2 43 eor r10,r10,r6,ror#2 @ F_00_19(B,C,D) 45 add r7,r7,r10 @ E+=F_00_19(B,C,D) 47 ldrb r10,[r [all...] |
H A D | sha512-ia64.pl | 150 { .mib; add r10=2*$SZ,ctx 159 { .mmi; $LDW C_=[r10],4*$SZ 165 { .mmi; $LDW G_=[r10] 201 add r10=2-$SZ,input 204 ld1 r10=[r10] 211 dep r11=r10,r11,8,8 };; 221 { .mib; and r10=B,C 228 add r10=3*$SZ,input };; 233 $LDW X[12]=[r10], [all...] |
/external/valgrind/main/coregrind/m_mach/ |
H A D | mach_traps-amd64-darwin.S | 40 movq %rcx, %r10 51 // movq %rcx, %r10 61 movq %rcx, %r10 71 movq %rcx, %r10 81 movq %rcx, %r10 91 movq %rcx, %r10 101 movq %rcx, %r10 111 movq %rcx, %r10 121 movq %rcx, %r10 131 movq %rcx, %r10 [all...] |
/external/openssl/crypto/bn/asm/ |
H A D | ppc.pl | 233 #.set r10,10 275 # Freely use registers r5,r6,r7,r8,r9,r10,r11 as follows: 279 # r9,r10, r11 are the equivalents of c1,c2, c3. 289 $UMULH r10,r5,r5 #in first iteration. No need 305 addc r10,r7,r10 # now add to temp result. 309 $ST r10,`1*$BNSZ`(r3) #r[1]=c2; 315 addze r10,r0 323 addze r10,r10 [all...] |
/external/libvpx/vp8/common/arm/armv6/ |
H A D | loopfilter_v6.asm | 76 ldr r10, [src], pstep ; p2 89 uqsub8 r6, r9, r10 ; p3 - p2 90 uqsub8 r7, r10, r9 ; p2 - p3 91 uqsub8 r8, r10, r11 ; p2 - p1 92 uqsub8 r10, r11, r10 ; p1 - p2 95 orr r8, r8, r10 ; abs (p2-p1) 102 ldr r10, [src], pstep ; q1 108 uqsub8 r6, r11, r10 ; p1 - q1 109 uqsub8 r7, r10, r1 [all...] |
H A D | idct_v6.asm | 12 ; r0 r1 r2 r3 r4 r5 r6 r7 r8 r9 r10 r11 r12 r14 71 smulwb r10, r5, r12 ; ([4] * sinpi8sqrt2) >> 16 1, r10 un 2, r12/r5 ^1 t1 76 rsb r11, r11, r10 ; c1 = temp1 - temp2 1 c1 78 smulwb r10, r4, r12 ; ([4] * cospi8sqrt2minus1) >> 16 1, r10 un 2, r12/r4 ^1 t1 82 add r10, r12, r10 ; temp1 1 83 add r3, r10, r3 ; d1 = temp1 + temp2 1 d1 84 add r10, r [all...] |
H A D | iwalsh_v6.asm | 34 qadd16 r10, r2, r8 ; a1 [1+13 | 0+12] 39 qadd16 r2, r10, r11 ; a1 + b1 [1 | 0] 41 qsub16 r6, r10, r11 ; a1 - b1 [9 | 8] 44 qadd16 r10, r3, r9 ; a1 [3+15 | 2+14] 49 qadd16 r3, r10, r11 ; a1 + b1 [3 | 2] 51 qsub16 r7, r10, r11 ; a1 - b1 [11 | 10] 56 qsubaddx r10, r2, r3 ; [c1|a1] [1-2 | 0+3] 61 qaddsubx r2, r10, r11 ; [b2|c2] [c1+d1 | a1-b1] 62 qaddsubx r3, r11, r10 ; [a2|d2] [b1+a1 | d1-c1] 63 ldr r10, c0x0003000 [all...] |
H A D | sixtappredict8x4_v6.asm | 54 ldrb r10, [r0, #-1] 62 pkhbt r9, r9, r10, lsl #16 ; r10 | r9 74 pkhbt r10, r10, r6, lsl #16 ; r10 | r9 75 pkhbt r6, r6, r7, lsl #16 ; r11 | r10 76 smlad r11, r10, r5, r11 ; vp8_filter[4], vp8_filter[5] 95 movne r8, r10 97 movne r10, r1 [all...] |
/external/tremolo/Tremolo/ |
H A D | bitwiseARM.s | 45 STMFD r13!,{r10,r11,r14} 53 LDR r10,[r3] @ r10= ptr[0] 57 MOV r10,r10,LSR r14 @ r10= ptr[0]>>(32-bitsLeftInWord) 60 ORRLT r10,r10,r11,LSL r14 @ r10= Next 32 bits. 63 AND r0,r10,r1 [all...] |
H A D | mdctLARM.s | 187 LDMFD r12,{r8,r9,r10} @ r8 = step 189 @ r10= wR 197 LDRB r6, [r10,#-1]! @ r6 = *--wR 227 LDMFD r12,{r8,r9,r10} @ r8 = step 229 @ r10= wR 238 LDRB r6, [r10,#-1]! @ r6 = *--wR 322 LDRB r10,[r5],r2 @ r10= T[0] T += step 327 MUL r9, r6, r10 @ r9 = s0*T[0] 331 MUL r12,r7, r10 [all...] |
H A D | mdctARM.s | 187 LDMFD r12,{r8,r9,r10} @ r8 = step 189 @ r10= wR 198 LDR r6, [r10,#-4]! @ r6 = *--wR 226 LDMFD r12,{r8,r9,r10} @ r8 = step 228 @ r10= wR 237 LDR r6, [r10,#-4]! @ r6 = *--wR 319 LDR r10,[r5],r2,LSL #2 @ r10= T[0] T += step 325 SMLAL r8, r9, r6, r10 @ (r8, r9) += s0*T[0] 328 SMULL r8, r12,r7, r10 [all...] |
/external/libffi/src/x86/ |
H A D | unix64.S | 48 movq (%rsp), %r10 /* Load return address. */ 53 movq %r10, 24(%rax) /* Relocate return address. */ 56 movq %rdi, %r10 /* Save a copy of the register area. */ 61 movq (%r10), %rdi 62 movq 8(%r10), %rsi 63 movq 16(%r10), %rdx 64 movq 24(%r10), %rcx 65 movq 32(%r10), %r8 66 movq 40(%r10), %r9 72 leaq 176(%r10), [all...] |
H A D | darwin64.S | 48 movq (%rsp), %r10 /* Load return address. */ 53 movq %r10, 24(%rax) /* Relocate return address. */ 56 movq %rdi, %r10 /* Save a copy of the register area. */ 61 movq (%r10), %rdi 62 movq 8(%r10), %rsi 63 movq 16(%r10), %rdx 64 movq 24(%r10), %rcx 65 movq 32(%r10), %r8 66 movq 40(%r10), %r9 72 leaq 176(%r10), [all...] |
/external/libvpx/vp8/decoder/arm/armv6/ |
H A D | dequant_idct_v6.asm | 70 smulwt r10, r4, r6 74 pkhbt r8, r8, r10, lsl #16 78 smulwb r10, r4, r12 82 pkhbt r10, r10, r7, lsl #16 85 uadd16 r6, r6, r10 86 uadd16 r10, r11, r14 88 uadd16 r9, r10, r6 89 usub16 r10, r10, r [all...] |
H A D | dequant_dc_idct_v6.asm | 92 smulwt r10, r4, r6 96 pkhbt r8, r8, r10, lsl #16 100 smulwb r10, r4, r12 104 pkhbt r10, r10, r7, lsl #16 107 uadd16 r6, r6, r10 108 uadd16 r10, r11, r14 110 uadd16 r9, r10, r6 111 usub16 r10, r10, r [all...] |
/external/libvpx/vp8/common/ppc/ |
H A D | recon_altivec.asm | 65 stvx v2, 0, r10 ;# 2 rows to dst from buf 66 lwz r0, 0(r10) 72 lwz r0, 4(r10) 74 lwz r0, 8(r10) 76 lwz r0, 12(r10) 95 la r10, -48(r1) ;# buf 110 stw r0, 0(r10) 112 stw r0, 4(r10) 114 stw r0, 8(r10) 116 stw r0, 12(r10) [all...] |
/external/libvpx/vp8/encoder/arm/armv6/ |
H A D | vp8_subtract_armv6.asm | 94 uxtb16 r10, r6, ror #8 ; [s3 | s1] (A) 98 usub16 r7, r10, r11 ; [d3 | d1] (A) 100 ldr r10, [r1, #4] ; src (B) 107 uxtb16 r8, r10 ; [s2 | s0] (B) 111 uxtb16 r10, r10, ror #8 ; [s3 | s1] (B) 115 usub16 r7, r10, r11 ; [d3 | d1] (B) 137 uxtb16 r10, r6, ror #8 ; [s3 | s1] (A) 141 usub16 r7, r10, r11 ; [d3 | d1] (A) 143 ldr r10, [r [all...] |
H A D | vp8_variance8x8_armv6.asm | 25 push {r4-r10, lr} 38 sel r10, r8, lr ; select bytes with positive difference 43 usad8 r6, r10, lr ; calculate sum of positive differences 45 orr r8, r8, r10 ; differences of all 4 pixels 52 uxtb16 r10, r8, ror #8 ; another two pixels to halfwords 58 smlad r5, r10, r10, r5 ; dual signed multiply, add and accumulate (2) 62 sel r10, r8, lr ; select bytes with positive difference 68 usad8 r6, r10, lr ; calculate sum of positive differences 70 orr r8, r8, r10 ; difference [all...] |
/external/libvpx/vp8/encoder/arm/armv5te/ |
H A D | vp8_packtokens_armv5.asm | 60 ldr r10, [sp, #52] ; vp8_coef_tree 81 ldrsb lr, [r10, lr] ; i = vp8_coef_tree[i+bb] 106 mov r10, #0 107 strb r10, [r7, r4] ; w->buffer[x] =(unsigned char)0 117 ldrb r10, [r7, r4] ; w->buffer[x] 118 add r10, r10, #1 119 strb r10, [r7, r4] ; w->buffer[x] + 1 122 ldr r10, [r0, #vp8_writer_buffer] 131 strb r7, [r10, r [all...] |
H A D | vp8_packtokens_partitions_armv5.asm | 66 ldr r10, [sp, #40] ; ptr 81 str r10, [r0, #vp8_writer_buffer] 112 ldr r10, [sp, #88] ; vp8_coef_tree 133 ldrsb lr, [r10, lr] ; i = vp8_coef_tree[i+bb] 158 mov r10, #0 159 strb r10, [r7, r4] ; w->buffer[x] =(unsigned char)0 169 ldrb r10, [r7, r4] ; w->buffer[x] 170 add r10, r10, #1 171 strb r10, [r [all...] |
/external/clang/test/CXX/except/except.spec/ |
H A D | p3.cpp | 45 extern void (*r10)() noexcept; 46 extern void (*r10)() noexcept;
|
/external/libvpx/vpx_scale/arm/neon/ |
H A D | vp8_vpxyv12_copyframe_func_neon.asm | 32 ldr r10, [r0, #yv12_buffer_config_v_buffer] ;srcptr1 44 str r10, [sp, #8] 53 add r10, r2, r6 59 vld1.8 {q8, q9}, [r10]! 61 vld1.8 {q10, q11}, [r10]! 63 vld1.8 {q12, q13}, [r10]! 65 vld1.8 {q14, q15}, [r10]! 86 ands r10, r5, #0x7f ;check to see if extra copy is needed 87 sub r11, r5, r10 111 add r10, r [all...] |
/external/valgrind/main/none/tests/amd64/ |
H A D | sbbmisc.stdout.exp | 10 r10 = -267879790 -267879789
|
/external/valgrind/main/none/tests/x86/ |
H A D | sbbmisc.stdout.exp | 10 r10 = -267879790 -267879789
|