Searched refs:r10 (Results 1 - 25 of 167) sorted by relevance

1234567

/external/llvm/test/MC/X86/
H A Dx86_64-bmi-encoding.s7 // CHECK: blsmskq %r11, %r10
9 blsmskq %r11, %r10
15 // CHECK: blsmskq (%rax), %r10
17 blsmskq (%rax), %r10
23 // CHECK: blsiq %r11, %r10
25 blsiq %r11, %r10
31 // CHECK: blsiq (%rax), %r10
33 blsiq (%rax), %r10
39 // CHECK: blsrq %r11, %r10
41 blsrq %r11, %r10
[all...]
/external/openssl/crypto/sha/asm/
H A Dsha1-armv4-large.s22 ldrb r10,[r1,#2]
27 orr r9,r9,r10,lsl#8
28 eor r10,r5,r6 @ F_xx_xx
35 eor r10,r5,r6 @ F_xx_xx
41 and r10,r4,r10,ror#2
43 eor r10,r10,r6,ror#2 @ F_00_19(B,C,D)
45 add r7,r7,r10 @ E+=F_00_19(B,C,D)
47 ldrb r10,[r
[all...]
H A Dsha512-ia64.pl150 { .mib; add r10=2*$SZ,ctx
159 { .mmi; $LDW C_=[r10],4*$SZ
165 { .mmi; $LDW G_=[r10]
201 add r10=2-$SZ,input
204 ld1 r10=[r10]
211 dep r11=r10,r11,8,8 };;
221 { .mib; and r10=B,C
228 add r10=3*$SZ,input };;
233 $LDW X[12]=[r10],
[all...]
/external/valgrind/main/coregrind/m_mach/
H A Dmach_traps-amd64-darwin.S40 movq %rcx, %r10
51 // movq %rcx, %r10
61 movq %rcx, %r10
71 movq %rcx, %r10
81 movq %rcx, %r10
91 movq %rcx, %r10
101 movq %rcx, %r10
111 movq %rcx, %r10
121 movq %rcx, %r10
131 movq %rcx, %r10
[all...]
/external/openssl/crypto/bn/asm/
H A Dppc.pl233 #.set r10,10
275 # Freely use registers r5,r6,r7,r8,r9,r10,r11 as follows:
279 # r9,r10, r11 are the equivalents of c1,c2, c3.
289 $UMULH r10,r5,r5 #in first iteration. No need
305 addc r10,r7,r10 # now add to temp result.
309 $ST r10,`1*$BNSZ`(r3) #r[1]=c2;
315 addze r10,r0
323 addze r10,r10
[all...]
/external/libvpx/vp8/common/arm/armv6/
H A Dloopfilter_v6.asm76 ldr r10, [src], pstep ; p2
89 uqsub8 r6, r9, r10 ; p3 - p2
90 uqsub8 r7, r10, r9 ; p2 - p3
91 uqsub8 r8, r10, r11 ; p2 - p1
92 uqsub8 r10, r11, r10 ; p1 - p2
95 orr r8, r8, r10 ; abs (p2-p1)
102 ldr r10, [src], pstep ; q1
108 uqsub8 r6, r11, r10 ; p1 - q1
109 uqsub8 r7, r10, r1
[all...]
H A Didct_v6.asm12 ; r0 r1 r2 r3 r4 r5 r6 r7 r8 r9 r10 r11 r12 r14
71 smulwb r10, r5, r12 ; ([4] * sinpi8sqrt2) >> 16 1, r10 un 2, r12/r5 ^1 t1
76 rsb r11, r11, r10 ; c1 = temp1 - temp2 1 c1
78 smulwb r10, r4, r12 ; ([4] * cospi8sqrt2minus1) >> 16 1, r10 un 2, r12/r4 ^1 t1
82 add r10, r12, r10 ; temp1 1
83 add r3, r10, r3 ; d1 = temp1 + temp2 1 d1
84 add r10, r
[all...]
H A Diwalsh_v6.asm34 qadd16 r10, r2, r8 ; a1 [1+13 | 0+12]
39 qadd16 r2, r10, r11 ; a1 + b1 [1 | 0]
41 qsub16 r6, r10, r11 ; a1 - b1 [9 | 8]
44 qadd16 r10, r3, r9 ; a1 [3+15 | 2+14]
49 qadd16 r3, r10, r11 ; a1 + b1 [3 | 2]
51 qsub16 r7, r10, r11 ; a1 - b1 [11 | 10]
56 qsubaddx r10, r2, r3 ; [c1|a1] [1-2 | 0+3]
61 qaddsubx r2, r10, r11 ; [b2|c2] [c1+d1 | a1-b1]
62 qaddsubx r3, r11, r10 ; [a2|d2] [b1+a1 | d1-c1]
63 ldr r10, c0x0003000
[all...]
H A Dsixtappredict8x4_v6.asm54 ldrb r10, [r0, #-1]
62 pkhbt r9, r9, r10, lsl #16 ; r10 | r9
74 pkhbt r10, r10, r6, lsl #16 ; r10 | r9
75 pkhbt r6, r6, r7, lsl #16 ; r11 | r10
76 smlad r11, r10, r5, r11 ; vp8_filter[4], vp8_filter[5]
95 movne r8, r10
97 movne r10, r1
[all...]
/external/tremolo/Tremolo/
H A DbitwiseARM.s45 STMFD r13!,{r10,r11,r14}
53 LDR r10,[r3] @ r10= ptr[0]
57 MOV r10,r10,LSR r14 @ r10= ptr[0]>>(32-bitsLeftInWord)
60 ORRLT r10,r10,r11,LSL r14 @ r10= Next 32 bits.
63 AND r0,r10,r1
[all...]
H A DmdctLARM.s187 LDMFD r12,{r8,r9,r10} @ r8 = step
189 @ r10= wR
197 LDRB r6, [r10,#-1]! @ r6 = *--wR
227 LDMFD r12,{r8,r9,r10} @ r8 = step
229 @ r10= wR
238 LDRB r6, [r10,#-1]! @ r6 = *--wR
322 LDRB r10,[r5],r2 @ r10= T[0] T += step
327 MUL r9, r6, r10 @ r9 = s0*T[0]
331 MUL r12,r7, r10
[all...]
H A DmdctARM.s187 LDMFD r12,{r8,r9,r10} @ r8 = step
189 @ r10= wR
198 LDR r6, [r10,#-4]! @ r6 = *--wR
226 LDMFD r12,{r8,r9,r10} @ r8 = step
228 @ r10= wR
237 LDR r6, [r10,#-4]! @ r6 = *--wR
319 LDR r10,[r5],r2,LSL #2 @ r10= T[0] T += step
325 SMLAL r8, r9, r6, r10 @ (r8, r9) += s0*T[0]
328 SMULL r8, r12,r7, r10
[all...]
/external/libffi/src/x86/
H A Dunix64.S48 movq (%rsp), %r10 /* Load return address. */
53 movq %r10, 24(%rax) /* Relocate return address. */
56 movq %rdi, %r10 /* Save a copy of the register area. */
61 movq (%r10), %rdi
62 movq 8(%r10), %rsi
63 movq 16(%r10), %rdx
64 movq 24(%r10), %rcx
65 movq 32(%r10), %r8
66 movq 40(%r10), %r9
72 leaq 176(%r10),
[all...]
H A Ddarwin64.S48 movq (%rsp), %r10 /* Load return address. */
53 movq %r10, 24(%rax) /* Relocate return address. */
56 movq %rdi, %r10 /* Save a copy of the register area. */
61 movq (%r10), %rdi
62 movq 8(%r10), %rsi
63 movq 16(%r10), %rdx
64 movq 24(%r10), %rcx
65 movq 32(%r10), %r8
66 movq 40(%r10), %r9
72 leaq 176(%r10),
[all...]
/external/libvpx/vp8/decoder/arm/armv6/
H A Ddequant_idct_v6.asm70 smulwt r10, r4, r6
74 pkhbt r8, r8, r10, lsl #16
78 smulwb r10, r4, r12
82 pkhbt r10, r10, r7, lsl #16
85 uadd16 r6, r6, r10
86 uadd16 r10, r11, r14
88 uadd16 r9, r10, r6
89 usub16 r10, r10, r
[all...]
H A Ddequant_dc_idct_v6.asm92 smulwt r10, r4, r6
96 pkhbt r8, r8, r10, lsl #16
100 smulwb r10, r4, r12
104 pkhbt r10, r10, r7, lsl #16
107 uadd16 r6, r6, r10
108 uadd16 r10, r11, r14
110 uadd16 r9, r10, r6
111 usub16 r10, r10, r
[all...]
/external/libvpx/vp8/common/ppc/
H A Drecon_altivec.asm65 stvx v2, 0, r10 ;# 2 rows to dst from buf
66 lwz r0, 0(r10)
72 lwz r0, 4(r10)
74 lwz r0, 8(r10)
76 lwz r0, 12(r10)
95 la r10, -48(r1) ;# buf
110 stw r0, 0(r10)
112 stw r0, 4(r10)
114 stw r0, 8(r10)
116 stw r0, 12(r10)
[all...]
/external/libvpx/vp8/encoder/arm/armv6/
H A Dvp8_subtract_armv6.asm94 uxtb16 r10, r6, ror #8 ; [s3 | s1] (A)
98 usub16 r7, r10, r11 ; [d3 | d1] (A)
100 ldr r10, [r1, #4] ; src (B)
107 uxtb16 r8, r10 ; [s2 | s0] (B)
111 uxtb16 r10, r10, ror #8 ; [s3 | s1] (B)
115 usub16 r7, r10, r11 ; [d3 | d1] (B)
137 uxtb16 r10, r6, ror #8 ; [s3 | s1] (A)
141 usub16 r7, r10, r11 ; [d3 | d1] (A)
143 ldr r10, [r
[all...]
H A Dvp8_variance8x8_armv6.asm25 push {r4-r10, lr}
38 sel r10, r8, lr ; select bytes with positive difference
43 usad8 r6, r10, lr ; calculate sum of positive differences
45 orr r8, r8, r10 ; differences of all 4 pixels
52 uxtb16 r10, r8, ror #8 ; another two pixels to halfwords
58 smlad r5, r10, r10, r5 ; dual signed multiply, add and accumulate (2)
62 sel r10, r8, lr ; select bytes with positive difference
68 usad8 r6, r10, lr ; calculate sum of positive differences
70 orr r8, r8, r10 ; difference
[all...]
/external/libvpx/vp8/encoder/arm/armv5te/
H A Dvp8_packtokens_armv5.asm60 ldr r10, [sp, #52] ; vp8_coef_tree
81 ldrsb lr, [r10, lr] ; i = vp8_coef_tree[i+bb]
106 mov r10, #0
107 strb r10, [r7, r4] ; w->buffer[x] =(unsigned char)0
117 ldrb r10, [r7, r4] ; w->buffer[x]
118 add r10, r10, #1
119 strb r10, [r7, r4] ; w->buffer[x] + 1
122 ldr r10, [r0, #vp8_writer_buffer]
131 strb r7, [r10, r
[all...]
H A Dvp8_packtokens_partitions_armv5.asm66 ldr r10, [sp, #40] ; ptr
81 str r10, [r0, #vp8_writer_buffer]
112 ldr r10, [sp, #88] ; vp8_coef_tree
133 ldrsb lr, [r10, lr] ; i = vp8_coef_tree[i+bb]
158 mov r10, #0
159 strb r10, [r7, r4] ; w->buffer[x] =(unsigned char)0
169 ldrb r10, [r7, r4] ; w->buffer[x]
170 add r10, r10, #1
171 strb r10, [r
[all...]
/external/clang/test/CXX/except/except.spec/
H A Dp3.cpp45 extern void (*r10)() noexcept;
46 extern void (*r10)() noexcept;
/external/libvpx/vpx_scale/arm/neon/
H A Dvp8_vpxyv12_copyframe_func_neon.asm32 ldr r10, [r0, #yv12_buffer_config_v_buffer] ;srcptr1
44 str r10, [sp, #8]
53 add r10, r2, r6
59 vld1.8 {q8, q9}, [r10]!
61 vld1.8 {q10, q11}, [r10]!
63 vld1.8 {q12, q13}, [r10]!
65 vld1.8 {q14, q15}, [r10]!
86 ands r10, r5, #0x7f ;check to see if extra copy is needed
87 sub r11, r5, r10
111 add r10, r
[all...]
/external/valgrind/main/none/tests/amd64/
H A Dsbbmisc.stdout.exp10 r10 = -267879790 -267879789
/external/valgrind/main/none/tests/x86/
H A Dsbbmisc.stdout.exp10 r10 = -267879790 -267879789

Completed in 1428 milliseconds

1234567