Searched refs:r9 (Results 176 - 200 of 508) sorted by relevance

1234567891011>>

/external/chromium_org/third_party/libvpx/source/libvpx/vp8/encoder/ppc/
H A Dfdct_altivec.asm36 load_c v0, dct_tab, 0, r9, r10
42 load_c v4, ppc_dctperm_tab, 0, r9, r10
43 load_c v5, ppc_dctperm_tab, r6, r9, r10
45 load_c v6, round_tab, 0, r10, r9
125 lvx v6, r6, r9 ;# v6 = Vround
154 lvx v6, r6, r9 ;# v6 = Vround
165 lvx v6, 0, r9 ;# v6 = Hround
176 lvx v6, r6, r9 ;# v6 = Vround
/external/chromium_org/v8/src/compiler/arm/
H A Dlinkage-arm.cc25 return r4.bit() | r5.bit() | r6.bit() | r7.bit() | r8.bit() | r9.bit() |
/external/libvpx/libvpx/vp8/common/arm/armv6/
H A Dvp8_variance16x16_armv6.asm46 usub8 r9, r5, r4 ; calculate difference with reversed operands
48 sel r6, r9, lr ; select bytes with negative difference
70 usub8 r9, r5, r4 ; calculate difference with reversed operands
71 sel r6, r9, lr ; select bytes with negative difference
94 usub8 r9, r5, r4 ; calculate difference with reversed operands
95 sel r6, r9, lr ; select bytes with negative difference
119 usub8 r9, r5, r4 ; calculate difference with reversed operands
121 sel r6, r9, lr ; select bytes with negative difference
H A Dvp8_variance_halfpixvar16x16_hv_armv6.asm38 add r9, r0, r1 ; pointer to pixels on the next row
42 ldr r5, [r9, #0] ; load source pixels c, row N+1
43 ldr r7, [r9, #1] ; load source pixels d, row N+1
82 ldr r5, [r9, #4] ; load source pixels c, row N+1
86 ldr r7, [r9, #5] ; load source pixels d, row N+1
124 ldr r5, [r9, #8] ; load source pixels c, row N+1
128 ldr r7, [r9, #9] ; load source pixels d, row N+1
166 ldr r5, [r9, #12] ; load source pixels c, row N+1
168 ldr r7, [r9, #13] ; load source pixels d, row N+1
/external/libvpx/libvpx/vp8/encoder/ppc/
H A Dfdct_altivec.asm36 load_c v0, dct_tab, 0, r9, r10
42 load_c v4, ppc_dctperm_tab, 0, r9, r10
43 load_c v5, ppc_dctperm_tab, r6, r9, r10
45 load_c v6, round_tab, 0, r10, r9
125 lvx v6, r6, r9 ;# v6 = Vround
154 lvx v6, r6, r9 ;# v6 = Vround
165 lvx v6, 0, r9 ;# v6 = Hround
176 lvx v6, r6, r9 ;# v6 = Vround
/external/linux-tools-perf/perf-3.12.0/arch/hexagon/lib/
H A Dmemset.S48 r9 = sub(r6, r7) /* bytes until double alignment */ define
54 p0 = tstbit(r9, #0)
71 p0 = tstbit(r9, #1)
84 p0 = tstbit(r9, #2)
/external/llvm/lib/Target/X86/
H A DX86CompilationCallback_Win64.asm28 push r9
58 pop r9
/external/llvm/test/MC/ARM/
H A Dthumb.s13 blx r9
15 @ CHECK: blx r9 @ encoding: [0xc8,0x47]
H A Dthumb-diagnostics.s80 push {r8, r9}
85 @ CHECK-ERRORS: push {r8, r9}
91 stm r1!, {r2, r9}
92 stm r2!, {r2, r9}
98 @ CHECK-ERRORS: stm r1!, {r2, r9}
101 @ CHECK-ERRORS-V8: stm r2!, {r2, r9}
/external/openssl/crypto/
H A Dx86_64cpuid.S149 shlq $32,%r9
152 orq %r9,%rax
215 xorq %r9,%r9
/external/libhevc/common/arm/
H A Dihevc_weighted_pred_bi.s130 @ r9 => off1
160 ldr r9,[sp,#60] @load off1
161 add r5,r7,r9 @off0 + off1
172 lsl r9,r7,#1
252 rsb r11,r9,r3,lsl #2 @2*src_strd1 - wd
254 rsb r12,r9,r4,lsl #2 @2*src_strd2 - wd
256 asr r7,r9,#1
H A Dihevc_intra_pred_luma_dc.s117 mov r9, #0
118 vmov d17, r11, r9
134 sub r9, r7, #1 @&src[2nt-1]
139 ldrb r12, [r9]
218 @r8 and r9 (2nt+1+col 2nt-1-row)
224 sub r9, r9, #7 @&src[2nt-1-row]
226 vld1.s8 d1, [r9] @row 7::1 (0 also) load (prol)
227 sub r9, r9, #
[all...]
H A Dihevc_itrans_recon_16x16.s152 add r9,r0,r6, lsl #1 @ 2 rows
218 add r9,r9,#8
223 vld1.16 d11,[r9],r6
225 vld1.16 d7,[r9],r10
230 vld1.16 d5,[r9],r6
232 vld1.16 d9,[r9],r8
306 vld1.16 d11,[r9],r6
308 vld1.16 d7,[r9],r10
310 vld1.16 d5,[r9],r
[all...]
H A Dihevc_inter_pred_chroma_horz.s140 sub r9,r0,#2
153 mov r9,#10
167 vld1.u32 {q3},[r12],r9 @vector load pu1_src
177 vld1.u32 {q7},[r4],r9 @vector load pu1_src
232 vld1.u32 {q3},[r12],r9 @vector load pu1_src
251 vld1.u32 {q7},[r4],r9 @vector load pu1_src
314 vld1.u32 {q3},[r12],r9 @vector load pu1_src
327 vld1.u32 {q7},[r4],r9 @vector load pu1_src
446 mov r12,r9
495 add r9,r
[all...]
H A Dihevc_itrans_recon_8x8.s161 add r9,r0,r6, lsl #1 @ 2 rows
185 vld1.16 d3,[r9]!
188 vld1.16 d5,[r9],r5
191 vld1.16 d7,[r9]!
195 vld1.16 d9,[r9],r10
199 vld1.16 d11,[r9]!
203 vld1.16 d13,[r9],r5
207 vld1.16 d15,[r9]!
211 vld1.16 d17,[r9],r10
286 vld1.16 d3,[r9]!
[all...]
H A Dihevc_sao_edge_offset_class0_chroma.s60 @r9 => wd
76 LDR r9,[sp,#64] @Loads wd
79 ADD r11,r3,r9 @pu1_src_top[wd]
105 MOV r12,r9 @Move wd to r12 for loop count
116 CMP r9,#16 @Compare wd with 16
120 MOV r8,r9 @move wd to r8 for loop count
123 CMP r8,r9 @if(col == wd)
150 SUB r5,r9,r8 @wd - col
244 SUB r5,r9,r8 @II wd - col
285 AND r8,r9,#
[all...]
/external/chromium_org/third_party/libvpx/source/libvpx/vp8/common/ppc/
H A Didctllm_altivec.asm29 load_c v8, sinpi8sqrt2, 0, r9, r10
30 load_c v9, cospi8sqrt2minus1, 0, r9, r10
31 load_c v10, hi_hi, 0, r9, r10
32 load_c v11, lo_lo, 0, r9, r10
33 load_c v12, shift_16, 0, r9, r10
/external/libvpx/libvpx/vp8/common/ppc/
H A Didctllm_altivec.asm29 load_c v8, sinpi8sqrt2, 0, r9, r10
30 load_c v9, cospi8sqrt2minus1, 0, r9, r10
31 load_c v10, hi_hi, 0, r9, r10
32 load_c v11, lo_lo, 0, r9, r10
33 load_c v12, shift_16, 0, r9, r10
/external/chromium_org/v8/src/x64/
H A Dcodegen-x64.cc243 __ SmiToInteger32(r9, FieldOperand(r8, FixedDoubleArray::kLengthOffset));
266 // r9 : elements array length
295 __ leap(rdi, Operand(r9, times_8, FixedArray::kHeaderSize));
311 __ Integer32ToSmi(r11, r9);
330 FieldOperand(r8, r9, times_pointer_size, FixedArray::kHeaderSize));
331 // r9 : current element's index
336 __ movsd(FieldOperand(r14, r9, times_8, FixedDoubleArray::kHeaderSize),
346 __ movq(FieldOperand(r14, r9, times_8, FixedDoubleArray::kHeaderSize), r15);
348 __ decp(r9);
384 __ SmiToInteger32(r9, FieldOperan
[all...]
/external/chromium_org/third_party/boringssl/src/crypto/bn/asm/
H A Dx86_64-mont5.pl62 $num="%r9"; # int num,
857 mov %r9,%rcx
912 my $num ="%r9"; # int num, has to be divisible by 8
1579 mov 8*1($tptr),%r9
1605 add %rax,%r9
1608 add %r9,%r8
1610 mov %rdx,%r9
1611 adc \$0,%r9
1617 add %r10,%r9
1675 adc 8*1($tptr),%r9
[all...]
/external/flac/libFLAC/ppc/as/
H A Dlpc_asm.s62 addi r9,r1,-28
64 andc r9,r9,r31 ; for quadword-aligned stack data
85 stw r31,-4(r9)
86 lvewx v0,r10,r9
228 stw r7,-4(r9)
229 lvewx v19,r10,r9 ; v19: lp_quantization vector
311 addi r9,r1,-28
313 andc r9,r9,r3
[all...]
/external/flac/libFLAC/ppc/gas/
H A Dlpc_asm.s64 addi r9,r1,-28
66 andc r9,r9,r31 # for quadword-aligned stack data
87 stw r31,-4(r9)
88 lvewx v0,r10,r9
230 stw r7,-4(r9)
231 lvewx v19,r10,r9 # v19: lp_quantization vector
313 addi r9,r1,-28
315 andc r9,r9,r3
[all...]
/external/openssl/crypto/bn/asm/
H A Ds390x.S42 lg %r9,8(%r2,%r3)
44 alcgr %r9,%r6
46 alg %r9,8(%r2,%r1)
47 stg %r9,8(%r2,%r1)
56 lg %r9,24(%r2,%r3)
58 alcgr %r9,%r6
60 alg %r9,24(%r2,%r1)
61 stg %r9,24(%r2,%r1)
115 lg %r9,8(%r2,%r3)
117 alcgr %r9,
[all...]
/external/chromium_org/third_party/libvpx/source/libvpx/vp8/common/arm/armv6/
H A Dvp8_variance_halfpixvar16x16_hv_armv6.asm38 add r9, r0, r1 ; pointer to pixels on the next row
42 ldr r5, [r9, #0] ; load source pixels c, row N+1
43 ldr r7, [r9, #1] ; load source pixels d, row N+1
82 ldr r5, [r9, #4] ; load source pixels c, row N+1
86 ldr r7, [r9, #5] ; load source pixels d, row N+1
124 ldr r5, [r9, #8] ; load source pixels c, row N+1
128 ldr r7, [r9, #9] ; load source pixels d, row N+1
166 ldr r5, [r9, #12] ; load source pixels c, row N+1
168 ldr r7, [r9, #13] ; load source pixels d, row N+1
/external/chromium_org/third_party/boringssl/win-x86_64/crypto/aes/
H A Dvpaes-x86_64.asm22 mov r9,rdx
28 movdqu xmm5,XMMWORD PTR[r9]
35 add r9,16
58 add r9,16
88 movdqu xmm5,XMMWORD PTR[r9]
112 mov r9,rdx
119 movdqu xmm5,XMMWORD PTR[r9]
131 add r9,16
169 add r9,16
194 movdqu xmm0,XMMWORD PTR[r9]
[all...]

Completed in 1574 milliseconds

1234567891011>>