/external/chromium_org/third_party/libvpx/source/libvpx/vp8/encoder/ppc/ |
H A D | fdct_altivec.asm | 36 load_c v0, dct_tab, 0, r9, r10 42 load_c v4, ppc_dctperm_tab, 0, r9, r10 43 load_c v5, ppc_dctperm_tab, r6, r9, r10 45 load_c v6, round_tab, 0, r10, r9 125 lvx v6, r6, r9 ;# v6 = Vround 154 lvx v6, r6, r9 ;# v6 = Vround 165 lvx v6, 0, r9 ;# v6 = Hround 176 lvx v6, r6, r9 ;# v6 = Vround
|
/external/chromium_org/v8/src/compiler/arm/ |
H A D | linkage-arm.cc | 25 return r4.bit() | r5.bit() | r6.bit() | r7.bit() | r8.bit() | r9.bit() |
|
/external/libvpx/libvpx/vp8/common/arm/armv6/ |
H A D | vp8_variance16x16_armv6.asm | 46 usub8 r9, r5, r4 ; calculate difference with reversed operands 48 sel r6, r9, lr ; select bytes with negative difference 70 usub8 r9, r5, r4 ; calculate difference with reversed operands 71 sel r6, r9, lr ; select bytes with negative difference 94 usub8 r9, r5, r4 ; calculate difference with reversed operands 95 sel r6, r9, lr ; select bytes with negative difference 119 usub8 r9, r5, r4 ; calculate difference with reversed operands 121 sel r6, r9, lr ; select bytes with negative difference
|
H A D | vp8_variance_halfpixvar16x16_hv_armv6.asm | 38 add r9, r0, r1 ; pointer to pixels on the next row 42 ldr r5, [r9, #0] ; load source pixels c, row N+1 43 ldr r7, [r9, #1] ; load source pixels d, row N+1 82 ldr r5, [r9, #4] ; load source pixels c, row N+1 86 ldr r7, [r9, #5] ; load source pixels d, row N+1 124 ldr r5, [r9, #8] ; load source pixels c, row N+1 128 ldr r7, [r9, #9] ; load source pixels d, row N+1 166 ldr r5, [r9, #12] ; load source pixels c, row N+1 168 ldr r7, [r9, #13] ; load source pixels d, row N+1
|
/external/libvpx/libvpx/vp8/encoder/ppc/ |
H A D | fdct_altivec.asm | 36 load_c v0, dct_tab, 0, r9, r10 42 load_c v4, ppc_dctperm_tab, 0, r9, r10 43 load_c v5, ppc_dctperm_tab, r6, r9, r10 45 load_c v6, round_tab, 0, r10, r9 125 lvx v6, r6, r9 ;# v6 = Vround 154 lvx v6, r6, r9 ;# v6 = Vround 165 lvx v6, 0, r9 ;# v6 = Hround 176 lvx v6, r6, r9 ;# v6 = Vround
|
/external/linux-tools-perf/perf-3.12.0/arch/hexagon/lib/ |
H A D | memset.S | 48 r9 = sub(r6, r7) /* bytes until double alignment */ define 54 p0 = tstbit(r9, #0) 71 p0 = tstbit(r9, #1) 84 p0 = tstbit(r9, #2)
|
/external/llvm/lib/Target/X86/ |
H A D | X86CompilationCallback_Win64.asm | 28 push r9 58 pop r9
|
/external/llvm/test/MC/ARM/ |
H A D | thumb.s | 13 blx r9 15 @ CHECK: blx r9 @ encoding: [0xc8,0x47]
|
H A D | thumb-diagnostics.s | 80 push {r8, r9} 85 @ CHECK-ERRORS: push {r8, r9} 91 stm r1!, {r2, r9} 92 stm r2!, {r2, r9} 98 @ CHECK-ERRORS: stm r1!, {r2, r9} 101 @ CHECK-ERRORS-V8: stm r2!, {r2, r9}
|
/external/openssl/crypto/ |
H A D | x86_64cpuid.S | 149 shlq $32,%r9 152 orq %r9,%rax 215 xorq %r9,%r9
|
/external/libhevc/common/arm/ |
H A D | ihevc_weighted_pred_bi.s | 130 @ r9 => off1 160 ldr r9,[sp,#60] @load off1 161 add r5,r7,r9 @off0 + off1 172 lsl r9,r7,#1 252 rsb r11,r9,r3,lsl #2 @2*src_strd1 - wd 254 rsb r12,r9,r4,lsl #2 @2*src_strd2 - wd 256 asr r7,r9,#1
|
H A D | ihevc_intra_pred_luma_dc.s | 117 mov r9, #0 118 vmov d17, r11, r9 134 sub r9, r7, #1 @&src[2nt-1] 139 ldrb r12, [r9] 218 @r8 and r9 (2nt+1+col 2nt-1-row) 224 sub r9, r9, #7 @&src[2nt-1-row] 226 vld1.s8 d1, [r9] @row 7::1 (0 also) load (prol) 227 sub r9, r9, # [all...] |
H A D | ihevc_itrans_recon_16x16.s | 152 add r9,r0,r6, lsl #1 @ 2 rows 218 add r9,r9,#8 223 vld1.16 d11,[r9],r6 225 vld1.16 d7,[r9],r10 230 vld1.16 d5,[r9],r6 232 vld1.16 d9,[r9],r8 306 vld1.16 d11,[r9],r6 308 vld1.16 d7,[r9],r10 310 vld1.16 d5,[r9],r [all...] |
H A D | ihevc_inter_pred_chroma_horz.s | 140 sub r9,r0,#2 153 mov r9,#10 167 vld1.u32 {q3},[r12],r9 @vector load pu1_src 177 vld1.u32 {q7},[r4],r9 @vector load pu1_src 232 vld1.u32 {q3},[r12],r9 @vector load pu1_src 251 vld1.u32 {q7},[r4],r9 @vector load pu1_src 314 vld1.u32 {q3},[r12],r9 @vector load pu1_src 327 vld1.u32 {q7},[r4],r9 @vector load pu1_src 446 mov r12,r9 495 add r9,r [all...] |
H A D | ihevc_itrans_recon_8x8.s | 161 add r9,r0,r6, lsl #1 @ 2 rows 185 vld1.16 d3,[r9]! 188 vld1.16 d5,[r9],r5 191 vld1.16 d7,[r9]! 195 vld1.16 d9,[r9],r10 199 vld1.16 d11,[r9]! 203 vld1.16 d13,[r9],r5 207 vld1.16 d15,[r9]! 211 vld1.16 d17,[r9],r10 286 vld1.16 d3,[r9]! [all...] |
H A D | ihevc_sao_edge_offset_class0_chroma.s | 60 @r9 => wd 76 LDR r9,[sp,#64] @Loads wd 79 ADD r11,r3,r9 @pu1_src_top[wd] 105 MOV r12,r9 @Move wd to r12 for loop count 116 CMP r9,#16 @Compare wd with 16 120 MOV r8,r9 @move wd to r8 for loop count 123 CMP r8,r9 @if(col == wd) 150 SUB r5,r9,r8 @wd - col 244 SUB r5,r9,r8 @II wd - col 285 AND r8,r9,# [all...] |
/external/chromium_org/third_party/libvpx/source/libvpx/vp8/common/ppc/ |
H A D | idctllm_altivec.asm | 29 load_c v8, sinpi8sqrt2, 0, r9, r10 30 load_c v9, cospi8sqrt2minus1, 0, r9, r10 31 load_c v10, hi_hi, 0, r9, r10 32 load_c v11, lo_lo, 0, r9, r10 33 load_c v12, shift_16, 0, r9, r10
|
/external/libvpx/libvpx/vp8/common/ppc/ |
H A D | idctllm_altivec.asm | 29 load_c v8, sinpi8sqrt2, 0, r9, r10 30 load_c v9, cospi8sqrt2minus1, 0, r9, r10 31 load_c v10, hi_hi, 0, r9, r10 32 load_c v11, lo_lo, 0, r9, r10 33 load_c v12, shift_16, 0, r9, r10
|
/external/chromium_org/v8/src/x64/ |
H A D | codegen-x64.cc | 243 __ SmiToInteger32(r9, FieldOperand(r8, FixedDoubleArray::kLengthOffset)); 266 // r9 : elements array length 295 __ leap(rdi, Operand(r9, times_8, FixedArray::kHeaderSize)); 311 __ Integer32ToSmi(r11, r9); 330 FieldOperand(r8, r9, times_pointer_size, FixedArray::kHeaderSize)); 331 // r9 : current element's index 336 __ movsd(FieldOperand(r14, r9, times_8, FixedDoubleArray::kHeaderSize), 346 __ movq(FieldOperand(r14, r9, times_8, FixedDoubleArray::kHeaderSize), r15); 348 __ decp(r9); 384 __ SmiToInteger32(r9, FieldOperan [all...] |
/external/chromium_org/third_party/boringssl/src/crypto/bn/asm/ |
H A D | x86_64-mont5.pl | 62 $num="%r9"; # int num, 857 mov %r9,%rcx 912 my $num ="%r9"; # int num, has to be divisible by 8 1579 mov 8*1($tptr),%r9 1605 add %rax,%r9 1608 add %r9,%r8 1610 mov %rdx,%r9 1611 adc \$0,%r9 1617 add %r10,%r9 1675 adc 8*1($tptr),%r9 [all...] |
/external/flac/libFLAC/ppc/as/ |
H A D | lpc_asm.s | 62 addi r9,r1,-28 64 andc r9,r9,r31 ; for quadword-aligned stack data 85 stw r31,-4(r9) 86 lvewx v0,r10,r9 228 stw r7,-4(r9) 229 lvewx v19,r10,r9 ; v19: lp_quantization vector 311 addi r9,r1,-28 313 andc r9,r9,r3 [all...] |
/external/flac/libFLAC/ppc/gas/ |
H A D | lpc_asm.s | 64 addi r9,r1,-28 66 andc r9,r9,r31 # for quadword-aligned stack data 87 stw r31,-4(r9) 88 lvewx v0,r10,r9 230 stw r7,-4(r9) 231 lvewx v19,r10,r9 # v19: lp_quantization vector 313 addi r9,r1,-28 315 andc r9,r9,r3 [all...] |
/external/openssl/crypto/bn/asm/ |
H A D | s390x.S | 42 lg %r9,8(%r2,%r3) 44 alcgr %r9,%r6 46 alg %r9,8(%r2,%r1) 47 stg %r9,8(%r2,%r1) 56 lg %r9,24(%r2,%r3) 58 alcgr %r9,%r6 60 alg %r9,24(%r2,%r1) 61 stg %r9,24(%r2,%r1) 115 lg %r9,8(%r2,%r3) 117 alcgr %r9, [all...] |
/external/chromium_org/third_party/libvpx/source/libvpx/vp8/common/arm/armv6/ |
H A D | vp8_variance_halfpixvar16x16_hv_armv6.asm | 38 add r9, r0, r1 ; pointer to pixels on the next row 42 ldr r5, [r9, #0] ; load source pixels c, row N+1 43 ldr r7, [r9, #1] ; load source pixels d, row N+1 82 ldr r5, [r9, #4] ; load source pixels c, row N+1 86 ldr r7, [r9, #5] ; load source pixels d, row N+1 124 ldr r5, [r9, #8] ; load source pixels c, row N+1 128 ldr r7, [r9, #9] ; load source pixels d, row N+1 166 ldr r5, [r9, #12] ; load source pixels c, row N+1 168 ldr r7, [r9, #13] ; load source pixels d, row N+1
|
/external/chromium_org/third_party/boringssl/win-x86_64/crypto/aes/ |
H A D | vpaes-x86_64.asm | 22 mov r9,rdx 28 movdqu xmm5,XMMWORD PTR[r9] 35 add r9,16 58 add r9,16 88 movdqu xmm5,XMMWORD PTR[r9] 112 mov r9,rdx 119 movdqu xmm5,XMMWORD PTR[r9] 131 add r9,16 169 add r9,16 194 movdqu xmm0,XMMWORD PTR[r9] [all...] |