/external/sonivox/arm-hybrid-22k/lib_src/ |
H A D | ARM-E_interpolate_noloop_gnu.s | 51 tmp2 .req r9
54 @SaveRegs RLIST {r4-r9,lr}
55 @RestoreRegs RLIST {r4-r9,pc}
60 STMFD sp!, {r4-r9,lr}
125 LDMFD sp!,{r4-r9,lr}
|
/external/sonivox/arm-wt-22k/lib_src/ |
H A D | ARM-E_interpolate_noloop_gnu.s | 51 tmp2 .req r9
54 @SaveRegs RLIST {r4-r9,lr}
55 @RestoreRegs RLIST {r4-r9,pc}
60 STMFD sp!, {r4-r9,lr}
125 LDMFD sp!,{r4-r9,lr}
|
/external/valgrind/main/none/tests/arm/ |
H A D | vfp.stdout.exp | 81 vmov.32 r9, d20[0] :: Rd 0x11223344 Qm (i32)0x11223344 88 vmov.u16 r9, d26[3] :: Rd 0x00001122 Qm (i32)0x11223344 95 vmov.u8 r9, d24[2] :: Rd 0x00000080 Qm (i32)0xff800000 104 vmov.s16 r9, d26[3] :: Rd 0x00001122 Qm (i32)0x11223344 111 vmov.s8 r9, d24[2] :: Rd 0xffffff80 Qm (i32)0xff800000 872 vldr d16, [r9, #-4] :: Dd 0x00000cc2 0x00000bb1 *(int*) (Rn + shift) 0x0bb1 874 vldr d22, [r9, #+8] :: Dd 0x0000022b 0x0000011a *(int*) (Rn + shift) 0x011a 881 vldr d9, [r9, #-4] :: Dd 0x00000cc2 0x00000bb1 *(int*) (Rn + shift) 0x0bb1 889 vldr s16, [r9, #-4] :: Sd 0x00000bb1 *(int*) (Rn + shift) 0x0bb1 891 vldr s22, [r9, # [all...] |
H A D | vfp.c | 1091 TESTINSN_scalar_to_core("vmov.32 r9, d20[0]", r9, d20, i32, 0x11223344); 1098 TESTINSN_scalar_to_core("vmov.u16 r9, d26[3]", r9, d26, i32, 0x11223344); 1105 TESTINSN_scalar_to_core("vmov.u8 r9, d24[2]", r9, d24, i32, f2u(-INFINITY)); 1114 TESTINSN_scalar_to_core("vmov.s16 r9, d26[3]", r9, d26, i32, 0x11223344); 1121 TESTINSN_scalar_to_core("vmov.s8 r9, d24[2]", r9, d2 [all...] |
/external/llvm/test/MC/ARM/ |
H A D | basic-thumb2-instructions.s | 25 adc r9, r7, #0xa5a5a5a5 35 @ CHECK: adc r9, r7, #2779096485 @ encoding: [0x47,0xf1,0xa5,0x39] 45 adc.w r9, r1, r3 46 adcs.w r9, r1, r3 54 @ CHECK: adc.w r9, r1, r3 @ encoding: [0x41,0xeb,0x03,0x09] 55 @ CHECK: adcs.w r9, r1, r3 @ encoding: [0x51,0xeb,0x03,0x09] 102 add r5, r9, r2, asr #32 110 @ CHECK: add.w r5, r9, r2, asr #32 @ encoding: [0x09,0xeb,0x22,0x05] 145 and r4, r9, r8 149 and.w r9, r1 [all...] |
H A D | thumb.s | 13 blx r9 15 @ CHECK: blx r9 @ encoding: [0xc8,0x47]
|
/external/libvpx/vp8/encoder/arm/armv5te/ |
H A D | vp8_packtokens_partitions_armv5.asm | 86 ldr r9, [r7, #tokenlist_stop] 87 str r9, [sp, #0] ; save stop for later comparison 99 ldr r9, [r1, #tokenextra_context_tree] ; pp 119 ldrb r4, [r9, lr, asr #1] ; pp [i>>1] 212 ldr r9, [r12, #vp8_extra_bit_struct_prob] 224 ldrb r4, [r9, lr, asr #1] ; pp[i>>1] 415 ldr r9, [sp, #20] ; num_parts 416 sub r9, r9, #1 418 cmp r10, r9 ; i [all...] |
/external/openssl/crypto/ |
H A D | ppccpuid.pl | 62 xor r9,r9,r9
|
H A D | x86_64cpuid.pl | 16 ($arg1,$arg2,$arg3,$arg4)=$win64?("%rcx","%rdx","%r8", "%r9") : # Win64 order 168 shl \$32,%r9 171 or %r9,%rax 237 xorq %r9,%r9 258 xorq %r9,%r9
|
/external/zlib/contrib/gcc_gvmat64/ |
H A D | gvmat64.S | 68 ; register used : rax,rbx,rcx,rdx,rsi,rdi,r8,r9,r10,r11,r12
176 ; rax, rcx, rdx, r8, r9, r10, and r11, which are scratch.
259 mov r9,r13
289 movzx r12d,word ptr [r9]
290 movzx ebx, word ptr [r9 + r11 - 1]
421 lea rdi, [r9 + r13 + 0x0108] //;MAX_MATCH_8]
492 sub rax, r9
528 movzx ebx, word ptr [r9 + rax - 1]
|
/external/libvpx/vp8/encoder/arm/armv6/ |
H A D | vp8_sad16x16_armv6.asm | 35 ldr r9, [r2, #0x4] ; load 4 ref pixels (1A) 40 usad8 r8, r7, r9 ; calculate sad for 4 pixels 57 ldr r9, [r2, #0x4] ; load 4 ref pixels (2A) 62 usad8 r8, r7, r9 ; calculate sad for 4 pixels
|
H A D | vp8_variance16x16_armv6.asm | 41 usub8 r9, r5, r4 ; calculate difference with reversed operands 42 sel r6, r9, lr ; select bytes with negative difference 64 usub8 r9, r5, r4 ; calculate difference with reversed operands 65 sel r6, r9, lr ; select bytes with negative difference 88 usub8 r9, r5, r4 ; calculate difference with reversed operands 89 sel r6, r9, lr ; select bytes with negative difference 113 usub8 r9, r5, r4 ; calculate difference with reversed operands 115 sel r6, r9, lr ; select bytes with negative difference
|
H A D | vp8_variance_halfpixvar16x16_hv_armv6.asm | 34 add r9, r0, r1 ; pointer to pixels on the next row 38 ldr r5, [r9, #0] ; load source pixels c, row N+1 39 ldr r7, [r9, #1] ; load source pixels d, row N+1 76 ldr r5, [r9, #4] ; load source pixels c, row N+1 80 ldr r7, [r9, #5] ; load source pixels d, row N+1 118 ldr r5, [r9, #8] ; load source pixels c, row N+1 122 ldr r7, [r9, #9] ; load source pixels d, row N+1 160 ldr r5, [r9, #12] ; load source pixels c, row N+1 162 ldr r7, [r9, #13] ; load source pixels d, row N+1
|
H A D | vp8_variance8x8_armv6.asm | 39 usub8 r9, r7, r6 ; calculate difference with reversed operands 40 sel r8, r9, lr ; select bytes with negative difference 63 usub8 r9, r7, r6 ; calculate difference with reversed operands 65 sel r8, r9, lr ; select bytes with negative difference
|
/external/webkit/Source/JavaScriptCore/tests/mozilla/ecma/Array/ |
H A D | 15.4.4.4-2.js | 140 var r9 = array[r7]; 144 array[r8] = r9;
|
/external/kernel-headers/original/asm-x86/ |
H A D | sigcontext.h | 107 unsigned long r9; member in struct:sigcontext
|
/external/libvpx/vp8/encoder/ppc/ |
H A D | fdct_altivec.asm | 36 load_c v0, dct_tab, 0, r9, r10 42 load_c v4, ppc_dctperm_tab, 0, r9, r10 43 load_c v5, ppc_dctperm_tab, r6, r9, r10 45 load_c v6, round_tab, 0, r10, r9 125 lvx v6, r6, r9 ;# v6 = Vround 154 lvx v6, r6, r9 ;# v6 = Vround 165 lvx v6, 0, r9 ;# v6 = Hround 176 lvx v6, r6, r9 ;# v6 = Vround
|
/external/llvm/lib/Target/X86/ |
H A D | X86CompilationCallback_Win64.asm | 28 push r9 58 pop r9
|
/external/v8/test/cctest/ |
H A D | test-macro-assembler-x64.cc | 71 using v8::internal::r9; 199 __ movq(r9, rdx); 216 __ cmpq(rdx, r9); 679 __ SmiNeg(r9, rcx, exit); 697 __ SmiNeg(r9, rcx, &smi_ok); 701 __ cmpq(r9, r8); 775 __ SmiAdd(r9, rcx, rdx, exit); 776 __ cmpq(r9, r8); 788 __ SmiAddConstant(r9, rcx, Smi::FromInt(second)); 789 __ cmpq(r9, r [all...] |
H A D | test-disasm-arm.cc | 122 COMPARE(eor(r4, r5, Operand(r9, LSL, 3), SetCC, cs), 123 "20354189 eorcss r4, r5, r9, lsl #3"); 163 COMPARE(sbc(r7, r9, Operand(ip, ROR, 4)), 164 "e0c9726c sbc r7, r9, ip, ror #4"); 235 COMPARE(bic(r0, r9, Operand(r2), LeaveCC, vc), 236 "71c90002 bicvc r0, r9, r2"); 244 COMPARE(mvn(r9, Operand(r2)), 245 "e1e09002 mvn r9, r2"); 495 COMPARE(vstr(s16, r9, 12), 496 "ed898a03 vstr s16, [r9 [all...] |
/external/libvpx/vp8/common/ppc/ |
H A D | idctllm_altivec.asm | 29 load_c v8, sinpi8sqrt2, 0, r9, r10 30 load_c v9, cospi8sqrt2minus1, 0, r9, r10 31 load_c v10, hi_hi, 0, r9, r10 32 load_c v11, lo_lo, 0, r9, r10 33 load_c v12, shift_16, 0, r9, r10
|
H A D | filter_bilinear_altivec.asm | 24 load_c \V0, vfilter_b, r6, r9, r10 45 load_c v20, hfilter_b, r5, r9, r0 49 load_c v28, b_hperm_b, 0, r9, r0 178 load_c v10, b_0123_b, 0, r9, r12 179 load_c v11, b_4567_b, 0, r9, r12 263 load_c v10, b_0123_b, 0, r9, r12 264 load_c v11, b_4567_b, 0, r9, r12 315 load_c v10, b_hilo_b, 0, r9, r10 349 load_c v10, b_0123_b, 0, r9, r12 350 load_c v11, b_4567_b, 0, r9, r1 [all...] |
/external/flac/libFLAC/ppc/as/ |
H A D | lpc_asm.s | 62 addi r9,r1,-28 64 andc r9,r9,r31 ; for quadword-aligned stack data 85 stw r31,-4(r9) 86 lvewx v0,r10,r9 228 stw r7,-4(r9) 229 lvewx v19,r10,r9 ; v19: lp_quantization vector 311 addi r9,r1,-28 313 andc r9,r9,r3 [all...] |
/external/flac/libFLAC/ppc/gas/ |
H A D | lpc_asm.s | 64 addi r9,r1,-28 66 andc r9,r9,r31 # for quadword-aligned stack data 87 stw r31,-4(r9) 88 lvewx v0,r10,r9 230 stw r7,-4(r9) 231 lvewx v19,r10,r9 # v19: lp_quantization vector 313 addi r9,r1,-28 315 andc r9,r9,r3 [all...] |
/external/openssl/crypto/bn/asm/ |
H A D | s390x.S | 42 lg %r9,8(%r2,%r3) 44 alcgr %r9,%r6 46 alg %r9,8(%r2,%r1) 47 stg %r9,8(%r2,%r1) 56 lg %r9,24(%r2,%r3) 58 alcgr %r9,%r6 60 alg %r9,24(%r2,%r1) 61 stg %r9,24(%r2,%r1) 115 lg %r9,8(%r2,%r3) 117 alcgr %r9, [all...] |