/external/chromium_org/third_party/yasm/source/patched-yasm/modules/arch/x86/tests/ |
H A D | bswap64.asm | 3 bswap r8
|
H A D | mem64-err.asm | 7 mov ah, [r8]
|
H A D | sse-prefix.asm | 32 cmpss xmm0, [r8], 0 37 cmpsd xmm0, [r8], 0
|
/external/chromium_org/third_party/openssl/openssl/crypto/bn/asm/ |
H A D | ppc.pl | 231 #.set r8,8 275 # Freely use registers r5,r6,r7,r8,r9,r10,r11 as follows: 278 # r7,r8 are the results of the 32x32 giving 64 bit multiply. 298 $UMULH r8,r5,r6 300 addc r7,r7,r7 # compute (r7,r8)=2*(r7,r8) 301 adde r8,r8,r8 306 addze r11,r8 # r [all...] |
/external/openssl/crypto/bn/asm/ |
H A D | ppc.pl | 231 #.set r8,8 275 # Freely use registers r5,r6,r7,r8,r9,r10,r11 as follows: 278 # r7,r8 are the results of the 32x32 giving 64 bit multiply. 298 $UMULH r8,r5,r6 300 addc r7,r7,r7 # compute (r7,r8)=2*(r7,r8) 301 adde r8,r8,r8 306 addze r11,r8 # r [all...] |
/external/libvpx/libvpx/vp8/common/arm/armv6/ |
H A D | dequant_idct_v6.asm | 68 smulwb r8, r4, r6 71 pkhbt r8, r8, r10, lsl #16 81 usub16 r7, r8, r7 84 usub16 r8, r11, r14 87 uadd16 r6, r8, r7 88 usub16 r7, r8, r7 101 ldr r8, [r0], #4 105 smulwt lr, r3, r8 106 smulwt r10, r4, r8 [all...] |
H A D | loopfilter_v6.asm | 92 uqsub8 r8, r10, r11 ; p2 - p1 96 orr r8, r8, r10 ; abs (p2-p1) 98 uqsub8 r8, r8, r2 ; compare to limit 100 orr lr, lr, r8 106 uqsub8 r8, r6, r3 ; compare to thresh -- save r8 for later 159 orr r10, r6, r8 ; calculate vp8_hevmask 167 ldr r8, [sr [all...] |
H A D | intra4x4_predict_v6.asm | 52 ldr r8, [r0] ; Above 57 usad8 r12, r8, r9 82 ldr r8, [r0] ; Above 92 uxtb16 r10, r8 ; a[2|0] 93 uxtb16 r11, r8, ror #8 ; a[3|1] 141 ldr r8, [r0] ; a[3|2|1|0] 148 uxtb16 r4, r8 ; a[2|0] 149 uxtb16 r5, r8, ror #8 ; a[3|1] 178 ldrb r8, [sp, #48] ; top_left 183 add r8, r [all...] |
H A D | vp8_variance8x8_armv6.asm | 41 usub8 r8, r6, r7 ; calculate difference 43 sel r10, r8, lr ; select bytes with positive difference 46 sel r8, r9, lr ; select bytes with negative difference 50 usad8 r7, r8, lr ; calculate sum of negative differences 51 orr r8, r8, r10 ; differences of all 4 pixels 57 uxtb16 r7, r8 ; byte (two pixels) to halfwords 58 uxtb16 r10, r8, ror #8 ; another two pixels to halfwords 66 usub8 r8, r6, r7 ; calculate difference 68 sel r10, r8, l [all...] |
H A D | vp8_sad16x16_armv6.asm | 39 ldr r8, [r2, #0x0] ; load 4 ref pixels (1A) 45 usada8 r4, r8, r6, r4 ; calculate sad for 4 pixels 46 usad8 r8, r7, r9 ; calculate sad for 4 pixels 58 usada8 r8, r11, lr, r8 ; calculate sad for 4 pixels 62 add r4, r4, r8 ; add partial sad values 65 ldr r8, [r2, #0x0] ; load 4 ref pixels (2A) 70 usada8 r4, r6, r8, r4 ; calculate sad for 4 pixels 71 usad8 r8, r7, r9 ; calculate sad for 4 pixels 80 usada8 r8, r1 [all...] |
H A D | simpleloopfilter_v6.asm | 76 uqsub8 r8, r6, r3 ; q1 - p1 79 orr r8, r8, r7 ; abs(p1 - q1) 82 uhadd8 r8, r8, lr ; abs(p1 - q2) >> 1 83 uqadd8 r10, r10, r8 ; abs(p0 - q0)*2 + abs(p1 - q1)/2 84 mvn r8, #0 86 sel r10, r8, lr ; filter mask: F or 0 102 ldr r8, c0x03030303 108 qadd8 r8 , r [all...] |
/external/valgrind/main/none/tests/x86/ |
H A D | incdec_alt.c | 7 int r1,r2,r3,r4,r5,r6,r7,r8,a1,a2; variable 49 "\tpopl " VG_SYM(r8) "\n" 58 r1=r2=r3=r4=r5=r6=r7=r8=0; 67 printf("0x%08x\n",r8);
|
/external/aac/libFDK/src/arm/ |
H A D | dct_arm.cpp | 109 r8 val_tw 117 LDR r8, [r1], #4 // val_tw = *twiddle++; local 121 SMULWT r9, r5, r8 // accuX = accu2*val_tw.l 122 SMULWB r5, r5, r8 // accu2 = accu2*val_tw.h 124 SMLAWT r5, r4, r8, r5 // accu2 = accu2*val_tw.h + accu1*val_tw.l local 125 SMLAWB r4, r4, r8, r9 // accu1 = accu1*val_tw.h - accu2*val_tw.l local 127 LDR r8, [r1], #4 // val_tw = *twiddle++; local 131 SMULWB r9, r7, r8 // accuX = accu4*val_tw.h 132 SMULWT r7, r7, r8 // accu4 = accu4*val_tw.l 134 SMLAWB r7, r6, r8, r local 135 SMLAWT r6, r6, r8, r9 // accu3 = accu3*val_tw.l-accu4*val_tw.h local 150 SMLAWT r5, r4, r8, r5 // accu2 = accu2*val_tw.h + accu1*val_tw.l local 151 SMLAWB r4, r4, r8, r9 // accu1 = accu1*val_tw.h - accu2*val_tw.l local 153 LDR r8, [r1], #4 // val_tw = *twiddle++; local 160 SMLAWB r7, r6, r8, r7 // accu4 = accu4*val_tw.l+accu3*val_tw.h local 161 SMLAWT r6, r6, r8, r9 // accu3 = accu3*val_tw.l-accu4*val_tw.h local 294 LDR r8, [r1], #4 // val_tw = *twiddle++ local 301 SMLAWB r9, r4, r8, r9 // accuX = accu1*val_tw.h-(-accu2)*val_tw.l local 304 SMLAWB r5, r5, r8, r4 // accu2 = (-accu2)*val_tw.t+accu1*val_tw.l local 305 LDR r8, [r1], #4 // val_tw = *twiddle++ local 312 SMLAWT r6, r6, r8, r5 // accu3 = (-accu4)*val_tw.l-(-accu3)*val_tw.h local 313 SMLAWT r7, r7, r8, r4 // accu4 = (-accu3)*val_tw.l+(-accu4)*val_tw.h local 317 LDR r8, [r1], #4 // val_tw = *twiddle++ local 324 SMLAWB r9, r4, r8, r9 // accuX = accu1*val_tw.h-(-accu2)*val_tw.l local 327 SMLAWB r5, r5, r8, r4 // accu2 = (-accu2)*val_tw.t+accu1*val_tw.l local 328 LDR r8, [r1], #4 // val_tw = *twiddle++ local 335 SMLAWT r6, r6, r8, r5 // accu3 = (-accu4)*val_tw.l-(-accu3)*val_tw.h local 336 SMLAWT r7, r7, r8, r4 // accu4 = (-accu3)*val_tw.l+(-accu4)*val_tw.h local [all...] |
/external/libffi/src/ia64/ |
H A D | unix.S | 60 mov r8 = in1 125 zxt1 r8 = r8 127 st8 [in1] = r8 131 sxt1 r8 = r8 133 st8 [in1] = r8 137 zxt2 r8 = r8 139 st8 [in1] = r8 [all...] |
/external/jpeg/ |
H A D | armv6_idct.S | 36 * 16-bit constants; r12 and r14 hold two of the four arguments; and r8 hold 38 * first pass, r8 to r11 are used to hold quantization values, so the loop 77 stmdb sp!, {r4, r5, r6, r7, r8, r9, r10, r11, r12, r14} 89 ldmdb r12!, {r8, r9, r10, r11} 102 mul r4, r8, r4 106 orreqs r8, r1, r2 107 orreqs r8, r3, r5 108 orreqs r8, r6, r7 121 ldmdb r12!, {r8, r9, r10, r11} 132 mul r7, r8, r [all...] |
/external/qemu/distrib/jpeg-6b/ |
H A D | armv6_idct.S | 36 * 16-bit constants; r12 and r14 hold two of the four arguments; and r8 hold 38 * first pass, r8 to r11 are used to hold quantization values, so the loop 77 stmdb sp!, {r4, r5, r6, r7, r8, r9, r10, r11, r12, r14} 89 ldmdb r12!, {r8, r9, r10, r11} 102 mul r4, r8, r4 106 orreqs r8, r1, r2 107 orreqs r8, r3, r5 108 orreqs r8, r6, r7 121 ldmdb r12!, {r8, r9, r10, r11} 132 mul r7, r8, r [all...] |
/external/chromium_org/third_party/openssl/openssl/crypto/modes/asm/ |
H A D | ghash-x86_64.S | 12 movzbq 15(%rdi),%r8 20 movq 8(%rsi,%rax,1),%r8 23 movq %r8,%rdx 28 shrq $4,%r8 33 xorq 8(%rsi,%rbx,1),%r8 38 movq %r8,%rdx 40 xorq %r10,%r8 44 shrq $4,%r8 48 xorq 8(%rsi,%rax,1),%r8 53 movq %r8, [all...] |
/external/openssl/crypto/modes/asm/ |
H A D | ghash-x86_64.S | 12 movzbq 15(%rdi),%r8 20 movq 8(%rsi,%rax,1),%r8 23 movq %r8,%rdx 28 shrq $4,%r8 33 xorq 8(%rsi,%rbx,1),%r8 38 movq %r8,%rdx 40 xorq %r10,%r8 44 shrq $4,%r8 48 xorq 8(%rsi,%rax,1),%r8 53 movq %r8, [all...] |
/external/tremolo/Tremolo/ |
H A D | mdctLARM.s | 187 LDMFD r12,{r8,r9,r10} @ r8 = step 191 MOV r8, r8, LSL #1 209 STRH r6, [r0], r8 227 LDMFD r12,{r8,r9,r10} @ r8 = step 231 MOV r8, r8, LSL #1 250 STRH r6, [r0], r8 [all...] |
/external/libvpx/libvpx/vp9/common/arm/neon/ |
H A D | vp9_mb_lpf_neon.asm | 28 push {r4-r8, lr} 38 sub r8, r0, r1, lsl #3 ; move src pointer down by 8 lines 40 vld1.u8 {d0}, [r8@64], r1 ; p7 41 vld1.u8 {d1}, [r8@64], r1 ; p6 42 vld1.u8 {d2}, [r8@64], r1 ; p5 43 vld1.u8 {d3}, [r8@64], r1 ; p4 44 vld1.u8 {d4}, [r8@64], r1 ; p3 45 vld1.u8 {d5}, [r8@64], r1 ; p2 46 vld1.u8 {d6}, [r8@64], r1 ; p1 47 vld1.u8 {d7}, [r8 [all...] |
/external/libvpx/libvpx/vp8/encoder/arm/armv6/ |
H A D | vp8_subtract_armv6.asm | 38 ldr r8, [r1, #vp8_blockd_predictor] 45 ldr r1, [r8], r2 ; pred 93 uxtb16 r8, r6 ; [s2 | s0] (A) 98 usub16 r6, r8, r9 ; [d2 | d0] (A) 104 pkhbt r8, r6, r7, lsl #16 ; [d1 | d0] (A) 107 str r8, [r0], #4 ; diff (A) 108 uxtb16 r8, r10 ; [s2 | s0] (B) 115 usub16 r6, r8, r9 ; [d2 | d0] (B) 121 pkhbt r8, r6, r7, lsl #16 ; [d1 | d0] (B) 124 str r8, [r [all...] |
H A D | vp8_mse16x16_armv6.asm | 44 usub8 r8, r5, r6 ; calculate difference 46 sel r7, r8, lr ; select bytes with positive difference 49 sel r8, r9, lr ; select bytes with negative difference 53 usad8 r6, r8, lr ; calculate sum of negative differences 54 orr r8, r8, r7 ; differences of all 4 pixels 59 uxtb16 r6, r8 ; byte (two pixels) to halfwords 60 uxtb16 r7, r8, ror #8 ; another two pixels to halfwords 67 usub8 r8, r5, r6 ; calculate difference 68 sel r7, r8, l [all...] |
/external/llvm/test/MC/ARM/ |
H A D | thumb-nop.s | 8 @ CHECK-V6: mov r8, r8 @ encoding: [0xc0,0x46]
|
/external/libvpx/libvpx/vpx_scale/arm/neon/ |
H A D | vp8_vpxyv12_copyframe_func_neon.asm | 31 ldr r8, [r0, #yv12_buffer_config_u_buffer] ;srcptr1 43 str r8, [sp] 52 mov r8, r2 60 vld1.8 {q0, q1}, [r8]! 62 vld1.8 {q2, q3}, [r8]! 64 vld1.8 {q4, q5}, [r8]! 66 vld1.8 {q6, q7}, [r8]! 112 mov r8, r2 120 vld1.8 {q0, q1}, [r8]! 122 vld1.8 {q2, q3}, [r8]! [all...] |
/external/valgrind/main/coregrind/m_dispatch/ |
H A D | dispatch-s390x-linux.S | 204 larl %r8, VG_(stats__n_xindirs_32) 205 l %r10,0(%r8) 207 st %r10,0(%r8) 221 larl %r8, VG_(tt_fast) 228 lg %r11, 8(%r8,%r7) /* .host */ 229 cg %r2, 0(%r8,%r7) /* next guest address == .guest ? */ 240 larl %r8, VG_(stats__n_xindir_misses_32) 241 l %r10,0(%r8) 243 st %r10,0(%r8)
|