/external/libvpx/vp8/common/arm/armv6/ |
H A D | recon_v6.asm | 48 pkhbt r8, r6, r7, lsl #16 ; 2 | 0 51 uxtab16 r8, r8, r4 ; 2 | 0 + 3 | 2 | 2 | 0 54 usat16 r8, #8, r8 57 orr r8, r8, r9, lsl #8 59 str r8, [dst], stride 68 pkhbt r8, r6, r7, lsl #16 ; 2 | 0 71 uxtab16 r8, r [all...] |
H A D | loopfilter_v6.asm | 91 uqsub8 r8, r10, r11 ; p2 - p1 95 orr r8, r8, r10 ; abs (p2-p1) 97 uqsub8 r8, r8, r2 ; compare to limit 99 orr lr, lr, r8 105 uqsub8 r8, r6, r3 ; compare to thresh -- save r8 for later 158 orr r10, r6, r8 ; calculate vp8_hevmask 166 ldr r8, [sr [all...] |
H A D | idct_v6.asm | 12 ; r0 r1 r2 r3 r4 r5 r6 r7 r8 r9 r10 r11 r12 r14 69 ldrsh r8, [r0, #16] ; input[8] 1, r8 unavail 2 [8] 73 add r9, r7, r8 ; a1 = [0] + [8] 1 a1 74 sub r7, r7, r8 ; b1 = [0] - [8] 1 b1 79 add r8, r7, r11 ; b1 + c1 1 b+c 80 strh r8, [r1, r2] ; out[pitch] = b1+c1 1 86 add r8, r2, r2 ; pitch * 2 1 p*2 87 strh r7, [r1, r8] ; out[pitch*2] = b1-c1 1 100 ldrsh r8, [r [all...] |
H A D | simpleloopfilter_v6.asm | 83 uqsub8 r8, r6, r3 ; q1 - p1 86 orr r8, r8, r7 ; abs(p1 - q1) 89 uhadd8 r8, r8, lr ; abs(p1 - q2) >> 1 90 uqadd8 r10, r10, r8 ; abs(p0 - q0)*2 + abs(p1 - q1)/2 91 mvn r8, #0 93 sel r10, r8, lr ; filter mask: F or 0 109 ldr r8, c0x03030303 115 qadd8 r8 , r [all...] |
/external/openssl/crypto/bn/asm/ |
H A D | ppc.pl | 231 #.set r8,8 275 # Freely use registers r5,r6,r7,r8,r9,r10,r11 as follows: 278 # r7,r8 are the results of the 32x32 giving 64 bit multiply. 298 $UMULH r8,r5,r6 300 addc r7,r7,r7 # compute (r7,r8)=2*(r7,r8) 301 adde r8,r8,r8 306 addze r11,r8 # r [all...] |
/external/libvpx/vp8/decoder/arm/armv6/ |
H A D | dequant_idct_v6.asm | 71 smulwb r8, r4, r6 74 pkhbt r8, r8, r10, lsl #16 84 usub16 r7, r8, r7 87 usub16 r8, r11, r14 90 uadd16 r6, r8, r7 91 usub16 r7, r8, r7 104 ldr r8, [r0], #4 108 smulwt lr, r3, r8 109 smulwt r10, r4, r8 [all...] |
H A D | dequant_dc_idct_v6.asm | 93 smulwb r8, r4, r6 96 pkhbt r8, r8, r10, lsl #16 106 usub16 r7, r8, r7 109 usub16 r8, r11, r14 112 uadd16 r6, r8, r7 113 usub16 r7, r8, r7 126 ldr r8, [r0], #4 130 smulwt lr, r3, r8 131 smulwt r10, r4, r8 [all...] |
/external/valgrind/main/none/tests/x86/ |
H A D | incdec_alt.c | 7 int r1,r2,r3,r4,r5,r6,r7,r8,a1,a2; variable 49 "\tpopl " VG_SYM(r8) "\n" 58 r1=r2=r3=r4=r5=r6=r7=r8=0; 67 printf("0x%08x\n",r8);
|
H A D | sbbmisc.stdout.exp | 8 r8 = 104 105
|
H A D | insn_basic.def | 27 adcb eflags[0x1,0x0] : r8.ub[12] r8.ub[34] => 1.ub[46] 28 adcb eflags[0x1,0x1] : r8.ub[12] r8.ub[34] => 1.ub[47] 29 adcb eflags[0x1,0x0] : r8.ub[12] m8.ub[34] => 1.ub[46] 30 adcb eflags[0x1,0x1] : r8.ub[12] m8.ub[34] => 1.ub[47] 31 adcb eflags[0x1,0x0] : m8.ub[12] r8.ub[34] => 1.ub[46] 32 adcb eflags[0x1,0x1] : m8.ub[12] r8.ub[34] => 1.ub[47] 64 addb r8.ub[12] r8 [all...] |
/external/aac/libFDK/src/arm/ |
H A D | dct_arm.cpp | 109 r8 val_tw 117 LDR r8, [r1], #4 // val_tw = *twiddle++; local 121 SMULWT r9, r5, r8 // accuX = accu2*val_tw.l 122 SMULWB r5, r5, r8 // accu2 = accu2*val_tw.h 124 SMLAWT r5, r4, r8, r5 // accu2 = accu2*val_tw.h + accu1*val_tw.l local 125 SMLAWB r4, r4, r8, r9 // accu1 = accu1*val_tw.h - accu2*val_tw.l local 127 LDR r8, [r1], #4 // val_tw = *twiddle++; local 131 SMULWB r9, r7, r8 // accuX = accu4*val_tw.h 132 SMULWT r7, r7, r8 // accu4 = accu4*val_tw.l 134 SMLAWB r7, r6, r8, r local 135 SMLAWT r6, r6, r8, r9 // accu3 = accu3*val_tw.l-accu4*val_tw.h local 150 SMLAWT r5, r4, r8, r5 // accu2 = accu2*val_tw.h + accu1*val_tw.l local 151 SMLAWB r4, r4, r8, r9 // accu1 = accu1*val_tw.h - accu2*val_tw.l local 153 LDR r8, [r1], #4 // val_tw = *twiddle++; local 160 SMLAWB r7, r6, r8, r7 // accu4 = accu4*val_tw.l+accu3*val_tw.h local 161 SMLAWT r6, r6, r8, r9 // accu3 = accu3*val_tw.l-accu4*val_tw.h local 294 LDR r8, [r1], #4 // val_tw = *twiddle++ local 301 SMLAWB r9, r4, r8, r9 // accuX = accu1*val_tw.h-(-accu2)*val_tw.l local 304 SMLAWB r5, r5, r8, r4 // accu2 = (-accu2)*val_tw.t+accu1*val_tw.l local 305 LDR r8, [r1], #4 // val_tw = *twiddle++ local 312 SMLAWT r6, r6, r8, r5 // accu3 = (-accu4)*val_tw.l-(-accu3)*val_tw.h local 313 SMLAWT r7, r7, r8, r4 // accu4 = (-accu3)*val_tw.l+(-accu4)*val_tw.h local 317 LDR r8, [r1], #4 // val_tw = *twiddle++ local 324 SMLAWB r9, r4, r8, r9 // accuX = accu1*val_tw.h-(-accu2)*val_tw.l local 327 SMLAWB r5, r5, r8, r4 // accu2 = (-accu2)*val_tw.t+accu1*val_tw.l local 328 LDR r8, [r1], #4 // val_tw = *twiddle++ local 335 SMLAWT r6, r6, r8, r5 // accu3 = (-accu4)*val_tw.l-(-accu3)*val_tw.h local 336 SMLAWT r7, r7, r8, r4 // accu4 = (-accu3)*val_tw.l+(-accu4)*val_tw.h local [all...] |
/external/libffi/src/ia64/ |
H A D | unix.S | 60 mov r8 = in1 125 zxt1 r8 = r8 127 st8 [in1] = r8 131 sxt1 r8 = r8 133 st8 [in1] = r8 137 zxt2 r8 = r8 139 st8 [in1] = r8 [all...] |
/external/jpeg/ |
H A D | armv6_idct.S | 36 * 16-bit constants; r12 and r14 hold two of the four arguments; and r8 hold 38 * first pass, r8 to r11 are used to hold quantization values, so the loop 77 stmdb sp!, {r4, r5, r6, r7, r8, r9, r10, r11, r12, r14} 89 ldmdb r12!, {r8, r9, r10, r11} 102 mul r4, r8, r4 106 orreqs r8, r1, r2 107 orreqs r8, r3, r5 108 orreqs r8, r6, r7 121 ldmdb r12!, {r8, r9, r10, r11} 132 mul r7, r8, r [all...] |
/external/qemu/distrib/jpeg-6b/ |
H A D | armv6_idct.S | 36 * 16-bit constants; r12 and r14 hold two of the four arguments; and r8 hold 38 * first pass, r8 to r11 are used to hold quantization values, so the loop 77 stmdb sp!, {r4, r5, r6, r7, r8, r9, r10, r11, r12, r14} 89 ldmdb r12!, {r8, r9, r10, r11} 102 mul r4, r8, r4 106 orreqs r8, r1, r2 107 orreqs r8, r3, r5 108 orreqs r8, r6, r7 121 ldmdb r12!, {r8, r9, r10, r11} 132 mul r7, r8, r [all...] |
/external/tremolo/Tremolo/ |
H A D | mdctLARM.s | 187 LDMFD r12,{r8,r9,r10} @ r8 = step 191 MOV r8, r8, LSL #1 209 STRH r6, [r0], r8 227 LDMFD r12,{r8,r9,r10} @ r8 = step 231 MOV r8, r8, LSL #1 250 STRH r6, [r0], r8 [all...] |
H A D | mdctARM.s | 187 LDMFD r12,{r8,r9,r10} @ r8 = step 191 MOV r8, r8, LSL #1 208 STRH r6, [r0], r8 226 LDMFD r12,{r8,r9,r10} @ r8 = step 230 MOV r8, r8, LSL #1 247 STRH r6, [r0], r8 [all...] |
/external/libvpx/vp8/encoder/arm/armv6/ |
H A D | vp8_subtract_armv6.asm | 38 ldr r8, [r1, #vp8_blockd_predictor] 45 ldr r1, [r8], r2 ; pred 92 uxtb16 r8, r6 ; [s2 | s0] (A) 97 usub16 r6, r8, r9 ; [d2 | d0] (A) 103 pkhbt r8, r6, r7, lsl #16 ; [d1 | d0] (A) 106 str r8, [r0], #4 ; diff (A) 107 uxtb16 r8, r10 ; [s2 | s0] (B) 114 usub16 r6, r8, r9 ; [d2 | d0] (B) 119 pkhbt r8, r6, r7, lsl #16 ; [d1 | d0] (B) 122 str r8, [r [all...] |
H A D | vp8_mse16x16_armv6.asm | 41 usub8 r8, r5, r6 ; calculate difference 42 sel r7, r8, lr ; select bytes with positive difference 44 sel r8, r9, lr ; select bytes with negative difference 48 usad8 r6, r8, lr ; calculate sum of negative differences 49 orr r8, r8, r7 ; differences of all 4 pixels 54 uxtb16 r6, r8 ; byte (two pixels) to halfwords 55 uxtb16 r7, r8, ror #8 ; another two pixels to halfwords 62 usub8 r8, r5, r6 ; calculate difference 63 sel r7, r8, l [all...] |
H A D | vp8_variance8x8_armv6.asm | 37 usub8 r8, r6, r7 ; calculate difference 38 sel r10, r8, lr ; select bytes with positive difference 40 sel r8, r9, lr ; select bytes with negative difference 44 usad8 r7, r8, lr ; calculate sum of negative differences 45 orr r8, r8, r10 ; differences of all 4 pixels 51 uxtb16 r7, r8 ; byte (two pixels) to halfwords 52 uxtb16 r10, r8, ror #8 ; another two pixels to halfwords 60 usub8 r8, r6, r7 ; calculate difference 62 sel r10, r8, l [all...] |
H A D | vp8_sad16x16_armv6.asm | 33 ldr r8, [r2, #0x0] ; load 4 ref pixels (1A) 39 usada8 r4, r8, r6, r4 ; calculate sad for 4 pixels 40 usad8 r8, r7, r9 ; calculate sad for 4 pixels 49 usada8 r8, r11, lr, r8 ; calculate sad for 4 pixels 53 add r4, r4, r8 ; add partial sad values 56 ldr r8, [r2, #0x0] ; load 4 ref pixels (2A) 61 usada8 r4, r6, r8, r4 ; calculate sad for 4 pixels 62 usad8 r8, r7, r9 ; calculate sad for 4 pixels 71 usada8 r8, r1 [all...] |
/external/llvm/test/MC/ARM/ |
H A D | diagnostics.s | 166 ssat r8, #0, r10, lsl #8 167 ssat r8, #33, r10, lsl #8 168 ssat r8, #1, r10, lsl #-1 169 ssat r8, #1, r10, lsl #32 170 ssat r8, #1, r10, asr #0 171 ssat r8, #1, r10, asr #33 172 ssat r8, #1, r10, lsr #5 173 ssat r8, #1, r10, lsl fred 174 ssat r8, #1, r10, lsl #fred 177 @ CHECK-ERRORS: ssat r8, # [all...] |
/external/libvpx/vpx_scale/arm/neon/ |
H A D | vp8_vpxyv12_copyframe_func_neon.asm | 30 ldr r8, [r0, #yv12_buffer_config_u_buffer] ;srcptr1 42 str r8, [sp] 51 mov r8, r2 58 vld1.8 {q0, q1}, [r8]! 60 vld1.8 {q2, q3}, [r8]! 62 vld1.8 {q4, q5}, [r8]! 64 vld1.8 {q6, q7}, [r8]! 109 mov r8, r2 116 vld1.8 {q0, q1}, [r8]! 118 vld1.8 {q2, q3}, [r8]! [all...] |
/external/clang/test/CXX/except/except.spec/ |
H A D | p3.cpp | 36 extern void (*r8)() throw(int); 37 extern void (*r8)() throw(const int);
|
/external/compiler-rt/lib/tsan/rtl/ |
H A D | tsan_rtl_amd64.S | 22 push %r8 24 .cfi_rel_offset %r8, 0 57 pop %r8 75 .cfi_restore %r8 101 push %r8 103 .cfi_rel_offset %r8, 0 136 pop %r8 154 .cfi_restore %r8
|
/external/valgrind/main/none/tests/amd64/ |
H A D | sbbmisc.stdout.exp | 8 r8 = 104 105
|