/external/chromium_org/third_party/yasm/source/patched-yasm/modules/arch/lc3b/tests/ |
H A D | lc3b-ea-err.asm | 1 ld r5, [r6+5]
|
H A D | lc3b-basic.asm | 1 add r7, r6, r5
|
/external/chromium_org/third_party/openssl/openssl/crypto/bn/asm/ |
H A D | ppc.pl | 229 #.set r6,6 # ... 275 # Freely use registers r5,r6,r7,r8,r9,r10,r11 as follows: 277 # r5,r6 are the two BN_ULONGs being multiplied. 296 $LD r6,`1*$BNSZ`(r4) 297 $UMULL r7,r5,r6 298 $UMULH r8,r5,r6 311 $UMULL r7,r6,r6 312 $UMULH r8,r6,r6 [all...] |
/external/openssl/crypto/bn/asm/ |
H A D | ppc.pl | 229 #.set r6,6 # ... 275 # Freely use registers r5,r6,r7,r8,r9,r10,r11 as follows: 277 # r5,r6 are the two BN_ULONGs being multiplied. 296 $LD r6,`1*$BNSZ`(r4) 297 $UMULL r7,r5,r6 298 $UMULH r8,r5,r6 311 $UMULL r7,r6,r6 312 $UMULH r8,r6,r6 [all...] |
/external/libvpx/libvpx/vp8/common/arm/armv6/ |
H A D | vp8_variance_halfpixvar16x16_h_armv6.asm | 40 ldr r6, [r0, #1] ; load 4 src pixels with 1 byte offset 44 mvn r6, r6 45 uhsub8 r4, r4, r6 48 usub8 r6, r4, r5 ; calculate difference 50 sel r7, r6, lr ; select bytes with positive difference 51 usub8 r6, r5, r4 ; calculate difference with reversed operands 53 sel r6, r6, lr ; select bytes with negative difference 57 usad8 r5, r6, l [all...] |
H A D | vp8_variance_halfpixvar16x16_v_armv6.asm | 41 ldr r6, [r9, #0] ; load 4 src pixels from next row 45 mvn r6, r6 46 uhsub8 r4, r4, r6 49 usub8 r6, r4, r5 ; calculate difference 51 sel r7, r6, lr ; select bytes with positive difference 52 usub8 r6, r5, r4 ; calculate difference with reversed operands 54 sel r6, r6, lr ; select bytes with negative difference 58 usad8 r5, r6, l [all...] |
H A D | dc_only_idct_add_v6.asm | 30 ldr r6, [r1], r2 37 uxtab16 r7, r0, r6 38 uxtab16 r6, r0, r6, ror #8 42 usat16 r6, #8, r6 44 orr r7, r7, r6, lsl #8 47 ldr r6, [r1] 52 uxtab16 r7, r0, r6 53 uxtab16 r6, r [all...] |
H A D | dequant_idct_v6.asm | 32 smulbb r6, r4, r5 38 strh r6, [r0], #2 41 smulbb r6, r4, r5 49 strh r6, [r0], #2 60 ldr r6, [r0, #8] 65 smulwt r9, r3, r6 66 smulwb r7, r3, r6 67 smulwt r10, r4, r6 68 smulwb r8, r4, r6 72 uadd16 r6, r [all...] |
H A D | vp8_variance_halfpixvar16x16_hv_armv6.asm | 41 ldr r6, [r0, #1] ; load source pixels b, row N 46 mvn r6, r6 47 uhsub8 r4, r4, r6 59 usub8 r6, r4, r5 ; calculate difference 61 sel r7, r6, lr ; select bytes with positive difference 62 usub8 r6, r5, r4 ; calculate difference with reversed operands 64 sel r6, r6, lr ; select bytes with negative difference 68 usad8 r5, r6, l [all...] |
H A D | loopfilter_v6.asm | 68 ldr r6, [sp, #36] ; load thresh address 77 ldrb r3, [r6] ; thresh 90 uqsub8 r6, r9, r10 ; p3 - p2 95 orr r6, r6, r7 ; abs (p3-p2) 97 uqsub8 lr, r6, r2 ; compare to limit. lr: vp8_filter_mask 99 uqsub8 r6, r11, r12 ; p1 - p0 104 orr r6, r6, r7 ; abs (p1-p0) 105 uqsub8 r7, r6, r [all...] |
H A D | idct_v6.asm | 38 ldr r6, [r0, #(4*2)] ; i5 | i4 42 smulbt r9, r5, r6 ; (ip[5] * cospi8sqrt2minus1) >> 16 43 smulbb r7, r5, r6 ; (ip[4] * cospi8sqrt2minus1) >> 16 44 smulwt r10, r4, r6 ; (ip[5] * sinpi8sqrt2) >> 16 45 smulwb r8, r4, r6 ; (ip[4] * sinpi8sqrt2) >> 16 50 uadd16 r6, r6, r7 ; 5c+5 | 4c+4 64 uadd16 r6, r6, r10 ; d 68 uadd16 r9, r10, r6 ; [all...] |
/external/libvpx/libvpx/vp9/common/arm/neon/ |
H A D | vp9_avg_neon.asm | 19 push {r4-r6, lr} 21 mov r6, r2 39 vld1.8 {q8-q9}, [r6@128]! 40 vld1.8 {q10-q11}, [r6@128], r4 49 pop {r4-r6, pc} 54 vld1.8 {q8-q9}, [r6@128], r3 55 vld1.8 {q10-q11}, [r6@128], r3 60 pld [r6] 62 pld [r6, r3] 68 pop {r4-r6, p [all...] |
/external/valgrind/main/none/tests/x86/ |
H A D | incdec_alt.c | 7 int r1,r2,r3,r4,r5,r6,r7,r8,a1,a2; variable 40 "\tpopl " VG_SYM(r6) "\n" 58 r1=r2=r3=r4=r5=r6=r7=r8=0; 65 printf("0x%08x\n",r6);
|
/external/jpeg/ |
H A D | armv6_idct.S | 77 stmdb sp!, {r4, r5, r6, r7, r8, r9, r10, r11, r12, r14} 98 ldrsh r6, [r14, #96] 108 orreqs r8, r6, r7 115 // r6 = q[6] * c[6] + r2; 118 mla r6, r11, r6, r2 124 // r2 = r2 * 2 - r6 = -(r6 - r2 * 2); 126 rsb r2, r6, r2, lsl #1 160 // r0 = r0 + r6; [all...] |
/external/qemu/distrib/jpeg-6b/ |
H A D | armv6_idct.S | 77 stmdb sp!, {r4, r5, r6, r7, r8, r9, r10, r11, r12, r14} 98 ldrsh r6, [r14, #96] 108 orreqs r8, r6, r7 115 // r6 = q[6] * c[6] + r2; 118 mla r6, r11, r6, r2 124 // r2 = r2 * 2 - r6 = -(r6 - r2 * 2); 126 rsb r2, r6, r2, lsl #1 160 // r0 = r0 + r6; [all...] |
/external/tremolo/Tremolo/ |
H A D | mdctLARM.s | 61 LDMDB r2!,{r5,r6,r7,r12} 64 MOV r6, r6, ASR #9 @ r6 = (*--r)>>9 78 MOV r14,r6, ASR #15 80 EORNE r6, r4, r14,ASR #31 81 STRH r6, [r0], r3 123 LDR r6, [r2],#8 128 RSB r6, r6, # [all...] |
H A D | bitwiseARM.s | 67 STMFD r13!,{r5,r6} 79 LDRLT r6,[r3] @ r6 = ptr[1] 82 ORRLT r10,r10,r6,LSL r12 @ r10= first bitsLeftInSeg bits+crap 97 LDMIA r11,{r6,r12,r14} @ r6 = buffer 100 LDR r6,[r6] @ r6 = buffer->data 103 ADD r6,r [all...] |
/external/libvpx/libvpx/vpx_scale/arm/neon/ |
H A D | vp8_vpxyv12_extendframeborders_neon.asm | 35 add r6, r1, r3 ; dest_ptr2 = src_ptr2 + 1 (src_ptr1 + plane_width) 36 sub r2, r6, #1 ; src_ptr2 = src_ptr1 + plane_width - 1 63 vst1.8 {q2, q3}, [r6], lr 65 vst1.8 {q6, q7}, [r6], lr 67 vst1.8 {q10, q11}, [r6], lr 69 vst1.8 {q14, q15}, [r6], lr 81 add r6, r1, r8 ; dest_ptr2 = src_ptr2 - plane_stride (src_ptr1 + (plane_height * plane_stride)) 82 sub r2, r6, lr ; src_ptr2 = src_ptr1 + (plane_height * plane_stride) - plane_stride 102 vst1.8 {q8, q9}, [r6]! 104 vst1.8 {q10, q11}, [r6]! [all...] |
/external/llvm/test/MC/ARM/ |
H A D | basic-arm-instructions.s | 48 adc r4, r5, r6 50 adc r4, r5, r6, lsl #1 51 adc r4, r5, r6, lsl #31 52 adc r4, r5, r6, lsr #1 53 adc r4, r5, r6, lsr #31 54 adc r4, r5, r6, lsr #32 55 adc r4, r5, r6, asr #1 56 adc r4, r5, r6, asr #31 57 adc r4, r5, r6, asr #32 58 adc r4, r5, r6, ro [all...] |
H A D | dot-req.s | 7 fred .req r6 11 @ CHECK: mov r1, r6 @ encoding: [0x06,0x10,0xa0,0xe1]
|
/external/libvpx/libvpx/vp8/encoder/arm/armv6/ |
H A D | vp8_subtract_armv6.asm | 33 ldr r6, [r0, #vp8_block_src_diff] 60 str r0, [r6, #0] ; diff 61 str r1, [r6, #4] ; diff 63 add r6, r6, r2, lsl #1 ; update diff pointer 90 ldr r6, [r1] ; usrc (A) 93 uxtb16 r8, r6 ; [s2 | s0] (A) 95 uxtb16 r10, r6, ror #8 ; [s3 | s1] (A) 98 usub16 r6, r8, r9 ; [d2 | d0] (A) 104 pkhbt r8, r6, r [all...] |
/external/libvpx/libvpx/vp8/encoder/arm/armv5te/ |
H A D | vp8_packtokens_mbrow_armv5.asm | 50 ldr r6, _VP8_COMMON_MBrows_ 53 ldr r5, [r4, r6] ; load up mb_rows 81 ldrb r6, [r1, #tokenextra_token] ; t 84 add r4, r4, r6, lsl #3 ; a = vp8_coef_encodings + t 89 ldr r6, [r4, #vp8_token_value] ; v 101 lsl r12, r6, r4 ; r12 = v << 32 - n 112 mul r6, r4, r7 ; ((range-1) * pp[i>>1])) 120 add r4, r7, r6, lsr #8 ; 1 + (((range-1) * pp[i>>1]) >> 8) 126 clz r6, r4 127 sub r6, r [all...] |
/external/skia/src/core/asm/ |
H A D | s32a_d565_opaque.S | 30 stmdb sp!, {r4, r5, r6, r7, lr} 41 mov r6, r2 43 andeq r3, r6, #63488 // 0xf800 46 orreq r3, r3, r6, lsr #27 54 mov lr, r6, lsr #24 55 and r6, r1, #255 // 0xff 71 add r1, r6, r1, lsr #5 87 1: ldmia sp!, {r4, r5, r6, r7, pc}
|
/external/chromium_org/third_party/openssl/openssl/crypto/modes/asm/ |
H A D | ghash-armv4.S | 55 eor r5,r5,r6,lsl#28 56 eor r6,r10,r6,lsr#4 57 eor r6,r6,r7,lsl#28 73 eor r5,r5,r6,lsl#28 75 eor r6,r10,r6,lsr#4 77 eor r6,r6,r [all...] |
/external/openssl/crypto/modes/asm/ |
H A D | ghash-armv4.S | 55 eor r5,r5,r6,lsl#28 56 eor r6,r10,r6,lsr#4 57 eor r6,r6,r7,lsl#28 73 eor r5,r5,r6,lsl#28 75 eor r6,r10,r6,lsr#4 77 eor r6,r6,r [all...] |