/external/llvm/test/MC/ARM/ |
H A D | neon-bitwise-encoding.s | 64 veor q4, q7, q3 65 veor.8 q4, q7, q3 66 veor.16 q4, q7, q3 67 veor.32 q4, q7, q3 68 veor.64 q4, q7, q3 70 veor.i8 q4, q7, q3 71 veor.i16 q4, q7, q3 72 veor.i32 q4, q7, q3 73 veor.i64 q4, q7, q3 75 veor.s8 q4, q7, q [all...] |
H A D | neon-shiftaccum-encoding.s | 7 vsra.s8 q7, q2, #8 15 vsra.u8 q1, q7, #8 16 vsra.u16 q2, q7, #6 33 vsra.u8 q7, #8 34 vsra.u16 q7, #6 42 @ CHECK: vsra.s8 q7, q2, #8 @ encoding: [0x54,0xe1,0x88,0xf2] 50 @ CHECK: vsra.u8 q1, q7, #8 @ encoding: [0x5e,0x21,0x88,0xf3] 51 @ CHECK: vsra.u16 q2, q7, #6 @ encoding: [0x5e,0x41,0x9a,0xf3] 67 @ CHECK: vsra.u8 q7, q7, # [all...] |
H A D | neont2-shiftaccum-encoding.s | 9 vsra.s8 q7, q2, #8 17 vsra.u8 q1, q7, #8 18 vsra.u16 q2, q7, #6 35 vsra.u8 q7, #8 36 vsra.u16 q7, #6 44 @ CHECK: vsra.s8 q7, q2, #8 @ encoding: [0x88,0xef,0x54,0xe1] 52 @ CHECK: vsra.u8 q1, q7, #8 @ encoding: [0x88,0xff,0x5e,0x21] 53 @ CHECK: vsra.u16 q2, q7, #6 @ encoding: [0x9a,0xff,0x5e,0x41] 69 @ CHECK: vsra.u8 q7, q7, # [all...] |
H A D | neon-minmax-encoding.s | 21 vmax.s32 q7, q8, q9 24 vmax.u32 q6, q7, q8 32 vmax.u32 q7, q8 51 @ CHECK: vmax.s32 q7, q8, q9 @ encoding: [0xe2,0xe6,0x20,0xf2] 54 @ CHECK: vmax.u32 q6, q7, q8 @ encoding: [0x60,0xc6,0x2e,0xf3] 61 @ CHECK: vmax.u32 q7, q7, q8 @ encoding: [0x60,0xe6,0x2e,0xf3] 83 vmin.s32 q7, q8, q9 86 vmin.u32 q6, q7, q8 94 vmin.u32 q7, q [all...] |
H A D | neont2-minmax-encoding.s | 23 vmax.s32 q7, q8, q9 26 vmax.u32 q6, q7, q8 34 vmax.u32 q7, q8 53 @ CHECK: vmax.s32 q7, q8, q9 @ encoding: [0x20,0xef,0xe2,0xe6] 56 @ CHECK: vmax.u32 q6, q7, q8 @ encoding: [0x2e,0xff,0x60,0xc6] 63 @ CHECK: vmax.u32 q7, q7, q8 @ encoding: [0x2e,0xff,0x60,0xe6] 85 vmin.s32 q7, q8, q9 88 vmin.u32 q6, q7, q8 96 vmin.u32 q7, q [all...] |
H A D | neont2-pairwise-encoding.s | 21 vpaddl.s8 q4, q7 24 vpaddl.u8 q7, q4 34 @ CHECK: vpaddl.s8 q4, q7 @ encoding: [0xb0,0xff,0x4e,0x82] 37 @ CHECK: vpaddl.u8 q7, q4 @ encoding: [0xb0,0xff,0xc8,0xe2] 51 vpadal.u8 q7, q13 64 @ CHECK: vpadal.u8 q7, q13 @ encoding: [0xb0,0xff,0xea,0xe6]
|
H A D | neon-shift-encoding.s | 114 vsra.s16 q2, q7, #15 125 vsra.s64 q7, #63 132 @ CHECK: vsra.s16 q2, q7, #15 @ encoding: [0x5e,0x41,0x91,0xf2] 142 @ CHECK: vsra.s64 q7, q7, #63 @ encoding: [0xde,0xe1,0x81,0xf2] 150 vsra.u16 q2, q7, #15 161 vsra.u64 q7, #63 168 @ CHECK: vsra.u16 q2, q7, #15 @ encoding: [0x5e,0x41,0x91,0xf3] 178 @ CHECK: vsra.u64 q7, q7, #6 [all...] |
H A D | neont2-dup-encoding.s | 10 vdup.32 q7, r6 17 @ CHECK: vdup.32 q7, r6 @ encoding: [0xae,0xee,0x10,0x6b]
|
/external/libvpx/vp8/decoder/arm/neon/ |
H A D | idct_dequant_dc_full_2x_neon.asm | 68 ; q7: 12 * sinpi : d1/temp2 72 vqdmulh.s16 q7, q5, d0[2] 98 vqadd.s16 q3, q4, q7 107 vqsub.s16 q7, q10, q3 111 vtrn.32 q5, q7 113 vtrn.16 q6, q7 118 ; q7: l 3, 7,11,15 r 3, 7,11,15 125 vqdmulh.s16 q9, q7, d0[2] 127 vqdmulh.s16 q11, q7, d0[0] 139 vqadd.s16 q11, q7, q1 [all...] |
H A D | idct_dequant_full_2x_neon.asm | 60 ; q7: 12 * sinpi : d1/temp2 64 vqdmulh.s16 q7, q5, d0[2] 90 vqadd.s16 q3, q4, q7 99 vqsub.s16 q7, q10, q3 103 vtrn.32 q5, q7 105 vtrn.16 q6, q7 110 ; q7: l 3, 7,11,15 r 3, 7,11,15 117 vqdmulh.s16 q9, q7, d0[2] 119 vqdmulh.s16 q11, q7, d0[0] 131 vqadd.s16 q11, q7, q1 [all...] |
/external/libvpx/vp8/encoder/arm/neon/ |
H A D | vp8_mse16x16_neon.asm | 31 vmov.i8 q7, #0 ;q7, q8, q9, q10 - sse 49 vmlal.s16 q7, d22, d22 56 vmlal.s16 q7, d26, d26 63 vadd.u32 q7, q7, q8 68 vadd.u32 q10, q7, q9 156 vmull.s16 q7, d22, d22 161 vadd.u32 q7, q7, q [all...] |
H A D | fastquantizeb_neon.asm | 55 vld1.s16 {q6, q7}, [r0] ;load round_ptr [0-15] 59 vadd.s16 q5, q7 84 vld1.s16 {q6, q7}, [r12] ;load dequant_ptr[i] 99 vmul.s16 q7, q5 104 vst1.s16 {q6, q7}, [r3] ;store dqcoeff = x * Dequant
|
H A D | vp8_subpixelvariance8x8_neon.asm | 50 vmull.u8 q7, d4, d0 60 vmlal.u8 q7, d5, d1 67 vqrshrn.u16 d23, q7, #7 77 vmull.u8 q7, d4, d0 89 vmlal.u8 q7, d5, d1 95 vqrshrn.u16 d27, q7, #7 119 vmull.u8 q7, d28, d0 128 vmlal.u8 q7, d29, d1 137 vqrshrn.u16 d28, q7, #7 180 vsubl.u8 q7, d2 [all...] |
H A D | fastfdct4x4_neon.asm | 57 vqdmulh.s16 q7, q4, d0[2] 61 vshr.s16 q7, q7, #1 ;d14:temp1 = ( c1 * x_c3)>>16; d15:temp1 = (d1 * x_c3)>>16 87 vqdmulh.s16 q7, q4, d0[2] 91 vshr.s16 q7, q7, #1 ;d14:temp1 = ( c1 * x_c3)>>16; d15:temp1 = (d1 * x_c3)>>16
|
H A D | vp8_subpixelvariance16x16s_neon.asm | 58 vext.8 q7, q6, q7, #1 64 vrhadd.u8 q3, q6, q7 69 vsubl.u8 q7, d3, d25 87 vpadal.s16 q8, q7 154 vld1.8 {q7}, [r2], r3 256 vext.8 q7, q6, q7, #1 261 vrhadd.u8 q3, q6, q7 268 vld1.8 {q7}, [r [all...] |
H A D | vp8_memcpy_neon.asm | 36 vld1.8 {q6, q7}, [r1]! 39 vst1.8 {q6, q7}, [r0]!
|
/external/valgrind/main/none/tests/arm/ |
H A D | neon128.c | 359 TESTINSN_imm("vmov.i16 q7", q7, 0x700); 373 TESTINSN_imm("vmvn.i16 q7", q7, 0x700); 451 TESTINSN_bin("vorr q7, q3, q0", q7, q3, i8, 0x24, q0, i16, 0xff); 457 TESTINSN_bin("vorn q7, q3, q0", q7, q3, i8, 0x24, q0, i16, 0xff); 467 TESTINSN_bin("veor q7, q3, q0", q7, q [all...] |
/external/libvpx/vp8/common/arm/neon/ |
H A D | sixtappredict4x4_neon.asm | 86 vmull.u8 q7, d18, d5 ;(src_ptr[3] * vp8_filter[5]) 96 vmlal.u8 q7, d6, d0 ;+(src_ptr[-2] * vp8_filter[0]) 103 vmlsl.u8 q7, d18, d1 ;-(src_ptr[-1] * vp8_filter[1]) 110 vmlsl.u8 q7, d6, d4 ;-(src_ptr[2] * vp8_filter[4]) 117 vmlal.u8 q7, d18, d2 ;(src_ptr[0] * vp8_filter[2]) 128 vqadd.s16 q7, q9 ;sum of all (src_data*filter_parameters) 134 vqrshrun.s16 d27, q7, #7 ;shift/round/saturate to u8 151 vmull.u8 q7, d18, d5 ;(src_ptr[3] * vp8_filter[5]) 163 vmlal.u8 q7, d6, d0 ;+(src_ptr[-2] * vp8_filter[0]) 173 vmlsl.u8 q7, d1 [all...] |
H A D | loopfiltersimplehorizontaledge_neon.asm | 37 vld1.u8 {q7}, [r0], r1 ; q0 42 vabd.u8 q15, q6, q7 ; abs(p0 - q0) 49 veor q7, q7, q0 ; qs0: q0 offset to convert to a signed value 59 ;vqsub.s8 q2, q7, q6 ; ( qs0 - ps0) 93 vqsub.s8 q10, q7, q4 ; u = vp8_signed_char_clamp(qs0 - Filter1) 98 veor q7, q10, q0 ; *oq0 = u^0x80 101 vst1.u8 {q7}, [r3] ; store oq0
|
H A D | recon16x16mb_neon.asm | 43 vmovl.u8 q7, d31 57 vadd.s16 q7, q7, q15 68 vqmovun.s16 d7, q7 94 vmovl.u8 q7, d31 106 vadd.s16 q7, q7, q15 117 vqmovun.s16 d7, q7
|
H A D | sixtappredict8x4_neon.asm | 80 vmull.u8 q7, d6, d0 ;(src_ptr[-2] * vp8_filter[0]) 90 vmlsl.u8 q7, d28, d1 ;-(src_ptr[-1] * vp8_filter[1]) 100 vmlsl.u8 q7, d28, d4 ;-(src_ptr[2] * vp8_filter[4]) 110 vmlal.u8 q7, d28, d2 ;(src_ptr[0] * vp8_filter[2]) 120 vmlal.u8 q7, d28, d5 ;(src_ptr[3] * vp8_filter[5]) 135 vqadd.s16 q7, q3 ;sum of all (src_data*filter_parameters) 142 vqrshrun.s16 d22, q7, #7 ;shift/round/saturate to u8 153 vld1.u8 {q7}, [r0], r1 221 vmull.u8 q7, d31, d3 227 vqadd.s16 q12, q7 [all...] |
H A D | sixtappredict8x8_neon.asm | 72 vmull.u8 q7, d6, d0 ;(src_ptr[-2] * vp8_filter[0]) 82 vmlsl.u8 q7, d28, d1 ;-(src_ptr[-1] * vp8_filter[1]) 92 vmlsl.u8 q7, d28, d4 ;-(src_ptr[2] * vp8_filter[4]) 102 vmlal.u8 q7, d28, d2 ;(src_ptr[0] * vp8_filter[2]) 112 vmlal.u8 q7, d28, d5 ;(src_ptr[3] * vp8_filter[5]) 129 vqadd.s16 q7, q3 ;sum of all (src_data*filter_parameters) 136 vqrshrun.s16 d22, q7, #7 ;shift/round/saturate to u8 156 vld1.u8 {q7}, [r0], r1 222 vmull.u8 q7, d31, d3 228 vqadd.s16 q12, q7 [all...] |
H A D | bilinearpredict8x8_neon.asm | 47 vmull.u8 q7, d4, d0 57 vmlal.u8 q7, d5, d1 64 vqrshrn.u16 d23, q7, #7 74 vmull.u8 q7, d4, d0 86 vmlal.u8 q7, d5, d1 92 vqrshrn.u16 d27, q7, #7 117 vmull.u8 q7, d28, d0 126 vmlal.u8 q7, d29, d1 135 vqrshrn.u16 d8, q7, #7
|
H A D | mbloopfilter_neon.asm | 47 vld1.u8 {q7}, [r0], r1 ; q0 63 vst1.u8 {q7}, [r3] ; store oq0 168 vtrn.32 q3, q7 175 vtrn.16 q7, q9 180 vtrn.8 q7, q8 201 vtrn.32 q3, q7 209 vtrn.16 q7, q9 214 vtrn.8 q7, q8 282 vtrn.32 q3, q7 289 vtrn.16 q7, q [all...] |
H A D | loopfilter_neon.asm | 47 vld1.u8 {q7}, [r2], r1 ; q0 58 vst1.u8 {q7}, [r0], r1 ; store oq0 162 vtrn.32 q3, q7 169 vtrn.16 q7, q9 174 vtrn.8 q7, q8 246 vtrn.32 q3, q7 253 vtrn.16 q7, q9 258 vtrn.8 q7, q8 306 ; q7 q0 317 vabd.u8 q14, q8, q7 ; ab [all...] |