/external/chromium_org/third_party/libvpx/source/libvpx/vp8/common/arm/neon/ |
H A D | idct_dequant_full_2x_neon.c | 24 int16x8_t q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10, q11; local 44 q4 = vld1q_s16(q); 71 q4 = vmulq_s16(q4, q0); 77 dLow1 = vget_low_s16(q4); 78 dHigh1 = vget_high_s16(q4); 80 q4 = vcombine_s16(dHigh0, dHigh1); 89 q6 = vqdmulhq_n_s16(q4, sinpi8sqrt2); 91 q8 = vqdmulhq_n_s16(q4, cospi8sqrt2minus1); 100 q4 [all...] |
H A D | dequant_idct_neon.c | 24 int16x8_t q1, q2, q3, q4, q5, q6; local 35 q4 = vld1q_s16(input); 55 q2 = vreinterpretq_s16_u16(vmulq_u16(vreinterpretq_u16_s16(q4), 64 q4 = vqdmulhq_n_s16(q2, cospi8sqrt2minus1); 67 q4 = vshrq_n_s16(q4, 1); 70 q4 = vqaddq_s16(q4, q2); 72 d10 = vqsub_s16(vget_low_s16(q3), vget_high_s16(q4)); 73 d11 = vqadd_s16(vget_high_s16(q3), vget_low_s16(q4)); [all...] |
H A D | sad_neon.c | 117 uint8x16_t q0, q4; local 126 q4 = vld1q_u8(ref_ptr); 128 q12 = vabdl_u8(vget_low_u8(q0), vget_low_u8(q4)); 129 q13 = vabdl_u8(vget_high_u8(q0), vget_high_u8(q4)); 134 q4 = vld1q_u8(ref_ptr); 136 q12 = vabal_u8(q12, vget_low_u8(q0), vget_low_u8(q4)); 137 q13 = vabal_u8(q13, vget_high_u8(q0), vget_high_u8(q4)); 154 uint8x16_t q0, q4; local 163 q4 = vld1q_u8(ref_ptr); 165 q12 = vabdl_u8(vget_low_u8(q0), vget_low_u8(q4)); [all...] |
H A D | mbloopfilter_neon.c | 19 uint8x16_t q4, // p2 39 q11u8 = vabdq_u8(q3, q4); 40 q12u8 = vabdq_u8(q4, q5); 71 q4 = veorq_u8(q4, q0u8); 139 q0s8 = vqaddq_s8(vreinterpretq_s8_u8(q4), q0s8); 161 uint8x16_t qblimit, qlimit, qthresh, q3, q4; local 172 q4 = vld1q_u8(src); 186 vp8_mbloop_filter_neon(qblimit, qlimit, qthresh, q3, q4, 188 &q4, 14 vp8_mbloop_filter_neon( uint8x16_t qblimit, uint8x16_t qlimit, uint8x16_t qthresh, uint8x16_t q3, uint8x16_t q4, uint8x16_t q5, uint8x16_t q6, uint8x16_t q7, uint8x16_t q8, uint8x16_t q9, uint8x16_t q10, uint8x16_t *q4r, uint8x16_t *q5r, uint8x16_t *q6r, uint8x16_t *q7r, uint8x16_t *q8r, uint8x16_t *q9r) argument 212 uint8x16_t qblimit, qlimit, qthresh, q3, q4; local 302 uint8x16_t qblimit, qlimit, qthresh, q3, q4; local 469 uint8x16_t qblimit, qlimit, qthresh, q3, q4; local [all...] |
/external/llvm/test/MC/ARM/ |
H A D | neon-bitwise-encoding.s | 65 veor q4, q7, q3 66 veor.8 q4, q7, q3 67 veor.16 q4, q7, q3 68 veor.32 q4, q7, q3 69 veor.64 q4, q7, q3 71 veor.i8 q4, q7, q3 72 veor.i16 q4, q7, q3 73 veor.i32 q4, q7, q3 74 veor.i64 q4, q7, q3 76 veor.s8 q4, q [all...] |
H A D | neon-v8.s | 5 vmaxnm.f32 q2, q4, q6 6 @ CHECK: vmaxnm.f32 q2, q4, q6 @ encoding: [0x5c,0x4f,0x08,0xf3] 16 vcvta.s32.f32 q4, q6 17 @ CHECK: vcvta.s32.f32 q4, q6 @ encoding: [0x4c,0x80,0xbb,0xf3] 18 vcvta.u32.f32 q4, q10 19 @ CHECK: vcvta.u32.f32 q4, q10 @ encoding: [0xe4,0x80,0xbb,0xf3] 43 vcvtp.s32.f32 q4, q15 44 @ CHECK: vcvtp.s32.f32 q4, q15 @ encoding: [0x6e,0x82,0xbb,0xf3] 50 vrintn.f32 q1, q4 51 @ CHECK: vrintn.f32 q1, q4 [all...] |
H A D | thumb-neon-v8.s | 5 vmaxnm.f32 q2, q4, q6 6 @ CHECK: vmaxnm.f32 q2, q4, q6 @ encoding: [0x08,0xff,0x5c,0x4f] 16 vcvta.s32.f32 q4, q6 17 @ CHECK: vcvta.s32.f32 q4, q6 @ encoding: [0xbb,0xff,0x4c,0x80] 18 vcvta.u32.f32 q4, q10 19 @ CHECK: vcvta.u32.f32 q4, q10 @ encoding: [0xbb,0xff,0xe4,0x80] 43 vcvtp.s32.f32 q4, q15 44 @ CHECK: vcvtp.s32.f32 q4, q15 @ encoding: [0xbb,0xff,0x6e,0x82] 50 vrintn.f32 q1, q4 51 @ CHECK: vrintn.f32 q1, q4 [all...] |
H A D | diagnostics-noneon.s | 5 vmov q4, q5
|
H A D | neon-shiftaccum-encoding.s | 10 vsra.s64 q8, q4, #64 18 vsra.u64 q4, q5, #25 28 vsra.s64 q4, #64 45 @ CHECK: vsra.s64 q8, q4, #64 @ encoding: [0xd8,0x01,0xc0,0xf2] 53 @ CHECK: vsra.u64 q4, q5, #25 @ encoding: [0xda,0x81,0xa7,0xf3] 62 @ CHECK: vsra.s64 q4, q4, #64 @ encoding: [0xd8,0x81,0x80,0xf2] 82 vrsra.s32 q3, q4, #32 83 vrsra.s64 q4, q5, #64 100 vrsra.s32 q4, #3 [all...] |
H A D | neont2-shiftaccum-encoding.s | 12 vsra.s64 q8, q4, #64 20 vsra.u64 q4, q5, #25 30 vsra.s64 q4, #64 47 @ CHECK: vsra.s64 q8, q4, #64 @ encoding: [0xc0,0xef,0xd8,0x01] 55 @ CHECK: vsra.u64 q4, q5, #25 @ encoding: [0xa7,0xff,0xda,0x81] 64 @ CHECK: vsra.s64 q4, q4, #64 @ encoding: [0x80,0xef,0xd8,0x81] 85 vrsra.s32 q3, q4, #32 86 vrsra.s64 q4, q5, #64 103 vrsra.s32 q4, #3 [all...] |
H A D | neon-shift-encoding.s | 116 vsra.s64 q4, q5, #63 122 vsra.s8 q4, #7 134 @ CHECK: vsra.s64 q4, q5, #63 @ encoding: [0xda,0x81,0x81,0xf2] 139 @ CHECK: vsra.s8 q4, q4, #7 @ encoding: [0x58,0x81,0x89,0xf2] 152 vsra.u64 q4, q5, #63 158 vsra.u8 q4, #7 170 @ CHECK: vsra.u64 q4, q5, #63 @ encoding: [0xda,0x81,0x81,0xf3] 175 @ CHECK: vsra.u8 q4, q4, # [all...] |
/external/libvpx/libvpx/vp8/common/arm/neon/ |
H A D | dequant_idct_neon.c | 24 int16x8_t q1, q2, q3, q4, q5, q6; local 35 q4 = vld1q_s16(input); 55 q2 = vreinterpretq_s16_u16(vmulq_u16(vreinterpretq_u16_s16(q4), 64 q4 = vqdmulhq_n_s16(q2, cospi8sqrt2minus1); 67 q4 = vshrq_n_s16(q4, 1); 70 q4 = vqaddq_s16(q4, q2); 72 d10 = vqsub_s16(vget_low_s16(q3), vget_high_s16(q4)); 73 d11 = vqadd_s16(vget_high_s16(q3), vget_low_s16(q4)); [all...] |
H A D | idct_dequant_full_2x_neon.asm | 28 vld1.16 {q4, q5}, [r0] ; r q 46 vmul.i16 q4, q4, q0 52 ; q4: l4r4 q5: l12r12 61 vqdmulh.s16 q6, q4, d0[2] ; sinpi8sqrt2 63 vqdmulh.s16 q8, q4, d0[0] ; cospi8sqrt2minus1 80 ; q4: 4 + 4 * cospi : d1/temp1 82 vqadd.s16 q4, q4, q8 88 vqadd.s16 q3, q4, q [all...] |
H A D | sixtappredict4x4_neon.asm | 62 vld1.u8 {q4}, [r0], r1 88 vmov q4, q3 ;keep original src data in q4 q6 93 vshr.u64 q9, q4, #8 ;construct src_ptr[-1] 100 vshr.u64 q3, q4, #32 ;construct src_ptr[2] 107 vshr.u64 q9, q4, #16 ;construct src_ptr[0] 114 vshr.u64 q3, q4, #24 ;construct src_ptr[1] 125 vld1.u8 {q4}, [r0], r1 154 vmov q4, q3 ;keep original src data in q4 q [all...] |
H A D | shortidct4x4llm_neon.asm | 48 vqdmulh.s16 q4, q2, d0[0] 54 vshr.s16 q4, q4, #1 57 vqadd.s16 q4, q4, q2 80 vqdmulh.s16 q4, q2, d0[0] 86 vshr.s16 q4, q4, #1 89 vqadd.s16 q4, q4, q [all...] |
/external/chromium_org/third_party/freetype/src/base/ |
H A D | ftbbox.c | 227 FT_Pos q1, q2, q3, q4; 233 q4 = p4; 240 if ( q1 + q2 > q3 + q4 ) /* first half */ 242 q4 = q4 + q3; 245 q4 = q4 + q3; 247 q4 = ( q4 + q3 ) / 8; 255 q3 = q3 + q4; [all...] |
/external/pdfium/core/src/fxge/fx_freetype/fxft2.5.01/src/base/ |
H A D | ftbbox.c | 227 FT_Pos q1, q2, q3, q4; 233 q4 = p4; 240 if ( q1 + q2 > q3 + q4 ) /* first half */ 242 q4 = q4 + q3; 245 q4 = q4 + q3; 247 q4 = ( q4 + q3 ) / 8; 255 q3 = q3 + q4; [all...] |
/external/valgrind/main/none/tests/arm/ |
H A D | neon128.c | 439 TESTINSN_bin("vand q4, q6, q5", q4, q6, i8, 0xff, q5, i16, 0x57); 445 TESTINSN_bin("vbic q4, q6, q5", q4, q6, i8, 0xff, q5, i16, 0x57); 452 TESTINSN_bin("vorr q4, q4, q4", q4, q4, i16, 0xff, q4, i1 [all...] |
/external/freetype/src/base/ |
H A D | ftbbox.c | 209 FT_Pos q4, 217 if ( q1 + q2 > q3 + q4 ) /* first half */ 219 q4 = q4 + q3; 222 q4 = q4 + q3; 224 q4 = ( q4 + q3 ) / 8; 232 q3 = q3 + q4; 246 if ( q3 == q4 206 update_cubic_max( FT_Pos q1, FT_Pos q2, FT_Pos q3, FT_Pos q4, FT_Pos max ) argument [all...] |
/external/libhevc/common/arm/ |
H A D | ihevc_intra_pred_chroma_horz.s | 130 vdup.16 q4,d1[0] 139 vst1.16 {q4},[r2],r3 140 vst1.16 {q4},[r9],r3 146 vdup.16 q4,d0[0] 155 vst1.16 {q4},[r2],r3 156 vst1.16 {q4},[r9],r3 162 vdup.16 q4,d11[0] 171 vst1.16 {q4},[r2],r3 172 vst1.16 {q4},[r9],r3 179 vdup.16 q4,d1 [all...] |
H A D | ihevc_intra_pred_luma_horz.s | 128 vdup.8 q4,d1[4] 137 vst1.8 {q4},[r2],r3 138 vst1.8 {q4},[r9],r3 144 vdup.8 q4,d1[0] 153 vst1.8 {q4},[r2],r3 154 vst1.8 {q4},[r9],r3 160 vdup.8 q4,d0[4] 169 vst1.8 {q4},[r2],r3 170 vst1.8 {q4},[r9],r3 177 vdup.8 q4,d [all...] |
H A D | ihevc_inter_pred_filters_luma_horz.s | 216 vmull.u8 q4,d1,d25 @mul_res = vmlal_u8(src[0_1], coeffabs_1)@ 218 vmlal.u8 q4,d3,d27 @mul_res = vmull_u8(src[0_3], coeffabs_3)@ 220 vmlsl.u8 q4,d0,d24 @mul_res = vmlsl_u8(src[0_0], coeffabs_0)@ 222 vmlsl.u8 q4,d2,d26 @mul_res = vmlsl_u8(src[0_2], coeffabs_2)@ 224 vmlal.u8 q4,d4,d28 @mul_res = vmlal_u8(src[0_4], coeffabs_4)@ 226 vmlsl.u8 q4,d5,d29 @mul_res = vmlsl_u8(src[0_5], coeffabs_5)@ 228 vmlal.u8 q4,d6,d30 @mul_res = vmlal_u8(src[0_6], coeffabs_6)@ 230 vmlsl.u8 q4,d7,d31 @mul_res = vmlsl_u8(src[0_7], coeffabs_7)@ 240 vqrshrun.s16 d20,q4,#6 @right shift and saturating narrow result 1 296 vmull.u8 q4,d [all...] |
H A D | ihevc_inter_pred_luma_vert_w16inp_w16out.s | 158 vmull.s16 q4,d1,d23 @mul_res1 = vmull_u8(src_tmp2, coeffabs_1)@ 160 vmlal.s16 q4,d0,d22 @mul_res1 = vmlal_u8(mul_res1, src_tmp1, coeffabs_0)@ 162 vmlal.s16 q4,d2,d24 @mul_res1 = vmlal_u8(mul_res1, src_tmp3, coeffabs_2)@ 164 vmlal.s16 q4,d3,d25 @mul_res1 = vmlal_u8(mul_res1, src_tmp4, coeffabs_3)@ 166 vmlal.s16 q4,d4,d26 @mul_res1 = vmlal_u8(mul_res1, src_tmp1, coeffabs_4)@ 168 vmlal.s16 q4,d5,d27 @mul_res1 = vmlal_u8(mul_res1, src_tmp2, coeffabs_5)@ 169 vmlal.s16 q4,d6,d28 @mul_res1 = vmlal_u8(mul_res1, src_tmp3, coeffabs_6)@ 170 vmlal.s16 q4,d7,d29 @mul_res1 = vmlal_u8(mul_res1, src_tmp4, coeffabs_7)@ 187 vsub.s32 q4, q4, q1 [all...] |
H A D | ihevc_inter_pred_luma_horz_w16out.s | 253 vmull.u8 q4,d1,d25 @arithmetic operations for ii iteration in the same time 254 vmlsl.u8 q4,d0,d24 255 vmlsl.u8 q4,d2,d26 256 vmlal.u8 q4,d3,d27 257 vmlal.u8 q4,d4,d28 258 vmlsl.u8 q4,d5,d29 259 vmlal.u8 q4,d6,d30 260 vmlsl.u8 q4,d7,d31 262 @ vqrshrun.s16 d8,q4,#6 @narrow right shift and saturating the result 310 vmull.u8 q4,d [all...] |
H A D | ihevc_itrans_recon_4x4_ttype1.s | 148 vmull.s16 q4,d1,d4[2] @74 * pi2_src[1] 149 vmlal.s16 q4,d0,d4[1] @74 * pi2_src[1] + 55 * pi2_src[0] 150 vmlsl.s16 q4,d2,d4[0] @74 * pi2_src[1] + 55 * pi2_src[0] - 29 * pi2_src[2] 151 vmlsl.s16 q4,d3,d4[3] @pi2_out[1] = 74 * pi2_src[1] + 55 * pi2_src[0] - 29 * pi2_src[2] - 84 * pi2_src[3]) 163 vqrshrn.s32 d15,q4,#shift_stage1_idct @ (pi2_out[1] + rounding ) >> shift_stage1_idct 187 vmull.s16 q4,d15,d4[2] @74 * pi2_src[1] 188 vmlal.s16 q4,d14,d4[1] @74 * pi2_src[1] + 55 * pi2_src[0] 189 vmlsl.s16 q4,d16,d4[0] @74 * pi2_src[1] + 55 * pi2_src[0] - 29 * pi2_src[2] 190 vmlsl.s16 q4,d17,d4[3] @pi2_out[1] = 74 * pi2_src[1] + 55 * pi2_src[0] - 29 * pi2_src[2] - 84 * pi2_src[3]) 203 vqrshrn.s32 d1,q4,#shift_stage2_idc [all...] |