Searched refs:q4 (Results 1 - 25 of 131) sorted by relevance

123456

/external/chromium_org/third_party/libvpx/source/libvpx/vp8/common/arm/neon/
H A Didct_dequant_full_2x_neon.c24 int16x8_t q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10, q11; local
44 q4 = vld1q_s16(q);
71 q4 = vmulq_s16(q4, q0);
77 dLow1 = vget_low_s16(q4);
78 dHigh1 = vget_high_s16(q4);
80 q4 = vcombine_s16(dHigh0, dHigh1);
89 q6 = vqdmulhq_n_s16(q4, sinpi8sqrt2);
91 q8 = vqdmulhq_n_s16(q4, cospi8sqrt2minus1);
100 q4
[all...]
H A Ddequant_idct_neon.c24 int16x8_t q1, q2, q3, q4, q5, q6; local
35 q4 = vld1q_s16(input);
55 q2 = vreinterpretq_s16_u16(vmulq_u16(vreinterpretq_u16_s16(q4),
64 q4 = vqdmulhq_n_s16(q2, cospi8sqrt2minus1);
67 q4 = vshrq_n_s16(q4, 1);
70 q4 = vqaddq_s16(q4, q2);
72 d10 = vqsub_s16(vget_low_s16(q3), vget_high_s16(q4));
73 d11 = vqadd_s16(vget_high_s16(q3), vget_low_s16(q4));
[all...]
H A Dsad_neon.c117 uint8x16_t q0, q4; local
126 q4 = vld1q_u8(ref_ptr);
128 q12 = vabdl_u8(vget_low_u8(q0), vget_low_u8(q4));
129 q13 = vabdl_u8(vget_high_u8(q0), vget_high_u8(q4));
134 q4 = vld1q_u8(ref_ptr);
136 q12 = vabal_u8(q12, vget_low_u8(q0), vget_low_u8(q4));
137 q13 = vabal_u8(q13, vget_high_u8(q0), vget_high_u8(q4));
154 uint8x16_t q0, q4; local
163 q4 = vld1q_u8(ref_ptr);
165 q12 = vabdl_u8(vget_low_u8(q0), vget_low_u8(q4));
[all...]
H A Dmbloopfilter_neon.c19 uint8x16_t q4, // p2
39 q11u8 = vabdq_u8(q3, q4);
40 q12u8 = vabdq_u8(q4, q5);
71 q4 = veorq_u8(q4, q0u8);
139 q0s8 = vqaddq_s8(vreinterpretq_s8_u8(q4), q0s8);
161 uint8x16_t qblimit, qlimit, qthresh, q3, q4; local
172 q4 = vld1q_u8(src);
186 vp8_mbloop_filter_neon(qblimit, qlimit, qthresh, q3, q4,
188 &q4,
14 vp8_mbloop_filter_neon( uint8x16_t qblimit, uint8x16_t qlimit, uint8x16_t qthresh, uint8x16_t q3, uint8x16_t q4, uint8x16_t q5, uint8x16_t q6, uint8x16_t q7, uint8x16_t q8, uint8x16_t q9, uint8x16_t q10, uint8x16_t *q4r, uint8x16_t *q5r, uint8x16_t *q6r, uint8x16_t *q7r, uint8x16_t *q8r, uint8x16_t *q9r) argument
212 uint8x16_t qblimit, qlimit, qthresh, q3, q4; local
302 uint8x16_t qblimit, qlimit, qthresh, q3, q4; local
469 uint8x16_t qblimit, qlimit, qthresh, q3, q4; local
[all...]
/external/llvm/test/MC/ARM/
H A Dneon-bitwise-encoding.s65 veor q4, q7, q3
66 veor.8 q4, q7, q3
67 veor.16 q4, q7, q3
68 veor.32 q4, q7, q3
69 veor.64 q4, q7, q3
71 veor.i8 q4, q7, q3
72 veor.i16 q4, q7, q3
73 veor.i32 q4, q7, q3
74 veor.i64 q4, q7, q3
76 veor.s8 q4, q
[all...]
H A Dneon-v8.s5 vmaxnm.f32 q2, q4, q6
6 @ CHECK: vmaxnm.f32 q2, q4, q6 @ encoding: [0x5c,0x4f,0x08,0xf3]
16 vcvta.s32.f32 q4, q6
17 @ CHECK: vcvta.s32.f32 q4, q6 @ encoding: [0x4c,0x80,0xbb,0xf3]
18 vcvta.u32.f32 q4, q10
19 @ CHECK: vcvta.u32.f32 q4, q10 @ encoding: [0xe4,0x80,0xbb,0xf3]
43 vcvtp.s32.f32 q4, q15
44 @ CHECK: vcvtp.s32.f32 q4, q15 @ encoding: [0x6e,0x82,0xbb,0xf3]
50 vrintn.f32 q1, q4
51 @ CHECK: vrintn.f32 q1, q4
[all...]
H A Dthumb-neon-v8.s5 vmaxnm.f32 q2, q4, q6
6 @ CHECK: vmaxnm.f32 q2, q4, q6 @ encoding: [0x08,0xff,0x5c,0x4f]
16 vcvta.s32.f32 q4, q6
17 @ CHECK: vcvta.s32.f32 q4, q6 @ encoding: [0xbb,0xff,0x4c,0x80]
18 vcvta.u32.f32 q4, q10
19 @ CHECK: vcvta.u32.f32 q4, q10 @ encoding: [0xbb,0xff,0xe4,0x80]
43 vcvtp.s32.f32 q4, q15
44 @ CHECK: vcvtp.s32.f32 q4, q15 @ encoding: [0xbb,0xff,0x6e,0x82]
50 vrintn.f32 q1, q4
51 @ CHECK: vrintn.f32 q1, q4
[all...]
H A Ddiagnostics-noneon.s5 vmov q4, q5
H A Dneon-shiftaccum-encoding.s10 vsra.s64 q8, q4, #64
18 vsra.u64 q4, q5, #25
28 vsra.s64 q4, #64
45 @ CHECK: vsra.s64 q8, q4, #64 @ encoding: [0xd8,0x01,0xc0,0xf2]
53 @ CHECK: vsra.u64 q4, q5, #25 @ encoding: [0xda,0x81,0xa7,0xf3]
62 @ CHECK: vsra.s64 q4, q4, #64 @ encoding: [0xd8,0x81,0x80,0xf2]
82 vrsra.s32 q3, q4, #32
83 vrsra.s64 q4, q5, #64
100 vrsra.s32 q4, #3
[all...]
H A Dneont2-shiftaccum-encoding.s12 vsra.s64 q8, q4, #64
20 vsra.u64 q4, q5, #25
30 vsra.s64 q4, #64
47 @ CHECK: vsra.s64 q8, q4, #64 @ encoding: [0xc0,0xef,0xd8,0x01]
55 @ CHECK: vsra.u64 q4, q5, #25 @ encoding: [0xa7,0xff,0xda,0x81]
64 @ CHECK: vsra.s64 q4, q4, #64 @ encoding: [0x80,0xef,0xd8,0x81]
85 vrsra.s32 q3, q4, #32
86 vrsra.s64 q4, q5, #64
103 vrsra.s32 q4, #3
[all...]
H A Dneon-shift-encoding.s116 vsra.s64 q4, q5, #63
122 vsra.s8 q4, #7
134 @ CHECK: vsra.s64 q4, q5, #63 @ encoding: [0xda,0x81,0x81,0xf2]
139 @ CHECK: vsra.s8 q4, q4, #7 @ encoding: [0x58,0x81,0x89,0xf2]
152 vsra.u64 q4, q5, #63
158 vsra.u8 q4, #7
170 @ CHECK: vsra.u64 q4, q5, #63 @ encoding: [0xda,0x81,0x81,0xf3]
175 @ CHECK: vsra.u8 q4, q4, #
[all...]
/external/libvpx/libvpx/vp8/common/arm/neon/
H A Ddequant_idct_neon.c24 int16x8_t q1, q2, q3, q4, q5, q6; local
35 q4 = vld1q_s16(input);
55 q2 = vreinterpretq_s16_u16(vmulq_u16(vreinterpretq_u16_s16(q4),
64 q4 = vqdmulhq_n_s16(q2, cospi8sqrt2minus1);
67 q4 = vshrq_n_s16(q4, 1);
70 q4 = vqaddq_s16(q4, q2);
72 d10 = vqsub_s16(vget_low_s16(q3), vget_high_s16(q4));
73 d11 = vqadd_s16(vget_high_s16(q3), vget_low_s16(q4));
[all...]
H A Didct_dequant_full_2x_neon.asm28 vld1.16 {q4, q5}, [r0] ; r q
46 vmul.i16 q4, q4, q0
52 ; q4: l4r4 q5: l12r12
61 vqdmulh.s16 q6, q4, d0[2] ; sinpi8sqrt2
63 vqdmulh.s16 q8, q4, d0[0] ; cospi8sqrt2minus1
80 ; q4: 4 + 4 * cospi : d1/temp1
82 vqadd.s16 q4, q4, q8
88 vqadd.s16 q3, q4, q
[all...]
H A Dsixtappredict4x4_neon.asm62 vld1.u8 {q4}, [r0], r1
88 vmov q4, q3 ;keep original src data in q4 q6
93 vshr.u64 q9, q4, #8 ;construct src_ptr[-1]
100 vshr.u64 q3, q4, #32 ;construct src_ptr[2]
107 vshr.u64 q9, q4, #16 ;construct src_ptr[0]
114 vshr.u64 q3, q4, #24 ;construct src_ptr[1]
125 vld1.u8 {q4}, [r0], r1
154 vmov q4, q3 ;keep original src data in q4 q
[all...]
H A Dshortidct4x4llm_neon.asm48 vqdmulh.s16 q4, q2, d0[0]
54 vshr.s16 q4, q4, #1
57 vqadd.s16 q4, q4, q2
80 vqdmulh.s16 q4, q2, d0[0]
86 vshr.s16 q4, q4, #1
89 vqadd.s16 q4, q4, q
[all...]
/external/chromium_org/third_party/freetype/src/base/
H A Dftbbox.c227 FT_Pos q1, q2, q3, q4;
233 q4 = p4;
240 if ( q1 + q2 > q3 + q4 ) /* first half */
242 q4 = q4 + q3;
245 q4 = q4 + q3;
247 q4 = ( q4 + q3 ) / 8;
255 q3 = q3 + q4;
[all...]
/external/pdfium/core/src/fxge/fx_freetype/fxft2.5.01/src/base/
H A Dftbbox.c227 FT_Pos q1, q2, q3, q4;
233 q4 = p4;
240 if ( q1 + q2 > q3 + q4 ) /* first half */
242 q4 = q4 + q3;
245 q4 = q4 + q3;
247 q4 = ( q4 + q3 ) / 8;
255 q3 = q3 + q4;
[all...]
/external/valgrind/main/none/tests/arm/
H A Dneon128.c439 TESTINSN_bin("vand q4, q6, q5", q4, q6, i8, 0xff, q5, i16, 0x57);
445 TESTINSN_bin("vbic q4, q6, q5", q4, q6, i8, 0xff, q5, i16, 0x57);
452 TESTINSN_bin("vorr q4, q4, q4", q4, q4, i16, 0xff, q4, i1
[all...]
/external/freetype/src/base/
H A Dftbbox.c209 FT_Pos q4,
217 if ( q1 + q2 > q3 + q4 ) /* first half */
219 q4 = q4 + q3;
222 q4 = q4 + q3;
224 q4 = ( q4 + q3 ) / 8;
232 q3 = q3 + q4;
246 if ( q3 == q4
206 update_cubic_max( FT_Pos q1, FT_Pos q2, FT_Pos q3, FT_Pos q4, FT_Pos max ) argument
[all...]
/external/libhevc/common/arm/
H A Dihevc_intra_pred_chroma_horz.s130 vdup.16 q4,d1[0]
139 vst1.16 {q4},[r2],r3
140 vst1.16 {q4},[r9],r3
146 vdup.16 q4,d0[0]
155 vst1.16 {q4},[r2],r3
156 vst1.16 {q4},[r9],r3
162 vdup.16 q4,d11[0]
171 vst1.16 {q4},[r2],r3
172 vst1.16 {q4},[r9],r3
179 vdup.16 q4,d1
[all...]
H A Dihevc_intra_pred_luma_horz.s128 vdup.8 q4,d1[4]
137 vst1.8 {q4},[r2],r3
138 vst1.8 {q4},[r9],r3
144 vdup.8 q4,d1[0]
153 vst1.8 {q4},[r2],r3
154 vst1.8 {q4},[r9],r3
160 vdup.8 q4,d0[4]
169 vst1.8 {q4},[r2],r3
170 vst1.8 {q4},[r9],r3
177 vdup.8 q4,d
[all...]
H A Dihevc_inter_pred_filters_luma_horz.s216 vmull.u8 q4,d1,d25 @mul_res = vmlal_u8(src[0_1], coeffabs_1)@
218 vmlal.u8 q4,d3,d27 @mul_res = vmull_u8(src[0_3], coeffabs_3)@
220 vmlsl.u8 q4,d0,d24 @mul_res = vmlsl_u8(src[0_0], coeffabs_0)@
222 vmlsl.u8 q4,d2,d26 @mul_res = vmlsl_u8(src[0_2], coeffabs_2)@
224 vmlal.u8 q4,d4,d28 @mul_res = vmlal_u8(src[0_4], coeffabs_4)@
226 vmlsl.u8 q4,d5,d29 @mul_res = vmlsl_u8(src[0_5], coeffabs_5)@
228 vmlal.u8 q4,d6,d30 @mul_res = vmlal_u8(src[0_6], coeffabs_6)@
230 vmlsl.u8 q4,d7,d31 @mul_res = vmlsl_u8(src[0_7], coeffabs_7)@
240 vqrshrun.s16 d20,q4,#6 @right shift and saturating narrow result 1
296 vmull.u8 q4,d
[all...]
H A Dihevc_inter_pred_luma_vert_w16inp_w16out.s158 vmull.s16 q4,d1,d23 @mul_res1 = vmull_u8(src_tmp2, coeffabs_1)@
160 vmlal.s16 q4,d0,d22 @mul_res1 = vmlal_u8(mul_res1, src_tmp1, coeffabs_0)@
162 vmlal.s16 q4,d2,d24 @mul_res1 = vmlal_u8(mul_res1, src_tmp3, coeffabs_2)@
164 vmlal.s16 q4,d3,d25 @mul_res1 = vmlal_u8(mul_res1, src_tmp4, coeffabs_3)@
166 vmlal.s16 q4,d4,d26 @mul_res1 = vmlal_u8(mul_res1, src_tmp1, coeffabs_4)@
168 vmlal.s16 q4,d5,d27 @mul_res1 = vmlal_u8(mul_res1, src_tmp2, coeffabs_5)@
169 vmlal.s16 q4,d6,d28 @mul_res1 = vmlal_u8(mul_res1, src_tmp3, coeffabs_6)@
170 vmlal.s16 q4,d7,d29 @mul_res1 = vmlal_u8(mul_res1, src_tmp4, coeffabs_7)@
187 vsub.s32 q4, q4, q1
[all...]
H A Dihevc_inter_pred_luma_horz_w16out.s253 vmull.u8 q4,d1,d25 @arithmetic operations for ii iteration in the same time
254 vmlsl.u8 q4,d0,d24
255 vmlsl.u8 q4,d2,d26
256 vmlal.u8 q4,d3,d27
257 vmlal.u8 q4,d4,d28
258 vmlsl.u8 q4,d5,d29
259 vmlal.u8 q4,d6,d30
260 vmlsl.u8 q4,d7,d31
262 @ vqrshrun.s16 d8,q4,#6 @narrow right shift and saturating the result
310 vmull.u8 q4,d
[all...]
H A Dihevc_itrans_recon_4x4_ttype1.s148 vmull.s16 q4,d1,d4[2] @74 * pi2_src[1]
149 vmlal.s16 q4,d0,d4[1] @74 * pi2_src[1] + 55 * pi2_src[0]
150 vmlsl.s16 q4,d2,d4[0] @74 * pi2_src[1] + 55 * pi2_src[0] - 29 * pi2_src[2]
151 vmlsl.s16 q4,d3,d4[3] @pi2_out[1] = 74 * pi2_src[1] + 55 * pi2_src[0] - 29 * pi2_src[2] - 84 * pi2_src[3])
163 vqrshrn.s32 d15,q4,#shift_stage1_idct @ (pi2_out[1] + rounding ) >> shift_stage1_idct
187 vmull.s16 q4,d15,d4[2] @74 * pi2_src[1]
188 vmlal.s16 q4,d14,d4[1] @74 * pi2_src[1] + 55 * pi2_src[0]
189 vmlsl.s16 q4,d16,d4[0] @74 * pi2_src[1] + 55 * pi2_src[0] - 29 * pi2_src[2]
190 vmlsl.s16 q4,d17,d4[3] @pi2_out[1] = 74 * pi2_src[1] + 55 * pi2_src[0] - 29 * pi2_src[2] - 84 * pi2_src[3])
203 vqrshrn.s32 d1,q4,#shift_stage2_idc
[all...]

Completed in 296 milliseconds

123456