Searched refs:q4 (Results 1 - 25 of 144) sorted by relevance

123456

/external/libvpx/libvpx/vp8/common/arm/neon/
H A Didct_dequant_full_2x_neon.c24 int16x8_t q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10, q11; local
44 q4 = vld1q_s16(q);
71 q4 = vmulq_s16(q4, q0);
77 dLow1 = vget_low_s16(q4);
78 dHigh1 = vget_high_s16(q4);
80 q4 = vcombine_s16(dHigh0, dHigh1);
89 q6 = vqdmulhq_n_s16(q4, sinpi8sqrt2);
91 q8 = vqdmulhq_n_s16(q4, cospi8sqrt2minus1);
100 q4
[all...]
H A Ddequant_idct_neon.c24 int16x8_t q1, q2, q3, q4, q5, q6; local
35 q4 = vld1q_s16(input);
55 q2 = vreinterpretq_s16_u16(vmulq_u16(vreinterpretq_u16_s16(q4),
64 q4 = vqdmulhq_n_s16(q2, cospi8sqrt2minus1);
67 q4 = vshrq_n_s16(q4, 1);
70 q4 = vqaddq_s16(q4, q2);
72 d10 = vqsub_s16(vget_low_s16(q3), vget_high_s16(q4));
73 d11 = vqadd_s16(vget_high_s16(q3), vget_low_s16(q4));
[all...]
H A Dmbloopfilter_neon.c19 uint8x16_t q4, // p2
39 q11u8 = vabdq_u8(q3, q4);
40 q12u8 = vabdq_u8(q4, q5);
71 q4 = veorq_u8(q4, q0u8);
139 q0s8 = vqaddq_s8(vreinterpretq_s8_u8(q4), q0s8);
161 uint8x16_t qblimit, qlimit, qthresh, q3, q4; local
172 q4 = vld1q_u8(src);
186 vp8_mbloop_filter_neon(qblimit, qlimit, qthresh, q3, q4,
188 &q4,
14 vp8_mbloop_filter_neon( uint8x16_t qblimit, uint8x16_t qlimit, uint8x16_t qthresh, uint8x16_t q3, uint8x16_t q4, uint8x16_t q5, uint8x16_t q6, uint8x16_t q7, uint8x16_t q8, uint8x16_t q9, uint8x16_t q10, uint8x16_t *q4r, uint8x16_t *q5r, uint8x16_t *q6r, uint8x16_t *q7r, uint8x16_t *q8r, uint8x16_t *q9r) argument
212 uint8x16_t qblimit, qlimit, qthresh, q3, q4; local
302 uint8x16_t qblimit, qlimit, qthresh, q3, q4; local
469 uint8x16_t qblimit, qlimit, qthresh, q3, q4; local
[all...]
/external/llvm/test/MC/ARM/
H A Dneon-bitwise-encoding.s110 veor q4, q7, q3
111 veor.8 q4, q7, q3
112 veor.16 q4, q7, q3
113 veor.32 q4, q7, q3
114 veor.64 q4, q7, q3
116 veor.i8 q4, q7, q3
117 veor.i16 q4, q7, q3
118 veor.i32 q4, q7, q3
119 veor.i64 q4, q7, q3
121 veor.s8 q4, q
[all...]
H A Dneon-v8.s5 vmaxnm.f32 q2, q4, q6
6 @ CHECK: vmaxnm.f32 q2, q4, q6 @ encoding: [0x5c,0x4f,0x08,0xf3]
16 vcvta.s32.f32 q4, q6
17 @ CHECK: vcvta.s32.f32 q4, q6 @ encoding: [0x4c,0x80,0xbb,0xf3]
18 vcvta.u32.f32 q4, q10
19 @ CHECK: vcvta.u32.f32 q4, q10 @ encoding: [0xe4,0x80,0xbb,0xf3]
43 vcvtp.s32.f32 q4, q15
44 @ CHECK: vcvtp.s32.f32 q4, q15 @ encoding: [0x6e,0x82,0xbb,0xf3]
50 vrintn.f32 q1, q4
51 @ CHECK: vrintn.f32 q1, q4
[all...]
H A Dthumb-neon-v8.s5 vmaxnm.f32 q2, q4, q6
6 @ CHECK: vmaxnm.f32 q2, q4, q6 @ encoding: [0x08,0xff,0x5c,0x4f]
16 vcvta.s32.f32 q4, q6
17 @ CHECK: vcvta.s32.f32 q4, q6 @ encoding: [0xbb,0xff,0x4c,0x80]
18 vcvta.u32.f32 q4, q10
19 @ CHECK: vcvta.u32.f32 q4, q10 @ encoding: [0xbb,0xff,0xe4,0x80]
43 vcvtp.s32.f32 q4, q15
44 @ CHECK: vcvtp.s32.f32 q4, q15 @ encoding: [0xbb,0xff,0x6e,0x82]
50 vrintn.f32 q1, q4
51 @ CHECK: vrintn.f32 q1, q4
[all...]
H A Ddiagnostics-noneon.s5 vmov q4, q5
H A Dneon-shiftaccum-encoding.s10 vsra.s64 q8, q4, #64
18 vsra.u64 q4, q5, #25
28 vsra.s64 q4, #64
45 @ CHECK: vsra.s64 q8, q4, #64 @ encoding: [0xd8,0x01,0xc0,0xf2]
53 @ CHECK: vsra.u64 q4, q5, #25 @ encoding: [0xda,0x81,0xa7,0xf3]
62 @ CHECK: vsra.s64 q4, q4, #64 @ encoding: [0xd8,0x81,0x80,0xf2]
82 vrsra.s32 q3, q4, #32
83 vrsra.s64 q4, q5, #64
100 vrsra.s32 q4, #3
[all...]
H A Dneont2-shiftaccum-encoding.s12 vsra.s64 q8, q4, #64
20 vsra.u64 q4, q5, #25
30 vsra.s64 q4, #64
47 @ CHECK: vsra.s64 q8, q4, #64 @ encoding: [0xc0,0xef,0xd8,0x01]
55 @ CHECK: vsra.u64 q4, q5, #25 @ encoding: [0xa7,0xff,0xda,0x81]
64 @ CHECK: vsra.s64 q4, q4, #64 @ encoding: [0x80,0xef,0xd8,0x81]
85 vrsra.s32 q3, q4, #32
86 vrsra.s64 q4, q5, #64
103 vrsra.s32 q4, #3
[all...]
H A Dbasic-arm-instructions-v8.1a.s15 vqrdmlsh.f32 q3, q4, q5
25 //CHECK-ERROR: vqrdmlsh.f32 q3, q4, q5
37 //CHECK-V8: vqrdmlsh.f32 q3, q4, q5
93 vqrdmlsh.s32 q3, q4, q5
94 //CHECK-V81aARM: vqrdmlsh.s32 q3, q4, q5 @ encoding: [0x5a,0x6c,0x28,0xf3]
95 //CHECK-V81aTHUMB: vqrdmlsh.s32 q3, q4, q5 @ encoding: [0x28,0xff,0x5a,0x6c]
97 //CHECK-V8: vqrdmlsh.s32 q3, q4, q5
103 vqrdmlsh.f32 q3, q4, d5[1]
113 //CHECK-ERROR: vqrdmlsh.f32 q3, q4, d5[1]
H A Dneon-shift-encoding.s116 vsra.s64 q4, q5, #63
122 vsra.s8 q4, #7
134 @ CHECK: vsra.s64 q4, q5, #63 @ encoding: [0xda,0x81,0x81,0xf2]
139 @ CHECK: vsra.s8 q4, q4, #7 @ encoding: [0x58,0x81,0x89,0xf2]
152 vsra.u64 q4, q5, #63
158 vsra.u8 q4, #7
170 @ CHECK: vsra.u64 q4, q5, #63 @ encoding: [0xda,0x81,0x81,0xf3]
175 @ CHECK: vsra.u8 q4, q4, #
[all...]
H A Ddirective-fpu-instrs.s10 vmov q4, q11 @ v4si
/external/freetype/src/base/
H A Dftbbox.c253 FT_Pos q4 )
271 FT_ABS( q4 ) ) );
282 q4 <<= shift;
289 q4 >>= -shift;
297 if ( q1 + q2 > q3 + q4 ) /* first half */
299 q4 = q4 + q3;
302 q4 = q4 + q3;
304 q4
[all...]
/external/libgdx/extensions/gdx-freetype/jni/freetype-2.6.2/src/base/
H A Dftbbox.c253 FT_Pos q4 )
271 FT_ABS( q4 ) ) );
282 q4 <<= shift;
289 q4 >>= -shift;
297 if ( q1 + q2 > q3 + q4 ) /* first half */
299 q4 = q4 + q3;
302 q4 = q4 + q3;
304 q4
[all...]
/external/pdfium/third_party/freetype/src/base/
H A Dftbbox.c253 FT_Pos q4 )
271 FT_ABS( q4 ) ) );
282 q4 <<= shift;
289 q4 >>= -shift;
297 if ( q1 + q2 > q3 + q4 ) /* first half */
299 q4 = q4 + q3;
302 q4 = q4 + q3;
304 q4
[all...]
/external/clang/test/SemaTemplate/
H A Dinstantiate-exception-spec.cpp24 void (*q4)() throw(T) = p2; // ok
/external/libavc/common/arm/
H A Dih264_inter_pred_filters_luma_horz_a9q.s127 vaddl.u8 q4, d31, d2 @// a0 + a5 (column1,row0)
135 vmlal.u8 q4, d31, d1 @// a0 + a5 + 20a2 (column1,row0)
143 vmlal.u8 q4, d31, d1 @// a0 + a5 + 20a2 + 20a3 (column1,row0)
151 vmlsl.u8 q4, d31, d0 @// a0 + a5 + 20a2 + 20a3 - 5a1 (column1,row0)
159 vmlsl.u8 q4, d31, d0 @// a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 (column1,row0)
165 vqrshrun.s16 d20, q4, #5 @// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column1,row0)
195 vaddl.u8 q4, d31, d2 @// a0 + a5 (column1,row0)
198 vmlal.u8 q4, d29, d1 @// a0 + a5 + 20a2 + 20a3 (column1,row0)
199 vmlal.u8 q4, d30, d1 @// a0 + a5 + 20a2 (column1,row0)
200 vmlsl.u8 q4, d2
[all...]
H A Dih264_inter_pred_luma_horz_qpel_a9q.s134 vaddl.u8 q4, d31, d2 @// a0 + a5 (column1,row0)
142 vmlal.u8 q4, d31, d1 @// a0 + a5 + 20a2 (column1,row0)
150 vmlal.u8 q4, d31, d1 @// a0 + a5 + 20a2 + 20a3 (column1,row0)
158 vmlsl.u8 q4, d31, d0 @// a0 + a5 + 20a2 + 20a3 - 5a1 (column1,row0)
166 vmlsl.u8 q4, d31, d0 @// a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 (column1,row0)
173 vqrshrun.s16 d20, q4, #5 @// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column1,row0)
207 vaddl.u8 q4, d31, d2 @// a0 + a5 (column1,row0)
210 vmlal.u8 q4, d29, d1 @// a0 + a5 + 20a2 + 20a3 (column1,row0)
211 vmlal.u8 q4, d30, d1 @// a0 + a5 + 20a2 (column1,row0)
212 vmlsl.u8 q4, d2
[all...]
/external/valgrind/none/tests/arm/
H A Dneon128.c439 TESTINSN_bin("vand q4, q6, q5", q4, q6, i8, 0xff, q5, i16, 0x57);
445 TESTINSN_bin("vbic q4, q6, q5", q4, q6, i8, 0xff, q5, i16, 0x57);
452 TESTINSN_bin("vorr q4, q4, q4", q4, q4, i16, 0xff, q4, i1
[all...]
/external/libhevc/common/arm/
H A Dihevc_inter_pred_filters_luma_horz.s216 vmull.u8 q4,d1,d25 @mul_res = vmlal_u8(src[0_1], coeffabs_1)@
218 vmlal.u8 q4,d3,d27 @mul_res = vmull_u8(src[0_3], coeffabs_3)@
220 vmlsl.u8 q4,d0,d24 @mul_res = vmlsl_u8(src[0_0], coeffabs_0)@
222 vmlsl.u8 q4,d2,d26 @mul_res = vmlsl_u8(src[0_2], coeffabs_2)@
224 vmlal.u8 q4,d4,d28 @mul_res = vmlal_u8(src[0_4], coeffabs_4)@
226 vmlsl.u8 q4,d5,d29 @mul_res = vmlsl_u8(src[0_5], coeffabs_5)@
228 vmlal.u8 q4,d6,d30 @mul_res = vmlal_u8(src[0_6], coeffabs_6)@
230 vmlsl.u8 q4,d7,d31 @mul_res = vmlsl_u8(src[0_7], coeffabs_7)@
240 vqrshrun.s16 d20,q4,#6 @right shift and saturating narrow result 1
296 vmull.u8 q4,d
[all...]
H A Dihevc_inter_pred_luma_vert_w16inp_w16out.s158 vmull.s16 q4,d1,d23 @mul_res1 = vmull_u8(src_tmp2, coeffabs_1)@
160 vmlal.s16 q4,d0,d22 @mul_res1 = vmlal_u8(mul_res1, src_tmp1, coeffabs_0)@
162 vmlal.s16 q4,d2,d24 @mul_res1 = vmlal_u8(mul_res1, src_tmp3, coeffabs_2)@
164 vmlal.s16 q4,d3,d25 @mul_res1 = vmlal_u8(mul_res1, src_tmp4, coeffabs_3)@
166 vmlal.s16 q4,d4,d26 @mul_res1 = vmlal_u8(mul_res1, src_tmp1, coeffabs_4)@
168 vmlal.s16 q4,d5,d27 @mul_res1 = vmlal_u8(mul_res1, src_tmp2, coeffabs_5)@
169 vmlal.s16 q4,d6,d28 @mul_res1 = vmlal_u8(mul_res1, src_tmp3, coeffabs_6)@
170 vmlal.s16 q4,d7,d29 @mul_res1 = vmlal_u8(mul_res1, src_tmp4, coeffabs_7)@
187 vsub.s32 q4, q4, q1
[all...]
H A Dihevc_intra_pred_chroma_horz.s130 vdup.16 q4,d1[0]
139 vst1.16 {q4},[r2],r3
140 vst1.16 {q4},[r9],r3
146 vdup.16 q4,d0[0]
155 vst1.16 {q4},[r2],r3
156 vst1.16 {q4},[r9],r3
162 vdup.16 q4,d11[0]
171 vst1.16 {q4},[r2],r3
172 vst1.16 {q4},[r9],r3
179 vdup.16 q4,d1
[all...]
H A Dihevc_intra_pred_luma_horz.s128 vdup.8 q4,d1[4]
137 vst1.8 {q4},[r2],r3
138 vst1.8 {q4},[r9],r3
144 vdup.8 q4,d1[0]
153 vst1.8 {q4},[r2],r3
154 vst1.8 {q4},[r9],r3
160 vdup.8 q4,d0[4]
169 vst1.8 {q4},[r2],r3
170 vst1.8 {q4},[r9],r3
177 vdup.8 q4,d
[all...]
H A Dihevc_inter_pred_luma_horz_w16out.s253 vmull.u8 q4,d1,d25 @arithmetic operations for ii iteration in the same time
254 vmlsl.u8 q4,d0,d24
255 vmlsl.u8 q4,d2,d26
256 vmlal.u8 q4,d3,d27
257 vmlal.u8 q4,d4,d28
258 vmlsl.u8 q4,d5,d29
259 vmlal.u8 q4,d6,d30
260 vmlsl.u8 q4,d7,d31
262 @ vqrshrun.s16 d8,q4,#6 @narrow right shift and saturating the result
310 vmull.u8 q4,d
[all...]
/external/boringssl/src/crypto/curve25519/asm/
H A Dx25519-asm-arm.S28 vpush {q4,q5,q6,q7}
105 vand q4,q4,q2
116 vadd.i64 q12,q4,q1
126 vsub.i64 q4,q4,q12
162 vadd.i64 q4,q4,q14
167 vadd.i64 q4,q4,q
[all...]

Completed in 1248 milliseconds

123456