Searched refs:q9 (Results 76 - 100 of 143) sorted by relevance

123456

/external/llvm/test/MC/ARM/
H A Dneon-shiftaccum-encoding.s9 vsra.s32 q9, q5, #32
44 @ CHECK: vsra.s32 q9, q5, #32 @ encoding: [0x5a,0x21,0xe0,0xf2]
87 vrsra.u64 q8, q9, #64
105 vrsra.u64 q9, #64
122 @ CHECK: vrsra.u64 q8, q9, #64 @ encoding: [0xf2,0x03,0xc0,0xf3]
139 @ CHECK: vrsra.u64 q9, q9, #64 @ encoding: [0xf2,0x23,0xc0,0xf3]
157 vsri.64 q9, q6, #64
192 @ CHECK: vsri.64 q9, q6, #64 @ encoding: [0xdc,0x24,0xc0,0xf3]
H A Dneont2-shiftaccum-encoding.s11 vsra.s32 q9, q5, #32
46 @ CHECK: vsra.s32 q9, q5, #32 @ encoding: [0xe0,0xef,0x5a,0x21]
90 vrsra.u64 q8, q9, #64
108 vrsra.u64 q9, #64
125 @ CHECK: vrsra.u64 q8, q9, #64 @ encoding: [0xc0,0xff,0xf2,0x03]
142 @ CHECK: vrsra.u64 q9, q9, #64 @ encoding: [0xc0,0xff,0xf2,0x23]
160 vsri.64 q9, q6, #64
195 @ CHECK: vsri.64 q9, q6, #64 @ encoding: [0xc0,0xff,0xdc,0x24]
H A Dneon-bitwise-encoding.s5 vand q8, q8, q9
8 @ CHECK: vand q8, q8, q9 @ encoding: [0xf2,0x01,0x40,0xf2]
11 veor q8, q8, q9
14 @ CHECK: veor q8, q8, q9 @ encoding: [0xf2,0x01,0x40,0xf3]
17 vorr q8, q8, q9
20 @ CHECK: vorr q8, q8, q9 @ encoding: [0xf2,0x01,0x60,0xf2]
31 vbic q8, q8, q9
38 @ CHECK: vbic q8, q8, q9 @ encoding: [0xf2,0x01,0x50,0xf2]
46 vorn q8, q8, q9
49 @ CHECK: vorn q8, q8, q9
[all...]
H A Dneont2-pairwise-encoding.s26 vpaddl.u32 q9, q2
39 @ CHECK: vpaddl.u32 q9, q2 @ encoding: [0xf8,0xff,0xc4,0x22]
53 vpadal.u32 q9, q15
66 @ CHECK: vpadal.u32 q9, q15 @ encoding: [0xf8,0xff,0xee,0x26]
/external/libhevc/common/arm/
H A Dihevc_inter_pred_luma_copy_w16out.s153 vmovl.u8 q9,d10 @vmovl_u8(vld1_u8(pu1_src_tmp)
158 vshl.i16 q1,q9,#6 @vshlq_n_s16(tmp, 6)
185 vmovl.u8 q9,d10 @vmovl_u8(vld1_u8(pu1_src_tmp)
201 vshl.i16 q1,q9,#6 @vshlq_n_s16(tmp, 6)
224 vmovl.u8 q9,d10 @vmovl_u8(vld1_u8(pu1_src_tmp)
233 vshl.i16 q1,q9,#6 @vshlq_n_s16(tmp, 6)
H A Dihevc_deblk_chroma_vert.s96 vmovl.u8 q9,d17
140 vsub.i16 q1,q9,q1
H A Dihevc_inter_pred_chroma_copy_w16out.s213 vmovl.u8 q9,d10 @vmovl_u8(vld1_u8(pu1_src_tmp)
218 vshl.i16 q1,q9,#6 @vshlq_n_s16(tmp, 6)
245 vmovl.u8 q9,d10 @vmovl_u8(vld1_u8(pu1_src_tmp)
261 vshl.i16 q1,q9,#6 @vshlq_n_s16(tmp, 6)
284 vmovl.u8 q9,d10 @vmovl_u8(vld1_u8(pu1_src_tmp)
293 vshl.i16 q1,q9,#6 @vshlq_n_s16(tmp, 6)
311 vmovl.u8 q9,d10 @vmovl_u8(vld1_u8(pu1_src_tmp)
314 vshl.i16 q1,q9,#6 @vshlq_n_s16(tmp, 6)
H A Dihevc_itrans_recon_4x4.s165 vsub.s32 q9,q6,q4 @((e[1] - o[1])
170 vqrshrn.s32 d2,q9,#shift_stage1_idct @pi2_out[2] = clip_s16((e[0] - o[0] + add)>>shift) )
196 vsub.s32 q9,q6,q4 @((e[1] - o[1])
201 vqrshrn.s32 d2,q9,#shift_stage2_idct @pi2_out[2] = clip_s16((e[0] - o[0] + add)>>shift) )
H A Dihevc_deblk_luma_vert.s439 vsubl.u8 q9,d4,d2
440 vmul.i16 q9,q9,q0
446 vsub.i16 q8,q9,q8
469 vmovl.u8 q9,d2
470 vadd.i16 q9,q9,q8
472 vqmovun.s16 d22,q9
473 vmovl.u8 q9,d4
474 vsub.i16 q8,q9,q
[all...]
H A Dihevc_intra_pred_chroma_mode_27_to_33.s201 vmull.u8 q9,d16,d26 @(iii)vmull_u8(ref_main_idx, dup_const_32_fract)
204 vmlal.u8 q9,d17,d27 @(iii)vmull_u8(ref_main_idx_1, dup_const_fract)
225 vrshrn.i16 d18,q9,#5 @(iii)shift_res = vrshrn_n_u16(add_res, 5)
265 vmull.u8 q9,d16,d26 @(vii)vmull_u8(ref_main_idx, dup_const_32_fract)
268 vmlal.u8 q9,d17,d27 @(vii)vmull_u8(ref_main_idx_1, dup_const_fract)
305 vrshrn.i16 d18,q9,#5 @(vii)shift_res = vrshrn_n_u16(add_res, 5)
344 vmull.u8 q9,d16,d26 @(iii)vmull_u8(ref_main_idx, dup_const_32_fract)
348 vmlal.u8 q9,d17,d27 @(iii)vmull_u8(ref_main_idx_1, dup_const_fract)
368 vrshrn.i16 d18,q9,#5 @(iii)shift_res = vrshrn_n_u16(add_res, 5)
414 vmull.u8 q9,d1
[all...]
H A Dihevc_intra_pred_filters_chroma_mode_19_to_25.s309 vmull.u8 q9,d16,d26 @(iii)vmull_u8(ref_main_idx, dup_const_32_fract)
312 vmlal.u8 q9,d17,d27 @(iii)vmull_u8(ref_main_idx_1, dup_const_fract)
332 vrshrn.i16 d18,q9,#5 @(iii)shift_res = vrshrn_n_u16(add_res, 5)
373 vmull.u8 q9,d16,d26 @(vii)vmull_u8(ref_main_idx, dup_const_32_fract)
376 vmlal.u8 q9,d17,d27 @(vii)vmull_u8(ref_main_idx_1, dup_const_fract)
413 vrshrn.i16 d18,q9,#5 @(vii)shift_res = vrshrn_n_u16(add_res, 5)
454 vmull.u8 q9,d16,d26 @(iii)vmull_u8(ref_main_idx, dup_const_32_fract)
458 vmlal.u8 q9,d17,d27 @(iii)vmull_u8(ref_main_idx_1, dup_const_fract)
482 vrshrn.i16 d18,q9,#5 @(iii)shift_res = vrshrn_n_u16(add_res, 5)
530 vmull.u8 q9,d1
[all...]
H A Dihevc_intra_pred_filters_luma_mode_19_to_25.s312 vmull.u8 q9,d16,d26 @(iii)vmull_u8(ref_main_idx, dup_const_32_fract)
315 vmlal.u8 q9,d17,d27 @(iii)vmull_u8(ref_main_idx_1, dup_const_fract)
334 vrshrn.i16 d18,q9,#5 @(iii)shift_res = vrshrn_n_u16(add_res, 5)
372 vmull.u8 q9,d16,d26 @(vii)vmull_u8(ref_main_idx, dup_const_32_fract)
375 vmlal.u8 q9,d17,d27 @(vii)vmull_u8(ref_main_idx_1, dup_const_fract)
410 vrshrn.i16 d18,q9,#5 @(vii)shift_res = vrshrn_n_u16(add_res, 5)
451 vmull.u8 q9,d16,d26 @(iii)vmull_u8(ref_main_idx, dup_const_32_fract)
454 vmlal.u8 q9,d17,d27 @(iii)vmull_u8(ref_main_idx_1, dup_const_fract)
478 vrshrn.i16 d18,q9,#5 @(iii)shift_res = vrshrn_n_u16(add_res, 5)
524 vmull.u8 q9,d1
[all...]
H A Dihevc_intra_pred_luma_mode_27_to_33.s204 vmull.u8 q9,d16,d26 @(iii)vmull_u8(ref_main_idx, dup_const_32_fract)
207 vmlal.u8 q9,d17,d27 @(iii)vmull_u8(ref_main_idx_1, dup_const_fract)
227 vrshrn.i16 d18,q9,#5 @(iii)shift_res = vrshrn_n_u16(add_res, 5)
267 vmull.u8 q9,d16,d26 @(vii)vmull_u8(ref_main_idx, dup_const_32_fract)
270 vmlal.u8 q9,d17,d27 @(vii)vmull_u8(ref_main_idx_1, dup_const_fract)
306 vrshrn.i16 d18,q9,#5 @(vii)shift_res = vrshrn_n_u16(add_res, 5)
345 vmull.u8 q9,d16,d26 @(iii)vmull_u8(ref_main_idx, dup_const_32_fract)
349 vmlal.u8 q9,d17,d27 @(iii)vmull_u8(ref_main_idx_1, dup_const_fract)
369 vrshrn.i16 d18,q9,#5 @(iii)shift_res = vrshrn_n_u16(add_res, 5)
414 vmull.u8 q9,d1
[all...]
H A Dihevc_intra_pred_filters_chroma_mode_11_to_17.s353 vmull.u8 q9, d10, d7 @mul (row 3)
354 vmlal.u8 q9, d11, d6 @mul (row 3)
361 vrshrn.i16 d18, q9, #5 @round shft (row 3)
392 vmull.u8 q9, d10, d7 @mul (row 7)
393 vmlal.u8 q9, d11, d6 @mul (row 7)
397 vrshrn.i16 d18, q9, #5 @round shft (row 7)
478 vmull.u8 q9, d10, d7 @mul (row 7)
480 vmlal.u8 q9, d11, d6 @mul (row 7)
504 vrshrn.i16 d18, q9, #5 @(from previous loop)round shft (row 7)
537 vmull.u8 q9, d1
[all...]
/external/libvpx/libvpx/vp8/common/arm/neon/
H A Dsixtappredict4x4_neon.asm93 vshr.u64 q9, q4, #8 ;construct src_ptr[-1]
107 vshr.u64 q9, q4, #16 ;construct src_ptr[0]
121 vmull.u8 q9, d6, d3 ;(src_ptr[1] * vp8_filter[3])
127 vqadd.s16 q7, q9 ;sum of all (src_data*filter_parameters)
159 vshr.u64 q9, q4, #8 ;construct src_ptr[-1]
178 vshr.u64 q9, q4, #16 ;construct src_ptr[0]
199 vmull.u8 q9, d6, d3 ;(src_ptr[1] * vp8_filter[3])
205 vqadd.s16 q7, q9 ;sum of all (src_data*filter_parameters)
305 vshr.u64 q9, q4, #8 ;construct src_ptr[-1]
319 vshr.u64 q9, q
[all...]
/external/chromium_org/third_party/libvpx/source/libvpx/vp9/common/arm/neon/
H A Dvp9_reconintra_neon.asm371 vdup.16 q9, d20[3]
373 vadd.s16 q9, q3, q9
378 vqmovun.s16 d3, q9
393 vdup.16 q9, d21[3]
395 vadd.s16 q9, q3, q9
400 vqmovun.s16 d3, q9
527 vsubl.u8 q9, d3, d1
542 vadd.s16 q13, q0, q9
[all...]
/external/libvpx/libvpx/vp9/common/arm/neon/
H A Dvp9_reconintra_neon.asm371 vdup.16 q9, d20[3]
373 vadd.s16 q9, q3, q9
378 vqmovun.s16 d3, q9
393 vdup.16 q9, d21[3]
395 vadd.s16 q9, q3, q9
400 vqmovun.s16 d3, q9
527 vsubl.u8 q9, d3, d1
542 vadd.s16 q13, q0, q9
[all...]
/external/libvpx/libvpx/vpx_scale/arm/neon/
H A Dvp8_vpxyv12_copyframe_func_neon.asm61 vld1.8 {q8, q9}, [r10]!
72 vst1.8 {q8, q9}, [r11]!
121 vld1.8 {q8, q9}, [r10]!
128 vst1.8 {q8, q9}, [r11]!
H A Dvp8_vpxyv12_extendframeborders_neon.asm55 vmov q9, q8
66 vst1.8 {q8, q9}, [r5], lr
88 vld1.8 {q8, q9}, [r2]!
102 vst1.8 {q8, q9}, [r6]!
176 vst1.8 {q9}, [r6], lr
199 vld1.8 {q8, q9}, [r2]!
209 vst1.8 {q8, q9}, [r6]!
/external/valgrind/main/none/tests/arm/
H A Dneon128.c707 TESTINSN_bin("vshl.s32 q9, q12, q15", q9, q12, i32, (1 << 31) + 2, q15, i32, 2);
719 TESTINSN_bin("vshl.u32 q9, q12, q15", q9, q12, i32, (1 << 31) + 2, q15, i32, 2);
737 TESTINSN_bin_q("vqshl.s32 q9, q10, q11", q9, q10, i32, (1 << 31) + 8, q11, i32, -1);
761 TESTINSN_bin_q("vqshl.u32 q9, q10, q11", q9, q10, i32, (1 << 31) + 8, q11, i32, -1);
799 TESTINSN_un_q("vqshl.s16 q9, q8, #1", q9, q
[all...]
/external/libvpx/libvpx/vp8/encoder/arm/neon/
H A Dsubtract_neon.asm84 vsubl.u8 q9, d1, d3
93 vst1.16 {q9}, [r5], r6
141 vsubl.u8 q9, d2, d3
150 vst1.16 {q9}, [r7], r12
177 vsubl.u8 q9, d2, d3
186 vst1.16 {q9}, [r7], r12
H A Dfastquantizeb_neon.asm45 vld1.s16 {q8, q9}, [r5@128] ; load quant_ptr [0-15]
55 vqdmulh.s16 q5, q9
72 vqdmulh.s16 q11, q9
179 vld1.s16 {q8, q9}, [r4@128] ; load quant_ptr [0-15]
187 vqdmulh.s16 q13, q9
/external/chromium_org/third_party/libvpx/source/libvpx/vp8/encoder/arm/neon/
H A Dfastquantizeb_neon.asm45 vld1.s16 {q8, q9}, [r5@128] ; load quant_ptr [0-15]
55 vqdmulh.s16 q5, q9
72 vqdmulh.s16 q11, q9
179 vld1.s16 {q8, q9}, [r4@128] ; load quant_ptr [0-15]
187 vqdmulh.s16 q13, q9
/external/chromium_org/third_party/boringssl/src/crypto/poly1305/
H A Dpoly1305_arm_asm.S760 # asm 2: vand.i32 <v23=q9,#0x03ffffff
761 vand.i32 q9,#0x03ffffff
990 # asm 2: vshr.u64 >t0=q9,<x4=q7,#26
991 vshr.u64 q9,q7,#26
1005 # asm 2: vadd.i64 >x01=q5,<r0=q5,<t0=q9
1006 vadd.i64 q5,q5,q9
1020 # asm 2: vshl.i64 >t0=q9,<t0=q9,#2
1021 vshl.i64 q9,q9,#
[all...]
/external/jpeg/
H A Djsimd_arm_neon.S208 vmul.s16 q9, q9, q11
211 idct_helper q2, q3, q4, q5, q6, q7, q8, q9, q10, q11, q12, q13, q14
219 vtrn.16 q8, q9
221 vtrn.32 q7, q9
229 idct_helper q2, q3, q4, q5, q6, q7, q8, q9, q10, q11, q12, q13, q14
238 vtrn.16 q8, q9
240 vtrn.32 q7, q9
256 vqadd.s16 q9, q9, q1
[all...]

Completed in 3378 milliseconds

123456