/external/llvm/test/MC/ARM/ |
H A D | neon-shiftaccum-encoding.s | 9 vsra.s32 q9, q5, #32 44 @ CHECK: vsra.s32 q9, q5, #32 @ encoding: [0x5a,0x21,0xe0,0xf2] 87 vrsra.u64 q8, q9, #64 105 vrsra.u64 q9, #64 122 @ CHECK: vrsra.u64 q8, q9, #64 @ encoding: [0xf2,0x03,0xc0,0xf3] 139 @ CHECK: vrsra.u64 q9, q9, #64 @ encoding: [0xf2,0x23,0xc0,0xf3] 157 vsri.64 q9, q6, #64 192 @ CHECK: vsri.64 q9, q6, #64 @ encoding: [0xdc,0x24,0xc0,0xf3]
|
H A D | neont2-shiftaccum-encoding.s | 11 vsra.s32 q9, q5, #32 46 @ CHECK: vsra.s32 q9, q5, #32 @ encoding: [0xe0,0xef,0x5a,0x21] 90 vrsra.u64 q8, q9, #64 108 vrsra.u64 q9, #64 125 @ CHECK: vrsra.u64 q8, q9, #64 @ encoding: [0xc0,0xff,0xf2,0x03] 142 @ CHECK: vrsra.u64 q9, q9, #64 @ encoding: [0xc0,0xff,0xf2,0x23] 160 vsri.64 q9, q6, #64 195 @ CHECK: vsri.64 q9, q6, #64 @ encoding: [0xc0,0xff,0xdc,0x24]
|
H A D | neon-bitwise-encoding.s | 5 vand q8, q8, q9 8 @ CHECK: vand q8, q8, q9 @ encoding: [0xf2,0x01,0x40,0xf2] 11 veor q8, q8, q9 14 @ CHECK: veor q8, q8, q9 @ encoding: [0xf2,0x01,0x40,0xf3] 17 vorr q8, q8, q9 20 @ CHECK: vorr q8, q8, q9 @ encoding: [0xf2,0x01,0x60,0xf2] 31 vbic q8, q8, q9 38 @ CHECK: vbic q8, q8, q9 @ encoding: [0xf2,0x01,0x50,0xf2] 46 vorn q8, q8, q9 49 @ CHECK: vorn q8, q8, q9 [all...] |
H A D | neont2-pairwise-encoding.s | 26 vpaddl.u32 q9, q2 39 @ CHECK: vpaddl.u32 q9, q2 @ encoding: [0xf8,0xff,0xc4,0x22] 53 vpadal.u32 q9, q15 66 @ CHECK: vpadal.u32 q9, q15 @ encoding: [0xf8,0xff,0xee,0x26]
|
/external/libhevc/common/arm/ |
H A D | ihevc_inter_pred_luma_copy_w16out.s | 153 vmovl.u8 q9,d10 @vmovl_u8(vld1_u8(pu1_src_tmp) 158 vshl.i16 q1,q9,#6 @vshlq_n_s16(tmp, 6) 185 vmovl.u8 q9,d10 @vmovl_u8(vld1_u8(pu1_src_tmp) 201 vshl.i16 q1,q9,#6 @vshlq_n_s16(tmp, 6) 224 vmovl.u8 q9,d10 @vmovl_u8(vld1_u8(pu1_src_tmp) 233 vshl.i16 q1,q9,#6 @vshlq_n_s16(tmp, 6)
|
H A D | ihevc_deblk_chroma_vert.s | 96 vmovl.u8 q9,d17 140 vsub.i16 q1,q9,q1
|
H A D | ihevc_inter_pred_chroma_copy_w16out.s | 213 vmovl.u8 q9,d10 @vmovl_u8(vld1_u8(pu1_src_tmp) 218 vshl.i16 q1,q9,#6 @vshlq_n_s16(tmp, 6) 245 vmovl.u8 q9,d10 @vmovl_u8(vld1_u8(pu1_src_tmp) 261 vshl.i16 q1,q9,#6 @vshlq_n_s16(tmp, 6) 284 vmovl.u8 q9,d10 @vmovl_u8(vld1_u8(pu1_src_tmp) 293 vshl.i16 q1,q9,#6 @vshlq_n_s16(tmp, 6) 311 vmovl.u8 q9,d10 @vmovl_u8(vld1_u8(pu1_src_tmp) 314 vshl.i16 q1,q9,#6 @vshlq_n_s16(tmp, 6)
|
H A D | ihevc_itrans_recon_4x4.s | 165 vsub.s32 q9,q6,q4 @((e[1] - o[1]) 170 vqrshrn.s32 d2,q9,#shift_stage1_idct @pi2_out[2] = clip_s16((e[0] - o[0] + add)>>shift) ) 196 vsub.s32 q9,q6,q4 @((e[1] - o[1]) 201 vqrshrn.s32 d2,q9,#shift_stage2_idct @pi2_out[2] = clip_s16((e[0] - o[0] + add)>>shift) )
|
H A D | ihevc_deblk_luma_vert.s | 439 vsubl.u8 q9,d4,d2 440 vmul.i16 q9,q9,q0 446 vsub.i16 q8,q9,q8 469 vmovl.u8 q9,d2 470 vadd.i16 q9,q9,q8 472 vqmovun.s16 d22,q9 473 vmovl.u8 q9,d4 474 vsub.i16 q8,q9,q [all...] |
H A D | ihevc_intra_pred_chroma_mode_27_to_33.s | 201 vmull.u8 q9,d16,d26 @(iii)vmull_u8(ref_main_idx, dup_const_32_fract) 204 vmlal.u8 q9,d17,d27 @(iii)vmull_u8(ref_main_idx_1, dup_const_fract) 225 vrshrn.i16 d18,q9,#5 @(iii)shift_res = vrshrn_n_u16(add_res, 5) 265 vmull.u8 q9,d16,d26 @(vii)vmull_u8(ref_main_idx, dup_const_32_fract) 268 vmlal.u8 q9,d17,d27 @(vii)vmull_u8(ref_main_idx_1, dup_const_fract) 305 vrshrn.i16 d18,q9,#5 @(vii)shift_res = vrshrn_n_u16(add_res, 5) 344 vmull.u8 q9,d16,d26 @(iii)vmull_u8(ref_main_idx, dup_const_32_fract) 348 vmlal.u8 q9,d17,d27 @(iii)vmull_u8(ref_main_idx_1, dup_const_fract) 368 vrshrn.i16 d18,q9,#5 @(iii)shift_res = vrshrn_n_u16(add_res, 5) 414 vmull.u8 q9,d1 [all...] |
H A D | ihevc_intra_pred_filters_chroma_mode_19_to_25.s | 309 vmull.u8 q9,d16,d26 @(iii)vmull_u8(ref_main_idx, dup_const_32_fract) 312 vmlal.u8 q9,d17,d27 @(iii)vmull_u8(ref_main_idx_1, dup_const_fract) 332 vrshrn.i16 d18,q9,#5 @(iii)shift_res = vrshrn_n_u16(add_res, 5) 373 vmull.u8 q9,d16,d26 @(vii)vmull_u8(ref_main_idx, dup_const_32_fract) 376 vmlal.u8 q9,d17,d27 @(vii)vmull_u8(ref_main_idx_1, dup_const_fract) 413 vrshrn.i16 d18,q9,#5 @(vii)shift_res = vrshrn_n_u16(add_res, 5) 454 vmull.u8 q9,d16,d26 @(iii)vmull_u8(ref_main_idx, dup_const_32_fract) 458 vmlal.u8 q9,d17,d27 @(iii)vmull_u8(ref_main_idx_1, dup_const_fract) 482 vrshrn.i16 d18,q9,#5 @(iii)shift_res = vrshrn_n_u16(add_res, 5) 530 vmull.u8 q9,d1 [all...] |
H A D | ihevc_intra_pred_filters_luma_mode_19_to_25.s | 312 vmull.u8 q9,d16,d26 @(iii)vmull_u8(ref_main_idx, dup_const_32_fract) 315 vmlal.u8 q9,d17,d27 @(iii)vmull_u8(ref_main_idx_1, dup_const_fract) 334 vrshrn.i16 d18,q9,#5 @(iii)shift_res = vrshrn_n_u16(add_res, 5) 372 vmull.u8 q9,d16,d26 @(vii)vmull_u8(ref_main_idx, dup_const_32_fract) 375 vmlal.u8 q9,d17,d27 @(vii)vmull_u8(ref_main_idx_1, dup_const_fract) 410 vrshrn.i16 d18,q9,#5 @(vii)shift_res = vrshrn_n_u16(add_res, 5) 451 vmull.u8 q9,d16,d26 @(iii)vmull_u8(ref_main_idx, dup_const_32_fract) 454 vmlal.u8 q9,d17,d27 @(iii)vmull_u8(ref_main_idx_1, dup_const_fract) 478 vrshrn.i16 d18,q9,#5 @(iii)shift_res = vrshrn_n_u16(add_res, 5) 524 vmull.u8 q9,d1 [all...] |
H A D | ihevc_intra_pred_luma_mode_27_to_33.s | 204 vmull.u8 q9,d16,d26 @(iii)vmull_u8(ref_main_idx, dup_const_32_fract) 207 vmlal.u8 q9,d17,d27 @(iii)vmull_u8(ref_main_idx_1, dup_const_fract) 227 vrshrn.i16 d18,q9,#5 @(iii)shift_res = vrshrn_n_u16(add_res, 5) 267 vmull.u8 q9,d16,d26 @(vii)vmull_u8(ref_main_idx, dup_const_32_fract) 270 vmlal.u8 q9,d17,d27 @(vii)vmull_u8(ref_main_idx_1, dup_const_fract) 306 vrshrn.i16 d18,q9,#5 @(vii)shift_res = vrshrn_n_u16(add_res, 5) 345 vmull.u8 q9,d16,d26 @(iii)vmull_u8(ref_main_idx, dup_const_32_fract) 349 vmlal.u8 q9,d17,d27 @(iii)vmull_u8(ref_main_idx_1, dup_const_fract) 369 vrshrn.i16 d18,q9,#5 @(iii)shift_res = vrshrn_n_u16(add_res, 5) 414 vmull.u8 q9,d1 [all...] |
H A D | ihevc_intra_pred_filters_chroma_mode_11_to_17.s | 353 vmull.u8 q9, d10, d7 @mul (row 3) 354 vmlal.u8 q9, d11, d6 @mul (row 3) 361 vrshrn.i16 d18, q9, #5 @round shft (row 3) 392 vmull.u8 q9, d10, d7 @mul (row 7) 393 vmlal.u8 q9, d11, d6 @mul (row 7) 397 vrshrn.i16 d18, q9, #5 @round shft (row 7) 478 vmull.u8 q9, d10, d7 @mul (row 7) 480 vmlal.u8 q9, d11, d6 @mul (row 7) 504 vrshrn.i16 d18, q9, #5 @(from previous loop)round shft (row 7) 537 vmull.u8 q9, d1 [all...] |
/external/libvpx/libvpx/vp8/common/arm/neon/ |
H A D | sixtappredict4x4_neon.asm | 93 vshr.u64 q9, q4, #8 ;construct src_ptr[-1] 107 vshr.u64 q9, q4, #16 ;construct src_ptr[0] 121 vmull.u8 q9, d6, d3 ;(src_ptr[1] * vp8_filter[3]) 127 vqadd.s16 q7, q9 ;sum of all (src_data*filter_parameters) 159 vshr.u64 q9, q4, #8 ;construct src_ptr[-1] 178 vshr.u64 q9, q4, #16 ;construct src_ptr[0] 199 vmull.u8 q9, d6, d3 ;(src_ptr[1] * vp8_filter[3]) 205 vqadd.s16 q7, q9 ;sum of all (src_data*filter_parameters) 305 vshr.u64 q9, q4, #8 ;construct src_ptr[-1] 319 vshr.u64 q9, q [all...] |
/external/chromium_org/third_party/libvpx/source/libvpx/vp9/common/arm/neon/ |
H A D | vp9_reconintra_neon.asm | 371 vdup.16 q9, d20[3] 373 vadd.s16 q9, q3, q9 378 vqmovun.s16 d3, q9 393 vdup.16 q9, d21[3] 395 vadd.s16 q9, q3, q9 400 vqmovun.s16 d3, q9 527 vsubl.u8 q9, d3, d1 542 vadd.s16 q13, q0, q9 [all...] |
/external/libvpx/libvpx/vp9/common/arm/neon/ |
H A D | vp9_reconintra_neon.asm | 371 vdup.16 q9, d20[3] 373 vadd.s16 q9, q3, q9 378 vqmovun.s16 d3, q9 393 vdup.16 q9, d21[3] 395 vadd.s16 q9, q3, q9 400 vqmovun.s16 d3, q9 527 vsubl.u8 q9, d3, d1 542 vadd.s16 q13, q0, q9 [all...] |
/external/libvpx/libvpx/vpx_scale/arm/neon/ |
H A D | vp8_vpxyv12_copyframe_func_neon.asm | 61 vld1.8 {q8, q9}, [r10]! 72 vst1.8 {q8, q9}, [r11]! 121 vld1.8 {q8, q9}, [r10]! 128 vst1.8 {q8, q9}, [r11]!
|
H A D | vp8_vpxyv12_extendframeborders_neon.asm | 55 vmov q9, q8 66 vst1.8 {q8, q9}, [r5], lr 88 vld1.8 {q8, q9}, [r2]! 102 vst1.8 {q8, q9}, [r6]! 176 vst1.8 {q9}, [r6], lr 199 vld1.8 {q8, q9}, [r2]! 209 vst1.8 {q8, q9}, [r6]!
|
/external/valgrind/main/none/tests/arm/ |
H A D | neon128.c | 707 TESTINSN_bin("vshl.s32 q9, q12, q15", q9, q12, i32, (1 << 31) + 2, q15, i32, 2); 719 TESTINSN_bin("vshl.u32 q9, q12, q15", q9, q12, i32, (1 << 31) + 2, q15, i32, 2); 737 TESTINSN_bin_q("vqshl.s32 q9, q10, q11", q9, q10, i32, (1 << 31) + 8, q11, i32, -1); 761 TESTINSN_bin_q("vqshl.u32 q9, q10, q11", q9, q10, i32, (1 << 31) + 8, q11, i32, -1); 799 TESTINSN_un_q("vqshl.s16 q9, q8, #1", q9, q [all...] |
/external/libvpx/libvpx/vp8/encoder/arm/neon/ |
H A D | subtract_neon.asm | 84 vsubl.u8 q9, d1, d3 93 vst1.16 {q9}, [r5], r6 141 vsubl.u8 q9, d2, d3 150 vst1.16 {q9}, [r7], r12 177 vsubl.u8 q9, d2, d3 186 vst1.16 {q9}, [r7], r12
|
H A D | fastquantizeb_neon.asm | 45 vld1.s16 {q8, q9}, [r5@128] ; load quant_ptr [0-15] 55 vqdmulh.s16 q5, q9 72 vqdmulh.s16 q11, q9 179 vld1.s16 {q8, q9}, [r4@128] ; load quant_ptr [0-15] 187 vqdmulh.s16 q13, q9
|
/external/chromium_org/third_party/libvpx/source/libvpx/vp8/encoder/arm/neon/ |
H A D | fastquantizeb_neon.asm | 45 vld1.s16 {q8, q9}, [r5@128] ; load quant_ptr [0-15] 55 vqdmulh.s16 q5, q9 72 vqdmulh.s16 q11, q9 179 vld1.s16 {q8, q9}, [r4@128] ; load quant_ptr [0-15] 187 vqdmulh.s16 q13, q9
|
/external/chromium_org/third_party/boringssl/src/crypto/poly1305/ |
H A D | poly1305_arm_asm.S | 760 # asm 2: vand.i32 <v23=q9,#0x03ffffff 761 vand.i32 q9,#0x03ffffff 990 # asm 2: vshr.u64 >t0=q9,<x4=q7,#26 991 vshr.u64 q9,q7,#26 1005 # asm 2: vadd.i64 >x01=q5,<r0=q5,<t0=q9 1006 vadd.i64 q5,q5,q9 1020 # asm 2: vshl.i64 >t0=q9,<t0=q9,#2 1021 vshl.i64 q9,q9,# [all...] |
/external/jpeg/ |
H A D | jsimd_arm_neon.S | 208 vmul.s16 q9, q9, q11 211 idct_helper q2, q3, q4, q5, q6, q7, q8, q9, q10, q11, q12, q13, q14 219 vtrn.16 q8, q9 221 vtrn.32 q7, q9 229 idct_helper q2, q3, q4, q5, q6, q7, q8, q9, q10, q11, q12, q13, q14 238 vtrn.16 q8, q9 240 vtrn.32 q7, q9 256 vqadd.s16 q9, q9, q1 [all...] |