/external/llvm/test/MC/ARM/ |
H A D | neont2-shuffle-encoding.s | 9 @ CHECK: vext.8 q8, q9, q8, #3 @ encoding: [0xf2,0xef,0xe0,0x03] 10 vext.8 q8, q9, q8, #3 11 @ CHECK: vext.8 q8, q9, q8, #7 @ encoding: [0xf2,0xef,0xe0,0x07] 12 vext.8 q8, q9, q8, #7 15 @ CHECK: vext.32 q8, q9, q8, #3 @ encoding: [0xf2,0xef,0xe0,0x0c] 16 vext.32 q8, q9, q8, #3 23 @ CHECK: vtrn.8 q9, q8 @ encoding: [0xf2,0xff,0xe0,0x20] 24 vtrn.8 q9, q8 25 @ CHECK: vtrn.16 q9, q8 @ encoding: [0xf6,0xff,0xe0,0x20] 26 vtrn.16 q9, q [all...] |
H A D | neont2-sub-encoding.s | 7 @ CHECK: vext.8 q8, q9, q8, #3 @ encoding: [0xf2,0xef,0xe0,0x03] 8 vext.8 q8, q9, q8, #3 9 @ CHECK: vext.8 q8, q9, q8, #7 @ encoding: [0xf2,0xef,0xe0,0x07] 10 vext.8 q8, q9, q8, #7 13 @ CHECK: vext.32 q8, q9, q8, #3 @ encoding: [0xf2,0xef,0xe0,0x0c] 14 vext.32 q8, q9, q8, #3 21 @ CHECK: vtrn.8 q9, q8 @ encoding: [0xf2,0xff,0xe0,0x20] 22 vtrn.8 q9, q8 23 @ CHECK: vtrn.16 q9, q8 @ encoding: [0xf6,0xff,0xe0,0x20] 24 vtrn.16 q9, q [all...] |
H A D | neont2-bitwise-encoding.s | 6 vand q8, q8, q9 9 @ CHECK: vand q8, q8, q9 @ encoding: [0x40,0xef,0xf2,0x01] 12 veor q8, q8, q9 15 @ CHECK: veor q8, q8, q9 @ encoding: [0x40,0xff,0xf2,0x01] 19 vorr q8, q8, q9 25 @ CHECK: vorr q8, q8, q9 @ encoding: [0x60,0xef,0xf2,0x01] 29 vbic q8, q8, q9 34 @ CHECK: vbic q8, q8, q9 @ encoding: [0x50,0xef,0xf2,0x01] 38 vorn q8, q8, q9 41 @ CHECK: vorn q8, q8, q9 [all...] |
H A D | neont2-add-encoding.s | 15 @ CHECK: vadd.f32 q8, q8, q9 @ encoding: [0x40,0xef,0xe2,0x0d] 16 vadd.f32 q8, q8, q9 56 @ CHECK: vhadd.s8 q8, q8, q9 @ encoding: [0x40,0xef,0xe2,0x00] 57 vhadd.s8 q8, q8, q9 58 @ CHECK: vhadd.s16 q8, q8, q9 @ encoding: [0x50,0xef,0xe2,0x00] 59 vhadd.s16 q8, q8, q9 60 @ CHECK: vhadd.s32 q8, q8, q9 @ encoding: [0x60,0xef,0xe2,0x00] 61 vhadd.s32 q8, q8, q9 62 @ CHECK: vhadd.u8 q8, q8, q9 @ encoding: [0x40,0xff,0xe2,0x00] 63 vhadd.u8 q8, q8, q9 [all...] |
H A D | neon-absdiff-encoding.s | 17 @ CHECK: vabd.s8 q8, q8, q9 @ encoding: [0xe2,0x07,0x40,0xf2] 18 vabd.s8 q8, q8, q9 19 @ CHECK: vabd.s16 q8, q8, q9 @ encoding: [0xe2,0x07,0x50,0xf2] 20 vabd.s16 q8, q8, q9 21 @ CHECK: vabd.s32 q8, q8, q9 @ encoding: [0xe2,0x07,0x60,0xf2] 22 vabd.s32 q8, q8, q9 23 @ CHECK: vabd.u8 q8, q8, q9 @ encoding: [0xe2,0x07,0x40,0xf3] 24 vabd.u8 q8, q8, q9 25 @ CHECK: vabd.u16 q8, q8, q9 @ encoding: [0xe2,0x07,0x50,0xf3] 26 vabd.u16 q8, q8, q9 [all...] |
H A D | neont2-absdiff-encoding.s | 12 vabd.s8 q8, q8, q9 13 vabd.s16 q8, q8, q9 14 vabd.s32 q8, q8, q9 15 vabd.u8 q8, q8, q9 16 vabd.u16 q8, q8, q9 17 vabd.u32 q8, q8, q9 18 vabd.f32 q8, q8, q9 27 @ CHECK: vabd.s8 q8, q8, q9 @ encoding: [0x40,0xef,0xe2,0x07] 28 @ CHECK: vabd.s16 q8, q8, q9 @ encoding: [0x50,0xef,0xe2,0x07] 29 @ CHECK: vabd.s32 q8, q8, q9 [all...] |
H A D | neon-sub-encoding.s | 8 vsub.i8 q8, q8, q9 9 vsub.i16 q8, q8, q9 10 vsub.i32 q8, q8, q9 11 vsub.i64 q8, q8, q9 12 vsub.f32 q8, q8, q9 20 vsub.i16 q2, q9 30 @ CHECK: vsub.i8 q8, q8, q9 @ encoding: [0xe2,0x08,0x40,0xf3] 31 @ CHECK: vsub.i16 q8, q8, q9 @ encoding: [0xe2,0x08,0x50,0xf3] 32 @ CHECK: vsub.i32 q8, q8, q9 @ encoding: [0xe2,0x08,0x60,0xf3] 33 @ CHECK: vsub.i64 q8, q8, q9 [all...] |
H A D | neon-cmp-encoding.s | 7 vceq.i8 q8, q8, q9 8 vceq.i16 q8, q8, q9 9 vceq.i32 q8, q8, q9 10 vceq.f32 q8, q8, q9 16 @ CHECK: vceq.i8 q8, q8, q9 @ encoding: [0xf2,0x08,0x40,0xf3] 17 @ CHECK: vceq.i16 q8, q8, q9 @ encoding: [0xf2,0x08,0x50,0xf3] 18 @ CHECK: vceq.i32 q8, q8, q9 @ encoding: [0xf2,0x08,0x60,0xf3] 19 @ CHECK: vceq.f32 q8, q8, q9 @ encoding: [0xe2,0x0e,0x40,0xf2] 28 vcge.s8 q8, q8, q9 29 vcge.s16 q8, q8, q9 [all...] |
H A D | neon-add-encoding.s | 14 @ CHECK: vadd.f32 q8, q8, q9 @ encoding: [0xe2,0x0d,0x40,0xf2] 15 vadd.f32 q8, q8, q9 55 @ CHECK: vhadd.s8 q8, q8, q9 @ encoding: [0xe2,0x00,0x40,0xf2] 56 vhadd.s8 q8, q8, q9 57 @ CHECK: vhadd.s16 q8, q8, q9 @ encoding: [0xe2,0x00,0x50,0xf2] 58 vhadd.s16 q8, q8, q9 59 @ CHECK: vhadd.s32 q8, q8, q9 @ encoding: [0xe2,0x00,0x60,0xf2] 60 vhadd.s32 q8, q8, q9 61 @ CHECK: vhadd.u8 q8, q8, q9 @ encoding: [0xe2,0x00,0x40,0xf3] 62 vhadd.u8 q8, q8, q9 [all...] |
H A D | neon-shuffle-encoding.s | 5 vext.8 q8, q9, q8, #3 6 vext.8 q8, q9, q8, #7 8 vext.32 q8, q9, q8, #3 9 vext.64 q8, q9, q8, #1 14 vext.8 q9, q4, #7 22 @ CHECK: vext.8 q8, q9, q8, #3 @ encoding: [0xe0,0x03,0xf2,0xf2] 23 @ CHECK: vext.8 q8, q9, q8, #7 @ encoding: [0xe0,0x07,0xf2,0xf2] 25 @ CHECK: vext.32 q8, q9, q8, #3 @ encoding: [0xe0,0x0c,0xf2,0xf2] 26 @ CHECK: vext.64 q8, q9, q8, #1 @ encoding: [0xe0,0x08,0xf2,0xf2] 31 @ CHECK: vext.8 q9, q [all...] |
H A D | neont2-mul-encoding.s | 9 vmul.i8 q8, q8, q9 10 vmul.i16 q8, q8, q9 11 vmul.i32 q8, q8, q9 12 vmul.f32 q8, q8, q9 14 vmul.p8 q8, q8, q9 21 @ CHECK: vmul.i8 q8, q8, q9 @ encoding: [0x40,0xef,0xf2,0x09] 22 @ CHECK: vmul.i16 q8, q8, q9 @ encoding: [0x50,0xef,0xf2,0x09] 23 @ CHECK: vmul.i32 q8, q8, q9 @ encoding: [0x60,0xef,0xf2,0x09] 24 @ CHECK: vmul.f32 q8, q8, q9 @ encoding: [0x40,0xff,0xf2,0x0d] 26 @ CHECK: vmul.p8 q8, q8, q9 [all...] |
H A D | neon-mul-encoding.s | 7 vmul.i8 q8, q8, q9 8 vmul.i16 q8, q8, q9 9 vmul.i32 q8, q8, q9 10 vmul.f32 q8, q8, q9 12 vmul.p8 q8, q8, q9 19 vmul.i8 q8, q9 20 vmul.i16 q8, q9 21 vmul.i32 q8, q9 22 vmul.f32 q8, q9 24 vmul.p8 q8, q9 [all...] |
H A D | neon-mul-accum-encoding.s | 7 vmla.i8 q9, q8, q10 8 vmla.i16 q9, q8, q10 9 vmla.i32 q9, q8, q10 10 vmla.f32 q9, q8, q10 17 @ CHECK: vmla.i8 q9, q8, q10 @ encoding: [0xe4,0x29,0x40,0xf2] 18 @ CHECK: vmla.i16 q9, q8, q10 @ encoding: [0xe4,0x29,0x50,0xf2] 19 @ CHECK: vmla.i32 q9, q8, q10 @ encoding: [0xe4,0x29,0x60,0xf2] 20 @ CHECK: vmla.f32 q9, q8, q10 @ encoding: [0xf4,0x2d,0x40,0xf2] 58 vmls.i8 q9, q8, q10 59 vmls.i16 q9, q [all...] |
H A D | neon-satshift-encoding.s | 19 @ CHECK: vqshl.s8 q8, q8, q9 @ encoding: [0xf0,0x04,0x42,0xf2] 20 vqshl.s8 q8, q8, q9 21 @ CHECK: vqshl.s16 q8, q8, q9 @ encoding: [0xf0,0x04,0x52,0xf2] 22 vqshl.s16 q8, q8, q9 23 @ CHECK: vqshl.s32 q8, q8, q9 @ encoding: [0xf0,0x04,0x62,0xf2] 24 vqshl.s32 q8, q8, q9 25 @ CHECK: vqshl.s64 q8, q8, q9 @ encoding: [0xf0,0x04,0x72,0xf2] 26 vqshl.s64 q8, q8, q9 27 @ CHECK: vqshl.u8 q8, q8, q9 @ encoding: [0xf0,0x04,0x42,0xf3] 28 vqshl.u8 q8, q8, q9 [all...] |
H A D | neont2-satshift-encoding.s | 21 @ CHECK: vqshl.s8 q8, q8, q9 @ encoding: [0x42,0xef,0xf0,0x04] 22 vqshl.s8 q8, q8, q9 23 @ CHECK: vqshl.s16 q8, q8, q9 @ encoding: [0x52,0xef,0xf0,0x04] 24 vqshl.s16 q8, q8, q9 25 @ CHECK: vqshl.s32 q8, q8, q9 @ encoding: [0x62,0xef,0xf0,0x04] 26 vqshl.s32 q8, q8, q9 27 @ CHECK: vqshl.s64 q8, q8, q9 @ encoding: [0x72,0xef,0xf0,0x04] 28 vqshl.s64 q8, q8, q9 29 @ CHECK: vqshl.u8 q8, q8, q9 @ encoding: [0x42,0xff,0xf0,0x04] 30 vqshl.u8 q8, q8, q9 [all...] |
H A D | neont2-dup-encoding.s | 8 vdup.8 q9, r4 15 @ CHECK: vdup.8 q9, r4 @ encoding: [0xe2,0xee,0x90,0x4b] 23 vdup.16 q9, d9[0] 29 vdup.16 q9, d9[1] 36 @ CHECK: vdup.16 q9, d9[0] @ encoding: [0xf2,0xff,0x49,0x2c] 42 @ CHECK: vdup.16 q9, d9[1] @ encoding: [0xf6,0xff,0x49,0x2c]
|
/external/openssl/crypto/modes/asm/ |
H A D | ghashv8-armx.S | 10 vld1.64 {q9},[r1] @ load H 12 vext.8 q3,q9,q9,#8 16 vdup.32 q9,d18[1] 18 vshr.s32 q9,q9,#31 @ broadcast carry bit 22 vand q8,q8,q9 34 vld1.64 {q9},[r0] @ load Xi 39 vrev64.8 q9,q9 [all...] |
/external/libvpx/libvpx/vp8/encoder/arm/neon/ |
H A D | vp8_mse16x16_neon.asm | 30 vmov.i8 q7, #0 ;q7, q8, q9, q10 - sse 32 vmov.i8 q9, #0 53 vmlal.s16 q9, d24, d24 57 vmlal.s16 q9, d28, d28 63 vadd.u32 q9, q9, q10 67 vadd.u32 q10, q7, q9 101 vmull.s16 q9, d26, d26 105 vadd.u32 q9, q9, q1 [all...] |
/external/libhevc/common/arm/ |
H A D | ihevc_itrans_recon_8x8.s | 189 vmull.s16 q9,d3,d1[2] @// y2 * sin2 (q3 is freed by this time)(part of d1) 239 vmlsl.s16 q9,d11,d0[2] @// d1 = y2 * sin2 - y6 * cos2(part of a0 and a1) 252 vsub.s32 q11,q10,q9 @// a2 = c1 - d1(part of r2,r5) 253 vadd.s32 q9,q10,q9 @// a1 = c1 + d1(part of r1,r6) 261 vadd.s32 q14,q9,q13 @// a1 + b1(part of r1) 262 vsub.s32 q9,q9,q13 @// a1 - b1(part of r6) 272 vqrshrn.s32 d11,q9,#shift_stage1_idct @// r6 = (a1 - b1 + rnd) >> 7(shift_stage1_idct) 312 vmull.s16 q9,d [all...] |
H A D | ihevc_itrans_recon_32x32.s | 235 vmull.s16 q9,d10,d0[0] 236 vmlal.s16 q9,d11,d3[2] 269 vmlal.s16 q9,d12,d7[0] 270 vmlsl.s16 q9,d13,d5[2] 305 vmlsl.s16 q9,d10,d2[0] 306 vmlsl.s16 q9,d11,d1[2] 343 vmlsl.s16 q9,d12,d5[0] 344 vmlal.s16 q9,d13,d7[2] 381 vmlal.s16 q9,d10,d0[0] 382 vmlal.s16 q9,d1 [all...] |
/external/libvpx/libvpx/vp8/common/arm/neon/ |
H A D | variance_neon.asm | 30 vmov.i8 q9, #0 ;q9, q10 - sse 50 vmlal.s16 q9, d22, d22 ;calculate sse 56 vmlal.s16 q9, d24, d24 59 vmlal.s16 q9, d26, d26 62 vmlal.s16 q9, d28, d28 67 vadd.u32 q10, q9, q10 ;accumulate sse 103 vmov.i8 q9, #0 ;q9, q10 - sse 120 vmlal.s16 q9, d2 [all...] |
/external/chromium_org/third_party/boringssl/linux-arm/crypto/aes/ |
H A D | bsaes-armv7.S | 84 vldmia r4!, {q9} @ round 0 key 88 veor q10, q0, q9 @ xor with round0 key 89 veor q11, q1, q9 92 veor q12, q2, q9 95 veor q13, q3, q9 98 veor q14, q4, q9 101 veor q15, q5, q9 104 veor q10, q6, q9 107 veor q11, q7, q9 113 vmov.i8 q9,# [all...] |
/external/openssl/crypto/aes/asm/ |
H A D | bsaes-armv7.S | 83 vldmia r4!, {q9} @ round 0 key 87 veor q10, q0, q9 @ xor with round0 key 88 veor q11, q1, q9 91 veor q12, q2, q9 94 veor q13, q3, q9 97 veor q14, q4, q9 100 veor q15, q5, q9 103 veor q10, q6, q9 106 veor q11, q7, q9 112 vmov.i8 q9,# [all...] |
H A D | aesv8-armx.S | 34 vext.8 q9,q0,q3,#12 39 veor q3,q3,q9 40 vext.8 q9,q0,q9,#12 41 veor q3,q3,q9 42 vext.8 q9,q0,q9,#12 44 veor q3,q3,q9 53 vext.8 q9,q0,q3,#12 57 veor q3,q3,q9 [all...] |
/external/chromium_org/third_party/libvpx/source/libvpx/vp8/common/arm/neon/ |
H A D | idct_dequant_full_2x_neon.c | 24 int16x8_t q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10, q11; local 92 q9 = vqdmulhq_n_s16(q5, cospi8sqrt2minus1); 98 q9 = vshrq_n_s16(q9, 1); 101 q5 = vqaddq_s16(q5, q9); 120 q9 = vqdmulhq_n_s16(q2tmp3.val[1], sinpi8sqrt2); 134 q9 = vqaddq_s16(q9, q10); 136 q4 = vqaddq_s16(q2, q9); 139 q7 = vqsubq_s16(q2, q9); [all...] |