/external/clang/test/CodeGen/ |
H A D | builtins-systemz-error2.c | 4 typedef __attribute__((vector_size(16))) char v16i8; typedef 6 v16i8 f0(v16i8 a, v16i8 b) { 8 v16i8 tmp = __builtin_s390_vaq(a, b); // expected-error {{'__builtin_s390_vaq' needs target feature vector}}
|
H A D | systemz-abi-vector.c | 27 typedef __attribute__((vector_size(16))) char v16i8; typedef 38 unsigned int align = __alignof__ (v16i8); 58 v16i8 pass_v16i8(v16i8 arg) { return arg; } 153 struct agg_v16i8 { v16i8 a; }; 317 v16i8 va_v16i8(__builtin_va_list l) { return __builtin_va_arg(l, v16i8); }
|
/external/libyuv/files/source/ |
H A D | scale_msa.cc | 33 src0 = (v16u8)__msa_ld_b((v16i8*)src_argb, 0); 34 src1 = (v16u8)__msa_ld_b((v16i8*)src_argb, 16); 51 src0 = (v16u8)__msa_ld_b((v16i8*)src_argb, 0); 52 src1 = (v16u8)__msa_ld_b((v16i8*)src_argb, 16); 71 v16i8 shuffler = {0, 4, 1, 5, 2, 6, 3, 7, 8, 12, 9, 13, 10, 14, 11, 15}; 74 src0 = (v16u8)__msa_ld_b((v16i8*)s, 0); 75 src1 = (v16u8)__msa_ld_b((v16i8*)s, 16); 76 src2 = (v16u8)__msa_ld_b((v16i8*)t, 0); 77 src3 = (v16u8)__msa_ld_b((v16i8*)t, 16); 78 vec0 = (v16u8)__msa_vshf_b(shuffler, (v16i8)src [all...] |
H A D | row_msa.cc | 78 v16i8 zero_m = {0}; \ 80 vec0_m = (v8i16)__msa_ilvr_b((v16i8)in_y, (v16i8)in_y); \ 81 vec1_m = (v8i16)__msa_ilvr_b((v16i8)zero_m, (v16i8)in_uv); \ 128 vec0_m = (v8i16)__msa_ilvev_b((v16i8)in1, (v16i8)in0); \ 129 vec1_m = (v8i16)__msa_ilvev_b((v16i8)in3, (v16i8)in2); \ 154 y_out = (v16u8)__msa_pckev_b((v16i8)reg1_ [all...] |
H A D | rotate_msa.cc | 24 out0 = (v16u8)__msa_ilvr_b((v16i8)in1, (v16i8)in0); \ 25 out1 = (v16u8)__msa_ilvl_b((v16i8)in1, (v16i8)in0); \ 26 out2 = (v16u8)__msa_ilvr_b((v16i8)in3, (v16i8)in2); \ 27 out3 = (v16u8)__msa_ilvl_b((v16i8)in3, (v16i8)in2); \ 90 src0 = (v16u8)__msa_ld_b((v16i8*)s, 0); 92 src1 = (v16u8)__msa_ld_b((v16i8*) [all...] |
/external/libvpx/libvpx/vpx_dsp/mips/ |
H A D | loopfilter_16_msa.c | 82 v16i8 zero = { 0 }; 112 q0_r_in = (v8u16)__msa_ilvr_b(zero, (v16i8)q0); 131 q0_l_in = (v8u16)__msa_ilvl_b(zero, (v16i8)q0); 146 r_out = (v8i16)__msa_pckev_b((v16i8)l_out, (v16i8)r_out); 152 q1_r_in = (v8u16)__msa_ilvr_b(zero, (v16i8)q1); 159 q1_l_in = (v8u16)__msa_ilvl_b(zero, (v16i8)q1); 166 r_out = (v8i16)__msa_pckev_b((v16i8)l_out, (v16i8)r_out); 172 q2_r_in = (v8u16)__msa_ilvr_b(zero, (v16i8)q [all...] |
H A D | vpx_convolve_msa.h | 24 tmp_dpadd_0 = __msa_dotp_s_h((v16i8)vec0, (v16i8)filt0); \ 25 tmp_dpadd_0 = __msa_dpadd_s_h(tmp_dpadd_0, (v16i8)vec1, (v16i8)filt1); \ 26 tmp_dpadd_1 = __msa_dotp_s_h((v16i8)vec2, (v16i8)filt2); \ 27 tmp_dpadd_1 = __msa_dpadd_s_h(tmp_dpadd_1, (v16i8)vec3, (v16i8)filt3); \ 36 v16i8 vec0_m, vec1_m, vec2_m, vec3_m; \ 54 v16i8 vec0_ [all...] |
H A D | vpx_convolve8_msa.c | 29 v16i8 src0, src1, src2, src3, src4, src5, src6, src7, src8, src9, src10; 30 v16i8 filt_hz0, filt_hz1, filt_hz2, filt_hz3; 65 out2 = (v8i16)__msa_ilvev_b((v16i8)hz_out5, (v16i8)hz_out4); 74 hz_out6 = (v8i16)__msa_sldi_b((v16i8)hz_out7, (v16i8)hz_out5, 8); 75 out3 = (v8i16)__msa_ilvev_b((v16i8)hz_out7, (v16i8)hz_out6); 81 hz_out8 = (v8i16)__msa_sldi_b((v16i8)hz_out9, (v16i8)hz_out [all...] |
H A D | macros_msa.h | 21 #define LD_SB(...) LD_B(v16i8, __VA_ARGS__) 32 #define ST_SB(...) ST_B(v16i8, __VA_ARGS__) 299 #define LD_SB2(...) LD_B2(v16i8, __VA_ARGS__) 314 #define LD_SB4(...) LD_B4(v16i8, __VA_ARGS__) 322 #define LD_SB5(...) LD_B5(v16i8, __VA_ARGS__) 329 #define LD_SB7(...) LD_B7(v16i8, __VA_ARGS__) 338 #define LD_SB8(...) LD_B8(v16i8, __VA_ARGS__) 631 v16i8 zero_m = { 0 }; \ 632 out0 = (RTYPE)__msa_sldi_b((v16i8)zero_m, (v16i8)in [all...] |
H A D | vpx_convolve8_avg_msa.c | 19 v16i8 src0, src1, src2, src3, src4, src5, src6, src7, src8, src9, src10; 21 v16i8 filt_hz0, filt_hz1, filt_hz2, filt_hz3; 55 vec2 = (v8i16)__msa_ilvev_b((v16i8)hz_out5, (v16i8)hz_out4); 65 hz_out6 = (v8i16)__msa_sldi_b((v16i8)hz_out7, (v16i8)hz_out5, 8); 66 vec3 = (v8i16)__msa_ilvev_b((v16i8)hz_out7, (v16i8)hz_out6); 72 hz_out8 = (v8i16)__msa_sldi_b((v16i8)hz_out9, (v16i8)hz_out [all...] |
H A D | loopfilter_msa.h | 19 v16i8 p1_m, p0_m, q0_m, q1_m, filt, q0_sub_p0, t1, t2; \ 20 const v16i8 cnst4b = __msa_ldi_b(4); \ 21 const v16i8 cnst3b = __msa_ldi_b(3); \ 23 p1_m = (v16i8)__msa_xori_b(p1_in, 0x80); \ 24 p0_m = (v16i8)__msa_xori_b(p0_in, 0x80); \ 25 q0_m = (v16i8)__msa_xori_b(q0_in, 0x80); \ 26 q1_m = (v16i8)__msa_xori_b(q1_in, 0x80); \
|
H A D | subtract_msa.c | 19 v16i8 src = { 0 }; 20 v16i8 pred = { 0 }; 38 v16i8 src = { 0 }; 39 v16i8 pred = { 0 }; 62 v16i8 src0, src1, src2, src3, src4, src5, src6, src7; 63 v16i8 pred0, pred1, pred2, pred3, pred4, pred5, pred6, pred7; 121 v16i8 src0, src1, src2, src3, src4, src5, src6, src7; 122 v16i8 pred0, pred1, pred2, pred3, pred4, pred5, pred6, pred7; 183 v16i8 src0, src1, src2, src3, src4, src5, src6, src7; 184 v16i8 pred [all...] |
/external/swiftshader/third_party/LLVM/test/CodeGen/CellSPU/useful-harnesses/ |
H A D | vecoperations.c | 3 typedef unsigned char v16i8 __attribute__((ext_vector_type(16))); typedef 10 void print_v16i8(const char *str, const v16i8 v) { 13 v16i8 vec; 24 void print_v16i8_hex(const char *str, const v16i8 v) { 27 v16i8 vec; 68 v16i8 v16i8_mpy(v16i8 v1, v16i8 v2) { 72 v16i8 v16i8_add(v16i8 v [all...] |
/external/libvpx/libvpx/vp8/common/mips/msa/ |
H A D | sixtap_filter_msa.c | 38 v16i8 vec0_m, vec1_m, vec2_m; \ 55 v16i8 vec0_m, vec1_m, vec2_m, vec3_m, vec4_m, vec5_m; \ 69 v16i8 vec0_m, vec1_m, vec2_m, vec3_m, vec4_m, vec5_m, vec6_m, vec7_m; \ 89 tmp0 = __msa_dotp_s_h((v16i8)vec0, (v16i8)filt0); \ 90 tmp0 = __msa_dpadd_s_h(tmp0, (v16i8)vec1, (v16i8)filt1); \ 97 v16i8 vec0_m, vec1_m; \ 112 v16i8 vec0_m, vec1_m, vec2_m, vec3_m; \ 123 v16i8 vec0_ [all...] |
H A D | mfqe_msa.c | 21 v16i8 src0 = { 0 }; 22 v16i8 src1 = { 0 }; 23 v16i8 dst0 = { 0 }; 24 v16i8 dst1 = { 0 }; 50 dst0 = (v16i8)__msa_pckev_b((v16i8)res_h_l, (v16i8)res_h_r); 61 dst1 = (v16i8)__msa_pckev_b((v16i8)res_h_l, (v16i8)res_h_ [all...] |
H A D | vp8_macros_msa.h | 21 #define LD_SB(...) LD_B(v16i8, __VA_ARGS__) 33 #define ST_SB(...) ST_B(v16i8, __VA_ARGS__) 274 #define LD_SB2(...) LD_B2(v16i8, __VA_ARGS__) 282 #define LD_SB3(...) LD_B3(v16i8, __VA_ARGS__) 290 #define LD_SB4(...) LD_B4(v16i8, __VA_ARGS__) 298 #define LD_SB5(...) LD_B5(v16i8, __VA_ARGS__) 307 #define LD_SB8(...) LD_B8(v16i8, __VA_ARGS__) 358 #define ST_SB4(...) ST_B4(v16i8, __VA_ARGS__) 514 v16i8 zero_m = { 0 }; \ 516 out0 = (RTYPE)__msa_sldi_b((v16i8)zero_ [all...] |
H A D | loopfilter_filters_msa.c | 21 p1_a_sub_q1 = (v16u8)__msa_srli_b((v16i8)p1_a_sub_q1, 1); \ 29 v16i8 p1_m, p0_m, q0_m, q1_m, filt, q0_sub_p0, t1, t2; \ 30 const v16i8 cnst4b = __msa_ldi_b(4); \ 31 const v16i8 cnst3b = __msa_ldi_b(3); \ 33 p1_m = (v16i8)__msa_xori_b(p1, 0x80); \ 34 p0_m = (v16i8)__msa_xori_b(p0, 0x80); \ 35 q0_m = (v16i8)__msa_xori_b(q0, 0x80); \ 36 q1_m = (v16i8)__msa_xori_b(q1, 0x80); \ 64 v16i8 p1_m, p0_m, q0_m, q1_m, filt, filt1, filt2; \ 65 v16i8 q0_sub_p [all...] |
H A D | bilinear_filter_msa.c | 33 v16i8 src0, src1, src2, src3, mask; 54 v16i8 src0, src1, src2, src3, src4, src5, src6, src7, mask; 55 v16i8 res0, res1, res2, res3; 90 v16i8 src0, src1, src2, src3, mask; 112 v16i8 src0, src1, src2, src3, mask, out0, out1; 184 v16i8 src0, src1, src2, src3, src4, src5, src6, src7, mask; 247 v16i8 src0, src1, src2, src3, src4; 248 v16i8 src10_r, src32_r, src21_r, src43_r, src2110, src4332; 264 src2110 = __msa_pckev_b((v16i8)tmp1, (v16i8)tmp [all...] |
/external/libpng/mips/ |
H A D | filter_msa_intrinsics.c | 42 #define MSA_SRLI_B(a, b) __msa_srli_b((v16i8) a, b) 300 out0 = (RTYPE) __msa_ilvr_b((v16i8) in0, (v16i8) in1); \ 301 out1 = (RTYPE) __msa_ilvr_b((v16i8) in2, (v16i8) in3); \ 314 v16i8 zero_m = { 0 }; \ 315 out0 = (RTYPE) __msa_sldi_b((v16i8) zero_m, (v16i8) in0, slide_val); \ 316 out1 = (RTYPE) __msa_sldi_b((v16i8) zero_m, (v16i8) in [all...] |
/external/libvpx/libvpx/vp8/encoder/mips/msa/ |
H A D | denoising_msa.c | 116 temp2_h = (v8i16)__msa_pckev_b((v16i8)temp3_h, (v16i8)temp2_h); 117 running_avg_y = (v16u8)__msa_pckev_b((v16i8)temp1_h, (v16i8)temp0_h); 164 temp2_h = (v8i16)__msa_pckev_b((v16i8)temp3_h, (v16i8)temp2_h); 165 running_avg_y = (v16u8)__msa_pckev_b((v16i8)temp1_h, (v16i8)temp0_h); 238 running_avg_y = (v16u8)__msa_pckev_b((v16i8)adjust3, (v16i8)adjust [all...] |
/external/libvpx/libvpx/vp9/common/mips/msa/ |
H A D | vp9_mfqe_msa.c | 21 v16i8 src0 = { 0 }; 22 v16i8 src1 = { 0 }; 23 v16i8 dst0 = { 0 }; 24 v16i8 dst1 = { 0 }; 50 dst0 = (v16i8)__msa_pckev_b((v16i8)res_h_l, (v16i8)res_h_r); 61 dst1 = (v16i8)__msa_pckev_b((v16i8)res_h_l, (v16i8)res_h_ [all...] |
/external/libyuv/files/include/libyuv/ |
H A D | macros_msa.h | 212 out0 = (RTYPE)__msa_vshf_b((v16i8)mask0, (v16i8)in1, (v16i8)in0); \ 213 out1 = (RTYPE)__msa_vshf_b((v16i8)mask1, (v16i8)in3, (v16i8)in2); \ 226 out0 = (RTYPE)__msa_ilvr_b((v16i8)in0, (v16i8)in1); \ 227 out1 = (RTYPE)__msa_ilvl_b((v16i8)in0, (v16i8)in [all...] |
/external/webp/src/dsp/ |
H A D | msa_macro.h | 27 #define SRAI_B(a, b) __msa_srai_b((v16i8)a, b) 48 #define LD_SB(...) LD_B(v16i8, __VA_ARGS__) 60 #define ST_SB(...) ST_B(v16i8, __VA_ARGS__) 223 #define LD_SB2(...) LD_B2(v16i8, __VA_ARGS__) 230 #define LD_SB3(...) LD_B3(v16i8, __VA_ARGS__) 237 #define LD_SB4(...) LD_B4(v16i8, __VA_ARGS__) 245 #define LD_SB8(...) LD_B8(v16i8, __VA_ARGS__) 299 #define ST_SB2(...) ST_B2(v16i8, __VA_ARGS__) 306 #define ST_SB4(...) ST_B4(v16i8, __VA_ARGS__) 415 (RTYPE)__msa_sldi_b((v16i8)in [all...] |
H A D | lossless_enc_msa.c | 38 const v16i8 g0 = VSHF_SB(src, src, mask0); \ 45 t1 = __msa_dotp_s_h(c1, (v16i8)t2); \ 54 const v16i8 g2br = (v16i8)__msa_fill_w(m->green_to_blue_ | 56 const v16i8 r2b = (v16i8)__msa_fill_w(m->red_to_blue_);
|
/external/llvm/lib/Target/AMDGPU/ |
H A D | SITypeRewriter.cpp | 14 /// v16i8 => i128 15 /// - v16i8 is used for constant memory resource descriptors. This type is 17 /// in the backend, because we want the legalizer to expand all v16i8 38 Type *v16i8; member in class:__anon12890::SITypeRewriter 59 v16i8 = VectorType::get(Type::getInt8Ty(M.getContext()), 16); 79 if (ElemTy == v16i8) { 107 if (Arg->getType() == v16i8) {
|