/external/clang/test/CodeGen/ |
H A D | sse.c | 31 return _mm_srli_si128(a, 5); 37 return _mm_srli_si128(a, 0); 43 return _mm_srli_si128(a, 16);
|
/external/clang/test/Sema/ |
H A D | warn-shadow-intrinsics.c | 9 qf = _mm_slli_si128(_mm_add_epi64(qf, _mm_srli_si128(qf, 8)), 8); // no-warning
|
/external/libavc/common/x86/ |
H A D | ih264_luma_intra_pred_filters_ssse3.c | 342 top_sh_8x16b = _mm_srli_si128(top_8x16b, 2); //t1 t2 t3 t4 t5 t6 t7 0 347 res2_8x16b = _mm_srli_si128(res1_8x16b, 2); 358 res_16x8b = _mm_srli_si128(res_16x8b, 1); 360 res_16x8b = _mm_srli_si128(res_16x8b, 1); 362 res_16x8b = _mm_srli_si128(res_16x8b, 1); 426 top_left_sh_16x8b = _mm_srli_si128(top_left_16x8b, 1); //l2 l1 l0 tl t0 t1 t2 t3... 433 res2_8x16b = _mm_srli_si128(res1_8x16b, 2); //l2+l1 l1+l0 l0+tl tl+t0 t0+t1 t1+t2 t2+t3... 443 res2_16x8b = _mm_srli_si128(res1_16x8b, 3); 446 res2_16x8b = _mm_srli_si128(res1_16x8b, 2); 448 res2_16x8b = _mm_srli_si128(res1_16x8 [all...] |
H A D | ih264_inter_pred_filters_ssse3.c | 255 src_r0_sht_16x8b = _mm_srli_si128(src_r0_16x8b, 1); //a1 a2 a3 a4 a5 a6 a7 a8 a9....a15 0 256 src_r1_sht_16x8b = _mm_srli_si128(src_r1_16x8b, 1); //b1 b2 b3 b4 b5 b6 b7 b8 b9....b15 0 265 src_r0_16x8b = _mm_srli_si128(src_r0_16x8b, 4); //a2 a3 a3 a4 a4 a5 a5 a6 a6 a7 a7 a8 0 0 0 0 266 src_r1_16x8b = _mm_srli_si128(src_r1_16x8b, 4); //b2 b3 b3 b4 b4 b5 b5 b6 b6 b7 b7 b8 0 0 0 0 272 src_r0_16x8b = _mm_srli_si128(src_r0_16x8b, 4); //a4 a5 a5 a6 a6 a7 a7 a8 0 0 0 0 0 0 0 0 273 src_r1_16x8b = _mm_srli_si128(src_r1_16x8b, 4); //b4 b5 b5 b6 b6 b7 b7 b8 0 0 0 0 0 0 0 0 295 res_r0r1_16x8b = _mm_srli_si128(res_r0r1_16x8b, 4); 320 src_r0_sht_16x8b = _mm_srli_si128(src_r0_16x8b, 1); //a1 a2 a3 a4 a5 a6 a7 a8 a9....a15 0 321 src_r1_sht_16x8b = _mm_srli_si128(src_r1_16x8b, 1); //b1 b2 b3 b4 b5 b6 b7 b8 b9....b15 0 331 src_r0_16x8b = _mm_srli_si128(src_r0_16x8 [all...] |
H A D | ih264_iquant_itrans_recon_dc_ssse3.c | 158 temp5 = _mm_srli_si128(temp4,4); 159 temp6 = _mm_srli_si128(temp5,4); 160 temp7 = _mm_srli_si128(temp6,4); 439 pred_r1 = _mm_srli_si128(pred_r0, 4); 440 pred_r2 = _mm_srli_si128(pred_r1, 4); 441 pred_r3 = _mm_srli_si128(pred_r2, 4);
|
H A D | ih264_deblk_chroma_ssse3.c | 224 lineb = _mm_srli_si128(linea, 8); 226 lined = _mm_srli_si128(linec, 8); 228 linef = _mm_srli_si128(linee, 8); 230 lineh = _mm_srli_si128(lineg, 8); 592 lineb = _mm_srli_si128(linea, 8); 594 lined = _mm_srli_si128(linec, 8); 596 linef = _mm_srli_si128(linee, 8); 598 lineh = _mm_srli_si128(lineg, 8); 862 p0_uv_16x8 = _mm_srli_si128(p1_uv_16x8, 8); 864 q1_uv_16x8 = _mm_srli_si128(q0_uv_16x [all...] |
/external/libhevc/decoder/x86/ |
H A D | ihevcd_it_rec_dc_sse42_intr.c | 115 m_temp_reg_1 = _mm_srli_si128(m_temp_reg_8, 4); 116 m_temp_reg_2 = _mm_srli_si128(m_temp_reg_8, 8); 117 m_temp_reg_3 = _mm_srli_si128(m_temp_reg_8, 12); 163 m_temp_reg_13 = _mm_srli_si128(m_temp_reg_12, 8); 173 m_temp_reg_15 = _mm_srli_si128(m_temp_reg_14, 8); 277 m_temp_reg_8 = _mm_srli_si128(m_temp_reg_8, 4); 279 m_temp_reg_8 = _mm_srli_si128(m_temp_reg_8, 4); 281 m_temp_reg_8 = _mm_srli_si128(m_temp_reg_8, 4); 351 m_temp_reg_4 = _mm_srli_si128(m_temp_reg_4, 8); 355 m_temp_reg_5 = _mm_srli_si128(m_temp_reg_ [all...] |
H A D | ihevcd_it_rec_dc_ssse3_intr.c | 116 m_temp_reg_1 = _mm_srli_si128(m_temp_reg_8, 4); 117 m_temp_reg_2 = _mm_srli_si128(m_temp_reg_8, 8); 118 m_temp_reg_3 = _mm_srli_si128(m_temp_reg_8, 12); 164 m_temp_reg_13 = _mm_srli_si128(m_temp_reg_12, 8); 174 m_temp_reg_15 = _mm_srli_si128(m_temp_reg_14, 8); 277 m_temp_reg_8 = _mm_srli_si128(m_temp_reg_8, 4); 279 m_temp_reg_8 = _mm_srli_si128(m_temp_reg_8, 4); 281 m_temp_reg_8 = _mm_srli_si128(m_temp_reg_8, 4); 351 m_temp_reg_4 = _mm_srli_si128(m_temp_reg_4, 8); 355 m_temp_reg_5 = _mm_srli_si128(m_temp_reg_ [all...] |
/external/libmpeg2/common/x86/ |
H A D | ideint_cac_ssse3.c | 151 sum_1 = _mm_cvtsi128_si32(_mm_srli_si128(sum_t[i + 0], 8)); 152 sum_2 = _mm_cvtsi128_si32(_mm_srli_si128(sum_b[i + 0], 8)); 157 sum_3 = _mm_cvtsi128_si32(_mm_srli_si128(sum_t[i + 1], 8)); 158 sum_4 = _mm_cvtsi128_si32(_mm_srli_si128(sum_b[i + 1], 8)); 224 diff_sum = _mm_cvtsi128_si32(_mm_srli_si128(diff, 8));
|
H A D | icv_variance_ssse3.c | 129 vsum = _mm_add_epi64(vsum, _mm_srli_si128(vsum, 8)); 149 vsum_sqr = _mm_add_epi32(vsum_sqr, _mm_srli_si128(vsum_sqr, 8)); 150 vsum_sqr = _mm_add_epi32(vsum_sqr, _mm_srli_si128(vsum_sqr, 4));
|
H A D | ideint_spatial_filter_ssse3.c | 164 adiff[3] = _mm_cvtsi128_si32(_mm_srli_si128(diffs[0], 8)); 165 adiff[4] = _mm_cvtsi128_si32(_mm_srli_si128(diffs[1], 8)); 166 adiff[5] = _mm_cvtsi128_si32(_mm_srli_si128(diffs[2], 8)); 233 _mm_storel_epi64((__m128i *)pu1_out, _mm_srli_si128(dst, 8));
|
H A D | icv_sad_ssse3.c | 140 res_r0 = _mm_add_epi64(res_r0, _mm_srli_si128(res_r0, 8));
|
/external/libhevc/common/x86/ |
H A D | ihevc_inter_pred_filters_ssse3_intr.c | 297 mask_high_96b = _mm_srli_si128(mask_low_32b, 12); 333 src_temp2_16x8b = _mm_srli_si128(src_temp1_16x8b, 1); /* row = 0 */ 338 src_temp1_16x8b = _mm_srli_si128(src_temp1_16x8b, 2); /* row = 0 */ 339 src_temp2_16x8b = _mm_srli_si128(src_temp2_16x8b, 2); /* row = 0 */ 344 src_temp1_16x8b = _mm_srli_si128(src_temp1_16x8b, 2); /* row = 0 */ 345 src_temp2_16x8b = _mm_srli_si128(src_temp2_16x8b, 2); /* row = 0 */ 350 src_temp1_16x8b = _mm_srli_si128(src_temp1_16x8b, 2); /* row = 0 */ 351 src_temp2_16x8b = _mm_srli_si128(src_temp2_16x8b, 2); /* row = 0 */ 366 src_temp12_16x8b = _mm_srli_si128(src_temp11_16x8b, 1); /* row =1 */ 371 src_temp11_16x8b = _mm_srli_si128(src_temp11_16x8 [all...] |
H A D | ihevc_chroma_intra_pred_filters_sse42_intr.c | 333 src_temp3 = _mm_srli_si128(src_temp3, 8); 334 src_temp4 = _mm_srli_si128(src_temp4, 8); 335 src_temp7 = _mm_srli_si128(src_temp7, 8); 336 src_temp8 = _mm_srli_si128(src_temp8, 8); 357 temp_sad = _mm_srli_si128(src_temp4, 4); /* Next 32 bits */ 371 src_temp3 = _mm_srli_si128(src_temp3, 8); 372 src_temp4 = _mm_srli_si128(src_temp4, 8); 386 temp_sad = _mm_srli_si128(src_temp4, 4); /* Next 32 bits */ 397 src_temp4 = _mm_srli_si128(src_temp3, 8); 407 temp_sad = _mm_srli_si128(src_temp [all...] |
H A D | ihevc_deblk_ssse3_intr.c | 183 temp_coef0_8x16b = _mm_srli_si128(mask_d_result_4x32b, 4); 184 temp_coef1_8x16b = _mm_srli_si128(mask_d_result_4x32b, 8); 185 mask_16x8b = _mm_srli_si128(mask_d_result_4x32b, 12); 195 temp_coef0_8x16b = _mm_srli_si128(mask_de_result_8x16b, 4); 196 temp_coef1_8x16b = _mm_srli_si128(mask_de_result_8x16b, 8); 197 mask_16x8b = _mm_srli_si128(mask_de_result_8x16b, 12); 379 src_row2_8x16b = _mm_srli_si128(src_row0_8x16b, 8); 380 src_row3_8x16b = _mm_srli_si128(src_row1_8x16b, 8); 525 src_row1_8x16b = _mm_srli_si128(src_row0_8x16b, 8); 526 src_row3_8x16b = _mm_srli_si128(src_row2_8x16 [all...] |
H A D | ihevc_chroma_intra_pred_filters_ssse3_intr.c | 354 src_temp3 = _mm_srli_si128(src_temp3, 8); 355 src_temp4 = _mm_srli_si128(src_temp4, 8); 356 src_temp7 = _mm_srli_si128(src_temp7, 8); 357 src_temp8 = _mm_srli_si128(src_temp8, 8); 380 temp_sad = _mm_srli_si128(src_temp4, 4); /* Next 32 bits */ 394 src_temp3 = _mm_srli_si128(src_temp3, 8); 395 src_temp4 = _mm_srli_si128(src_temp4, 8); 411 temp_sad = _mm_srli_si128(src_temp4, 4); /* Next 32 bits */ 422 src_temp4 = _mm_srli_si128(src_temp3, 8); 435 temp_sad = _mm_srli_si128(src_temp [all...] |
H A D | ihevc_intra_pred_filters_sse42_intr.c | 257 src_temp2 = _mm_srli_si128(src_temp1, 1); 258 src_temp3 = _mm_srli_si128(src_temp2, 1); 273 src_temp5 = _mm_srli_si128(src_temp4, 1); 274 src_temp6 = _mm_srli_si128(src_temp5, 1); 956 src_temp2 = _mm_srli_si128(src_temp1, 1); 957 src_temp3 = _mm_srli_si128(src_temp1, 2); 958 src_temp4 = _mm_srli_si128(src_temp1, 3); 959 src_temp5 = _mm_srli_si128(src_temp1, 4); 960 src_temp6 = _mm_srli_si128(src_temp1, 5); 961 src_temp7 = _mm_srli_si128(src_temp [all...] |
H A D | ihevc_intra_pred_filters_ssse3_intr.c | 543 src_temp2 = _mm_srli_si128(src_temp1, 1); 544 src_temp3 = _mm_srli_si128(src_temp2, 1); 559 src_temp5 = _mm_srli_si128(src_temp4, 1); 560 src_temp6 = _mm_srli_si128(src_temp5, 1); 1466 src_temp2 = _mm_srli_si128(src_temp1, 1); 1467 src_temp3 = _mm_srli_si128(src_temp1, 2); 1468 src_temp4 = _mm_srli_si128(src_temp1, 3); 1469 src_temp5 = _mm_srli_si128(src_temp1, 4); 1470 src_temp6 = _mm_srli_si128(src_temp1, 5); 1471 src_temp7 = _mm_srli_si128(src_temp [all...] |
/external/libvpx/libvpx/vpx_dsp/x86/ |
H A D | variance_sse2.c | 32 vsum = _mm_add_epi32(vsum, _mm_srli_si128(vsum, 8)); 33 vsum = _mm_add_epi32(vsum, _mm_srli_si128(vsum, 4)); 54 vsum = _mm_add_epi16(vsum, _mm_srli_si128(vsum, 8)); 55 vsum = _mm_add_epi16(vsum, _mm_srli_si128(vsum, 4)); 56 vsum = _mm_add_epi16(vsum, _mm_srli_si128(vsum, 2)); 62 vsum = _mm_add_epi32(vsum, _mm_srli_si128(vsum, 8)); 63 vsum = _mm_add_epi32(vsum, _mm_srli_si128(vsum, 4)); 95 vsum = _mm_add_epi16(vsum, _mm_srli_si128(vsum, 8)); 96 vsum = _mm_add_epi16(vsum, _mm_srli_si128(vsum, 4)); 97 vsum = _mm_add_epi16(vsum, _mm_srli_si128(vsu [all...] |
/external/libavc/encoder/x86/ |
H A D | ih264e_half_pel_ssse3.c | 135 src_r0_sht_16x8b = _mm_srli_si128(src_r0_16x8b, 1); //a1 a2 a3 a4 a5 a6 a7 a8 a9....a15 0 136 src_r1_sht_16x8b = _mm_srli_si128(src_r1_16x8b, 1); //b1 b2 b3 b4 b5 b6 b7 b8 b9....b15 0 146 src_r0_16x8b = _mm_srli_si128(src_r0_16x8b, 2); //a2 a3 a4 a5 a6 a7 a8 a9....a15 0 0 147 src_r1_16x8b = _mm_srli_si128(src_r1_16x8b, 2); //b2 b3 b4 b5 b6 b7 b8 b9....b15 0 0 149 src_r0_sht_16x8b = _mm_srli_si128(src_r0_sht_16x8b, 2); //a3 a4 a5 a6 a7 a8 a9....a15 0 0 0 150 src_r1_sht_16x8b = _mm_srli_si128(src_r1_sht_16x8b, 2); //b3 b4 b5 b6 b7 b8 b9....b15 0 0 0 160 src_r0_16x8b = _mm_srli_si128(src_r0_16x8b, 2); //a4 a5 a6 a7 a8 a9....a15 0 0 0 0 161 src_r1_16x8b = _mm_srli_si128(src_r1_16x8b, 2); //b4 b5 b6 b7 b8 b9....b15 0 0 0 0 163 src_r0_sht_16x8b = _mm_srli_si128(src_r0_sht_16x8b, 2); //a5 a6 a7 a8 a9....a15 0 0 0 0 0 164 src_r1_sht_16x8b = _mm_srli_si128(src_r1_sht_16x8 [all...] |
H A D | ih264e_intra_modes_eval_ssse3.c | 528 pred0_16x8b = _mm_srli_si128(left_top_16x8b, 5); 559 temp_16x8b = _mm_srli_si128(left_top_16x8b, 5); 609 temp1_16x8b = _mm_srli_si128(left_top_16x8b, 1); 612 temp2_16x8b = _mm_srli_si128(left_top_16x8b, 6); 614 temp3_8x16b = _mm_srli_si128(temp2_8x16b, 2); //t2 t3 t4 t5 t6 t7 0 0 618 temp3_8x16b = _mm_srli_si128(temp2_8x16b, 2); //t2+t3 t3+t4... t7+t7 0 0 655 temp1_16x8b = _mm_srli_si128(w121_16x8b, 1); 684 temp1_16x8b = _mm_srli_si128(w121_16x8b, 5); 685 temp2_16x8b = _mm_srli_si128(w11_16x8b, 5); 769 mask_low_32b = _mm_srli_si128(mask_low_32 [all...] |
/external/webrtc/webrtc/modules/video_processing/util/ |
H A D | denoiser_filter_sse2.cc | 47 vsum = _mm_add_epi16(vsum, _mm_srli_si128(vsum, 8)); 48 vsum = _mm_add_epi16(vsum, _mm_srli_si128(vsum, 4)); 49 vsum = _mm_add_epi16(vsum, _mm_srli_si128(vsum, 2)); 53 vsse = _mm_add_epi32(vsse, _mm_srli_si128(vsse, 8)); 54 vsse = _mm_add_epi32(vsse, _mm_srli_si128(vsse, 4)); 93 _mm_add_epi32(hg_fe_dc_ba, _mm_srli_si128(hg_fe_dc_ba, 8)); 95 _mm_add_epi32(hgfe_dcba, _mm_srli_si128(hgfe_dcba, 4));
|
/external/libvpx/libvpx/vp9/common/x86/ |
H A D | vp9_idct_intrin_sse2.c | 69 d0 = _mm_srli_si128(d0, 4); 72 d0 = _mm_srli_si128(d0, 4); 75 d0 = _mm_srli_si128(d0, 4);
|
/external/neven/Embedded/common/src/b_BasicEm/ |
H A D | MathSSE2.c | 192 m_XMM0 = _mm_srli_si128( m_XMM0, 4 ); 196 m_XMM0 = _mm_srli_si128( m_XMM0, 4 ); 200 m_XMM0 = _mm_srli_si128( m_XMM0, 4 ); 296 m_XMM0 = _mm_srli_si128( m_XMM0, 4 ); 300 m_XMM0 = _mm_srli_si128( m_XMM0, 4 ); 304 m_XMM0 = _mm_srli_si128( m_XMM0, 4 );
|
/external/fec/ |
H A D | viterbi39_sse2.c | 176 adjustv = _mm_min_epi16(adjustv,_mm_srli_si128(adjustv,8)); 177 adjustv = _mm_min_epi16(adjustv,_mm_srli_si128(adjustv,4)); 178 adjustv = _mm_min_epi16(adjustv,_mm_srli_si128(adjustv,2));
|