Searched refs:_mm_srli_si128 (Results 1 - 25 of 76) sorted by relevance

1234

/external/clang/test/CodeGen/
H A Dsse.c31 return _mm_srli_si128(a, 5);
37 return _mm_srli_si128(a, 0);
43 return _mm_srli_si128(a, 16);
/external/clang/test/Sema/
H A Dwarn-shadow-intrinsics.c9 qf = _mm_slli_si128(_mm_add_epi64(qf, _mm_srli_si128(qf, 8)), 8); // no-warning
/external/libavc/common/x86/
H A Dih264_luma_intra_pred_filters_ssse3.c342 top_sh_8x16b = _mm_srli_si128(top_8x16b, 2); //t1 t2 t3 t4 t5 t6 t7 0
347 res2_8x16b = _mm_srli_si128(res1_8x16b, 2);
358 res_16x8b = _mm_srli_si128(res_16x8b, 1);
360 res_16x8b = _mm_srli_si128(res_16x8b, 1);
362 res_16x8b = _mm_srli_si128(res_16x8b, 1);
426 top_left_sh_16x8b = _mm_srli_si128(top_left_16x8b, 1); //l2 l1 l0 tl t0 t1 t2 t3...
433 res2_8x16b = _mm_srli_si128(res1_8x16b, 2); //l2+l1 l1+l0 l0+tl tl+t0 t0+t1 t1+t2 t2+t3...
443 res2_16x8b = _mm_srli_si128(res1_16x8b, 3);
446 res2_16x8b = _mm_srli_si128(res1_16x8b, 2);
448 res2_16x8b = _mm_srli_si128(res1_16x8
[all...]
H A Dih264_inter_pred_filters_ssse3.c255 src_r0_sht_16x8b = _mm_srli_si128(src_r0_16x8b, 1); //a1 a2 a3 a4 a5 a6 a7 a8 a9....a15 0
256 src_r1_sht_16x8b = _mm_srli_si128(src_r1_16x8b, 1); //b1 b2 b3 b4 b5 b6 b7 b8 b9....b15 0
265 src_r0_16x8b = _mm_srli_si128(src_r0_16x8b, 4); //a2 a3 a3 a4 a4 a5 a5 a6 a6 a7 a7 a8 0 0 0 0
266 src_r1_16x8b = _mm_srli_si128(src_r1_16x8b, 4); //b2 b3 b3 b4 b4 b5 b5 b6 b6 b7 b7 b8 0 0 0 0
272 src_r0_16x8b = _mm_srli_si128(src_r0_16x8b, 4); //a4 a5 a5 a6 a6 a7 a7 a8 0 0 0 0 0 0 0 0
273 src_r1_16x8b = _mm_srli_si128(src_r1_16x8b, 4); //b4 b5 b5 b6 b6 b7 b7 b8 0 0 0 0 0 0 0 0
295 res_r0r1_16x8b = _mm_srli_si128(res_r0r1_16x8b, 4);
320 src_r0_sht_16x8b = _mm_srli_si128(src_r0_16x8b, 1); //a1 a2 a3 a4 a5 a6 a7 a8 a9....a15 0
321 src_r1_sht_16x8b = _mm_srli_si128(src_r1_16x8b, 1); //b1 b2 b3 b4 b5 b6 b7 b8 b9....b15 0
331 src_r0_16x8b = _mm_srli_si128(src_r0_16x8
[all...]
H A Dih264_iquant_itrans_recon_dc_ssse3.c158 temp5 = _mm_srli_si128(temp4,4);
159 temp6 = _mm_srli_si128(temp5,4);
160 temp7 = _mm_srli_si128(temp6,4);
439 pred_r1 = _mm_srli_si128(pred_r0, 4);
440 pred_r2 = _mm_srli_si128(pred_r1, 4);
441 pred_r3 = _mm_srli_si128(pred_r2, 4);
H A Dih264_deblk_chroma_ssse3.c224 lineb = _mm_srli_si128(linea, 8);
226 lined = _mm_srli_si128(linec, 8);
228 linef = _mm_srli_si128(linee, 8);
230 lineh = _mm_srli_si128(lineg, 8);
592 lineb = _mm_srli_si128(linea, 8);
594 lined = _mm_srli_si128(linec, 8);
596 linef = _mm_srli_si128(linee, 8);
598 lineh = _mm_srli_si128(lineg, 8);
862 p0_uv_16x8 = _mm_srli_si128(p1_uv_16x8, 8);
864 q1_uv_16x8 = _mm_srli_si128(q0_uv_16x
[all...]
/external/libhevc/decoder/x86/
H A Dihevcd_it_rec_dc_sse42_intr.c115 m_temp_reg_1 = _mm_srli_si128(m_temp_reg_8, 4);
116 m_temp_reg_2 = _mm_srli_si128(m_temp_reg_8, 8);
117 m_temp_reg_3 = _mm_srli_si128(m_temp_reg_8, 12);
163 m_temp_reg_13 = _mm_srli_si128(m_temp_reg_12, 8);
173 m_temp_reg_15 = _mm_srli_si128(m_temp_reg_14, 8);
277 m_temp_reg_8 = _mm_srli_si128(m_temp_reg_8, 4);
279 m_temp_reg_8 = _mm_srli_si128(m_temp_reg_8, 4);
281 m_temp_reg_8 = _mm_srli_si128(m_temp_reg_8, 4);
351 m_temp_reg_4 = _mm_srli_si128(m_temp_reg_4, 8);
355 m_temp_reg_5 = _mm_srli_si128(m_temp_reg_
[all...]
H A Dihevcd_it_rec_dc_ssse3_intr.c116 m_temp_reg_1 = _mm_srli_si128(m_temp_reg_8, 4);
117 m_temp_reg_2 = _mm_srli_si128(m_temp_reg_8, 8);
118 m_temp_reg_3 = _mm_srli_si128(m_temp_reg_8, 12);
164 m_temp_reg_13 = _mm_srli_si128(m_temp_reg_12, 8);
174 m_temp_reg_15 = _mm_srli_si128(m_temp_reg_14, 8);
277 m_temp_reg_8 = _mm_srli_si128(m_temp_reg_8, 4);
279 m_temp_reg_8 = _mm_srli_si128(m_temp_reg_8, 4);
281 m_temp_reg_8 = _mm_srli_si128(m_temp_reg_8, 4);
351 m_temp_reg_4 = _mm_srli_si128(m_temp_reg_4, 8);
355 m_temp_reg_5 = _mm_srli_si128(m_temp_reg_
[all...]
/external/libmpeg2/common/x86/
H A Dideint_cac_ssse3.c151 sum_1 = _mm_cvtsi128_si32(_mm_srli_si128(sum_t[i + 0], 8));
152 sum_2 = _mm_cvtsi128_si32(_mm_srli_si128(sum_b[i + 0], 8));
157 sum_3 = _mm_cvtsi128_si32(_mm_srli_si128(sum_t[i + 1], 8));
158 sum_4 = _mm_cvtsi128_si32(_mm_srli_si128(sum_b[i + 1], 8));
224 diff_sum = _mm_cvtsi128_si32(_mm_srli_si128(diff, 8));
H A Dicv_variance_ssse3.c129 vsum = _mm_add_epi64(vsum, _mm_srli_si128(vsum, 8));
149 vsum_sqr = _mm_add_epi32(vsum_sqr, _mm_srli_si128(vsum_sqr, 8));
150 vsum_sqr = _mm_add_epi32(vsum_sqr, _mm_srli_si128(vsum_sqr, 4));
H A Dideint_spatial_filter_ssse3.c164 adiff[3] = _mm_cvtsi128_si32(_mm_srli_si128(diffs[0], 8));
165 adiff[4] = _mm_cvtsi128_si32(_mm_srli_si128(diffs[1], 8));
166 adiff[5] = _mm_cvtsi128_si32(_mm_srli_si128(diffs[2], 8));
233 _mm_storel_epi64((__m128i *)pu1_out, _mm_srli_si128(dst, 8));
H A Dicv_sad_ssse3.c140 res_r0 = _mm_add_epi64(res_r0, _mm_srli_si128(res_r0, 8));
/external/libhevc/common/x86/
H A Dihevc_inter_pred_filters_ssse3_intr.c297 mask_high_96b = _mm_srli_si128(mask_low_32b, 12);
333 src_temp2_16x8b = _mm_srli_si128(src_temp1_16x8b, 1); /* row = 0 */
338 src_temp1_16x8b = _mm_srli_si128(src_temp1_16x8b, 2); /* row = 0 */
339 src_temp2_16x8b = _mm_srli_si128(src_temp2_16x8b, 2); /* row = 0 */
344 src_temp1_16x8b = _mm_srli_si128(src_temp1_16x8b, 2); /* row = 0 */
345 src_temp2_16x8b = _mm_srli_si128(src_temp2_16x8b, 2); /* row = 0 */
350 src_temp1_16x8b = _mm_srli_si128(src_temp1_16x8b, 2); /* row = 0 */
351 src_temp2_16x8b = _mm_srli_si128(src_temp2_16x8b, 2); /* row = 0 */
366 src_temp12_16x8b = _mm_srli_si128(src_temp11_16x8b, 1); /* row =1 */
371 src_temp11_16x8b = _mm_srli_si128(src_temp11_16x8
[all...]
H A Dihevc_chroma_intra_pred_filters_sse42_intr.c333 src_temp3 = _mm_srli_si128(src_temp3, 8);
334 src_temp4 = _mm_srli_si128(src_temp4, 8);
335 src_temp7 = _mm_srli_si128(src_temp7, 8);
336 src_temp8 = _mm_srli_si128(src_temp8, 8);
357 temp_sad = _mm_srli_si128(src_temp4, 4); /* Next 32 bits */
371 src_temp3 = _mm_srli_si128(src_temp3, 8);
372 src_temp4 = _mm_srli_si128(src_temp4, 8);
386 temp_sad = _mm_srli_si128(src_temp4, 4); /* Next 32 bits */
397 src_temp4 = _mm_srli_si128(src_temp3, 8);
407 temp_sad = _mm_srli_si128(src_temp
[all...]
H A Dihevc_deblk_ssse3_intr.c183 temp_coef0_8x16b = _mm_srli_si128(mask_d_result_4x32b, 4);
184 temp_coef1_8x16b = _mm_srli_si128(mask_d_result_4x32b, 8);
185 mask_16x8b = _mm_srli_si128(mask_d_result_4x32b, 12);
195 temp_coef0_8x16b = _mm_srli_si128(mask_de_result_8x16b, 4);
196 temp_coef1_8x16b = _mm_srli_si128(mask_de_result_8x16b, 8);
197 mask_16x8b = _mm_srli_si128(mask_de_result_8x16b, 12);
379 src_row2_8x16b = _mm_srli_si128(src_row0_8x16b, 8);
380 src_row3_8x16b = _mm_srli_si128(src_row1_8x16b, 8);
525 src_row1_8x16b = _mm_srli_si128(src_row0_8x16b, 8);
526 src_row3_8x16b = _mm_srli_si128(src_row2_8x16
[all...]
H A Dihevc_chroma_intra_pred_filters_ssse3_intr.c354 src_temp3 = _mm_srli_si128(src_temp3, 8);
355 src_temp4 = _mm_srli_si128(src_temp4, 8);
356 src_temp7 = _mm_srli_si128(src_temp7, 8);
357 src_temp8 = _mm_srli_si128(src_temp8, 8);
380 temp_sad = _mm_srli_si128(src_temp4, 4); /* Next 32 bits */
394 src_temp3 = _mm_srli_si128(src_temp3, 8);
395 src_temp4 = _mm_srli_si128(src_temp4, 8);
411 temp_sad = _mm_srli_si128(src_temp4, 4); /* Next 32 bits */
422 src_temp4 = _mm_srli_si128(src_temp3, 8);
435 temp_sad = _mm_srli_si128(src_temp
[all...]
H A Dihevc_intra_pred_filters_sse42_intr.c257 src_temp2 = _mm_srli_si128(src_temp1, 1);
258 src_temp3 = _mm_srli_si128(src_temp2, 1);
273 src_temp5 = _mm_srli_si128(src_temp4, 1);
274 src_temp6 = _mm_srli_si128(src_temp5, 1);
956 src_temp2 = _mm_srli_si128(src_temp1, 1);
957 src_temp3 = _mm_srli_si128(src_temp1, 2);
958 src_temp4 = _mm_srli_si128(src_temp1, 3);
959 src_temp5 = _mm_srli_si128(src_temp1, 4);
960 src_temp6 = _mm_srli_si128(src_temp1, 5);
961 src_temp7 = _mm_srli_si128(src_temp
[all...]
H A Dihevc_intra_pred_filters_ssse3_intr.c543 src_temp2 = _mm_srli_si128(src_temp1, 1);
544 src_temp3 = _mm_srli_si128(src_temp2, 1);
559 src_temp5 = _mm_srli_si128(src_temp4, 1);
560 src_temp6 = _mm_srli_si128(src_temp5, 1);
1466 src_temp2 = _mm_srli_si128(src_temp1, 1);
1467 src_temp3 = _mm_srli_si128(src_temp1, 2);
1468 src_temp4 = _mm_srli_si128(src_temp1, 3);
1469 src_temp5 = _mm_srli_si128(src_temp1, 4);
1470 src_temp6 = _mm_srli_si128(src_temp1, 5);
1471 src_temp7 = _mm_srli_si128(src_temp
[all...]
/external/libvpx/libvpx/vpx_dsp/x86/
H A Dvariance_sse2.c32 vsum = _mm_add_epi32(vsum, _mm_srli_si128(vsum, 8));
33 vsum = _mm_add_epi32(vsum, _mm_srli_si128(vsum, 4));
54 vsum = _mm_add_epi16(vsum, _mm_srli_si128(vsum, 8));
55 vsum = _mm_add_epi16(vsum, _mm_srli_si128(vsum, 4));
56 vsum = _mm_add_epi16(vsum, _mm_srli_si128(vsum, 2));
62 vsum = _mm_add_epi32(vsum, _mm_srli_si128(vsum, 8));
63 vsum = _mm_add_epi32(vsum, _mm_srli_si128(vsum, 4));
95 vsum = _mm_add_epi16(vsum, _mm_srli_si128(vsum, 8));
96 vsum = _mm_add_epi16(vsum, _mm_srli_si128(vsum, 4));
97 vsum = _mm_add_epi16(vsum, _mm_srli_si128(vsu
[all...]
/external/libavc/encoder/x86/
H A Dih264e_half_pel_ssse3.c135 src_r0_sht_16x8b = _mm_srli_si128(src_r0_16x8b, 1); //a1 a2 a3 a4 a5 a6 a7 a8 a9....a15 0
136 src_r1_sht_16x8b = _mm_srli_si128(src_r1_16x8b, 1); //b1 b2 b3 b4 b5 b6 b7 b8 b9....b15 0
146 src_r0_16x8b = _mm_srli_si128(src_r0_16x8b, 2); //a2 a3 a4 a5 a6 a7 a8 a9....a15 0 0
147 src_r1_16x8b = _mm_srli_si128(src_r1_16x8b, 2); //b2 b3 b4 b5 b6 b7 b8 b9....b15 0 0
149 src_r0_sht_16x8b = _mm_srli_si128(src_r0_sht_16x8b, 2); //a3 a4 a5 a6 a7 a8 a9....a15 0 0 0
150 src_r1_sht_16x8b = _mm_srli_si128(src_r1_sht_16x8b, 2); //b3 b4 b5 b6 b7 b8 b9....b15 0 0 0
160 src_r0_16x8b = _mm_srli_si128(src_r0_16x8b, 2); //a4 a5 a6 a7 a8 a9....a15 0 0 0 0
161 src_r1_16x8b = _mm_srli_si128(src_r1_16x8b, 2); //b4 b5 b6 b7 b8 b9....b15 0 0 0 0
163 src_r0_sht_16x8b = _mm_srli_si128(src_r0_sht_16x8b, 2); //a5 a6 a7 a8 a9....a15 0 0 0 0 0
164 src_r1_sht_16x8b = _mm_srli_si128(src_r1_sht_16x8
[all...]
H A Dih264e_intra_modes_eval_ssse3.c528 pred0_16x8b = _mm_srli_si128(left_top_16x8b, 5);
559 temp_16x8b = _mm_srli_si128(left_top_16x8b, 5);
609 temp1_16x8b = _mm_srli_si128(left_top_16x8b, 1);
612 temp2_16x8b = _mm_srli_si128(left_top_16x8b, 6);
614 temp3_8x16b = _mm_srli_si128(temp2_8x16b, 2); //t2 t3 t4 t5 t6 t7 0 0
618 temp3_8x16b = _mm_srli_si128(temp2_8x16b, 2); //t2+t3 t3+t4... t7+t7 0 0
655 temp1_16x8b = _mm_srli_si128(w121_16x8b, 1);
684 temp1_16x8b = _mm_srli_si128(w121_16x8b, 5);
685 temp2_16x8b = _mm_srli_si128(w11_16x8b, 5);
769 mask_low_32b = _mm_srli_si128(mask_low_32
[all...]
/external/webrtc/webrtc/modules/video_processing/util/
H A Ddenoiser_filter_sse2.cc47 vsum = _mm_add_epi16(vsum, _mm_srli_si128(vsum, 8));
48 vsum = _mm_add_epi16(vsum, _mm_srli_si128(vsum, 4));
49 vsum = _mm_add_epi16(vsum, _mm_srli_si128(vsum, 2));
53 vsse = _mm_add_epi32(vsse, _mm_srli_si128(vsse, 8));
54 vsse = _mm_add_epi32(vsse, _mm_srli_si128(vsse, 4));
93 _mm_add_epi32(hg_fe_dc_ba, _mm_srli_si128(hg_fe_dc_ba, 8));
95 _mm_add_epi32(hgfe_dcba, _mm_srli_si128(hgfe_dcba, 4));
/external/libvpx/libvpx/vp9/common/x86/
H A Dvp9_idct_intrin_sse2.c69 d0 = _mm_srli_si128(d0, 4);
72 d0 = _mm_srli_si128(d0, 4);
75 d0 = _mm_srli_si128(d0, 4);
/external/neven/Embedded/common/src/b_BasicEm/
H A DMathSSE2.c192 m_XMM0 = _mm_srli_si128( m_XMM0, 4 );
196 m_XMM0 = _mm_srli_si128( m_XMM0, 4 );
200 m_XMM0 = _mm_srli_si128( m_XMM0, 4 );
296 m_XMM0 = _mm_srli_si128( m_XMM0, 4 );
300 m_XMM0 = _mm_srli_si128( m_XMM0, 4 );
304 m_XMM0 = _mm_srli_si128( m_XMM0, 4 );
/external/fec/
H A Dviterbi39_sse2.c176 adjustv = _mm_min_epi16(adjustv,_mm_srli_si128(adjustv,8));
177 adjustv = _mm_min_epi16(adjustv,_mm_srli_si128(adjustv,4));
178 adjustv = _mm_min_epi16(adjustv,_mm_srli_si128(adjustv,2));

Completed in 1343 milliseconds

1234