/external/chromium_org/third_party/libvpx/source/libvpx/vp9/common/x86/ |
H A D | vp9_subpixel_8t_intrin_ssse3.c | 82 srcRegFilt1 = _mm_maddubs_epi16(srcRegFilt1, firstFilters); 83 srcRegFilt2 = _mm_maddubs_epi16(srcRegFilt2, secondFilters); 157 srcRegFilt1 = _mm_maddubs_epi16(srcRegFilt1, firstFilters); 158 srcRegFilt2 = _mm_maddubs_epi16(srcRegFilt2, secondFilters); 165 srcRegFilt3 = _mm_maddubs_epi16(srcRegFilt3, thirdFilters); 166 srcRegFilt4 = _mm_maddubs_epi16(srcRegFilt4, forthFilters); 237 srcRegFilt1_1 = _mm_maddubs_epi16(srcRegFilt1_1, firstFilters); 238 srcRegFilt2 = _mm_maddubs_epi16(srcRegFilt2, forthFilters); 248 srcRegFilt3 = _mm_maddubs_epi16(srcRegFilt3, secondFilters); 249 srcRegFilt2 = _mm_maddubs_epi16(srcRegFilt [all...] |
H A D | vp9_subpixel_8t_intrin_avx2.c | 214 srcRegFilt1_1 = _mm_maddubs_epi16(srcRegFilt1_1, 216 srcRegFilt2 = _mm_maddubs_epi16(srcRegFilt2, 229 srcRegFilt3 = _mm_maddubs_epi16(srcRegFilt3, 231 srcRegFilt2 = _mm_maddubs_epi16(srcRegFilt2, 253 srcRegFilt2_1 = _mm_maddubs_epi16(srcRegFilt2_1, 255 srcRegFilt2 = _mm_maddubs_epi16(srcRegFilt2, 268 srcRegFilt3 = _mm_maddubs_epi16(srcRegFilt3, 270 srcRegFilt2 = _mm_maddubs_epi16(srcRegFilt2, 488 srcRegFilt1 = _mm_maddubs_epi16(_mm256_castsi256_si128(srcReg32b10), 490 srcRegFilt4 = _mm_maddubs_epi16(srcRegFilt [all...] |
/external/libvpx/libvpx/vp9/common/x86/ |
H A D | vp9_subpixel_8t_intrin_ssse3.c | 80 srcRegFilt1 = _mm_maddubs_epi16(srcRegFilt1, firstFilters); 81 srcRegFilt2 = _mm_maddubs_epi16(srcRegFilt2, secondFilters); 155 srcRegFilt1 = _mm_maddubs_epi16(srcRegFilt1, firstFilters); 156 srcRegFilt2 = _mm_maddubs_epi16(srcRegFilt2, secondFilters); 163 srcRegFilt3 = _mm_maddubs_epi16(srcRegFilt3, thirdFilters); 164 srcRegFilt4 = _mm_maddubs_epi16(srcRegFilt4, forthFilters); 235 srcRegFilt1_1 = _mm_maddubs_epi16(srcRegFilt1_1, firstFilters); 236 srcRegFilt2 = _mm_maddubs_epi16(srcRegFilt2, secondFilters); 246 srcRegFilt3 = _mm_maddubs_epi16(srcRegFilt3, forthFilters); 247 srcRegFilt2 = _mm_maddubs_epi16(srcRegFilt [all...] |
H A D | vp9_subpixel_8t_intrin_avx2.c | 214 srcRegFilt1_1 = _mm_maddubs_epi16(srcRegFilt1_1, 216 srcRegFilt2 = _mm_maddubs_epi16(srcRegFilt2, 229 srcRegFilt3 = _mm_maddubs_epi16(srcRegFilt3, 231 srcRegFilt2 = _mm_maddubs_epi16(srcRegFilt2, 253 srcRegFilt2_1 = _mm_maddubs_epi16(srcRegFilt2_1, 255 srcRegFilt2 = _mm_maddubs_epi16(srcRegFilt2, 268 srcRegFilt3 = _mm_maddubs_epi16(srcRegFilt3, 270 srcRegFilt2 = _mm_maddubs_epi16(srcRegFilt2, 488 srcRegFilt1 = _mm_maddubs_epi16(_mm256_castsi256_si128(srcReg32b10), 490 srcRegFilt4 = _mm_maddubs_epi16(srcRegFilt [all...] |
/external/libhevc/common/x86/ |
H A D | ihevc_inter_pred_filters_ssse3_intr.c | 336 res_temp1_8x16b = _mm_maddubs_epi16(src_temp3_16x8b, coeff0_1_8x16b); /* row = 0 */ 342 res_temp2_8x16b = _mm_maddubs_epi16(src_temp4_16x8b, coeff2_3_8x16b); /* row = 0 */ 348 res_temp3_8x16b = _mm_maddubs_epi16(src_temp5_16x8b, coeff4_5_8x16b); /* row = 0 */ 354 res_temp4_8x16b = _mm_maddubs_epi16(src_temp6_16x8b, coeff6_7_8x16b); /* row = 0 */ 369 res_temp11_8x16b = _mm_maddubs_epi16(src_temp13_16x8b, coeff0_1_8x16b); /* row = 1 */ 375 res_temp12_8x16b = _mm_maddubs_epi16(src_temp14_16x8b, coeff2_3_8x16b); /* row = 1 */ 381 res_temp13_8x16b = _mm_maddubs_epi16(src_temp15_16x8b, coeff4_5_8x16b); /* row = 1 */ 387 res_temp14_8x16b = _mm_maddubs_epi16(src_temp16_16x8b, coeff6_7_8x16b); /* row = 1 */ 425 res_temp1_8x16b = _mm_maddubs_epi16(src_temp3_16x8b, coeff0_1_8x16b); /* row = 0 */ 431 res_temp2_8x16b = _mm_maddubs_epi16(src_temp4_16x8 [all...] |
H A D | ihevc_intra_pred_filters_sse42_intr.c | 1515 src_temp1_8x16b = _mm_maddubs_epi16(src_temp1_8x16b, temp1_8x16b); 1516 src_temp2_8x16b = _mm_maddubs_epi16(src_temp2_8x16b, temp2_8x16b); 1517 src_temp3_8x16b = _mm_maddubs_epi16(src_temp3_8x16b, temp3_8x16b); 1518 src_temp4_8x16b = _mm_maddubs_epi16(src_temp4_8x16b, temp4_8x16b); 1654 src_temp1_8x16b = _mm_maddubs_epi16(src_temp1_8x16b, temp1_8x16b); 1655 src_temp2_8x16b = _mm_maddubs_epi16(src_temp2_8x16b, temp2_8x16b); 1656 src_temp3_8x16b = _mm_maddubs_epi16(src_temp3_8x16b, temp3_8x16b); 1657 src_temp4_8x16b = _mm_maddubs_epi16(src_temp4_8x16b, temp4_8x16b); 1660 src_temp11_8x16b = _mm_maddubs_epi16(src_temp11_8x16b, temp11_8x16b); 1661 src_temp12_8x16b = _mm_maddubs_epi16(src_temp12_8x16 [all...] |
H A D | ihevc_intra_pred_filters_ssse3_intr.c | 754 res_temp_8x16b = _mm_maddubs_epi16(pu1_ref_16x8b, row_16x8b); 755 res_temp1_8x16b = _mm_maddubs_epi16(pu1_ref_16x8b, row1_16x8b); 756 res_temp2_8x16b = _mm_maddubs_epi16(pu1_ref_16x8b, row2_16x8b); 757 res_temp3_8x16b = _mm_maddubs_epi16(pu1_ref_16x8b, row3_16x8b); 759 res_temp4_8x16b = _mm_maddubs_epi16(res_temp4_8x16b, nt_row_16x8b); 760 res_temp5_8x16b = _mm_maddubs_epi16(res_temp5_8x16b, nt_row1_16x8b); 761 res_temp6_8x16b = _mm_maddubs_epi16(res_temp6_8x16b, nt_row2_16x8b); 762 res_temp7_8x16b = _mm_maddubs_epi16(res_temp7_8x16b, nt_row3_16x8b); 832 res_temp_8x16b = _mm_maddubs_epi16(pu1_ref_16x8b, row_16x8b); 833 res_temp1_8x16b = _mm_maddubs_epi16(pu1_ref_16x8 [all...] |
H A D | ihevc_deblk_ssse3_intr.c | 153 mask_d_result_4x32b = _mm_maddubs_epi16(src_tmp_8x16b, coef_8x16b); 166 mask_de_result_8x16b = _mm_maddubs_epi16(mask_de_result_8x16b, temp_coef0_8x16b); 274 temp_pq_str0_16x8b = _mm_maddubs_epi16(temp_pq_str0_16x8b, const2_8x16b); 290 temp_pq1_str0_16x8b = _mm_maddubs_epi16(temp_pq1_str0_16x8b, const2_8x16b); 291 temp_pq1_str1_16x8b = _mm_maddubs_epi16(temp_pq1_str1_16x8b, const2_8x16b); 319 temp_pq2_str0_16x8b = _mm_maddubs_epi16(temp_str3_16x8b, const2_8x16b); 411 tmp_delta3_8x16b = _mm_maddubs_epi16(tmp_delta2_8x16b, coefdelta_0_8x16b); 433 tmp_delta3_8x16b = _mm_maddubs_epi16(tmp_delta2_8x16b, coefdelta_0_8x16b); 613 mask_d_result_4x32b = _mm_maddubs_epi16(src_tmp_p_0_8x16b, coef_8x16b); 626 mask_de_result_8x16b = _mm_maddubs_epi16(mask_de_result_8x16 [all...] |
H A D | ihevc_chroma_intra_pred_filters_ssse3_intr.c | 1278 src_temp1_8x16b = _mm_maddubs_epi16(src_temp1_8x16b, temp1_8x16b); 1279 src_temp2_8x16b = _mm_maddubs_epi16(src_temp2_8x16b, temp2_8x16b); 1280 src_temp3_8x16b = _mm_maddubs_epi16(src_temp3_8x16b, temp3_8x16b); 1281 src_temp4_8x16b = _mm_maddubs_epi16(src_temp4_8x16b, temp4_8x16b); 1446 src_temp1_8x16b = _mm_maddubs_epi16(src_temp1_8x16b, temp1_8x16b); 1447 src_temp2_8x16b = _mm_maddubs_epi16(src_temp2_8x16b, temp2_8x16b); 1448 src_temp3_8x16b = _mm_maddubs_epi16(src_temp3_8x16b, temp3_8x16b); 1449 src_temp4_8x16b = _mm_maddubs_epi16(src_temp4_8x16b, temp4_8x16b); 1452 src_temp11_8x16b = _mm_maddubs_epi16(src_temp11_8x16b, temp11_8x16b); 1453 src_temp12_8x16b = _mm_maddubs_epi16(src_temp12_8x16 [all...] |
/external/chromium_org/third_party/skia/src/opts/ |
H A D | SkBitmapProcState_opts_SSSE3.cpp | 166 return _mm_maddubs_epi16(a0, scale_x); 234 __m128i sum = _mm_maddubs_epi16(a0, scale_x); 360 a0 = _mm_maddubs_epi16(a0, scale_x); 382 // _mm_maddubs_epi16 : this allows us to perform multiplications and additions
|
/external/skia/src/opts/ |
H A D | SkBitmapProcState_opts_SSSE3.cpp | 167 return _mm_maddubs_epi16(a0, scale_x); 235 __m128i sum = _mm_maddubs_epi16(a0, scale_x); 361 a0 = _mm_maddubs_epi16(a0, scale_x); 383 // _mm_maddubs_epi16 : this allows us to perform multiplications and additions
|
/external/clang/lib/Headers/ |
H A D | tmmintrin.h | 152 _mm_maddubs_epi16(__m128i __a, __m128i __b) function
|
/external/chromium_org/third_party/libvpx/source/libvpx/third_party/libyuv/source/ |
H A D | row_win.cc | 15 #include <tmmintrin.h> // For _mm_maddubs_epi16 94 xmm0 = _mm_maddubs_epi16(xmm0, *(__m128i*)kUVToB); 95 xmm1 = _mm_maddubs_epi16(xmm1, *(__m128i*)kUVToG); 96 xmm2 = _mm_maddubs_epi16(xmm2, *(__m128i*)kUVToR); 147 xmm0 = _mm_maddubs_epi16(xmm0, *(__m128i*)kUVToB); 148 xmm1 = _mm_maddubs_epi16(xmm1, *(__m128i*)kUVToG); 149 xmm2 = _mm_maddubs_epi16(xmm2, *(__m128i*)kUVToR);
|
/external/chromium_org/third_party/libyuv/source/ |
H A D | row_win.cc | 15 #include <tmmintrin.h> // For _mm_maddubs_epi16 94 xmm0 = _mm_maddubs_epi16(xmm0, *(__m128i*)kUVToB); 95 xmm1 = _mm_maddubs_epi16(xmm1, *(__m128i*)kUVToG); 96 xmm2 = _mm_maddubs_epi16(xmm2, *(__m128i*)kUVToR); 148 xmm0 = _mm_maddubs_epi16(xmm0, *(__m128i*)kUVToB); 149 xmm1 = _mm_maddubs_epi16(xmm1, *(__m128i*)kUVToG); 150 xmm2 = _mm_maddubs_epi16(xmm2, *(__m128i*)kUVToR);
|