Searched refs:_mm256_add_epi16 (Results 1 - 10 of 10) sorted by relevance
/external/chromium_org/third_party/libvpx/source/libvpx/vp9/common/x86/ |
H A D | vp9_loopfilter_intrin_avx2.c | 593 pixelFilter_p = _mm256_add_epi16(_mm256_add_epi16(p256_6, p256_5), 594 _mm256_add_epi16(p256_4, p256_3)); 595 pixelFilter_q = _mm256_add_epi16(_mm256_add_epi16(q256_6, q256_5), 596 _mm256_add_epi16(q256_4, q256_3)); 598 pixetFilter_p2p1p0 = _mm256_add_epi16(p256_0, 599 _mm256_add_epi16(p256_2, p256_1)); 600 pixelFilter_p = _mm256_add_epi16(pixelFilter_p, pixetFilter_p2p1p0); 602 pixetFilter_q2q1q0 = _mm256_add_epi16(q256_ [all...] |
/external/libvpx/libvpx/vp9/common/x86/ |
H A D | vp9_loopfilter_intrin_avx2.c | 593 pixelFilter_p = _mm256_add_epi16(_mm256_add_epi16(p256_6, p256_5), 594 _mm256_add_epi16(p256_4, p256_3)); 595 pixelFilter_q = _mm256_add_epi16(_mm256_add_epi16(q256_6, q256_5), 596 _mm256_add_epi16(q256_4, q256_3)); 598 pixetFilter_p2p1p0 = _mm256_add_epi16(p256_0, 599 _mm256_add_epi16(p256_2, p256_1)); 600 pixelFilter_p = _mm256_add_epi16(pixelFilter_p, pixetFilter_p2p1p0); 602 pixetFilter_q2q1q0 = _mm256_add_epi16(q256_ [all...] |
/external/chromium_org/third_party/libvpx/source/libvpx/vp9/encoder/x86/ |
H A D | vp9_dct32x32_avx2.c | 122 step1a[ 0] = _mm256_add_epi16(ina0, inb0); 123 step1a[ 1] = _mm256_add_epi16(ina1, inb1); 124 step1a[ 2] = _mm256_add_epi16(ina2, inb2); 125 step1a[ 3] = _mm256_add_epi16(ina3, inb3); 152 step1a[ 0] = _mm256_add_epi16(ina0, inb0); 153 step1a[ 1] = _mm256_add_epi16(ina1, inb1); 154 step1a[ 2] = _mm256_add_epi16(ina2, inb2); 155 step1a[ 3] = _mm256_add_epi16(ina3, inb3); 182 step1a[ 0] = _mm256_add_epi16(ina0, inb0); 183 step1a[ 1] = _mm256_add_epi16(ina [all...] |
H A D | vp9_variance_impl_intrin_avx2.c | 56 src_expand_low = _mm256_add_epi16(src_expand_low, src_expand_high); 61 sum_ref_src = _mm256_add_epi16(sum_ref_src, src_expand_low); 158 src_expand_low = _mm256_add_epi16(src_expand_low, src_expand_high); 163 sum_ref_src = _mm256_add_epi16(sum_ref_src, src_expand_low);
|
H A D | vp9_subpel_variance_impl_intrin_avx2.c | 56 exp_src_lo = _mm256_add_epi16(exp_src_lo, pw8); \ 57 exp_src_hi = _mm256_add_epi16(exp_src_hi, pw8); \ 91 sum_reg = _mm256_add_epi16(sum_reg, exp_src_lo); \ 93 sum_reg = _mm256_add_epi16(sum_reg, exp_src_hi); \
|
/external/libvpx/libvpx/vp9/encoder/x86/ |
H A D | vp9_dct32x32_avx2.c | 122 step1a[ 0] = _mm256_add_epi16(ina0, inb0); 123 step1a[ 1] = _mm256_add_epi16(ina1, inb1); 124 step1a[ 2] = _mm256_add_epi16(ina2, inb2); 125 step1a[ 3] = _mm256_add_epi16(ina3, inb3); 152 step1a[ 0] = _mm256_add_epi16(ina0, inb0); 153 step1a[ 1] = _mm256_add_epi16(ina1, inb1); 154 step1a[ 2] = _mm256_add_epi16(ina2, inb2); 155 step1a[ 3] = _mm256_add_epi16(ina3, inb3); 182 step1a[ 0] = _mm256_add_epi16(ina0, inb0); 183 step1a[ 1] = _mm256_add_epi16(ina [all...] |
H A D | vp9_variance_impl_intrin_avx2.c | 56 src_expand_low = _mm256_add_epi16(src_expand_low, src_expand_high); 61 sum_ref_src = _mm256_add_epi16(sum_ref_src, src_expand_low); 158 src_expand_low = _mm256_add_epi16(src_expand_low, src_expand_high); 163 sum_ref_src = _mm256_add_epi16(sum_ref_src, src_expand_low);
|
H A D | vp9_subpel_variance_impl_intrin_avx2.c | 56 exp_src_lo = _mm256_add_epi16(exp_src_lo, pw8); \ 57 exp_src_hi = _mm256_add_epi16(exp_src_hi, pw8); \ 91 sum_reg = _mm256_add_epi16(sum_reg, exp_src_lo); \ 93 sum_reg = _mm256_add_epi16(sum_reg, exp_src_hi); \
|
/external/clang/test/CodeGen/ |
H A D | avx2-builtins.c | 60 return _mm256_add_epi16(a, b);
|
/external/clang/lib/Headers/ |
H A D | avx2intrin.h | 83 _mm256_add_epi16(__m256i __a, __m256i __b) function
|
Completed in 4397 milliseconds