/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/common/arm/armv6/ |
H A D | simpleloopfilter_v6.asm | 49 src RN r0 label 62 ldr r3, [src, -pstep, lsl #1] ; p1 63 ldr r4, [src, -pstep] ; p0 64 ldr r5, [src] ; q0 65 ldr r6, [src, pstep] ; q1 120 str r5, [src] ; store oq0 result 122 str r4, [src, -pstep] ; store op0 result 126 addne src, src, #4 ; next row 128 ldrne r3, [src, [all...] |
H A D | loopfilter_v6.asm | 50 src RN r0 label 66 sub src, src, pstep, lsl #2 ; move src pointer down by 4 lines 71 ldr r9, [src], pstep ; p3 73 ldr r10, [src], pstep ; p2 75 ldr r11, [src], pstep ; p1 88 ldr r12, [src], pstep ; p0 102 ldr r9, [src], pstep ; q0 103 ldr r10, [src], pste [all...] |
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/common/arm/neon/ |
H A D | copymem_neon.c | 14 unsigned char *src, 22 vtmp = vld1_u8(src); 24 src += src_stride; 30 unsigned char *src, 38 vtmp = vld1_u8(src); 40 src += src_stride; 46 unsigned char *src, 54 qtmp = vld1q_u8(src); 56 src += src_stride; 13 vp8_copy_mem8x4_neon( unsigned char *src, int src_stride, unsigned char *dst, int dst_stride) argument 29 vp8_copy_mem8x8_neon( unsigned char *src, int src_stride, unsigned char *dst, int dst_stride) argument 45 vp8_copy_mem16x16_neon( unsigned char *src, int src_stride, unsigned char *dst, int dst_stride) argument
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/common/mips/dspr2/ |
H A D | vp9_common_dspr2.h | 54 static INLINE void vp9_prefetch_load(const unsigned char *src) { argument 56 "pref 0, 0(%[src]) \n\t" 58 : [src] "r" (src) 71 static INLINE void vp9_prefetch_load_streamed(const unsigned char *src) { argument 73 "pref 4, 0(%[src]) \n\t" 75 : [src] "r" (src) 91 void vp9_convolve2_horiz_dspr2(const uint8_t *src, ptrdiff_t src_stride, 97 void vp9_convolve2_avg_horiz_dspr2(const uint8_t *src, ptrdiff_ [all...] |
H A D | vp9_convolve2_avg_dspr2.c | 23 static void convolve_bi_avg_vert_4_dspr2(const uint8_t *src, argument 50 src_ptr = src + x; 123 src += src_stride; 128 static void convolve_bi_avg_vert_64_dspr2(const uint8_t *src, argument 155 src_ptr = src + x; 228 src += src_stride; 233 void vp9_convolve2_avg_vert_dspr2(const uint8_t *src, ptrdiff_t src_stride, argument 255 convolve_bi_avg_vert_4_dspr2(src, src_stride, 261 convolve_bi_avg_vert_64_dspr2(src, src_stride, 266 vp9_convolve8_avg_vert_c(src, src_strid [all...] |
H A D | vp9_convolve2_vert_dspr2.c | 23 static void convolve_bi_vert_4_dspr2(const uint8_t *src, argument 50 src_ptr = src + x; 116 src += src_stride; 121 static void convolve_bi_vert_64_dspr2(const uint8_t *src, argument 147 src_ptr = src + x; 213 src += src_stride; 218 void vp9_convolve2_vert_dspr2(const uint8_t *src, ptrdiff_t src_stride, argument 240 convolve_bi_vert_4_dspr2(src, src_stride, 246 convolve_bi_vert_64_dspr2(src, src_stride, 251 vp9_convolve8_vert_c(src, src_strid [all...] |
H A D | vp9_convolve8_vert_dspr2.c | 23 static void convolve_vert_4_dspr2(const uint8_t *src, argument 48 src -= 3 * src_stride; 55 src_ptr = src + x; 173 src += src_stride; 178 static void convolve_vert_64_dspr2(const uint8_t *src, argument 202 src -= 3 * src_stride; 210 src_ptr = src + x; 328 src += src_stride; 333 void vp9_convolve8_vert_dspr2(const uint8_t *src, ptrdiff_t src_stride, argument 339 vp9_convolve_copy(src, src_strid [all...] |
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/test/ |
H A D | vp8_decrypt_test.cc | 15 #include "third_party/googletest/src/include/gtest/gtest.h" 29 void encrypt_buffer(const uint8_t *src, uint8_t *dst, size_t size, argument 32 dst[i] = src[i] ^ test_key[(offset + i) & 15];
|
H A D | vp9_subtract_test.cc | 11 #include "third_party/googletest/src/include/gtest/gtest.h" 48 uint8_t *src = reinterpret_cast<uint8_t *>( local 54 src[r * block_width * 2 + c] = rnd.Rand8(); 60 src, block_width, pred, block_width); 65 (src[r * block_width + c] - 73 src, block_width * 2, pred, block_width * 2); 78 (src[r * block_width * 2 + c] - 87 vpx_free(src);
|
H A D | idct8x8_test.cc | 15 #include "third_party/googletest/src/include/gtest/gtest.h" 114 uint8_t dst[64], src[64]; local 117 src[j] = rnd.Rand8(); 122 input[j] = src[j] - dst[j]; 129 const int diff = dst[j] - src[j];
|
H A D | sixtap_predict_test.cc | 18 #include "third_party/googletest/src/include/gtest/gtest.h" 78 // The src stores the macroblock we will filter on, and makes it 1 byte larger 144 uint8_t *src = const_cast<uint8_t*>(test_data); local 146 REGISTER_STATE_CHECK(sixtap_predict_(&src[kSrcStride * 2 + 2 + 1], kSrcStride,
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/third_party/libmkv/ |
H A D | EbmlBufferWriter.c | 12 unsigned char *src = glob->buf; local 13 src += glob->offset; 14 memcpy(src, buffer_in, len);
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/common/mips/dspr2/ |
H A D | reconinter_dspr2.c | 17 inline void prefetch_load_int(unsigned char *src) argument 20 "pref 0, 0(%[src]) \n\t" 22 : [src] "r" (src) 28 unsigned char *RESTRICT src, 38 /* load src data in cache memory */ 39 prefetch_load_int(src + src_stride); 43 "ulw %[a0], 0(%[src]) \n\t" 44 "ulw %[a1], 4(%[src]) \n\t" 45 "ulw %[a2], 8(%[src]) \ 27 vp8_copy_mem16x16_dspr2( unsigned char *RESTRICT src, int src_stride, unsigned char *RESTRICT dst, int dst_stride) argument 62 vp8_copy_mem8x8_dspr2( unsigned char *RESTRICT src, int src_stride, unsigned char *RESTRICT dst, int dst_stride) argument 92 vp8_copy_mem8x4_dspr2( unsigned char *RESTRICT src, int src_stride, unsigned char *RESTRICT dst, int dst_stride) argument [all...] |
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/common/arm/neon/ |
H A D | vp9_convolve_neon.c | 15 void vp9_convolve8_neon(const uint8_t *src, ptrdiff_t src_stride, argument 29 return vp9_convolve8_c(src, src_stride, 40 vp9_convolve8_horiz_neon(src - src_stride * 3, src_stride, 52 void vp9_convolve8_avg_neon(const uint8_t *src, ptrdiff_t src_stride, argument 61 return vp9_convolve8_avg_c(src, src_stride, 70 vp9_convolve8_horiz_neon(src - src_stride * 3, src_stride,
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/common/ |
H A D | vp9_reconinter.h | 33 void vp9_build_inter_predictor(const uint8_t *src, int src_stride, 50 uint8_t *src, int stride, 56 dst->buf = src + scaled_buffer_offset(x, y, stride, scale); 60 void vp9_setup_dst_planes(MACROBLOCKD *xd, const YV12_BUFFER_CONFIG *src, 64 const YV12_BUFFER_CONFIG *src, int mi_row, int mi_col, 49 setup_pred_plane(struct buf_2d *dst, uint8_t *src, int stride, int mi_row, int mi_col, const struct scale_factors *scale, int subsampling_x, int subsampling_y) argument
|
/hardware/samsung_slsi/exynos5/exynos_omx/openmax/exynos_omx/osal/ |
H A D | Exynos_OSAL_Memory.c | 63 OMX_PTR Exynos_OSAL_Memcpy(OMX_PTR dest, OMX_PTR src, OMX_S32 n) argument 65 return memcpy(dest, src, n); 68 OMX_PTR Exynos_OSAL_Memmove(OMX_PTR dest, OMX_PTR src, OMX_S32 n) argument 70 return memmove(dest, src, n);
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vpx_scale/generic/ |
H A D | gen_scalers.c | 43 const unsigned char *src = source; local 48 a = src[0]; 49 b = src[1]; 50 c = src[2]; 51 d = src[3]; 52 e = src[4]; 59 src += 5; 75 unsigned char *src = source; local 79 a = src[0 * src_pitch]; 80 b = src[ 124 const unsigned char *src = source; local 153 unsigned char *src = source; local 199 const unsigned char *src = source; local [all...] |
H A D | yv12extend.c | 17 static void extend_plane(uint8_t *const src, int src_stride, argument 25 uint8_t *src_ptr1 = src; 26 uint8_t *src_ptr2 = src + width - 1; 27 uint8_t *dst_ptr1 = src - extend_left; 28 uint8_t *dst_ptr2 = src + width; 42 src_ptr1 = src - extend_left; 43 src_ptr2 = src + src_stride * (height - 1) - extend_left; 44 dst_ptr1 = src + src_stride * -extend_top - extend_left; 45 dst_ptr2 = src + src_stride * height - extend_left; 128 const uint8_t *src local 169 const uint8_t *src = src_ybc->y_buffer; local [all...] |
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/common/ |
H A D | extend.c | 75 void vp8_copy_and_extend_frame(YV12_BUFFER_CONFIG *src, argument 80 int eb = dst->border + dst->y_height - src->y_height; 81 int er = dst->border + dst->y_width - src->y_width; 83 copy_and_extend_plane(src->y_buffer, src->y_stride, 85 src->y_height, src->y_width, 90 eb = (dst->border >> 1) + dst->uv_height - src->uv_height; 91 er = (dst->border >> 1) + dst->uv_width - src->uv_width; 93 copy_and_extend_plane(src 105 vp8_copy_and_extend_frame_with_rect(YV12_BUFFER_CONFIG *src, YV12_BUFFER_CONFIG *dst, int srcy, int srcx, int srch, int srcw) argument [all...] |
H A D | findnearmv.c | 140 static void invert_and_clamp_mvs(int_mv *inv, int_mv *src, MACROBLOCKD *xd) argument 142 inv->as_mv.row = src->as_mv.row * -1; 143 inv->as_mv.col = src->as_mv.col * -1; 145 vp8_clamp_mv2(src, xd);
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/encoder/ |
H A D | lookahead.c | 105 YV12_BUFFER_CONFIG *src, 113 int mb_rows = (src->y_height + 15) >> 4; 114 int mb_cols = (src->y_width + 15) >> 4; 155 vp8_copy_and_extend_frame_with_rect(src, &buf->img, 169 vp8_copy_and_extend_frame(src, &buf->img); 104 vp8_lookahead_push(struct lookahead_ctx *ctx, YV12_BUFFER_CONFIG *src, int64_t ts_start, int64_t ts_end, unsigned int flags, unsigned char *active_map) argument
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/encoder/ |
H A D | vp9_lookahead.c | 93 int vp9_lookahead_push(struct lookahead_ctx *ctx, YV12_BUFFER_CONFIG *src, argument 98 int mb_rows = (src->y_height + 15) >> 4; 99 int mb_cols = (src->y_width + 15) >> 4; 139 vp9_copy_and_extend_frame_with_rect(src, &buf->img, 151 vp9_copy_and_extend_frame(src, &buf->img); 155 vp9_copy_and_extend_frame(src, &buf->img);
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/encoder/x86/ |
H A D | vp9_sad4d_intrin_avx2.c | 13 void vp9_sad32x32x4d_avx2(uint8_t *src, argument 33 // load src and all refs 34 src_reg = _mm256_load_si256((__m256i *)(src)); 39 // sum of the absolute differences between every ref-i to src 50 src+= src_stride; 83 void vp9_sad64x64x4d_avx2(uint8_t *src, argument 105 // load 64 bytes from src and all refs 106 src_reg = _mm256_load_si256((__m256i *)(src)); 107 srcnext_reg = _mm256_load_si256((__m256i *)(src + 32)); 116 // sum of the absolute differences between every ref-i to src [all...] |
H A D | vp9_variance_avx2.c | 47 const uint8_t *src, 59 const uint8_t *src, 185 unsigned int vp9_sub_pixel_variance64x64_avx2(const uint8_t *src, argument 194 int se = vp9_sub_pixel_variance32xh_avx2(src, src_stride, x_offset, 199 int se2 = vp9_sub_pixel_variance32xh_avx2(src + 32, src_stride, 209 unsigned int vp9_sub_pixel_variance32x32_avx2(const uint8_t *src, argument 218 int se = vp9_sub_pixel_variance32xh_avx2(src, src_stride, x_offset, 225 unsigned int vp9_sub_pixel_avg_variance64x64_avx2(const uint8_t *src, argument 236 int se = vp9_sub_pixel_avg_variance32xh_avx2(src, src_stride, x_offset, 241 int se2 = vp9_sub_pixel_avg_variance32xh_avx2(src 251 vp9_sub_pixel_avg_variance32x32_avx2(const uint8_t *src, int src_stride, int x_offset, int y_offset, const uint8_t *dst, int dst_stride, unsigned int *sseptr, const uint8_t *sec) argument [all...] |
H A D | vp9_variance_impl_intrin_avx2.c | 19 __m256i src, src_expand_low, src_expand_high, ref, ref_expand_low; local 31 src = _mm256_castsi128_si256( 33 src = _mm256_inserti128_si256(src, 42 src_expand_low = _mm256_unpacklo_epi8(src, zero_reg); 43 src_expand_high = _mm256_unpackhi_epi8(src, zero_reg); 48 // src-ref 52 // madd low (src - ref) 58 // madd high (src - ref) 130 __m256i src, src_expand_lo local [all...] |