/external/libvpx/libvpx/vp9/common/ |
H A D | vp9_common_data.c | 79 BLOCK_16X16, BLOCK_16X32, BLOCK_32X16, 98 BLOCK_16X16, BLOCK_INVALID, BLOCK_INVALID, 127 {{BLOCK_16X16, BLOCK_16X8}, {BLOCK_8X16, BLOCK_8X8}}, 128 {{BLOCK_16X32, BLOCK_16X16}, {BLOCK_INVALID, BLOCK_8X16}}, 129 {{BLOCK_32X16, BLOCK_INVALID}, {BLOCK_16X16, BLOCK_16X8}}, 130 {{BLOCK_32X32, BLOCK_32X16}, {BLOCK_16X32, BLOCK_16X16}},
|
H A D | vp9_enums.h | 53 BLOCK_16X16, enumerator in enum:BLOCK_SIZE
|
/external/chromium_org/third_party/libvpx/source/libvpx/vp9/common/ |
H A D | vp9_common_data.c | 79 BLOCK_16X16, BLOCK_16X32, BLOCK_32X16, 98 BLOCK_16X16, BLOCK_INVALID, BLOCK_INVALID, 113 BLOCK_16X16, // TX_16X16 134 {{BLOCK_16X16, BLOCK_16X8}, {BLOCK_8X16, BLOCK_8X8}}, 135 {{BLOCK_16X32, BLOCK_16X16}, {BLOCK_INVALID, BLOCK_8X16}}, 136 {{BLOCK_32X16, BLOCK_INVALID}, {BLOCK_16X16, BLOCK_16X8}}, 137 {{BLOCK_32X32, BLOCK_32X16}, {BLOCK_16X32, BLOCK_16X16}},
|
H A D | vp9_enums.h | 50 BLOCK_16X16, enumerator in enum:BLOCK_SIZE
|
/external/chromium_org/third_party/libvpx/source/libvpx/vp9/encoder/ |
H A D | vp9_mbgraph.c | 30 const vp9_variance_fn_ptr_t v_fn_ptr = cpi->fn_ptr[BLOCK_16X16]; 69 vp9_build_inter_predictors_sby(xd, mb_row, mb_col, BLOCK_16X16); 251 mi_local.mbmi.sb_type = BLOCK_16X16;
|
H A D | vp9_encodeframe.c | 107 return BLOCK_16X16; 123 return BLOCK_16X16; 278 case BLOCK_16X16: { 447 fill_variance_tree(&vt.split[i].split[j], BLOCK_16X16); 475 BLOCK_16X16); 488 BLOCK_16X16, 579 if (bsize < BLOCK_16X16) 743 const int energy = bsize <= BLOCK_16X16 ? x->mb_energy 1193 mi_8x8[index].src_mi->mbmi.sb_type = BLOCK_16X16; 1231 int bh = num_8x8_blocks_high_lookup[BLOCK_16X16]; [all...] |
H A D | vp9_context_tree.c | 16 BLOCK_16X16,
|
H A D | vp9_temporal_filter.c | 173 &cpi->fn_ptr[BLOCK_16X16], 0, &best_ref_mv1, ref_mv); 180 &cpi->fn_ptr[BLOCK_16X16],
|
H A D | vp9_speed_features.c | 389 sf->always_this_block_size = BLOCK_16X16;
|
/external/libvpx/libvpx/vp9/encoder/ |
H A D | vp9_mbgraph.c | 31 vp9_variance_fn_ptr_t v_fn_ptr = cpi->fn_ptr[BLOCK_16X16]; 67 vp9_build_inter_predictors_sby(xd, mb_row, mb_col, BLOCK_16X16); 256 mi_local.mbmi.sb_type = BLOCK_16X16;
|
H A D | vp9_encodeframe.c | 60 case BLOCK_16X16: 141 return BLOCK_16X16; 157 return BLOCK_16X16; 181 if (bsize <= BLOCK_16X16) { 356 case BLOCK_16X16: { 526 fill_variance_tree(&vt.split[i].split[j], BLOCK_16X16); 554 BLOCK_16X16); 567 BLOCK_16X16, 891 if (bsize < BLOCK_16X16) 1019 const int energy = bsize <= BLOCK_16X16 [all...] |
H A D | vp9_temporal_filter.c | 164 &cpi->fn_ptr[BLOCK_16X16], 0, &best_ref_mv1, ref_mv); 176 &cpi->fn_ptr[BLOCK_16X16],
|
H A D | vp9_block.h | 214 case BLOCK_16X16:
|
H A D | vp9_speed_features.c | 344 sf->always_this_block_size = BLOCK_16X16;
|
H A D | vp9_encodemb.c | 622 mbmi->tx_size = use_16x16_pred ? (mbmi->sb_type >= BLOCK_16X16 ? TX_16X16
|
/external/chromium_org/third_party/libvpx/source/libvpx/vp8/encoder/ |
H A D | rdopt.c | 1688 near_sad[1] = cpi->fn_ptr[BLOCK_16X16].sdf(src_y_ptr, b->src_stride, xd->dst.y_buffer - 16,xd->dst.y_stride, UINT_MAX); 1692 near_sad[0] = cpi->fn_ptr[BLOCK_16X16].sdf(src_y_ptr, b->src_stride, xd->dst.y_buffer - xd->dst.y_stride *16,xd->dst.y_stride, UINT_MAX); 1695 near_sad[0] = cpi->fn_ptr[BLOCK_16X16].sdf(src_y_ptr, b->src_stride, xd->dst.y_buffer - xd->dst.y_stride *16,xd->dst.y_stride, UINT_MAX); 1696 near_sad[1] = cpi->fn_ptr[BLOCK_16X16].sdf(src_y_ptr, b->src_stride, xd->dst.y_buffer - 16,xd->dst.y_stride, UINT_MAX); 1697 near_sad[2] = cpi->fn_ptr[BLOCK_16X16].sdf(src_y_ptr, b->src_stride, xd->dst.y_buffer - xd->dst.y_stride *16 -16,xd->dst.y_stride, UINT_MAX); 1712 near_sad[4] = cpi->fn_ptr[BLOCK_16X16].sdf(src_y_ptr, b->src_stride, pre_y_buffer - pre_y_stride *16, pre_y_stride, UINT_MAX); 1714 near_sad[5] = cpi->fn_ptr[BLOCK_16X16].sdf(src_y_ptr, b->src_stride, pre_y_buffer - 16, pre_y_stride, UINT_MAX); 1715 near_sad[3] = cpi->fn_ptr[BLOCK_16X16].sdf(src_y_ptr, b->src_stride, pre_y_buffer, pre_y_stride, UINT_MAX); 1717 near_sad[6] = cpi->fn_ptr[BLOCK_16X16].sdf(src_y_ptr, b->src_stride, pre_y_buffer + 16, pre_y_stride, UINT_MAX); 1719 near_sad[7] = cpi->fn_ptr[BLOCK_16X16] [all...] |
H A D | pickinter.c | 504 &cpi->fn_ptr[BLOCK_16X16], 951 &cpi->fn_ptr[BLOCK_16X16], 987 &cpi->fn_ptr[BLOCK_16X16], 995 &cpi->fn_ptr[BLOCK_16X16], 1016 &cpi->fn_ptr[BLOCK_16X16], 1039 &cpi->fn_ptr[BLOCK_16X16],
|
H A D | temporal_filter.c | 193 &cpi->fn_ptr[BLOCK_16X16], 206 &cpi->fn_ptr[BLOCK_16X16],
|
H A D | onyx_int.h | 229 BLOCK_16X16, enumerator in enum:__anon13095
|
H A D | onyx_if.c | 2065 cpi->fn_ptr[BLOCK_16X16].sdf = vp8_sad16x16; 2066 cpi->fn_ptr[BLOCK_16X16].vf = vp8_variance16x16; 2067 cpi->fn_ptr[BLOCK_16X16].svf = vp8_sub_pixel_variance16x16; 2068 cpi->fn_ptr[BLOCK_16X16].svf_halfpix_h = vp8_variance_halfpixvar16x16_h; 2069 cpi->fn_ptr[BLOCK_16X16].svf_halfpix_v = vp8_variance_halfpixvar16x16_v; 2070 cpi->fn_ptr[BLOCK_16X16].svf_halfpix_hv = vp8_variance_halfpixvar16x16_hv; 2071 cpi->fn_ptr[BLOCK_16X16].sdx3f = vp8_sad16x16x3; 2072 cpi->fn_ptr[BLOCK_16X16].sdx8f = vp8_sad16x16x8; 2073 cpi->fn_ptr[BLOCK_16X16].sdx4df = vp8_sad16x16x4d; 2116 cpi->fn_ptr[BLOCK_16X16] [all...] |
/external/libvpx/libvpx/vp8/encoder/ |
H A D | rdopt.c | 1689 near_sad[1] = cpi->fn_ptr[BLOCK_16X16].sdf(src_y_ptr, b->src_stride, xd->dst.y_buffer - 16,xd->dst.y_stride, UINT_MAX); 1693 near_sad[0] = cpi->fn_ptr[BLOCK_16X16].sdf(src_y_ptr, b->src_stride, xd->dst.y_buffer - xd->dst.y_stride *16,xd->dst.y_stride, UINT_MAX); 1696 near_sad[0] = cpi->fn_ptr[BLOCK_16X16].sdf(src_y_ptr, b->src_stride, xd->dst.y_buffer - xd->dst.y_stride *16,xd->dst.y_stride, UINT_MAX); 1697 near_sad[1] = cpi->fn_ptr[BLOCK_16X16].sdf(src_y_ptr, b->src_stride, xd->dst.y_buffer - 16,xd->dst.y_stride, UINT_MAX); 1698 near_sad[2] = cpi->fn_ptr[BLOCK_16X16].sdf(src_y_ptr, b->src_stride, xd->dst.y_buffer - xd->dst.y_stride *16 -16,xd->dst.y_stride, UINT_MAX); 1713 near_sad[4] = cpi->fn_ptr[BLOCK_16X16].sdf(src_y_ptr, b->src_stride, pre_y_buffer - pre_y_stride *16, pre_y_stride, UINT_MAX); 1715 near_sad[5] = cpi->fn_ptr[BLOCK_16X16].sdf(src_y_ptr, b->src_stride, pre_y_buffer - 16, pre_y_stride, UINT_MAX); 1716 near_sad[3] = cpi->fn_ptr[BLOCK_16X16].sdf(src_y_ptr, b->src_stride, pre_y_buffer, pre_y_stride, UINT_MAX); 1718 near_sad[6] = cpi->fn_ptr[BLOCK_16X16].sdf(src_y_ptr, b->src_stride, pre_y_buffer + 16, pre_y_stride, UINT_MAX); 1720 near_sad[7] = cpi->fn_ptr[BLOCK_16X16] [all...] |
H A D | pickinter.c | 503 &cpi->fn_ptr[BLOCK_16X16], 941 &cpi->fn_ptr[BLOCK_16X16], 977 &cpi->fn_ptr[BLOCK_16X16], 985 &cpi->fn_ptr[BLOCK_16X16], 1006 &cpi->fn_ptr[BLOCK_16X16], 1029 &cpi->fn_ptr[BLOCK_16X16],
|
H A D | temporal_filter.c | 192 &cpi->fn_ptr[BLOCK_16X16], 205 &cpi->fn_ptr[BLOCK_16X16],
|
H A D | onyx_int.h | 232 BLOCK_16X16, enumerator in enum:__anon24885
|
H A D | onyx_if.c | 2015 cpi->fn_ptr[BLOCK_16X16].sdf = vp8_sad16x16; 2016 cpi->fn_ptr[BLOCK_16X16].vf = vp8_variance16x16; 2017 cpi->fn_ptr[BLOCK_16X16].svf = vp8_sub_pixel_variance16x16; 2018 cpi->fn_ptr[BLOCK_16X16].svf_halfpix_h = vp8_variance_halfpixvar16x16_h; 2019 cpi->fn_ptr[BLOCK_16X16].svf_halfpix_v = vp8_variance_halfpixvar16x16_v; 2020 cpi->fn_ptr[BLOCK_16X16].svf_halfpix_hv = vp8_variance_halfpixvar16x16_hv; 2021 cpi->fn_ptr[BLOCK_16X16].sdx3f = vp8_sad16x16x3; 2022 cpi->fn_ptr[BLOCK_16X16].sdx8f = vp8_sad16x16x8; 2023 cpi->fn_ptr[BLOCK_16X16].sdx4df = vp8_sad16x16x4d; 2066 cpi->fn_ptr[BLOCK_16X16] [all...] |