/external/libvpx/libvpx/vp8/encoder/ |
H A D | segmentation.c | 15 int mb_row, mb_col; local 29 for (mb_col = 0; mb_col < cm->mb_cols; ++mb_col) {
|
H A D | mr_dissim.c | 77 int mb_col; local 84 for (mb_col = 0; mb_col < cm->mb_cols; ++mb_col) { 111 if (mb_col < (cm->mb_cols - 1)) { 125 if (mb_col < (cm->mb_cols - 1) && mb_row < (cm->mb_rows - 1)) { 135 if (mb_col < (cm->mb_cols - 1)) { 149 if (mb_col < (cm->mb_cols - 1) && mb_row < (cm->mb_rows - 1)) {
|
H A D | ethreading.c | 76 int mb_col; local 108 for (mb_col = 0; mb_col < cm->mb_cols; ++mb_col) { 109 if (((mb_col - 1) % nsync) == 0) { 111 protected_write(mutex, current_mb_col, mb_col - 1); 114 if (mb_row && !(mb_col & (nsync - 1))) { 116 sync_read(mutex, mb_col, last_row_current_mb_col, nsync); 127 xd->mb_to_left_edge = -((mb_col * 16) << 3); 128 xd->mb_to_right_edge = ((cm->mb_cols - 1 - mb_col) * 1 [all...] |
H A D | denoising.c | 474 loop_filter_info_n *lfi_n, int mb_row, int mb_col, 692 if (mb_col > 0) { 471 vp8_denoiser_denoise_mb(VP8_DENOISER *denoiser, MACROBLOCK *x, unsigned int best_sse, unsigned int zero_mv_sse, int recon_yoffset, int recon_uvoffset, loop_filter_info_n *lfi_n, int mb_row, int mb_col, int block_index, int consec_zero_last) argument
|
H A D | encodeframe.c | 107 int mb_row, int mb_col) { 111 int use_dc_pred = (mb_col || mb_row) && (!mb_col || !mb_row); 186 int mb_row, mb_col; local 203 for (mb_col = 0; mb_col < cm->mb_cols; ++mb_col) { 248 int mb_row, mb_col; local 260 for (mb_col = 0; mb_col < c 106 mb_activity_measure(VP8_COMP *cpi, MACROBLOCK *x, int mb_row, int mb_col) argument 332 int mb_col; local 1122 vp8cx_encode_inter_macroblock(VP8_COMP *cpi, MACROBLOCK *x, TOKENEXTRA **t, int recon_yoffset, int recon_uvoffset, int mb_row, int mb_col) argument [all...] |
H A D | temporal_filter.c | 190 int mb_col, mb_row; local 225 for (mb_col = 0; mb_col < mb_cols; ++mb_col) { 233 cpi->mb.mv_col_min = -((mb_col * 16) + (16 - 5)); 234 cpi->mb.mv_col_max = ((cpi->common.mb_cols - 1 - mb_col) * 16) + (16 - 5);
|
H A D | bitstream.c | 457 int mb_col = -1; local 459 while (++mb_col < pc->mb_cols) { 470 xd->mb_to_left_edge = -((mb_col * 16) << 3); 471 xd->mb_to_right_edge = ((pc->mb_cols - 1 - mb_col) * 16) << 3; 622 int mb_col = -1; local 624 while (++mb_col < c->mb_cols) {
|
H A D | pickinter.c | 128 int mb_col, int channel) { 134 int index = mb_row * cpi->common.mb_cols + mb_col; 470 int mb_col) { 485 parent_mb_col = mb_col * cpi->oxcf.mr_down_sampling_factor.den / 633 int mb_col) { 707 cpi, x, target_y, stride, plane[LAST_FRAME][0], mb_row, mb_col, 0); 711 cpi, x, target_u, stride_uv, plane[LAST_FRAME][1], mb_row, mb_col, 1); 714 cpi, x, target_v, stride_uv, plane[LAST_FRAME][2], mb_row, mb_col, 730 &parent_ref_mv, mb_row, mb_col); 781 int block_index = mb_row * cpi->common.mb_cols + mb_col; 125 check_dot_artifact_candidate(VP8_COMP *cpi, MACROBLOCK *x, unsigned char *target_last, int stride, unsigned char *last_ref, int mb_row, int mb_col, int channel) argument 466 get_lower_res_motion_info(VP8_COMP *cpi, MACROBLOCKD *xd, int *dissim, int *parent_ref_frame, MB_PREDICTION_MODE *parent_mode, int_mv *parent_ref_mv, int mb_row, int mb_col) argument 630 vp8_pick_inter_mode(VP8_COMP *cpi, MACROBLOCK *x, int recon_yoffset, int recon_uvoffset, int *returnrate, int *returndistortion, int *returnintra, int mb_row, int mb_col) argument [all...] |
H A D | firstpass.c | 473 int mb_row, mb_col; local 560 for (mb_col = 0; mb_col < cm->mb_cols; ++mb_col) { 563 int use_dc_pred = (mb_col || mb_row) && (!mb_col || !mb_row); 568 xd->left_available = (mb_col != 0); 593 x->mv_col_min = -((mb_col * 16) + (VP8BORDERINPIXELS - 16)); 595 ((cm->mb_cols - 1 - mb_col) * 16) + (VP8BORDERINPIXELS - 16); 709 if (mb_col < c [all...] |
H A D | onyx_if.c | 4314 int mb_col; local 4320 for (mb_col = 0; mb_col < cm->mb_cols + 1; ++mb_col) { 4322 cpi->lfmv[mb_col + mb_row * (cm->mode_info_stride + 1)].as_int = 4326 cpi->lf_ref_frame_sign_bias[mb_col + 4329 cpi->lf_ref_frame[mb_col + mb_row * (cm->mode_info_stride + 1)] = 4340 int mb_col; local 4348 for (mb_col = 0; mb_col < c [all...] |
H A D | rdopt.c | 1746 int mb_col) { 2271 int block_index = mb_row * cpi->common.mb_cols + mb_col; 2282 mb_row, mb_col, block_index, 0); 1743 vp8_rd_pick_inter_mode(VP8_COMP *cpi, MACROBLOCK *x, int recon_yoffset, int recon_uvoffset, int *returnrate, int *returndistortion, int *returnintra, int mb_row, int mb_col) argument
|
/external/libvpx/libvpx/vp8/common/ |
H A D | debugmodes.c | 17 int mb_col; local 26 for (mb_col = 0; mb_col < cols; ++mb_col) { 42 for (mb_col = 0; mb_col < cols; ++mb_col) { 59 for (mb_col = 0; mb_col < cols; ++mb_col) { [all...] |
H A D | threading.h | 214 static INLINE void sync_read(pthread_mutex_t *const mutex, int mb_col, argument 217 while (mb_col > (protected_read(mutex, last_row_current_mb_col) - nsync)) {
|
H A D | vp8_loopfilter.c | 173 int mb_col; local 179 for (mb_col = 0; mb_col < cm->mb_cols; ++mb_col) { 197 if (mb_col > 0) 227 int mb_col; local 232 for (mb_col = 0; mb_col < cm->mb_cols; ++mb_col) { 244 if (mb_col > 275 int mb_col; local 397 int mb_col; local 487 int mb_col; local [all...] |
H A D | mfqe.c | 242 int mb_col; local 260 for (mb_col = 0; mb_col < cm->mb_cols; ++mb_col) {
|
/external/webrtc/webrtc/modules/video_processing/util/ |
H A D | skin_detection.cc | 50 const int mb_col) { 51 const uint8_t* y = y_src + ((mb_row << 4) + 8) * stride_y + (mb_col << 4) + 8; 52 const uint8_t* u = u_src + ((mb_row << 3) + 4) * stride_u + (mb_col << 3) + 4; 53 const uint8_t* v = v_src + ((mb_row << 3) + 4) * stride_v + (mb_col << 3) + 4; 43 MbHasSkinColor(const uint8_t* y_src, const uint8_t* u_src, const uint8_t* v_src, const int stride_y, const int stride_u, const int stride_v, const int mb_row, const int mb_col) argument
|
/external/libvpx/libvpx/vp8/decoder/ |
H A D | onyxd_if.c | 430 int mb_row, mb_col; local 433 for (mb_col = 0; mb_col < oci->mb_cols; mb_col++, mi++) {
|
H A D | decodemv.c | 531 int mb_col = -1; local 536 while (++mb_col < pbi->common.mb_cols) { 538 int mb_num = mb_row * pbi->common.mb_cols + mb_col;
|
H A D | threading.c | 295 int mb_col; local 358 for (mb_col = 0; mb_col < pc->mb_cols; ++mb_col) { 359 if (((mb_col - 1) % nsync) == 0) { 361 protected_write(mutex, current_mb_col, mb_col - 1); 364 if (mb_row && !(mb_col & (nsync - 1))) { 366 sync_read(mutex, mb_col, last_row_current_mb_col, nsync); 373 xd->mb_to_left_edge = -((mb_col * 16) << 3); 374 xd->mb_to_right_edge = ((pc->mb_cols - 1 - mb_col) * 1 [all...] |
H A D | decodeframe.c | 446 int mb_row, mb_col; local 526 for (mb_col = 0; mb_col < pc->mb_cols; ++mb_col) { 531 xd->mb_to_left_edge = -((mb_col * 16) << 3); 532 xd->mb_to_right_edge = ((pc->mb_cols - 1 - mb_col) * 16) << 3; 551 vp8_interpolate_motion(xd, mb_row, mb_col, pc->mb_rows, pc->mb_cols);
|
H A D | error_concealment.c | 96 * (mb_row, mb_col) in macroblocks, which are being overlapped by a given 103 int mb_col, int first_blk_row, 105 /* Find the blocks within this MB (defined by mb_row, mb_col) which are 111 const int rel_ol_blk_col = first_blk_col - mb_col * 4; 124 int end_col = VPXMIN(4 + mb_col * 4 - first_blk_col, 2); 137 if (new_col < (mb_col * 16) << 3) end_col = 1; 277 int mb_row, int mb_col, int mb_rows, 285 4 * mb_row + sub_row, 4 * mb_col + sub_col); 295 int mb_row, mb_col; local 299 for (mb_col 101 calculate_overlaps_mb(B_OVERLAP *b_overlaps, union b_mode_info *bmi, int new_row, int new_col, int mb_row, int mb_col, int first_blk_row, int first_blk_col) argument 276 calc_prev_mb_overlaps(MB_OVERLAP *overlaps, MODE_INFO *prev_mi, int mb_row, int mb_col, int mb_rows, int mb_cols) argument 360 find_neighboring_blocks(MODE_INFO *mi, EC_BLOCK *neighbors, int mb_row, int mb_col, int mb_rows, int mb_cols, int mi_stride) argument 461 vp8_interpolate_motion(MACROBLOCKD *mb, int mb_row, int mb_col, int mb_rows, int mb_cols) argument [all...] |
/external/libvpx/libvpx/vp9/encoder/ |
H A D | vp9_mbgraph.c | 27 int mb_col) { 70 vp9_build_inter_predictors_sby(xd, mb_row, mb_col, BLOCK_16X16); 77 int_mv *dst_mv, int mb_row, int mb_col) { 91 tmp_err = do_16x16_motion_iteration(cpi, ref_mv, &tmp_mv, mb_row, mb_col); 104 do_16x16_motion_iteration(cpi, &zero_ref_mv, &tmp_mv, mb_row, mb_col); 163 int mb_col) { 188 &stats->ref[GOLDEN_FRAME].m.mv, mb_row, mb_col); 220 int mb_col, mb_row, offset = 0; local 255 for (mb_col = 0; mb_col < c 25 do_16x16_motion_iteration(VP9_COMP *cpi, const MV *ref_mv, MV *dst_mv, int mb_row, int mb_col) argument 76 do_16x16_motion_search(VP9_COMP *cpi, const MV *ref_mv, int_mv *dst_mv, int mb_row, int mb_col) argument 158 update_mbgraph_mb_stats(VP9_COMP *cpi, MBGRAPH_MB_STATS *stats, YV12_BUFFER_CONFIG *buf, int mb_y_offset, YV12_BUFFER_CONFIG *golden_ref, const MV *prev_golden_ref_mv, YV12_BUFFER_CONFIG *alt_ref, int mb_row, int mb_col) argument 289 int mb_col, mb_row, offset, i; local [all...] |
H A D | vp9_temporal_filter.c | 295 int mb_col; local 341 for (mb_col = mb_col_start; mb_col < mb_col_end; mb_col++) { 349 td->mb.mv_limits.col_min = -((mb_col * 16) + (17 - 2 * VP9_INTERP_EXTEND)); 351 ((mb_cols - 1 - mb_col) * 16) + (17 - 2 * VP9_INTERP_EXTEND); 384 mb_col * 16, mb_row * 16);
|
H A D | vp9_firstpass.c | 459 static BLOCK_SIZE get_bsize(const VP9_COMMON *cm, int mb_row, int mb_col) { argument 460 if (2 * mb_col + 1 < cm->mi_cols) { 688 int mb_row, mb_col; local 693 for (mb_col = 0; mb_col < cm->mb_cols; ++mb_col) { 694 const int mb_index = mb_row * cm->mb_cols + mb_col; 823 int mb_col; local 902 for (mb_col = tile.mi_col_start >> 1, c = 0; mb_col < (til [all...] |
H A D | vp9_encodeframe.c | 2992 int mb_col = mi_col >> 1; local 2996 VPXMIN(mb_col + num_16x16_blocks_wide_lookup[bsize], cm->mb_cols); 3002 for (c = mb_col; c < mb_col_end; c++) { 3087 int mb_col = mi_col >> 1; local 3091 VPXMIN(mb_col + num_16x16_blocks_wide_lookup[bsize], cm->mb_cols); 3096 for (c = mb_col; c < mb_col_end; c++) {
|