/external/libvpx/libvpx/vp8/encoder/ |
H A D | segmentation.c | 15 int mb_row, mb_col; local 29 for (mb_col = 0; mb_col < cm->mb_cols; ++mb_col) {
|
H A D | mr_dissim.c | 77 int mb_col; local 84 for (mb_col = 0; mb_col < cm->mb_cols; ++mb_col) { 111 if (mb_col < (cm->mb_cols - 1)) { 125 if (mb_col < (cm->mb_cols - 1) && mb_row < (cm->mb_rows - 1)) { 135 if (mb_col < (cm->mb_cols - 1)) { 149 if (mb_col < (cm->mb_cols - 1) && mb_row < (cm->mb_rows - 1)) {
|
H A D | ethreading.c | 77 int mb_col; local 109 for (mb_col = 0; mb_col < cm->mb_cols; ++mb_col) { 110 if (((mb_col - 1) % nsync) == 0) { 111 vpx_atomic_store_release(current_mb_col, mb_col - 1); 114 if (mb_row && !(mb_col & (nsync - 1))) { 115 vp8_atomic_spin_wait(mb_col, last_row_current_mb_col, nsync); 126 xd->mb_to_left_edge = -((mb_col * 16) << 3); 127 xd->mb_to_right_edge = ((cm->mb_cols - 1 - mb_col) * 1 [all...] |
H A D | denoising.c | 474 loop_filter_info_n *lfi_n, int mb_row, int mb_col, 692 if (mb_col > 0) { 471 vp8_denoiser_denoise_mb(VP8_DENOISER *denoiser, MACROBLOCK *x, unsigned int best_sse, unsigned int zero_mv_sse, int recon_yoffset, int recon_uvoffset, loop_filter_info_n *lfi_n, int mb_row, int mb_col, int block_index, int consec_zero_last) argument
|
H A D | encodeframe.c | 104 int mb_row, int mb_col) { 108 int use_dc_pred = (mb_col || mb_row) && (!mb_col || !mb_row); 183 int mb_row, mb_col; local 200 for (mb_col = 0; mb_col < cm->mb_cols; ++mb_col) { 245 int mb_row, mb_col; local 257 for (mb_col = 0; mb_col < c 103 mb_activity_measure(VP8_COMP *cpi, MACROBLOCK *x, int mb_row, int mb_col) argument 329 int mb_col; local 1119 vp8cx_encode_inter_macroblock(VP8_COMP *cpi, MACROBLOCK *x, TOKENEXTRA **t, int recon_yoffset, int recon_uvoffset, int mb_row, int mb_col) argument [all...] |
H A D | temporal_filter.c | 191 int mb_col, mb_row; local 226 for (mb_col = 0; mb_col < mb_cols; ++mb_col) { 234 cpi->mb.mv_col_min = -((mb_col * 16) + (16 - 5)); 235 cpi->mb.mv_col_max = ((cpi->common.mb_cols - 1 - mb_col) * 16) + (16 - 5);
|
H A D | bitstream.c | 457 int mb_col = -1; local 459 while (++mb_col < pc->mb_cols) { 470 xd->mb_to_left_edge = -((mb_col * 16) << 3); 471 xd->mb_to_right_edge = ((pc->mb_cols - 1 - mb_col) * 16) << 3; 621 int mb_col = -1; local 623 while (++mb_col < c->mb_cols) {
|
H A D | pickinter.c | 56 int mb_col, int channel) { 62 int index = mb_row * cpi->common.mb_cols + mb_col; 398 int mb_col) { 413 parent_mb_col = mb_col * cpi->oxcf.mr_down_sampling_factor.den / 561 int mb_col) { 635 cpi, x, target_y, stride, plane[LAST_FRAME][0], mb_row, mb_col, 0); 639 cpi, x, target_u, stride_uv, plane[LAST_FRAME][1], mb_row, mb_col, 1); 642 cpi, x, target_v, stride_uv, plane[LAST_FRAME][2], mb_row, mb_col, 658 &parent_ref_mv, mb_row, mb_col); 693 int block_index = mb_row * cpi->common.mb_cols + mb_col; 53 check_dot_artifact_candidate(VP8_COMP *cpi, MACROBLOCK *x, unsigned char *target_last, int stride, unsigned char *last_ref, int mb_row, int mb_col, int channel) argument 394 get_lower_res_motion_info(VP8_COMP *cpi, MACROBLOCKD *xd, int *dissim, int *parent_ref_frame, MB_PREDICTION_MODE *parent_mode, int_mv *parent_ref_mv, int mb_row, int mb_col) argument 558 vp8_pick_inter_mode(VP8_COMP *cpi, MACROBLOCK *x, int recon_yoffset, int recon_uvoffset, int *returnrate, int *returndistortion, int *returnintra, int mb_row, int mb_col) argument [all...] |
H A D | firstpass.c | 473 int mb_row, mb_col; local 560 for (mb_col = 0; mb_col < cm->mb_cols; ++mb_col) { 563 int use_dc_pred = (mb_col || mb_row) && (!mb_col || !mb_row); 568 xd->left_available = (mb_col != 0); 593 x->mv_col_min = -((mb_col * 16) + (VP8BORDERINPIXELS - 16)); 595 ((cm->mb_cols - 1 - mb_col) * 16) + (VP8BORDERINPIXELS - 16); 709 if (mb_col < c [all...] |
H A D | onyx_if.c | 612 int mb_row, mb_col, num_bl; local 625 for (mb_col = 0; mb_col < cm->mb_cols; mb_col++) { 626 const int bl_index = mb_row * cm->mb_cols + mb_col; 643 for (mb_col = 1; mb_col < cm->mb_cols - 1; mb_col++) { 644 const int bl_index = mb_row * cm->mb_cols + mb_col; 651 int bl_neighbor_index = (mb_row + mi) * cm->mb_cols + mb_col 4324 int mb_col; local 4350 int mb_col; local [all...] |
H A D | rdopt.c | 1741 int mb_col) { 2266 int block_index = mb_row * cpi->common.mb_cols + mb_col; 2277 mb_row, mb_col, block_index, 0); 1738 vp8_rd_pick_inter_mode(VP8_COMP *cpi, MACROBLOCK *x, int recon_yoffset, int recon_uvoffset, int *returnrate, int *returndistortion, int *returnintra, int mb_row, int mb_col) argument
|
/external/libvpx/libvpx/vp8/common/ |
H A D | debugmodes.c | 17 int mb_col; local 26 for (mb_col = 0; mb_col < cols; ++mb_col) { 42 for (mb_col = 0; mb_col < cols; ++mb_col) { 59 for (mb_col = 0; mb_col < cols; ++mb_col) { [all...] |
H A D | threading.h | 198 int mb_col, const vpx_atomic_int *last_row_current_mb_col, 200 while (mb_col > (vpx_atomic_load_acquire(last_row_current_mb_col) - nsync)) { 197 vp8_atomic_spin_wait( int mb_col, const vpx_atomic_int *last_row_current_mb_col, const int nsync) argument
|
H A D | vp8_skin_detection.c | 72 int i, j, mb_row, mb_col, num_bl; local 92 for (mb_col = 0; mb_col < cm->mb_cols; mb_col += 1) {
|
H A D | vp8_loopfilter.c | 171 int mb_col; local 177 for (mb_col = 0; mb_col < cm->mb_cols; ++mb_col) { 195 if (mb_col > 0) 225 int mb_col; local 230 for (mb_col = 0; mb_col < cm->mb_cols; ++mb_col) { 242 if (mb_col > 273 int mb_col; local 394 int mb_col; local 484 int mb_col; local [all...] |
H A D | mfqe.c | 238 int mb_col; local 256 for (mb_col = 0; mb_col < cm->mb_cols; ++mb_col) {
|
/external/webrtc/webrtc/modules/video_processing/util/ |
H A D | skin_detection.cc | 50 const int mb_col) { 51 const uint8_t* y = y_src + ((mb_row << 4) + 8) * stride_y + (mb_col << 4) + 8; 52 const uint8_t* u = u_src + ((mb_row << 3) + 4) * stride_u + (mb_col << 3) + 4; 53 const uint8_t* v = v_src + ((mb_row << 3) + 4) * stride_v + (mb_col << 3) + 4; 43 MbHasSkinColor(const uint8_t* y_src, const uint8_t* u_src, const uint8_t* v_src, const int stride_y, const int stride_u, const int stride_v, const int mb_row, const int mb_col) argument
|
/external/libvpx/libvpx/vp8/decoder/ |
H A D | onyxd_if.c | 429 int mb_row, mb_col; local 432 for (mb_col = 0; mb_col < oci->mb_cols; mb_col++, mi++) {
|
H A D | decodemv.c | 530 int mb_col = -1; local 535 while (++mb_col < pbi->common.mb_cols) { 537 int mb_num = mb_row * pbi->common.mb_cols + mb_col;
|
H A D | threading.c | 296 int mb_col; local 359 for (mb_col = 0; mb_col < pc->mb_cols; ++mb_col) { 360 if (((mb_col - 1) % nsync) == 0) { 361 vpx_atomic_store_release(current_mb_col, mb_col - 1); 364 if (mb_row && !(mb_col & (nsync - 1))) { 365 vp8_atomic_spin_wait(mb_col, last_row_current_mb_col, nsync); 372 xd->mb_to_left_edge = -((mb_col * 16) << 3); 373 xd->mb_to_right_edge = ((pc->mb_cols - 1 - mb_col) * 1 [all...] |
H A D | decodeframe.c | 446 int mb_row, mb_col; local 526 for (mb_col = 0; mb_col < pc->mb_cols; ++mb_col) { 531 xd->mb_to_left_edge = -((mb_col * 16) << 3); 532 xd->mb_to_right_edge = ((pc->mb_cols - 1 - mb_col) * 16) << 3; 551 vp8_interpolate_motion(xd, mb_row, mb_col, pc->mb_rows, pc->mb_cols);
|
H A D | error_concealment.c | 96 * (mb_row, mb_col) in macroblocks, which are being overlapped by a given 103 int mb_col, int first_blk_row, 105 /* Find the blocks within this MB (defined by mb_row, mb_col) which are 111 const int rel_ol_blk_col = first_blk_col - mb_col * 4; 124 int end_col = VPXMIN(4 + mb_col * 4 - first_blk_col, 2); 137 if (new_col < (mb_col * 16) << 3) end_col = 1; 277 int mb_row, int mb_col, int mb_rows, 285 4 * mb_row + sub_row, 4 * mb_col + sub_col); 295 int mb_row, mb_col; local 299 for (mb_col 101 calculate_overlaps_mb(B_OVERLAP *b_overlaps, union b_mode_info *bmi, int new_row, int new_col, int mb_row, int mb_col, int first_blk_row, int first_blk_col) argument 276 calc_prev_mb_overlaps(MB_OVERLAP *overlaps, MODE_INFO *prev_mi, int mb_row, int mb_col, int mb_rows, int mb_cols) argument 360 find_neighboring_blocks(MODE_INFO *mi, EC_BLOCK *neighbors, int mb_row, int mb_col, int mb_rows, int mb_cols, int mi_stride) argument 461 vp8_interpolate_motion(MACROBLOCKD *mb, int mb_row, int mb_col, int mb_rows, int mb_cols) argument [all...] |
/external/libvpx/libvpx/vp9/encoder/ |
H A D | vp9_mbgraph.c | 27 int mb_col) { 70 vp9_build_inter_predictors_sby(xd, mb_row, mb_col, BLOCK_16X16); 77 int_mv *dst_mv, int mb_row, int mb_col) { 91 tmp_err = do_16x16_motion_iteration(cpi, ref_mv, &tmp_mv, mb_row, mb_col); 104 do_16x16_motion_iteration(cpi, &zero_ref_mv, &tmp_mv, mb_row, mb_col); 163 int mb_col) { 188 &stats->ref[GOLDEN_FRAME].m.mv, mb_row, mb_col); 220 int mb_col, mb_row, offset = 0; local 255 for (mb_col = 0; mb_col < c 25 do_16x16_motion_iteration(VP9_COMP *cpi, const MV *ref_mv, MV *dst_mv, int mb_row, int mb_col) argument 76 do_16x16_motion_search(VP9_COMP *cpi, const MV *ref_mv, int_mv *dst_mv, int mb_row, int mb_col) argument 158 update_mbgraph_mb_stats(VP9_COMP *cpi, MBGRAPH_MB_STATS *stats, YV12_BUFFER_CONFIG *buf, int mb_y_offset, YV12_BUFFER_CONFIG *golden_ref, const MV *prev_golden_ref_mv, YV12_BUFFER_CONFIG *alt_ref, int mb_row, int mb_col) argument 289 int mb_col, mb_row, offset, i; local [all...] |
H A D | vp9_temporal_filter.c | 295 int mb_col; local 341 for (mb_col = mb_col_start; mb_col < mb_col_end; mb_col++) { 349 td->mb.mv_limits.col_min = -((mb_col * 16) + (17 - 2 * VP9_INTERP_EXTEND)); 351 ((mb_cols - 1 - mb_col) * 16) + (17 - 2 * VP9_INTERP_EXTEND); 405 mb_col * 16, mb_row * 16);
|
H A D | vp9_firstpass.c | 467 static BLOCK_SIZE get_bsize(const VP9_COMMON *cm, int mb_row, int mb_col) { argument 468 if (2 * mb_col + 1 < cm->mi_cols) { 696 int mb_row, mb_col; local 701 for (mb_col = 0; mb_col < cm->mb_cols; ++mb_col) { 702 const int mb_index = mb_row * cm->mb_cols + mb_col; 832 int mb_col; local 911 for (mb_col = tile.mi_col_start >> 1, c = 0; mb_col < (til [all...] |