/external/libvpx/libvpx/vp8/decoder/ |
D | error_concealment.c | 53 pbi->overlaps = vpx_calloc(pbi->common.mb_rows * pbi->common.mb_cols, in vp8_alloc_overlap_lists() 171 int mb_rows, int mb_cols, in vp8_calculate_overlaps() argument 190 if (new_row >= ((16*mb_rows) << 3) || new_col >= ((16*mb_cols) << 3)) in vp8_calculate_overlaps() 212 end_col = MIN(mb_cols - overlap_mb_col, 2); in vp8_calculate_overlaps() 230 mb_overlap = overlap_ul + (overlap_mb_row + rel_row) * mb_cols + in vp8_calculate_overlaps() 328 int mb_rows, int mb_cols) in calc_prev_mb_overlaps() argument 337 overlaps, mb_rows, mb_cols, in calc_prev_mb_overlaps() 349 int mb_rows, int mb_cols, in estimate_missing_mvs() argument 353 vpx_memset(overlaps, 0, sizeof(MB_OVERLAP) * mb_rows * mb_cols); in estimate_missing_mvs() 357 for (mb_col = 0; mb_col < mb_cols; ++mb_col) in estimate_missing_mvs() [all …]
|
D | error_concealment.h | 37 int mb_rows, int mb_cols,
|
D | decodeframe.c | 596 for (mb_col = 0; mb_col < pc->mb_cols; mb_col++) in decode_mb_rows() 603 xd->mb_to_right_edge = ((pc->mb_cols - 1 - mb_col) * 16) << 3; in decode_mb_rows() 625 pc->mb_rows, pc->mb_cols, in decode_mb_rows() 707 lf_mic += pc->mb_cols; in decode_mb_rows() 1335 pbi->mvs_corrupt_from_mb < (unsigned int)pc->mb_cols * pc->mb_rows) in vp8_decode_frame() 1343 vpx_memset(pc->above_context, 0, sizeof(ENTROPY_CONTEXT_PLANES) * pc->mb_cols); in vp8_decode_frame()
|
D | threading.c | 303 const int first_row_no_sync_above = pc->mb_cols + nsync; in mt_decode_mb_rows() 405 for (mb_col = 0; mb_col < pc->mb_cols; mb_col++) in mt_decode_mb_rows() 423 xd->mb_to_right_edge = ((pc->mb_cols - 1 - mb_col) * 16) << 3; in mt_decode_mb_rows() 449 pc->mb_rows, pc->mb_cols, in mt_decode_mb_rows() 506 if(mb_col != pc->mb_cols-1) in mt_decode_mb_rows()
|
D | decodemv.c | 628 mb_to_right_edge_start = ((pbi->common.mb_cols - 1) * 16) << 3; in vp8_decode_mode_mvs() 637 while (++mb_col < pbi->common.mb_cols) in vp8_decode_mode_mvs() 640 int mb_num = mb_row * pbi->common.mb_cols + mb_col; in vp8_decode_mode_mvs()
|
D | onyxd_if.c | 374 for (col = 0; col < pbi->common.mb_cols; ++col) in vp8dx_receive_compressed_data() 441 for (mb_col = 0; mb_col < oci->mb_cols; mb_col++,mi++) in vp8dx_references_buffer()
|
/external/libvpx/libvpx/test/ |
D | set_roi.cc | 54 cpi.common.mb_cols = 320 >> 4; in TEST() 55 const int mbs = (cpi.common.mb_rows * cpi.common.mb_cols); in TEST() 70 cpi.common.mb_cols, delta_q, delta_lf, in TEST() 138 cpi.common.mb_cols, rand_deltas, in TEST() 148 cpi.common.mb_cols, delta_q, in TEST() 160 cpi.common.mb_cols, delta_q, in TEST() 167 cpi.common.mb_cols, delta_q, in TEST() 172 cpi.common.mb_cols - 1, delta_q, in TEST()
|
/external/libvpx/libvpx/vp8/common/ |
D | alloccommon.c | 83 oci->mb_cols = width >> 4; in vp8_alloc_frame_buffers() 84 oci->MBs = oci->mb_rows * oci->mb_cols; in vp8_alloc_frame_buffers() 85 oci->mode_info_stride = oci->mb_cols + 1; in vp8_alloc_frame_buffers() 86 oci->mip = vpx_calloc((oci->mb_cols + 1) * (oci->mb_rows + 1), sizeof(MODE_INFO)); in vp8_alloc_frame_buffers() 96 oci->above_context = vpx_calloc(sizeof(ENTROPY_CONTEXT_PLANES) * oci->mb_cols, 1); in vp8_alloc_frame_buffers() 114 oci->pp_limits_buffer = vpx_memalign(16, 24 * ((oci->mb_cols + 1) & ~1)); in vp8_alloc_frame_buffers()
|
D | loopfilter.c | 209 for (mb_col = 0; mb_col < cm->mb_cols; mb_col++) in vp8_loop_filter_row_normal() 265 for (mb_col = 0; mb_col < cm->mb_cols; mb_col++) in vp8_loop_filter_row_simple() 316 int mb_cols = cm->mb_cols; in vp8_loop_filter_frame() local 340 for (mb_col = 0; mb_col < mb_cols; mb_col++) in vp8_loop_filter_frame() 396 for (mb_col = 0; mb_col < mb_cols; mb_col++) in vp8_loop_filter_frame() 482 for (mb_col = 0; mb_col < cm->mb_cols; mb_col++) in vp8_loop_filter_frame_yonly() 565 int mb_cols = post->y_width >> 4; in vp8_loop_filter_partial_frame() local 592 mode_info_context = cm->mi + (post->y_height >> 5) * (mb_cols + 1); in vp8_loop_filter_partial_frame() 597 for (mb_col = 0; mb_col < mb_cols; mb_col++) in vp8_loop_filter_partial_frame()
|
D | mfqe.c | 304 for (mb_col = 0; mb_col < cm->mb_cols; mb_col++) in vp8_multiframe_quality_enhance() 376 y_ptr += show->y_stride * 16 - 16 * cm->mb_cols; in vp8_multiframe_quality_enhance() 377 u_ptr += show->uv_stride * 8 - 8 * cm->mb_cols; in vp8_multiframe_quality_enhance() 378 v_ptr += show->uv_stride * 8 - 8 * cm->mb_cols; in vp8_multiframe_quality_enhance() 379 yd_ptr += dest->y_stride * 16 - 16 * cm->mb_cols; in vp8_multiframe_quality_enhance() 380 ud_ptr += dest->uv_stride * 8 - 8 * cm->mb_cols; in vp8_multiframe_quality_enhance() 381 vd_ptr += dest->uv_stride * 8 - 8 * cm->mb_cols; in vp8_multiframe_quality_enhance()
|
D | postproc.c | 339 unsigned char *uvlimits = cm->pp_limits_buffer + 16 * cm->mb_cols; in vp8_deblock() 349 for (mbc = 0; mbc < cm->mb_cols; mbc++) in vp8_deblock() 400 int mb_cols = cm->mb_cols; in vp8_de_noise() local 406 vpx_memset(limits, (unsigned char)ppl, 16 * mb_cols); in vp8_de_noise() 846 oci->mb_cols, oci->mb_rows); in vp8_post_proc_frame() 856 int mb_cols = post->y_width >> 4; in vp8_post_proc_frame() local 865 for (j = 0; j < mb_cols; j++) in vp8_post_proc_frame() 888 int mb_cols = post->y_width >> 4; in vp8_post_proc_frame() local 897 for (j = 0; j < mb_cols; j++) in vp8_post_proc_frame()
|
D | onyxc_int.h | 104 int mb_cols; member
|
/external/libvpx/libvpx/vp9/encoder/ |
D | vp9_lookahead.c | 90 int mb_cols = (src->y_width + 15) >> 4; in vp9_lookahead_push() local 112 for (; col < mb_cols; ++col) { in vp9_lookahead_push() 118 if (col == mb_cols) in vp9_lookahead_push() 124 for (; active_end < mb_cols; ++active_end) { in vp9_lookahead_push() 139 active_map += mb_cols; in vp9_lookahead_push()
|
D | vp9_mbgraph.c | 260 x->mv_col_max = (cm->mb_cols - 1) * 8 + BORDER_MV_PIXELS_B16; in update_mbgraph_frame_stats() 263 for (mb_col = 0; mb_col < cm->mb_cols; mb_col++) { in update_mbgraph_frame_stats() 287 offset += cm->mb_cols; in update_mbgraph_frame_stats() 302 vpx_calloc(cm->mb_rows * cm->mb_cols * sizeof(*arf_not_zz), in separate_arf_mbs() 314 offset += cm->mb_cols, mb_row++) { in separate_arf_mbs() 315 for (mb_col = 0; mb_col < cm->mb_cols; mb_col++) { in separate_arf_mbs() 338 if (arf_not_zz[mi_row / 2 * cm->mb_cols + mi_col / 2]) { in separate_arf_mbs() 387 cm->mb_rows * cm->mb_cols * in vp9_update_mbgraph_stats()
|
D | vp9_temporal_filter.c | 199 int mb_cols = cpi->common.mb_cols; in temporal_filter_iterate_c() local 235 for (mb_col = 0; mb_col < mb_cols; mb_col++) { in temporal_filter_iterate_c() 243 cpi->mb.mv_col_max = ((cpi->common.mb_cols - 1 - mb_col) * 16) in temporal_filter_iterate_c() 349 mb_y_offset += 16 * (f->y_stride - mb_cols); in temporal_filter_iterate_c() 350 mb_uv_offset += mb_uv_height * f->uv_stride - mb_uv_width * mb_cols; in temporal_filter_iterate_c()
|
D | vp9_encoder.h | 507 static INLINE int get_token_alloc(int mb_rows, int mb_cols) { in get_token_alloc() argument 513 return mb_rows * mb_cols * (16 * 16 * 3 + 4); in get_token_alloc()
|
D | vp9_firstpass.c | 529 for (mb_col = 0; mb_col < cm->mb_cols; ++mb_col) { in vp9_first_pass() 535 const int mb_index = mb_row * cm->mb_cols + mb_col; in vp9_first_pass() 591 x->mv_col_max = ((cm->mb_cols - 1 - mb_col) * 16) + BORDER_MV_PIXELS_B16; in vp9_first_pass() 791 if (mb_col < cm->mb_cols / 2) { in vp9_first_pass() 796 } else if (mb_col > cm->mb_cols / 2) { in vp9_first_pass() 819 x->plane[0].src.buf += 16 * x->plane[0].src.stride - 16 * cm->mb_cols; in vp9_first_pass() 821 uv_mb_height * cm->mb_cols; in vp9_first_pass() 823 uv_mb_height * cm->mb_cols; in vp9_first_pass()
|
/external/libvpx/libvpx/vp8/encoder/ |
D | lookahead.c | 114 int mb_cols = (src->y_width + 15) >> 4; in vp8_lookahead_push() local 135 for (; col < mb_cols; ++col) in vp8_lookahead_push() 142 if (col == mb_cols) in vp8_lookahead_push() 148 for (; active_end < mb_cols; ++active_end) in vp8_lookahead_push() 164 active_map += mb_cols; in vp8_lookahead_push()
|
D | segmentation.c | 26 vpx_memset(cpi->gf_active_flags, 1, (cm->mb_rows * cm->mb_cols)); in vp8_update_gf_useage_maps() 27 cpi->gf_active_count = cm->mb_rows * cm->mb_cols; in vp8_update_gf_useage_maps() 35 for (mb_col = 0; mb_col < cm->mb_cols; mb_col++) in vp8_update_gf_useage_maps()
|
D | mr_dissim.c | 93 for (mb_col = 0; mb_col < cm->mb_cols; mb_col ++) in vp8_cal_dissimilarity() 123 if(mb_col < (cm->mb_cols-1)) in vp8_cal_dissimilarity() 139 if(mb_col < (cm->mb_cols-1) in vp8_cal_dissimilarity() 152 if(mb_col < (cm->mb_cols-1)) in vp8_cal_dissimilarity() 168 if(mb_col < (cm->mb_cols-1) in vp8_cal_dissimilarity()
|
D | ethreading.c | 89 int map_index = (mb_row * cm->mb_cols); in thread_encoding_proc() 96 tp = cpi->tok + (mb_row * (cm->mb_cols * 16 * 24)); in thread_encoding_proc() 116 for (mb_col = 0; mb_col < cm->mb_cols; mb_col++) in thread_encoding_proc() 138 xd->mb_to_right_edge = ((cm->mb_cols - 1 - mb_col) * 16) << 3; in thread_encoding_proc() 146 x->mv_col_max = ((cm->mb_cols - 1 - mb_col) * 16) + (VP8BORDERINPIXELS - 16); in thread_encoding_proc() 303 … x->src.y_buffer += 16 * x->src.y_stride * (cpi->encoding_thread_count + 1) - 16 * cm->mb_cols; in thread_encoding_proc() 304 … x->src.u_buffer += 8 * x->src.uv_stride * (cpi->encoding_thread_count + 1) - 8 * cm->mb_cols; in thread_encoding_proc() 305 … x->src.v_buffer += 8 * x->src.uv_stride * (cpi->encoding_thread_count + 1) - 8 * cm->mb_cols; in thread_encoding_proc() 309 x->gf_active_ptr += cm->mb_cols * cpi->encoding_thread_count; in thread_encoding_proc() 532 if(th_count > ((cm->mb_cols / cpi->mt_sync_range) - 1)) in vp8cx_create_encoder_threads() [all …]
|
D | encodeframe.c | 225 for (mb_col = 0; mb_col < cm->mb_cols; mb_col++) in calc_activity_index() 287 for (mb_col = 0; mb_col < cm->mb_cols; mb_col++) in build_activity_map() 315 x->src.y_buffer += 16 * x->src.y_stride - 16 * cm->mb_cols; in build_activity_map() 376 int map_index = (mb_row * cpi->common.mb_cols); in encode_mb_row() 386 const int rightmost_col = cm->mb_cols + nsync; in encode_mb_row() 430 for (mb_col = 0; mb_col < cm->mb_cols; mb_col++) in encode_mb_row() 441 xd->mb_to_right_edge = ((cm->mb_cols - 1 - mb_col) * 16) << 3; in encode_mb_row() 447 x->mv_col_max = ((cm->mb_cols - 1 - mb_col) * 16) in encode_mb_row() 662 sizeof(ENTROPY_CONTEXT_PLANES) * cm->mb_cols); in init_encode_frame_mb_context() 839 tp = cpi->tok + mb_row * (cm->mb_cols * 16 * 24); in vp8_encode_frame() [all …]
|
D | temporal_filter.c | 235 int mb_cols = cpi->common.mb_cols; in vp8_temporal_filter_iterate_c() local 270 for (mb_col = 0; mb_col < mb_cols; mb_col++) in vp8_temporal_filter_iterate_c() 280 cpi->mb.mv_col_max = ((cpi->common.mb_cols - 1 - mb_col) * 16) in vp8_temporal_filter_iterate_c() 418 mb_y_offset += 16*(f->y_stride-mb_cols); in vp8_temporal_filter_iterate_c() 419 mb_uv_offset += 8*(f->uv_stride-mb_cols); in vp8_temporal_filter_iterate_c()
|
D | onyx_if.c | 510 … vpx_memcpy(cpi->segmentation_map, segmentation_map, (cpi->common.mb_rows * cpi->common.mb_cols)); in set_segmentation_map() 542 CHECK_MEM_ERROR(seg_map, vpx_calloc(cpi->common.mb_rows * cpi->common.mb_cols, 1)); in segmentation_test_function() 578 int mbs_in_frame = cpi->common.mb_rows * cpi->common.mb_cols; in cyclic_background_refresh() 1163 cpi->mb.pip = vpx_calloc((cpi->common.mb_cols + 1) * in vp8_alloc_partition_data() 1213 unsigned int tokens = cm->mb_rows * cm->mb_cols * 24 * 16; in vp8_alloc_compressor_data() 1226 cm->mb_rows * cm->mb_cols)); in vp8_alloc_compressor_data() 1227 cpi->gf_active_count = cm->mb_rows * cm->mb_cols; in vp8_alloc_compressor_data() 1232 cm->mb_rows * cm->mb_cols)); in vp8_alloc_compressor_data() 1236 CHECK_MEM_ERROR(cpi->lfmv, vpx_calloc((cm->mb_rows+2) * (cm->mb_cols+2), in vp8_alloc_compressor_data() 1240 vpx_calloc((cm->mb_rows+2) * (cm->mb_cols+2), in vp8_alloc_compressor_data() [all …]
|
/external/libvpx/libvpx/vp9/common/ |
D | vp9_alloccommon.c | 39 cm->mb_cols = (cm->mi_cols + 1) >> 1; in vp9_set_mb_mi() 41 cm->MBs = cm->mb_rows * cm->mb_cols; in vp9_set_mb_mi()
|