• Adrian Grange's avatar
    Superblock encoding order · 9daf3154
    Adrian Grange authored
    This is the first patch to add superblock (32x32) coding
    order capabilities. It does not yet do any mode selection
    at the SB level, that will follow in a further patch.
    
    This patch encodes rows of SBs rather than
    MBs, each SB contains 2x2 MBs.
    
    Two intra prediction modes have been disabled since they
    require reconstructed data for the above-right MB which
    may not have been encoded yet (e.g. for the bottom right
    MB in each SB).
    
    Results on the one test clip I have tried (720p GIPS clip)
    suggest that it is somewhere around 0.2dB worse than the
    baseline version, so there may be bugs.
    
    It has been tested with no experiments enabled and with
    the following 3 experiments enabled:
      --enable-enhanced_interp
      --enable-high_precision_mv
      --enable-sixteenth_subpel_uv
    in each case the decode buffer matches the recon buffer
    (using "cmp" to compare the dumped/decoded frames).
    Note: Testing these experiments individually created
    errors.
    
    Some problems were found with other experiments but it
    is unclear what state these experiments are in:
      --enable-comp_intra_pred
      --enable-newentropy
      --enable-uvintra
    
    This code has not been extensively tested yet, so there
    is every likelihood that further bugs remain. I also
    intend to do some code cleanup & refactoring in tandem
    with the next patch that adds the 32x32 modes.
    
    Change-Id: I1eba7f740a70b3510df58db53464535ef881b4d9
    9daf3154
encodeintra.c 9.83 KiB
/*
 *  Copyright (c) 2010 The WebM project authors. All Rights Reserved.
 *  Use of this source code is governed by a BSD-style license
 *  that can be found in the LICENSE file in the root of the source
 *  tree. An additional intellectual property rights grant can be found
 *  in the file PATENTS.  All contributing project authors may
 *  be found in the AUTHORS file in the root of the source tree.
 */
#include "vpx_ports/config.h"
#include "vp8/common/idct.h"
#include "quantize.h"
#include "vp8/common/reconintra.h"
#include "vp8/common/reconintra4x4.h"
#include "encodemb.h"
#include "vp8/common/invtrans.h"
#include "vp8/common/recon.h"
#include "dct.h"
#include "vp8/common/g_common.h"
#include "encodeintra.h"
#ifdef ENC_DEBUG
extern int enc_debug;
#endif
#if CONFIG_RUNTIME_CPU_DETECT
#define IF_RTCD(x) (x)
#else
#define IF_RTCD(x) NULL
#endif
int vp8_encode_intra(VP8_COMP *cpi, MACROBLOCK *x, int use_16x16_pred)
    int i;
    int intra_pred_var = 0;
    (void) cpi;
    if (use_16x16_pred)
        x->e_mbd.mode_info_context->mbmi.mode = DC_PRED;
#if CONFIG_COMP_INTRA_PRED
        x->e_mbd.mode_info_context->mbmi.second_mode = (MB_PREDICTION_MODE) (DC_PRED - 1);
#endif
        x->e_mbd.mode_info_context->mbmi.uv_mode = DC_PRED;
        x->e_mbd.mode_info_context->mbmi.ref_frame = INTRA_FRAME;
        vp8_encode_intra16x16mby(IF_RTCD(&cpi->rtcd), x);
    else
        for (i = 0; i < 16; i++)
            x->e_mbd.block[i].bmi.as_mode.first = B_DC_PRED;
            vp8_encode_intra4x4block(IF_RTCD(&cpi->rtcd), x, i);
    intra_pred_var = VARIANCE_INVOKE(&cpi->rtcd.variance, getmbss)(x->src_diff);
    return intra_pred_var;
void vp8_encode_intra4x4block(const VP8_ENCODER_RTCD *rtcd,
                              MACROBLOCK *x, int ib)
    BLOCKD *b = &x->e_mbd.block[ib];
7172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140
BLOCK *be = &x->block[ib]; #if CONFIG_COMP_INTRA_PRED if (b->bmi.as_mode.second == (B_PREDICTION_MODE) (B_DC_PRED - 1)) { #endif RECON_INVOKE(&rtcd->common->recon, intra4x4_predict) (b, b->bmi.as_mode.first, b->predictor); #if CONFIG_COMP_INTRA_PRED } else { RECON_INVOKE(&rtcd->common->recon, comp_intra4x4_predict) (b, b->bmi.as_mode.first, b->bmi.as_mode.second, b->predictor); } #endif ENCODEMB_INVOKE(&rtcd->encodemb, subb)(be, b, 16); x->vp8_short_fdct4x4(be->src_diff, be->coeff, 32); x->quantize_b(be, b); vp8_inverse_transform_b(IF_RTCD(&rtcd->common->idct), b, 32); RECON_INVOKE(&rtcd->common->recon, recon)(b->predictor, b->diff, *(b->base_dst) + b->dst, b->dst_stride); } void vp8_encode_intra4x4mby(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *mb) { int i; MACROBLOCKD *x = &mb->e_mbd; #if 0 // Intra modes requiring top-right MB reconstructed data have been disabled vp8_intra_prediction_down_copy(x); #endif for (i = 0; i < 16; i++) vp8_encode_intra4x4block(rtcd, mb, i); return; } void vp8_encode_intra16x16mby(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x) { BLOCK *b = &x->block[0]; int tx_type = x->e_mbd.mode_info_context->mbmi.txfm_size; #if CONFIG_COMP_INTRA_PRED if (x->e_mbd.mode_info_context->mbmi.second_mode == (MB_PREDICTION_MODE) (DC_PRED - 1)) #endif RECON_INVOKE(&rtcd->common->recon, build_intra_predictors_mby)(&x->e_mbd); #if CONFIG_COMP_INTRA_PRED else RECON_INVOKE(&rtcd->common->recon, build_comp_intra_predictors_mby)(&x->e_mbd); #endif ENCODEMB_INVOKE(&rtcd->encodemb, submby)(x->src_diff, *(b->base_src), x->e_mbd.predictor, b->src_stride); if( tx_type == TX_8X8 ) vp8_transform_intra_mby_8x8(x); else vp8_transform_intra_mby(x); if(tx_type == TX_8X8) vp8_quantize_mby_8x8(x); else vp8_quantize_mby(x);
141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210
if (x->optimize) { if( tx_type == TX_8X8 ) vp8_optimize_mby_8x8(x, rtcd); else vp8_optimize_mby(x, rtcd); } if(tx_type == TX_8X8) vp8_inverse_transform_mby_8x8(IF_RTCD(&rtcd->common->idct), &x->e_mbd); else vp8_inverse_transform_mby(IF_RTCD(&rtcd->common->idct), &x->e_mbd); #ifdef ENC_DEBUG if (enc_debug) { int i; printf("Intra qcoeff:\n"); printf("%d %d:\n", x->e_mbd.mb_to_left_edge, x->e_mbd.mb_to_top_edge); for (i =0; i<400; i++) { printf("%3d ", x->e_mbd.qcoeff[i]); if (i%16 == 15) printf("\n"); } printf("Intra dqcoeff:\n"); for (i =0; i<400; i++) { printf("%3d ", x->e_mbd.dqcoeff[i]); if (i%16 == 15) printf("\n"); } printf("Intra diff:\n"); for (i =0; i<400; i++) { printf("%3d ", x->e_mbd.diff[i]); if (i%16 == 15) printf("\n"); } printf("Intra predictor:\n"); for (i =0; i<400; i++) { printf("%3d ", x->e_mbd.predictor[i]); if (i%16 == 15) printf("\n"); } printf("eobs:\n"); for (i=0;i<25;i++) printf("%d ", x->e_mbd.block[i].eob); printf("\n"); } #endif RECON_INVOKE(&rtcd->common->recon, recon_mby) (IF_RTCD(&rtcd->common->recon), &x->e_mbd); } void vp8_encode_intra16x16mbuv(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x) { int tx_type = x->e_mbd.mode_info_context->mbmi.txfm_size; #if CONFIG_COMP_INTRA_PRED if (x->e_mbd.mode_info_context->mbmi.second_uv_mode == (MB_PREDICTION_MODE) (DC_PRED - 1)) { #endif RECON_INVOKE(&rtcd->common->recon, build_intra_predictors_mbuv)(&x->e_mbd); #if CONFIG_COMP_INTRA_PRED } else { RECON_INVOKE(&rtcd->common->recon, build_comp_intra_predictors_mbuv)(&x->e_mbd); } #endif ENCODEMB_INVOKE(&rtcd->encodemb, submbuv)(x->src_diff, x->src.u_buffer, x->src.v_buffer, x->e_mbd.predictor, x->src.uv_stride); if(tx_type == TX_8X8) vp8_transform_mbuv_8x8(x); else vp8_transform_mbuv(x);
211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280
if(tx_type == TX_8X8) vp8_quantize_mbuv_8x8(x); else vp8_quantize_mbuv(x); #ifdef ENC_DEBUG if (enc_debug) { int i; printf("vp8_encode_intra16x16mbuv\n"); printf("%d %d:\n", x->e_mbd.mb_to_left_edge, x->e_mbd.mb_to_top_edge); printf("qcoeff:\n"); for (i =0; i<400; i++) { printf("%3d ", x->e_mbd.qcoeff[i]); if (i%16 == 15) printf("\n"); } printf("dqcoeff:\n"); for (i =0; i<400; i++) { printf("%3d ", x->e_mbd.dqcoeff[i]); if (i%16 == 15) printf("\n"); } printf("diff:\n"); for (i =0; i<400; i++) { printf("%3d ", x->e_mbd.diff[i]); if (i%16 == 15) printf("\n"); } printf("predictor:\n"); for (i =0; i<400; i++) { printf("%3d ", x->e_mbd.predictor[i]); if (i%16 == 15) printf("\n"); } printf("eobs:\n"); for (i=0;i<25;i++) printf("%d ", x->e_mbd.block[i].eob); printf("\n"); } #endif if (x->optimize) { if(tx_type == TX_8X8) vp8_optimize_mbuv_8x8(x, rtcd); else vp8_optimize_mbuv(x, rtcd); } if(tx_type == TX_8X8) vp8_inverse_transform_mbuv_8x8(IF_RTCD(&rtcd->common->idct), &x->e_mbd); else vp8_inverse_transform_mbuv(IF_RTCD(&rtcd->common->idct), &x->e_mbd); vp8_recon_intra_mbuv(IF_RTCD(&rtcd->common->recon), &x->e_mbd); } void vp8_encode_intra8x8(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x, int ib) { BLOCKD *b = &x->e_mbd.block[ib]; BLOCK *be = &x->block[ib]; const int iblock[4]={0,1,4,5}; int i; #if CONFIG_COMP_INTRA_PRED if (b->bmi.as_mode.second == (MB_PREDICTION_MODE) (DC_PRED - 1)) { #endif RECON_INVOKE(&rtcd->common->recon, intra8x8_predict) (b, b->bmi.as_mode.first, b->predictor); #if CONFIG_COMP_INTRA_PRED } else
281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350
{ RECON_INVOKE(&rtcd->common->recon, comp_intra8x8_predict) (b, b->bmi.as_mode.first, b->bmi.as_mode.second, b->predictor); } #endif for(i=0;i<4;i++) { b = &x->e_mbd.block[ib + iblock[i]]; be = &x->block[ib + iblock[i]]; ENCODEMB_INVOKE(&rtcd->encodemb, subb)(be, b, 16); x->vp8_short_fdct4x4(be->src_diff, be->coeff, 32); x->quantize_b(be, b); vp8_inverse_transform_b(IF_RTCD(&rtcd->common->idct), b, 32); RECON_INVOKE(&rtcd->common->recon, recon)(b->predictor, b->diff, *(b->base_dst) + b->dst, b->dst_stride); } } extern const int vp8_i8x8_block[4]; void vp8_encode_intra8x8mby(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x) { int i, ib; for(i=0;i<4;i++) { ib = vp8_i8x8_block[i]; vp8_encode_intra8x8(rtcd, x, ib); } } void vp8_encode_intra_uv4x4(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x, int ib, int mode, int second) { BLOCKD *b = &x->e_mbd.block[ib]; BLOCK *be = &x->block[ib]; #if CONFIG_COMP_INTRA_PRED if (second == -1) { #endif RECON_INVOKE(&rtcd->common->recon, intra_uv4x4_predict) (b, mode, b->predictor); #if CONFIG_COMP_INTRA_PRED } else { RECON_INVOKE(&rtcd->common->recon, comp_intra_uv4x4_predict) (b, mode, second, b->predictor); } #endif ENCODEMB_INVOKE(&rtcd->encodemb, subb)(be, b, 8); x->vp8_short_fdct4x4(be->src_diff, be->coeff, 16); x->quantize_b(be, b); vp8_inverse_transform_b(IF_RTCD(&rtcd->common->idct), b, 16); RECON_INVOKE(&rtcd->common->recon, recon_uv)(b->predictor, b->diff, *(b->base_dst) + b->dst, b->dst_stride); } void vp8_encode_intra8x8mbuv(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x)
351352353354355356357358359360361362363364365366367368369370
{ int i, ib, mode, second; BLOCKD *b; for(i=0;i<4;i++) { ib = vp8_i8x8_block[i]; b = &x->e_mbd.block[ib]; mode = b->bmi.as_mode.first; #if CONFIG_COMP_INTRA_PRED second = b->bmi.as_mode.second; #else second = -1; #endif /*u */ vp8_encode_intra_uv4x4(rtcd, x, i+16, mode, second); /*v */ vp8_encode_intra_uv4x4(rtcd, x, i+20, mode, second); } }