• Yaowu Xu's avatar
    add 8x8 intra prediction modes · ca6b85aa
    Yaowu Xu authored
    Patch 1 to Patch 3 is an initial implementation of 8x8 intra prediction
    modes, here are with the following assumptions:
    a. 8x8 has 4 prediction modes DC, H, V and TM
    b. UV 4x4 block use the same mode as corresponding 8x8 area
    c. i8x8 modes are enabled for key frame only for now
    Patch 4:
    d. removed debug code from previous patches
    Patch 5:
    e. added stats code to collect entropy stats and further cleaned up
    Patch 6:
    f. changed mode stats code to collect finer stats of modes
    Patch 7:
    g. normalized i8x8 modes distribution to total at 256 (8bits).
    Patch 8:
    h. fixed a bug in decoder and removed debug printf output.
    Patch 9:
    i. more cleanups to address paul's comment
    Patch 10:
    j. messy rebase/merges to bring the commit up to date.
    
    Tests on HD clips encoded with all key frame showing consistent gain
    on all clips and all metrics:~0.5%(psnr) and 0.6%(ssim):
    http://www.corp.google.com/~yaowu/no_crawl/i8x8hd_allkey_fixedq.html
    
    To build and test, configure with:
    --enable-experimental --enable-i8x8
    
    Change-Id: I9813fe07ae48cab5fdb5d904bca022514ad01e7f
    ca6b85aa
encodeintra.c 8.38 KiB
/*
 *  Copyright (c) 2010 The WebM project authors. All Rights Reserved.
 *  Use of this source code is governed by a BSD-style license
 *  that can be found in the LICENSE file in the root of the source
 *  tree. An additional intellectual property rights grant can be found
 *  in the file PATENTS.  All contributing project authors may
 *  be found in the AUTHORS file in the root of the source tree.
 */
#include "vpx_ports/config.h"
#include "vp8/common/idct.h"
#include "quantize.h"
#include "vp8/common/reconintra.h"
#include "vp8/common/reconintra4x4.h"
#include "encodemb.h"
#include "vp8/common/invtrans.h"
#include "vp8/common/recon.h"
#include "dct.h"
#include "vp8/common/g_common.h"
#include "encodeintra.h"
#ifdef ENC_DEBUG
extern int enc_debug;
#endif
#if CONFIG_RUNTIME_CPU_DETECT
#define IF_RTCD(x) (x)
#else
#define IF_RTCD(x) NULL
#endif
int vp8_encode_intra(VP8_COMP *cpi, MACROBLOCK *x, int use_dc_pred)
    int i;
    int intra_pred_var = 0;
    (void) cpi;
    if (use_dc_pred)
        x->e_mbd.mode_info_context->mbmi.mode = DC_PRED;
        x->e_mbd.mode_info_context->mbmi.uv_mode = DC_PRED;
        x->e_mbd.mode_info_context->mbmi.ref_frame = INTRA_FRAME;
        vp8_encode_intra16x16mby(IF_RTCD(&cpi->rtcd), x);
    else
        for (i = 0; i < 16; i++)
            x->e_mbd.block[i].bmi.as_mode = B_DC_PRED;
            vp8_encode_intra4x4block(IF_RTCD(&cpi->rtcd), x, i);
    intra_pred_var = VARIANCE_INVOKE(&cpi->rtcd.variance, getmbss)(x->src_diff);
    return intra_pred_var;
void vp8_encode_intra4x4block(const VP8_ENCODER_RTCD *rtcd,
                              MACROBLOCK *x, int ib)
    BLOCKD *b = &x->e_mbd.block[ib];
    BLOCK *be = &x->block[ib];
    RECON_INVOKE(&rtcd->common->recon, intra4x4_predict)
7172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140
(b, b->bmi.as_mode, b->predictor); ENCODEMB_INVOKE(&rtcd->encodemb, subb)(be, b, 16); x->vp8_short_fdct4x4(be->src_diff, be->coeff, 32); x->quantize_b(be, b); vp8_inverse_transform_b(IF_RTCD(&rtcd->common->idct), b, 32); RECON_INVOKE(&rtcd->common->recon, recon)(b->predictor, b->diff, *(b->base_dst) + b->dst, b->dst_stride); } void vp8_encode_intra4x4mby(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *mb) { int i; MACROBLOCKD *x = &mb->e_mbd; vp8_intra_prediction_down_copy(x); for (i = 0; i < 16; i++) vp8_encode_intra4x4block(rtcd, mb, i); return; } void vp8_encode_intra16x16mby(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x) { BLOCK *b = &x->block[0]; RECON_INVOKE(&rtcd->common->recon, build_intra_predictors_mby)(&x->e_mbd); ENCODEMB_INVOKE(&rtcd->encodemb, submby)(x->src_diff, *(b->base_src), x->e_mbd.predictor, b->src_stride); #if CONFIG_T8X8 if(x->e_mbd.mode_info_context->mbmi.segment_id >= 2) vp8_transform_intra_mby_8x8(x); else #endif vp8_transform_intra_mby(x); #if CONFIG_T8X8 if(x->e_mbd.mode_info_context->mbmi.segment_id >= 2) vp8_quantize_mby_8x8(x); else #endif vp8_quantize_mby(x); if (x->optimize) { #if CONFIG_T8X8 if(x->e_mbd.mode_info_context->mbmi.segment_id >= 2) vp8_optimize_mby_8x8(x, rtcd); else #endif vp8_optimize_mby(x, rtcd); } #if CONFIG_T8X8 if(x->e_mbd.mode_info_context->mbmi.segment_id >= 2) vp8_inverse_transform_mby_8x8(IF_RTCD(&rtcd->common->idct), &x->e_mbd); else #endif vp8_inverse_transform_mby(IF_RTCD(&rtcd->common->idct), &x->e_mbd); #ifdef ENC_DEBUG if (enc_debug) { int i; printf("Intra qcoeff:\n"); printf("%d %d:\n", x->e_mbd.mb_to_left_edge, x->e_mbd.mb_to_top_edge); for (i =0; i<400; i++) { printf("%3d ", x->e_mbd.qcoeff[i]);
141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210
if (i%16 == 15) printf("\n"); } printf("Intra dqcoeff:\n"); for (i =0; i<400; i++) { printf("%3d ", x->e_mbd.dqcoeff[i]); if (i%16 == 15) printf("\n"); } printf("Intra diff:\n"); for (i =0; i<400; i++) { printf("%3d ", x->e_mbd.diff[i]); if (i%16 == 15) printf("\n"); } printf("Intra predictor:\n"); for (i =0; i<400; i++) { printf("%3d ", x->e_mbd.predictor[i]); if (i%16 == 15) printf("\n"); } printf("eobs:\n"); for (i=0;i<25;i++) printf("%d ", x->e_mbd.block[i].eob); printf("\n"); } #endif RECON_INVOKE(&rtcd->common->recon, recon_mby) (IF_RTCD(&rtcd->common->recon), &x->e_mbd); } void vp8_encode_intra16x16mbuv(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x) { RECON_INVOKE(&rtcd->common->recon, build_intra_predictors_mbuv)(&x->e_mbd); ENCODEMB_INVOKE(&rtcd->encodemb, submbuv)(x->src_diff, x->src.u_buffer, x->src.v_buffer, x->e_mbd.predictor, x->src.uv_stride); #if CONFIG_T8X8 if(x->e_mbd.mode_info_context->mbmi.segment_id >= 2) vp8_transform_mbuv_8x8(x); else #endif vp8_transform_mbuv(x); #if CONFIG_T8X8 if(x->e_mbd.mode_info_context->mbmi.segment_id >= 2) vp8_quantize_mbuv_8x8(x); else #endif vp8_quantize_mbuv(x); #ifdef ENC_DEBUG if (enc_debug) { int i; printf("vp8_encode_intra16x16mbuv\n"); printf("%d %d:\n", x->e_mbd.mb_to_left_edge, x->e_mbd.mb_to_top_edge); printf("qcoeff:\n"); for (i =0; i<400; i++) { printf("%3d ", x->e_mbd.qcoeff[i]); if (i%16 == 15) printf("\n"); } printf("dqcoeff:\n"); for (i =0; i<400; i++) { printf("%3d ", x->e_mbd.dqcoeff[i]); if (i%16 == 15) printf("\n"); } printf("diff:\n"); for (i =0; i<400; i++) { printf("%3d ", x->e_mbd.diff[i]); if (i%16 == 15) printf("\n"); } printf("predictor:\n"); for (i =0; i<400; i++) {
211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280
printf("%3d ", x->e_mbd.predictor[i]); if (i%16 == 15) printf("\n"); } printf("eobs:\n"); for (i=0;i<25;i++) printf("%d ", x->e_mbd.block[i].eob); printf("\n"); } #endif if (x->optimize) { #if CONFIG_T8X8 if(x->e_mbd.mode_info_context->mbmi.segment_id >= 2) vp8_optimize_mbuv_8x8(x, rtcd); else #endif vp8_optimize_mbuv(x, rtcd); } #if CONFIG_T8X8 if(x->e_mbd.mode_info_context->mbmi.segment_id >= 2) vp8_inverse_transform_mbuv_8x8(IF_RTCD(&rtcd->common->idct), &x->e_mbd); else #endif vp8_inverse_transform_mbuv(IF_RTCD(&rtcd->common->idct), &x->e_mbd); vp8_recon_intra_mbuv(IF_RTCD(&rtcd->common->recon), &x->e_mbd); } #if CONFIG_I8X8 void vp8_encode_intra8x8(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x, int ib) { BLOCKD *b = &x->e_mbd.block[ib]; BLOCK *be = &x->block[ib]; const int iblock[4]={0,1,4,5}; int i; RECON_INVOKE(&rtcd->common->recon, intra8x8_predict) (b, b->bmi.as_mode, b->predictor); for(i=0;i<4;i++) { b = &x->e_mbd.block[ib + iblock[i]]; be = &x->block[ib + iblock[i]]; ENCODEMB_INVOKE(&rtcd->encodemb, subb)(be, b, 16); x->vp8_short_fdct4x4(be->src_diff, be->coeff, 32); x->quantize_b(be, b); vp8_inverse_transform_b(IF_RTCD(&rtcd->common->idct), b, 32); RECON_INVOKE(&rtcd->common->recon, recon)(b->predictor, b->diff, *(b->base_dst) + b->dst, b->dst_stride); } } extern const int vp8_i8x8_block[4]; void vp8_encode_intra8x8mby(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x) { int i, ib; for(i=0;i<4;i++) { ib = vp8_i8x8_block[i]; vp8_encode_intra8x8(rtcd, x, ib); } } void vp8_encode_intra_uv4x4(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x, int ib,
281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319
int mode) { BLOCKD *b = &x->e_mbd.block[ib]; BLOCK *be = &x->block[ib]; RECON_INVOKE(&rtcd->common->recon, intra_uv4x4_predict) (b, mode, b->predictor); ENCODEMB_INVOKE(&rtcd->encodemb, subb)(be, b, 8); x->vp8_short_fdct4x4(be->src_diff, be->coeff, 16); x->quantize_b(be, b); vp8_inverse_transform_b(IF_RTCD(&rtcd->common->idct), b, 16); RECON_INVOKE(&rtcd->common->recon, recon_uv)(b->predictor, b->diff, *(b->base_dst) + b->dst, b->dst_stride); } void vp8_encode_intra8x8mbuv(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x) { int i, ib, mode; BLOCKD *b; for(i=0;i<4;i++) { ib = vp8_i8x8_block[i]; b = &x->e_mbd.block[ib]; mode = b->bmi.as_mode; /*u */ vp8_encode_intra_uv4x4(rtcd, x, i+16, mode); /*v */ vp8_encode_intra_uv4x4(rtcd, x, i+20, mode); } } #endif