• John Koleszar's avatar
    Remove diff from BLOCKD · 6c0c6b86
    John Koleszar authored
    The underlying storage for these buffers is in the per-plane MACROBLOCKD
    area, so read it from there directly.
    
    Change-Id: Id6bd835117fdd9dea07db95ad06eff9f12afaaf7
    6c0c6b86
vp9_encodeintra.c 10.35 KiB
/*
 *  Copyright (c) 2010 The WebM project authors. All Rights Reserved.
 *  Use of this source code is governed by a BSD-style license
 *  that can be found in the LICENSE file in the root of the source
 *  tree. An additional intellectual property rights grant can be found
 *  in the file PATENTS.  All contributing project authors may
 *  be found in the AUTHORS file in the root of the source tree.
 */
#include "./vpx_config.h"
#include "vp9_rtcd.h"
#include "vp9/encoder/vp9_quantize.h"
#include "vp9/common/vp9_reconintra.h"
#include "vp9/encoder/vp9_encodemb.h"
#include "vp9/common/vp9_invtrans.h"
#include "vp9/encoder/vp9_encodeintra.h"
static void encode_intra4x4block(MACROBLOCK *x, int ib);
int vp9_encode_intra(VP9_COMP *cpi, MACROBLOCK *x, int use_16x16_pred) {
  MB_MODE_INFO * mbmi = &x->e_mbd.mode_info_context->mbmi;
  (void) cpi;
  if (use_16x16_pred) {
    mbmi->mode = DC_PRED;
    mbmi->uv_mode = DC_PRED;
    mbmi->ref_frame = INTRA_FRAME;
    vp9_encode_intra16x16mby(&cpi->common, x);
  } else {
    int i;
    for (i = 0; i < 16; i++) {
      x->e_mbd.block[i].bmi.as_mode.first = B_DC_PRED;
      encode_intra4x4block(x, i);
  return vp9_get_mb_ss(x->plane[0].src_diff);
static void encode_intra4x4block(MACROBLOCK *x, int ib) {
  BLOCKD *b = &x->e_mbd.block[ib];
  MACROBLOCKD * const xd = &x->e_mbd;
  TX_TYPE tx_type;
  uint8_t* const src =
      raster_block_offset_uint8(xd, BLOCK_SIZE_MB16X16, 0, ib,
                                x->plane[0].src.buf, x->plane[0].src.stride);
  int16_t* const src_diff =
      raster_block_offset_int16(xd, BLOCK_SIZE_MB16X16, 0, ib,
                                x->plane[0].src_diff);
  int16_t* const diff =
      raster_block_offset_int16(xd, BLOCK_SIZE_MB16X16, 0, ib,
                                xd->plane[0].diff);
  int16_t* const coeff = BLOCK_OFFSET(x->plane[0].coeff, ib, 16);
  assert(ib < 16);
#if CONFIG_NEWBINTRAMODES
  b->bmi.as_mode.context = vp9_find_bpred_context(&x->e_mbd, b);
#endif
  vp9_intra4x4_predict(&x->e_mbd, b, b->bmi.as_mode.first,
                       *(b->base_dst) + b->dst, b->dst_stride);
  vp9_subtract_block(4, 4, src_diff, 16,
                     src, x->plane[0].src.stride,
                     *(b->base_dst) + b->dst, b->dst_stride);
  tx_type = get_tx_type_4x4(&x->e_mbd, ib);
7172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140
if (tx_type != DCT_DCT) { vp9_short_fht4x4(src_diff, coeff, 16, tx_type); vp9_ht_quantize_b_4x4(x, ib, tx_type); vp9_short_iht4x4(BLOCK_OFFSET(xd->plane[0].dqcoeff, ib, 16), diff, 16, tx_type); } else { x->fwd_txm4x4(src_diff, coeff, 32); x->quantize_b_4x4(x, ib, 16); vp9_inverse_transform_b_4x4(&x->e_mbd, xd->plane[0].eobs[ib], BLOCK_OFFSET(xd->plane[0].dqcoeff, ib, 16), diff, 32); } vp9_recon_b(*(b->base_dst) + b->dst, diff, *(b->base_dst) + b->dst, b->dst_stride); } void vp9_encode_intra4x4mby(MACROBLOCK *mb) { int i; for (i = 0; i < 16; i++) encode_intra4x4block(mb, i); } void vp9_encode_intra16x16mby(VP9_COMMON *const cm, MACROBLOCK *x) { MACROBLOCKD *xd = &x->e_mbd; TX_SIZE tx_size = xd->mode_info_context->mbmi.txfm_size; vp9_build_intra_predictors_sby_s(xd, BLOCK_SIZE_MB16X16); vp9_subtract_sby(x, BLOCK_SIZE_MB16X16); switch (tx_size) { case TX_16X16: vp9_transform_sby_16x16(x, BLOCK_SIZE_MB16X16); vp9_quantize_sby_16x16(x, BLOCK_SIZE_MB16X16); if (x->optimize) vp9_optimize_sby_16x16(cm, x, BLOCK_SIZE_MB16X16); vp9_inverse_transform_sby_16x16(xd, BLOCK_SIZE_MB16X16); break; case TX_8X8: vp9_transform_sby_8x8(x, BLOCK_SIZE_MB16X16); vp9_quantize_sby_8x8(x, BLOCK_SIZE_MB16X16); if (x->optimize) vp9_optimize_sby_8x8(cm, x, BLOCK_SIZE_MB16X16); vp9_inverse_transform_sby_8x8(xd, BLOCK_SIZE_MB16X16); break; default: vp9_transform_sby_4x4(x, BLOCK_SIZE_MB16X16); vp9_quantize_sby_4x4(x, BLOCK_SIZE_MB16X16); if (x->optimize) vp9_optimize_sby_4x4(cm, x, BLOCK_SIZE_MB16X16); vp9_inverse_transform_sby_4x4(xd, BLOCK_SIZE_MB16X16); break; } vp9_recon_sby(xd, BLOCK_SIZE_MB16X16); } void vp9_encode_intra16x16mbuv(VP9_COMMON *const cm, MACROBLOCK *x) { MACROBLOCKD *xd = &x->e_mbd; TX_SIZE tx_size = xd->mode_info_context->mbmi.txfm_size; vp9_build_intra_predictors_sbuv_s(xd, BLOCK_SIZE_MB16X16); vp9_subtract_sbuv(x, BLOCK_SIZE_MB16X16); switch (tx_size) { case TX_4X4: vp9_transform_sbuv_4x4(x, BLOCK_SIZE_MB16X16); vp9_quantize_sbuv_4x4(x, BLOCK_SIZE_MB16X16); if (x->optimize)
141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210
vp9_optimize_sbuv_4x4(cm, x, BLOCK_SIZE_MB16X16); vp9_inverse_transform_sbuv_4x4(xd, BLOCK_SIZE_MB16X16); break; default: // 16x16 or 8x8 vp9_transform_sbuv_8x8(x, BLOCK_SIZE_MB16X16); vp9_quantize_sbuv_8x8(x, BLOCK_SIZE_MB16X16); if (x->optimize) vp9_optimize_sbuv_8x8(cm, x, BLOCK_SIZE_MB16X16); vp9_inverse_transform_sbuv_8x8(xd, BLOCK_SIZE_MB16X16); break; } vp9_recon_sbuv(xd, BLOCK_SIZE_MB16X16); } void vp9_encode_intra8x8(MACROBLOCK *x, int ib) { MACROBLOCKD *xd = &x->e_mbd; BLOCKD *b = &xd->block[ib]; uint8_t* const src = raster_block_offset_uint8(xd, BLOCK_SIZE_MB16X16, 0, ib, x->plane[0].src.buf, x->plane[0].src.stride); int16_t* const src_diff = raster_block_offset_int16(xd, BLOCK_SIZE_MB16X16, 0, ib, x->plane[0].src_diff); int16_t* const diff = raster_block_offset_int16(xd, BLOCK_SIZE_MB16X16, 0, ib, xd->plane[0].diff); const int iblock[4] = {0, 1, 4, 5}; int i; TX_TYPE tx_type; vp9_intra8x8_predict(xd, b, b->bmi.as_mode.first, *(b->base_dst) + b->dst, b->dst_stride); // generate residual blocks vp9_subtract_block(8, 8, src_diff, 16, src, x->plane[0].src.stride, *(b->base_dst) + b->dst, b->dst_stride); if (xd->mode_info_context->mbmi.txfm_size == TX_8X8) { int idx = (ib & 0x02) ? (ib + 2) : ib; int16_t* const dqcoeff = BLOCK_OFFSET(xd->plane[0].dqcoeff, idx, 16); int16_t* const coeff = BLOCK_OFFSET(x->plane[0].coeff, idx, 16); assert(idx < 16); tx_type = get_tx_type_8x8(xd, ib); if (tx_type != DCT_DCT) { vp9_short_fht8x8(src_diff, coeff, 16, tx_type); x->quantize_b_8x8(x, idx, tx_type, 16); vp9_short_iht8x8(dqcoeff, diff, 16, tx_type); } else { x->fwd_txm8x8(src_diff, coeff, 32); x->quantize_b_8x8(x, idx, DCT_DCT, 16); vp9_short_idct8x8(dqcoeff, diff, 32); } } else { for (i = 0; i < 4; i++) { int idx = ib + iblock[i]; int16_t* const dqcoeff = BLOCK_OFFSET(xd->plane[0].dqcoeff, idx, 16); int16_t* const coeff = BLOCK_OFFSET(x->plane[0].coeff, idx, 16); int16_t* const src_diff = raster_block_offset_int16(xd, BLOCK_SIZE_MB16X16, 0, idx, x->plane[0].src_diff); int16_t* const diff = raster_block_offset_int16(xd, BLOCK_SIZE_MB16X16, 0, idx, xd->plane[0].diff); assert(idx < 16); b = &xd->block[ib + iblock[i]]; tx_type = get_tx_type_4x4(xd, ib + iblock[i]); if (tx_type != DCT_DCT) {
211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280
vp9_short_fht4x4(src_diff, coeff, 16, tx_type); vp9_ht_quantize_b_4x4(x, ib + iblock[i], tx_type); vp9_short_iht4x4(dqcoeff, diff, 16, tx_type); } else if (!(i & 1) && get_tx_type_4x4(xd, ib + iblock[i] + 1) == DCT_DCT) { x->fwd_txm8x4(src_diff, coeff, 32); x->quantize_b_4x4_pair(x, ib + iblock[i], ib + iblock[i] + 1, 16); vp9_inverse_transform_b_4x4(xd, xd->plane[0].eobs[ib + iblock[i]], dqcoeff, diff, 32); vp9_inverse_transform_b_4x4(xd, xd->plane[0].eobs[ib + iblock[i] + 1], dqcoeff + 16, diff + 4, 32); i++; } else { x->fwd_txm4x4(src_diff, coeff, 32); x->quantize_b_4x4(x, ib + iblock[i], 16); vp9_inverse_transform_b_4x4(xd, xd->plane[0].eobs[ib + iblock[i]], dqcoeff, diff, 32); } } } // reconstruct submacroblock for (i = 0; i < 4; i++) { int16_t* const diff = raster_block_offset_int16(xd, BLOCK_SIZE_MB16X16, 0, ib + iblock[i], xd->plane[0].diff); b = &xd->block[ib + iblock[i]]; vp9_recon_b_c(*(b->base_dst) + b->dst, diff, *(b->base_dst) + b->dst, b->dst_stride); } } void vp9_encode_intra8x8mby(MACROBLOCK *x) { int i; for (i = 0; i < 4; i++) vp9_encode_intra8x8(x, vp9_i8x8_block[i]); } static void encode_intra_uv4x4(MACROBLOCK *x, int ib, int mode) { MACROBLOCKD * const xd = &x->e_mbd; BLOCKD *b = &x->e_mbd.block[ib]; int16_t * const dqcoeff = MB_SUBBLOCK_FIELD(xd, dqcoeff, ib); int16_t* const coeff = MB_SUBBLOCK_FIELD(x, coeff, ib); const int plane = ib < 20 ? 1 : 2; const int block = ib < 20 ? ib - 16 : ib - 20; uint8_t* const src = raster_block_offset_uint8(xd, BLOCK_SIZE_MB16X16, plane, block, x->plane[plane].src.buf, x->plane[plane].src.stride); int16_t* const src_diff = raster_block_offset_int16(xd, BLOCK_SIZE_MB16X16, plane, block, x->plane[plane].src_diff); int16_t* const diff = raster_block_offset_int16(xd, BLOCK_SIZE_MB16X16, plane, block, xd->plane[plane].diff); assert(ib >= 16 && ib < 24); vp9_intra_uv4x4_predict(&x->e_mbd, b, mode, *(b->base_dst) + b->dst, b->dst_stride); assert(xd->plane[1].subsampling_x == 1); vp9_subtract_block(4, 4, src_diff, 8, src, x->plane[plane].src.stride, *(b->base_dst) + b->dst, b->dst_stride); x->fwd_txm4x4(src_diff, coeff, 16); x->quantize_b_4x4(x, ib, 16); vp9_inverse_transform_b_4x4(&x->e_mbd, xd->plane[plane].eobs[block], dqcoeff, diff, 16);
281282283284285286287288289290291292293294295296297
vp9_recon_uv_b_c(*(b->base_dst) + b->dst, diff, *(b->base_dst) + b->dst, b->dst_stride); } void vp9_encode_intra8x8mbuv(MACROBLOCK *x) { int i; for (i = 0; i < 4; i++) { BLOCKD *b = &x->e_mbd.block[vp9_i8x8_block[i]]; int mode = b->bmi.as_mode.first; encode_intra_uv4x4(x, i + 16, mode); // u encode_intra_uv4x4(x, i + 20, mode); // v } }