vp9_variance_c.c 41.64 KiB
/*
 *  Copyright (c) 2010 The WebM project authors. All Rights Reserved.
 *  Use of this source code is governed by a BSD-style license
 *  that can be found in the LICENSE file in the root of the source
 *  tree. An additional intellectual property rights grant can be found
 *  in the file PATENTS.  All contributing project authors may
 *  be found in the AUTHORS file in the root of the source tree.
 */
#include "vp9/encoder/vp9_variance.h"
#include "vp9/common/vp9_filter.h"
#include "vp9/common/vp9_subpelvar.h"
#include "vpx/vpx_integer.h"
#include "vpx_ports/mem.h"
unsigned int vp9_get_mb_ss_c(const int16_t *src_ptr) {
  unsigned int i, sum = 0;
  for (i = 0; i < 256; i++) {
    sum += (src_ptr[i] * src_ptr[i]);
  return sum;
unsigned int vp9_variance64x32_c(const uint8_t *src_ptr,
                                 int  source_stride,
                                 const uint8_t *ref_ptr,
                                 int  recon_stride,
                                 unsigned int *sse) {
  unsigned int var;
  int avg;
  variance(src_ptr, source_stride, ref_ptr, recon_stride, 64, 32, &var, &avg);
  *sse = var;
  return (var - (((int64_t)avg * avg) >> 11));
unsigned int vp9_sub_pixel_variance64x32_c(const uint8_t *src_ptr,
                                           int  src_pixels_per_line,
                                           int  xoffset,
                                           int  yoffset,
                                           const uint8_t *dst_ptr,
                                           int dst_pixels_per_line,
                                           unsigned int *sse) {
  uint16_t fdata3[65 * 64];  // Temp data bufffer used in filtering
  uint8_t temp2[68 * 64];
  const int16_t *hfilter, *vfilter;
  hfilter = VP9_BILINEAR_FILTERS_2TAP(xoffset);
  vfilter = VP9_BILINEAR_FILTERS_2TAP(yoffset);
  var_filter_block2d_bil_first_pass(src_ptr, fdata3, src_pixels_per_line,
                                    1, 33, 64, hfilter);
  var_filter_block2d_bil_second_pass(fdata3, temp2, 64, 64, 32, 64, vfilter);
  return vp9_variance64x32_c(temp2, 64, dst_ptr, dst_pixels_per_line, sse);
unsigned int vp9_sub_pixel_avg_variance64x32_c(const uint8_t *src_ptr,
                                               int  src_pixels_per_line,
                                               int  xoffset,
                                               int  yoffset,
                                               const uint8_t *dst_ptr,
                                               int dst_pixels_per_line,
                                               unsigned int *sse,
                                               const uint8_t *second_pred) {
  uint16_t fdata3[65 * 64];  // Temp data bufffer used in filtering
7172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140
uint8_t temp2[68 * 64]; DECLARE_ALIGNED_ARRAY(16, uint8_t, temp3, 64 * 64); // compound pred buffer const int16_t *hfilter, *vfilter; hfilter = VP9_BILINEAR_FILTERS_2TAP(xoffset); vfilter = VP9_BILINEAR_FILTERS_2TAP(yoffset); var_filter_block2d_bil_first_pass(src_ptr, fdata3, src_pixels_per_line, 1, 33, 64, hfilter); var_filter_block2d_bil_second_pass(fdata3, temp2, 64, 64, 32, 64, vfilter); comp_avg_pred(temp3, second_pred, 64, 32, temp2, 64); return vp9_variance64x32_c(temp3, 64, dst_ptr, dst_pixels_per_line, sse); } unsigned int vp9_variance32x64_c(const uint8_t *src_ptr, int source_stride, const uint8_t *ref_ptr, int recon_stride, unsigned int *sse) { unsigned int var; int avg; variance(src_ptr, source_stride, ref_ptr, recon_stride, 32, 64, &var, &avg); *sse = var; return (var - (((int64_t)avg * avg) >> 11)); } unsigned int vp9_sub_pixel_variance32x64_c(const uint8_t *src_ptr, int src_pixels_per_line, int xoffset, int yoffset, const uint8_t *dst_ptr, int dst_pixels_per_line, unsigned int *sse) { uint16_t fdata3[65 * 64]; // Temp data bufffer used in filtering uint8_t temp2[68 * 64]; const int16_t *hfilter, *vfilter; hfilter = VP9_BILINEAR_FILTERS_2TAP(xoffset); vfilter = VP9_BILINEAR_FILTERS_2TAP(yoffset); var_filter_block2d_bil_first_pass(src_ptr, fdata3, src_pixels_per_line, 1, 65, 32, hfilter); var_filter_block2d_bil_second_pass(fdata3, temp2, 32, 32, 64, 32, vfilter); return vp9_variance32x64_c(temp2, 32, dst_ptr, dst_pixels_per_line, sse); } unsigned int vp9_sub_pixel_avg_variance32x64_c(const uint8_t *src_ptr, int src_pixels_per_line, int xoffset, int yoffset, const uint8_t *dst_ptr, int dst_pixels_per_line, unsigned int *sse, const uint8_t *second_pred) { uint16_t fdata3[65 * 64]; // Temp data bufffer used in filtering uint8_t temp2[68 * 64]; DECLARE_ALIGNED_ARRAY(16, uint8_t, temp3, 32 * 64); // compound pred buffer const int16_t *hfilter, *vfilter; hfilter = VP9_BILINEAR_FILTERS_2TAP(xoffset); vfilter = VP9_BILINEAR_FILTERS_2TAP(yoffset); var_filter_block2d_bil_first_pass(src_ptr, fdata3, src_pixels_per_line, 1, 65, 32, hfilter); var_filter_block2d_bil_second_pass(fdata3, temp2, 32, 32, 64, 32, vfilter); comp_avg_pred(temp3, second_pred, 32, 64, temp2, 32); return vp9_variance32x64_c(temp3, 32, dst_ptr, dst_pixels_per_line, sse); }
141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210
unsigned int vp9_variance32x16_c(const uint8_t *src_ptr, int source_stride, const uint8_t *ref_ptr, int recon_stride, unsigned int *sse) { unsigned int var; int avg; variance(src_ptr, source_stride, ref_ptr, recon_stride, 32, 16, &var, &avg); *sse = var; return (var - (((int64_t)avg * avg) >> 9)); } unsigned int vp9_sub_pixel_variance32x16_c(const uint8_t *src_ptr, int src_pixels_per_line, int xoffset, int yoffset, const uint8_t *dst_ptr, int dst_pixels_per_line, unsigned int *sse) { uint16_t fdata3[33 * 32]; // Temp data bufffer used in filtering uint8_t temp2[36 * 32]; const int16_t *hfilter, *vfilter; hfilter = VP9_BILINEAR_FILTERS_2TAP(xoffset); vfilter = VP9_BILINEAR_FILTERS_2TAP(yoffset); var_filter_block2d_bil_first_pass(src_ptr, fdata3, src_pixels_per_line, 1, 17, 32, hfilter); var_filter_block2d_bil_second_pass(fdata3, temp2, 32, 32, 16, 32, vfilter); return vp9_variance32x16_c(temp2, 32, dst_ptr, dst_pixels_per_line, sse); } unsigned int vp9_sub_pixel_avg_variance32x16_c(const uint8_t *src_ptr, int src_pixels_per_line, int xoffset, int yoffset, const uint8_t *dst_ptr, int dst_pixels_per_line, unsigned int *sse, const uint8_t *second_pred) { uint16_t fdata3[33 * 32]; // Temp data bufffer used in filtering uint8_t temp2[36 * 32]; DECLARE_ALIGNED_ARRAY(16, uint8_t, temp3, 32 * 16); // compound pred buffer const int16_t *hfilter, *vfilter; hfilter = VP9_BILINEAR_FILTERS_2TAP(xoffset); vfilter = VP9_BILINEAR_FILTERS_2TAP(yoffset); var_filter_block2d_bil_first_pass(src_ptr, fdata3, src_pixels_per_line, 1, 17, 32, hfilter); var_filter_block2d_bil_second_pass(fdata3, temp2, 32, 32, 16, 32, vfilter); comp_avg_pred(temp3, second_pred, 32, 16, temp2, 32); return vp9_variance32x16_c(temp3, 32, dst_ptr, dst_pixels_per_line, sse); } unsigned int vp9_variance16x32_c(const uint8_t *src_ptr, int source_stride, const uint8_t *ref_ptr, int recon_stride, unsigned int *sse) { unsigned int var; int avg; variance(src_ptr, source_stride, ref_ptr, recon_stride, 16, 32, &var, &avg); *sse = var; return (var - (((int64_t)avg * avg) >> 9)); }
211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280
unsigned int vp9_sub_pixel_variance16x32_c(const uint8_t *src_ptr, int src_pixels_per_line, int xoffset, int yoffset, const uint8_t *dst_ptr, int dst_pixels_per_line, unsigned int *sse) { uint16_t fdata3[33 * 32]; // Temp data bufffer used in filtering uint8_t temp2[36 * 32]; const int16_t *hfilter, *vfilter; hfilter = VP9_BILINEAR_FILTERS_2TAP(xoffset); vfilter = VP9_BILINEAR_FILTERS_2TAP(yoffset); var_filter_block2d_bil_first_pass(src_ptr, fdata3, src_pixels_per_line, 1, 33, 16, hfilter); var_filter_block2d_bil_second_pass(fdata3, temp2, 16, 16, 32, 16, vfilter); return vp9_variance16x32_c(temp2, 16, dst_ptr, dst_pixels_per_line, sse); } unsigned int vp9_sub_pixel_avg_variance16x32_c(const uint8_t *src_ptr, int src_pixels_per_line, int xoffset, int yoffset, const uint8_t *dst_ptr, int dst_pixels_per_line, unsigned int *sse, const uint8_t *second_pred) { uint16_t fdata3[33 * 32]; // Temp data bufffer used in filtering uint8_t temp2[36 * 32]; DECLARE_ALIGNED_ARRAY(16, uint8_t, temp3, 16 * 32); // compound pred buffer const int16_t *hfilter, *vfilter; hfilter = VP9_BILINEAR_FILTERS_2TAP(xoffset); vfilter = VP9_BILINEAR_FILTERS_2TAP(yoffset); var_filter_block2d_bil_first_pass(src_ptr, fdata3, src_pixels_per_line, 1, 33, 16, hfilter); var_filter_block2d_bil_second_pass(fdata3, temp2, 16, 16, 32, 16, vfilter); comp_avg_pred(temp3, second_pred, 16, 32, temp2, 16); return vp9_variance16x32_c(temp3, 16, dst_ptr, dst_pixels_per_line, sse); } unsigned int vp9_variance64x64_c(const uint8_t *src_ptr, int source_stride, const uint8_t *ref_ptr, int recon_stride, unsigned int *sse) { unsigned int var; int avg; variance(src_ptr, source_stride, ref_ptr, recon_stride, 64, 64, &var, &avg); *sse = var; return (var - (((int64_t)avg * avg) >> 12)); } unsigned int vp9_variance32x32_c(const uint8_t *src_ptr, int source_stride, const uint8_t *ref_ptr, int recon_stride, unsigned int *sse) { unsigned int var; int avg; variance(src_ptr, source_stride, ref_ptr, recon_stride, 32, 32, &var, &avg); *sse = var; return (var - (((int64_t)avg * avg) >> 10)); }
281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350
unsigned int vp9_variance16x16_c(const uint8_t *src_ptr, int source_stride, const uint8_t *ref_ptr, int recon_stride, unsigned int *sse) { unsigned int var; int avg; variance(src_ptr, source_stride, ref_ptr, recon_stride, 16, 16, &var, &avg); *sse = var; return (var - (((unsigned int)avg * avg) >> 8)); } unsigned int vp9_variance8x16_c(const uint8_t *src_ptr, int source_stride, const uint8_t *ref_ptr, int recon_stride, unsigned int *sse) { unsigned int var; int avg; variance(src_ptr, source_stride, ref_ptr, recon_stride, 8, 16, &var, &avg); *sse = var; return (var - (((unsigned int)avg * avg) >> 7)); } unsigned int vp9_variance16x8_c(const uint8_t *src_ptr, int source_stride, const uint8_t *ref_ptr, int recon_stride, unsigned int *sse) { unsigned int var; int avg; variance(src_ptr, source_stride, ref_ptr, recon_stride, 16, 8, &var, &avg); *sse = var; return (var - (((unsigned int)avg * avg) >> 7)); } void vp9_get_sse_sum_8x8_c(const uint8_t *src_ptr, int source_stride, const uint8_t *ref_ptr, int ref_stride, unsigned int *sse, int *sum) { variance(src_ptr, source_stride, ref_ptr, ref_stride, 8, 8, sse, sum); } unsigned int vp9_variance8x8_c(const uint8_t *src_ptr, int source_stride, const uint8_t *ref_ptr, int recon_stride, unsigned int *sse) { unsigned int var; int avg; variance(src_ptr, source_stride, ref_ptr, recon_stride, 8, 8, &var, &avg); *sse = var; return (var - (((unsigned int)avg * avg) >> 6)); } unsigned int vp9_variance8x4_c(const uint8_t *src_ptr, int source_stride, const uint8_t *ref_ptr, int recon_stride, unsigned int *sse) { unsigned int var; int avg; variance(src_ptr, source_stride, ref_ptr, recon_stride, 8, 4, &var, &avg); *sse = var; return (var - (((unsigned int)avg * avg) >> 5));
351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420
} unsigned int vp9_variance4x8_c(const uint8_t *src_ptr, int source_stride, const uint8_t *ref_ptr, int recon_stride, unsigned int *sse) { unsigned int var; int avg; variance(src_ptr, source_stride, ref_ptr, recon_stride, 4, 8, &var, &avg); *sse = var; return (var - (((unsigned int)avg * avg) >> 5)); } unsigned int vp9_variance4x4_c(const uint8_t *src_ptr, int source_stride, const uint8_t *ref_ptr, int recon_stride, unsigned int *sse) { unsigned int var; int avg; variance(src_ptr, source_stride, ref_ptr, recon_stride, 4, 4, &var, &avg); *sse = var; return (var - (((unsigned int)avg * avg) >> 4)); } unsigned int vp9_mse16x16_c(const uint8_t *src_ptr, int source_stride, const uint8_t *ref_ptr, int recon_stride, unsigned int *sse) { unsigned int var; int avg; variance(src_ptr, source_stride, ref_ptr, recon_stride, 16, 16, &var, &avg); *sse = var; return var; } unsigned int vp9_mse16x8_c(const uint8_t *src_ptr, int source_stride, const uint8_t *ref_ptr, int recon_stride, unsigned int *sse) { unsigned int var; int avg; variance(src_ptr, source_stride, ref_ptr, recon_stride, 16, 8, &var, &avg); *sse = var; return var; } unsigned int vp9_mse8x16_c(const uint8_t *src_ptr, int source_stride, const uint8_t *ref_ptr, int recon_stride, unsigned int *sse) { unsigned int var; int avg; variance(src_ptr, source_stride, ref_ptr, recon_stride, 8, 16, &var, &avg); *sse = var; return var; } unsigned int vp9_mse8x8_c(const uint8_t *src_ptr, int source_stride,