From d12b39daab0a9dd1e3b9f0a4f04d2e456f0746f5 Mon Sep 17 00:00:00 2001
From: James Zern <jzern@google.com>
Date: Tue, 18 Feb 2014 20:42:49 -0800
Subject: [PATCH] vp9_subpel_variance_impl_intrin_avx2.c: make some tables
 static

+ fix formatting

Change-Id: I7b4ec11b7b46d8926750e0b69f7a606f3ab80895
---
 .../vp9_subpel_variance_impl_intrin_avx2.c    | 53 ++++++++++---------
 1 file changed, 27 insertions(+), 26 deletions(-)

diff --git a/vp9/encoder/x86/vp9_subpel_variance_impl_intrin_avx2.c b/vp9/encoder/x86/vp9_subpel_variance_impl_intrin_avx2.c
index a8f98e94b0..b8bfa89000 100644
--- a/vp9/encoder/x86/vp9_subpel_variance_impl_intrin_avx2.c
+++ b/vp9/encoder/x86/vp9_subpel_variance_impl_intrin_avx2.c
@@ -12,7 +12,7 @@
 #include "vpx_ports/mem.h"
 #include "vp9/encoder/vp9_variance.h"
 
-DECLARE_ALIGNED(32, const unsigned char, vp9_bilinear_filters_avx2[512])= {
+DECLARE_ALIGNED(32, static const uint8_t, bilinear_filters_avx2[512]) = {
   16, 0, 16, 0, 16, 0, 16, 0, 16, 0, 16, 0, 16, 0, 16, 0,
   16, 0, 16, 0, 16, 0, 16, 0, 16, 0, 16, 0, 16, 0, 16, 0,
   15, 1, 15, 1, 15, 1, 15, 1, 15, 1, 15, 1, 15, 1, 15, 1,
@@ -44,7 +44,8 @@ DECLARE_ALIGNED(32, const unsigned char, vp9_bilinear_filters_avx2[512])= {
   2, 14, 2, 14, 2, 14, 2, 14, 2, 14, 2, 14, 2, 14, 2, 14,
   2, 14, 2, 14, 2, 14, 2, 14, 2, 14, 2, 14, 2, 14, 2, 14,
   1, 15, 1, 15, 1, 15, 1, 15, 1, 15, 1, 15, 1, 15, 1, 15,
-  1, 15, 1, 15, 1, 15, 1, 15, 1, 15, 1, 15, 1, 15, 1, 15};
+  1, 15, 1, 15, 1, 15, 1, 15, 1, 15, 1, 15, 1, 15, 1, 15
+};
 
 unsigned int vp9_sub_pixel_variance32xh_avx2(const uint8_t *src,
                                              int src_stride,
@@ -137,12 +138,12 @@ unsigned int vp9_sub_pixel_variance32xh_avx2(const uint8_t *src,
         int64_t y_offset64;
         y_offset64 = y_offset;
         y_offset64 <<= 5;
-        filter = _mm256_load_si256((__m256i const *)
-                 (vp9_bilinear_filters_avx2 + y_offset64));
+        filter = _mm256_load_si256(
+            (__m256i const *)(bilinear_filters_avx2 + y_offset64));
 #else
         y_offset <<= 5;
-        filter = _mm256_load_si256((__m256i const *)
-                 (vp9_bilinear_filters_avx2 + y_offset));
+        filter = _mm256_load_si256(
+            (__m256i const *)(bilinear_filters_avx2 + y_offset));
 #endif
         pw8 = _mm256_set1_epi16(8);
         for (i = 0; i < height ; i++) {
@@ -283,12 +284,12 @@ unsigned int vp9_sub_pixel_variance32xh_avx2(const uint8_t *src,
           int64_t y_offset64;
           y_offset64 = y_offset;
           y_offset64 <<= 5;
-          filter = _mm256_load_si256((__m256i const *)
-                   (vp9_bilinear_filters_avx2+y_offset64));
+          filter = _mm256_load_si256(
+              (__m256i const *)(bilinear_filters_avx2 + y_offset64));
 #else
           y_offset <<= 5;
-          filter = _mm256_load_si256((__m256i const *)
-                   (vp9_bilinear_filters_avx2 + y_offset));
+          filter = _mm256_load_si256(
+              (__m256i const *)(bilinear_filters_avx2 + y_offset));
 #endif
           pw8 = _mm256_set1_epi16(8);
           // load source and another source starting from the next
@@ -354,12 +355,12 @@ unsigned int vp9_sub_pixel_variance32xh_avx2(const uint8_t *src,
         int64_t x_offset64;
         x_offset64 = x_offset;
         x_offset64 <<= 5;
-        filter = _mm256_load_si256((__m256i const *)
-                (vp9_bilinear_filters_avx2+x_offset64));
+        filter = _mm256_load_si256(
+            (__m256i const *)(bilinear_filters_avx2 + x_offset64));
 #else
         x_offset <<= 5;
-        filter = _mm256_load_si256((__m256i const *)
-                 (vp9_bilinear_filters_avx2 + x_offset));
+        filter = _mm256_load_si256(
+            (__m256i const *)(bilinear_filters_avx2 + x_offset));
 #endif
         pw8 = _mm256_set1_epi16(8);
         for (i = 0; i < height ; i++) {
@@ -413,12 +414,12 @@ unsigned int vp9_sub_pixel_variance32xh_avx2(const uint8_t *src,
           int64_t x_offset64;
           x_offset64 = x_offset;
           x_offset64 <<= 5;
-          filter = _mm256_load_si256((__m256i const *)
-                  (vp9_bilinear_filters_avx2+x_offset64));
+          filter = _mm256_load_si256(
+              (__m256i const *)(bilinear_filters_avx2 + x_offset64));
 #else
           x_offset <<= 5;
-          filter = _mm256_load_si256((__m256i const *)
-                  (vp9_bilinear_filters_avx2 + x_offset));
+          filter = _mm256_load_si256(
+              (__m256i const *)(bilinear_filters_avx2 + x_offset));
 #endif
           pw8 = _mm256_set1_epi16(8);
           // load source and another source starting from the next
@@ -508,17 +509,17 @@ unsigned int vp9_sub_pixel_variance32xh_avx2(const uint8_t *src,
           x_offset64 <<= 5;
           y_offset64 = y_offset;
           y_offset64 <<= 5;
-          xfilter = _mm256_load_si256((__m256i const *)
-                   (vp9_bilinear_filters_avx2+x_offset64));
-          yfilter = _mm256_load_si256((__m256i const *)
-                   (vp9_bilinear_filters_avx2+y_offset64));
+          xfilter = _mm256_load_si256(
+              (__m256i const *)(bilinear_filters_avx2 + x_offset64));
+          yfilter = _mm256_load_si256(
+              (__m256i const *)(bilinear_filters_avx2 + y_offset64));
 #else
           x_offset <<= 5;
-          xfilter = _mm256_load_si256((__m256i const *)
-                   (vp9_bilinear_filters_avx2 + x_offset));
+          xfilter = _mm256_load_si256(
+              (__m256i const *)(bilinear_filters_avx2 + x_offset));
           y_offset <<= 5;
-          yfilter = _mm256_load_si256((__m256i const *)
-                   (vp9_bilinear_filters_avx2 + y_offset));
+          yfilter = _mm256_load_si256(
+              (__m256i const *)(bilinear_filters_avx2 + y_offset));
 #endif
           pw8 = _mm256_set1_epi16(8);
           // load source and another source starting from the next
-- 
GitLab