diff --git a/vp8/common/maskingmv.c b/vp8/common/maskingmv.c
index 72109378d8bf4cf177d4f6500a3e9cf7f6745a00..05cfb4a46f22c69ae20f23f9c0700f3c8096f950 100644
--- a/vp8/common/maskingmv.c
+++ b/vp8/common/maskingmv.c
@@ -11,14 +11,14 @@
 #include <stdio.h>
 #include <stdlib.h>
 #include <string.h>
-extern unsigned int vp8_sad16x16_sse3(
+extern unsigned int vp9_sad16x16_sse3(
   unsigned char *src_ptr,
   int  src_stride,
   unsigned char *ref_ptr,
   int  ref_stride,
   int  max_err);
 
-extern void vp8_sad16x16x3_sse3(
+extern void vp9_sad16x16x3_sse3(
   unsigned char *src_ptr,
   int  src_stride,
   unsigned char *ref_ptr,
@@ -43,14 +43,14 @@ extern void vp8_makemask_sse3(
   int ut,
   int vt);
 
-unsigned int vp8_sad16x16_unmasked_wmt(
+unsigned int vp9_sad16x16_unmasked_wmt(
   unsigned char *src_ptr,
   int  src_stride,
   unsigned char *ref_ptr,
   int  ref_stride,
   unsigned char *mask);
 
-unsigned int vp8_sad16x16_masked_wmt(
+unsigned int vp9_sad16x16_masked_wmt(
   unsigned char *src_ptr,
   int  src_stride,
   unsigned char *ref_ptr,
@@ -503,7 +503,7 @@ int fast_masked_motion_search(unsigned char *y, unsigned char *u, unsigned char
 
       vp8_growmaskmb_sse3(dym, dym2);
 
-      e = vp8_sad16x16_unmasked_wmt(y, yp, dyz + j, dyp, dym2);
+      e = vp9_sad16x16_unmasked_wmt(y, yp, dyz + j, dyp, dym2);
 
       if (e < beste) {
         bui = i;
@@ -529,7 +529,7 @@ int fast_masked_motion_search(unsigned char *y, unsigned char *u, unsigned char
   for (i = -32; i < 32; i++) {
     unsigned char *dyz = i * dyp + dy;
     for (j = -32; j < 32; j++) {
-      e = vp8_sad16x16_masked_wmt(y, yp, dyz + j, dyp, dym2);
+      e = vp9_sad16x16_masked_wmt(y, yp, dyz + j, dyp, dym2);
       if (e < beste) {
         bmi = i;
         bmj = j;
@@ -581,7 +581,7 @@ int fast_masked_motion_search(unsigned char *y, unsigned char *u, unsigned char
 
   vp8_growmaskmb_sse3(dym, dym2);
 
-  obeste = vp8_sad16x16_masked_wmt(y, yp, dy + bmi * dyp + bmj, dyp, dym2);
+  obeste = vp9_sad16x16_masked_wmt(y, yp, dy + bmi * dyp + bmj, dyp, dym2);
 
   beste = 0xffffffff;
 
@@ -589,7 +589,7 @@ int fast_masked_motion_search(unsigned char *y, unsigned char *u, unsigned char
   for (i = -32; i < 32; i++) {
     unsigned char *dyz = i * dyp + dy;
     for (j = -32; j < 32; j++) {
-      e = vp8_sad16x16_unmasked_wmt(y, yp, dyz + j, dyp, dym2);
+      e = vp9_sad16x16_unmasked_wmt(y, yp, dyz + j, dyp, dym2);
 
       if (e < beste) {
         bui = i;
@@ -698,8 +698,8 @@ int mainz(int argc, char *argv[]) {
 
   vp8_growmaskmb_sse3(ym, ym3);
 
-  a = vp8_sad16x16_masked_wmt(str, 16, sts, 16, ym3);
-  b = vp8_sad16x16_unmasked_wmt(str, 16, sts, 16, ym3);
+  a = vp9_sad16x16_masked_wmt(str, 16, sts, 16, ym3);
+  b = vp9_sad16x16_unmasked_wmt(str, 16, sts, 16, ym3);
 
   vp8_masked_predictor_wmt(str, sts, 16, ym, 16, ym3);
 
@@ -738,7 +738,7 @@ int mainz(int argc, char *argv[]) {
         int bmi, bmj, bui, buj, bwm;
         unsigned char ym[256];
 
-        if (vp8_sad16x16_sse3(ys + c, y_stride, yd + c, y_stride, 0xffff) == 0)
+        if (vp9_sad16x16_sse3(ys + c, y_stride, yd + c, y_stride, 0xffff) == 0)
           bmi = bmj = bui = buj = bwm = 0;
         else {
           COLOR_SEG_ELEMENT cs[5];
diff --git a/vp8/common/onyx.h b/vp8/common/onyx.h
index 3d8f565e5c4615b4e9eafa4e0d7f16d4c464a771..703de7deda85f1e3369e63808adc0f749849b1d3 100644
--- a/vp8/common/onyx.h
+++ b/vp8/common/onyx.h
@@ -172,29 +172,29 @@ extern "C"
   } VP8_CONFIG;
 
 
-  void vp8_initialize();
+  void vp9_initialize();
 
-  VP8_PTR vp8_create_compressor(VP8_CONFIG *oxcf);
-  void vp8_remove_compressor(VP8_PTR *comp);
+  VP8_PTR vp9_create_compressor(VP8_CONFIG *oxcf);
+  void vp9_remove_compressor(VP8_PTR *comp);
 
   void vp8_init_config(VP8_PTR onyx, VP8_CONFIG *oxcf);
-  void vp8_change_config(VP8_PTR onyx, VP8_CONFIG *oxcf);
+  void vp9_change_config(VP8_PTR onyx, VP8_CONFIG *oxcf);
 
 // receive a frames worth of data caller can assume that a copy of this frame is made
 // and not just a copy of the pointer..
-  int vp8_receive_raw_frame(VP8_PTR comp, unsigned int frame_flags, YV12_BUFFER_CONFIG *sd, int64_t time_stamp, int64_t end_time_stamp);
-  int vp8_get_compressed_data(VP8_PTR comp, unsigned int *frame_flags, unsigned long *size, unsigned char *dest, int64_t *time_stamp, int64_t *time_end, int flush);
-  int vp8_get_preview_raw_frame(VP8_PTR comp, YV12_BUFFER_CONFIG *dest, vp8_ppflags_t *flags);
-
-  int vp8_use_as_reference(VP8_PTR comp, int ref_frame_flags);
-  int vp8_update_reference(VP8_PTR comp, int ref_frame_flags);
-  int vp8_get_reference(VP8_PTR comp, VP8_REFFRAME ref_frame_flag, YV12_BUFFER_CONFIG *sd);
-  int vp8_set_reference(VP8_PTR comp, VP8_REFFRAME ref_frame_flag, YV12_BUFFER_CONFIG *sd);
-  int vp8_update_entropy(VP8_PTR comp, int update);
-  int vp8_set_roimap(VP8_PTR comp, unsigned char *map, unsigned int rows, unsigned int cols, int delta_q[4], int delta_lf[4], unsigned int threshold[4]);
-  int vp8_set_active_map(VP8_PTR comp, unsigned char *map, unsigned int rows, unsigned int cols);
-  int vp8_set_internal_size(VP8_PTR comp, VPX_SCALING horiz_mode, VPX_SCALING vert_mode);
-  int vp8_get_quantizer(VP8_PTR c);
+  int vp9_receive_raw_frame(VP8_PTR comp, unsigned int frame_flags, YV12_BUFFER_CONFIG *sd, int64_t time_stamp, int64_t end_time_stamp);
+  int vp9_get_compressed_data(VP8_PTR comp, unsigned int *frame_flags, unsigned long *size, unsigned char *dest, int64_t *time_stamp, int64_t *time_end, int flush);
+  int vp9_get_preview_raw_frame(VP8_PTR comp, YV12_BUFFER_CONFIG *dest, vp8_ppflags_t *flags);
+
+  int vp9_use_as_reference(VP8_PTR comp, int ref_frame_flags);
+  int vp9_update_reference(VP8_PTR comp, int ref_frame_flags);
+  int vp9_get_reference(VP8_PTR comp, VP8_REFFRAME ref_frame_flag, YV12_BUFFER_CONFIG *sd);
+  int vp9_set_reference(VP8_PTR comp, VP8_REFFRAME ref_frame_flag, YV12_BUFFER_CONFIG *sd);
+  int vp9_update_entropy(VP8_PTR comp, int update);
+  int vp9_set_roimap(VP8_PTR comp, unsigned char *map, unsigned int rows, unsigned int cols, int delta_q[4], int delta_lf[4], unsigned int threshold[4]);
+  int vp9_set_active_map(VP8_PTR comp, unsigned char *map, unsigned int rows, unsigned int cols);
+  int vp9_set_internal_size(VP8_PTR comp, VPX_SCALING horiz_mode, VPX_SCALING vert_mode);
+  int vp9_get_quantizer(VP8_PTR c);
 
 #ifdef __cplusplus
 }
diff --git a/vp8/common/rtcd_defs.sh b/vp8/common/rtcd_defs.sh
index 9a9073697a5b7d6786ce9d492d3d1a19c63ecbcb..c3c76839292655d509f9e08fccabba33e0417082 100644
--- a/vp8/common/rtcd_defs.sh
+++ b/vp8/common/rtcd_defs.sh
@@ -236,203 +236,203 @@ if [ "$CONFIG_VP8_ENCODER" = "yes" ]; then
 # variance
 [ $arch = "x86_64" ] && mmx_x86_64=mmx && sse2_x86_64=sse2
 
-prototype unsigned int vp8_variance32x32 "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int ref_stride, unsigned int *sse"
-specialize vp8_variance32x32
+prototype unsigned int vp9_variance32x32 "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int ref_stride, unsigned int *sse"
+specialize vp9_variance32x32
 
-prototype unsigned int vp8_variance16x16 "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int ref_stride, unsigned int *sse"
-specialize vp8_variance16x16 mmx sse2
-vp8_variance16x16_sse2=vp8_variance16x16_wmt
-vp8_variance16x16_mmx=vp8_variance16x16_mmx
+prototype unsigned int vp9_variance16x16 "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int ref_stride, unsigned int *sse"
+specialize vp9_variance16x16 mmx sse2
+vp9_variance16x16_sse2=vp9_variance16x16_wmt
+vp9_variance16x16_mmx=vp9_variance16x16_mmx
 
-prototype unsigned int vp8_variance16x8 "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int ref_stride, unsigned int *sse"
-specialize vp8_variance16x8 mmx sse2
-vp8_variance16x8_sse2=vp8_variance16x8_wmt
-vp8_variance16x8_mmx=vp8_variance16x8_mmx
+prototype unsigned int vp9_variance16x8 "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int ref_stride, unsigned int *sse"
+specialize vp9_variance16x8 mmx sse2
+vp9_variance16x8_sse2=vp9_variance16x8_wmt
+vp9_variance16x8_mmx=vp9_variance16x8_mmx
 
-prototype unsigned int vp8_variance8x16 "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int ref_stride, unsigned int *sse"
-specialize vp8_variance8x16 mmx sse2
-vp8_variance8x16_sse2=vp8_variance8x16_wmt
-vp8_variance8x16_mmx=vp8_variance8x16_mmx
+prototype unsigned int vp9_variance8x16 "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int ref_stride, unsigned int *sse"
+specialize vp9_variance8x16 mmx sse2
+vp9_variance8x16_sse2=vp9_variance8x16_wmt
+vp9_variance8x16_mmx=vp9_variance8x16_mmx
 
-prototype unsigned int vp8_variance8x8 "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int ref_stride, unsigned int *sse"
-specialize vp8_variance8x8 mmx sse2
-vp8_variance8x8_sse2=vp8_variance8x8_wmt
-vp8_variance8x8_mmx=vp8_variance8x8_mmx
+prototype unsigned int vp9_variance8x8 "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int ref_stride, unsigned int *sse"
+specialize vp9_variance8x8 mmx sse2
+vp9_variance8x8_sse2=vp9_variance8x8_wmt
+vp9_variance8x8_mmx=vp9_variance8x8_mmx
 
-prototype unsigned int vp8_variance4x4 "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int ref_stride, unsigned int *sse"
-specialize vp8_variance4x4 mmx sse2
-vp8_variance4x4_sse2=vp8_variance4x4_wmt
-vp8_variance4x4_mmx=vp8_variance4x4_mmx
+prototype unsigned int vp9_variance4x4 "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int ref_stride, unsigned int *sse"
+specialize vp9_variance4x4 mmx sse2
+vp9_variance4x4_sse2=vp9_variance4x4_wmt
+vp9_variance4x4_mmx=vp9_variance4x4_mmx
 
-prototype unsigned int vp8_sub_pixel_variance32x32 "const unsigned char *src_ptr, int source_stride, int xoffset, int  yoffset, const unsigned char *ref_ptr, int Refstride, unsigned int *sse"
-specialize vp8_sub_pixel_variance32x32
+prototype unsigned int vp9_sub_pixel_variance32x32 "const unsigned char *src_ptr, int source_stride, int xoffset, int  yoffset, const unsigned char *ref_ptr, int Refstride, unsigned int *sse"
+specialize vp9_sub_pixel_variance32x32
 
-prototype unsigned int vp8_sub_pixel_variance16x16 "const unsigned char *src_ptr, int source_stride, int xoffset, int  yoffset, const unsigned char *ref_ptr, int Refstride, unsigned int *sse"
-specialize vp8_sub_pixel_variance16x16 sse2 mmx ssse3
-vp8_sub_pixel_variance16x16_sse2=vp8_sub_pixel_variance16x16_wmt
+prototype unsigned int vp9_sub_pixel_variance16x16 "const unsigned char *src_ptr, int source_stride, int xoffset, int  yoffset, const unsigned char *ref_ptr, int Refstride, unsigned int *sse"
+specialize vp9_sub_pixel_variance16x16 sse2 mmx ssse3
+vp9_sub_pixel_variance16x16_sse2=vp9_sub_pixel_variance16x16_wmt
 
-prototype unsigned int vp8_sub_pixel_variance8x16 "const unsigned char *src_ptr, int source_stride, int xoffset, int  yoffset, const unsigned char *ref_ptr, int Refstride, unsigned int *sse"
-specialize vp8_sub_pixel_variance8x16 sse2 mmx
-vp8_sub_pixel_variance8x16_sse2=vp8_sub_pixel_variance8x16_wmt
+prototype unsigned int vp9_sub_pixel_variance8x16 "const unsigned char *src_ptr, int source_stride, int xoffset, int  yoffset, const unsigned char *ref_ptr, int Refstride, unsigned int *sse"
+specialize vp9_sub_pixel_variance8x16 sse2 mmx
+vp9_sub_pixel_variance8x16_sse2=vp9_sub_pixel_variance8x16_wmt
 
-prototype unsigned int vp8_sub_pixel_variance16x8 "const unsigned char *src_ptr, int source_stride, int xoffset, int  yoffset, const unsigned char *ref_ptr, int Refstride, unsigned int *sse"
-specialize vp8_sub_pixel_variance16x8 sse2 mmx ssse3
-vp8_sub_pixel_variance16x8_sse2=vp8_sub_pixel_variance16x8_ssse3;
-vp8_sub_pixel_variance16x8_sse2=vp8_sub_pixel_variance16x8_wmt
+prototype unsigned int vp9_sub_pixel_variance16x8 "const unsigned char *src_ptr, int source_stride, int xoffset, int  yoffset, const unsigned char *ref_ptr, int Refstride, unsigned int *sse"
+specialize vp9_sub_pixel_variance16x8 sse2 mmx ssse3
+vp9_sub_pixel_variance16x8_sse2=vp9_sub_pixel_variance16x8_ssse3;
+vp9_sub_pixel_variance16x8_sse2=vp9_sub_pixel_variance16x8_wmt
 
-prototype unsigned int vp8_sub_pixel_variance8x8 "const unsigned char *src_ptr, int source_stride, int xoffset, int  yoffset, const unsigned char *ref_ptr, int Refstride, unsigned int *sse"
-specialize vp8_sub_pixel_variance8x8 sse2 mmx
-vp8_sub_pixel_variance8x8_sse2=vp8_sub_pixel_variance8x8_wmt
+prototype unsigned int vp9_sub_pixel_variance8x8 "const unsigned char *src_ptr, int source_stride, int xoffset, int  yoffset, const unsigned char *ref_ptr, int Refstride, unsigned int *sse"
+specialize vp9_sub_pixel_variance8x8 sse2 mmx
+vp9_sub_pixel_variance8x8_sse2=vp9_sub_pixel_variance8x8_wmt
 
-prototype unsigned int vp8_sub_pixel_variance4x4 "const unsigned char *src_ptr, int source_stride, int xoffset, int  yoffset, const unsigned char *ref_ptr, int Refstride, unsigned int *sse"
-specialize vp8_sub_pixel_variance4x4 sse2 mmx
-vp8_sub_pixel_variance4x4_sse2=vp8_sub_pixel_variance4x4_wmt
+prototype unsigned int vp9_sub_pixel_variance4x4 "const unsigned char *src_ptr, int source_stride, int xoffset, int  yoffset, const unsigned char *ref_ptr, int Refstride, unsigned int *sse"
+specialize vp9_sub_pixel_variance4x4 sse2 mmx
+vp9_sub_pixel_variance4x4_sse2=vp9_sub_pixel_variance4x4_wmt
 
-prototype unsigned int vp8_sad32x32 "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int  ref_stride, unsigned int max_sad"
-specialize vp8_sad32x32
+prototype unsigned int vp9_sad32x32 "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int  ref_stride, unsigned int max_sad"
+specialize vp9_sad32x32
 
-prototype unsigned int vp8_sad16x16 "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int  ref_stride, unsigned int max_sad"
-specialize vp8_sad16x16 mmx sse2 sse3
-vp8_sad16x16_sse2=vp8_sad16x16_wmt
+prototype unsigned int vp9_sad16x16 "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int  ref_stride, unsigned int max_sad"
+specialize vp9_sad16x16 mmx sse2 sse3
+vp9_sad16x16_sse2=vp9_sad16x16_wmt
 
-prototype unsigned int vp8_sad16x8 "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int  ref_stride, unsigned int max_sad"
-specialize vp8_sad16x8 mmx sse2
-vp8_sad16x8_sse2=vp8_sad16x8_wmt
+prototype unsigned int vp9_sad16x8 "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int  ref_stride, unsigned int max_sad"
+specialize vp9_sad16x8 mmx sse2
+vp9_sad16x8_sse2=vp9_sad16x8_wmt
 
-prototype unsigned int vp8_sad8x16 "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int  ref_stride, unsigned int max_sad"
-specialize vp8_sad8x16 mmx sse2
-vp8_sad8x16_sse2=vp8_sad8x16_wmt
+prototype unsigned int vp9_sad8x16 "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int  ref_stride, unsigned int max_sad"
+specialize vp9_sad8x16 mmx sse2
+vp9_sad8x16_sse2=vp9_sad8x16_wmt
 
-prototype unsigned int vp8_sad8x8 "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int  ref_stride, unsigned int max_sad"
-specialize vp8_sad8x8 mmx sse2
-vp8_sad8x8_sse2=vp8_sad8x8_wmt
+prototype unsigned int vp9_sad8x8 "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int  ref_stride, unsigned int max_sad"
+specialize vp9_sad8x8 mmx sse2
+vp9_sad8x8_sse2=vp9_sad8x8_wmt
 
-prototype unsigned int vp8_sad4x4 "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int  ref_stride, unsigned int max_sad"
-specialize vp8_sad4x4 mmx sse2
-vp8_sad4x4_sse2=vp8_sad4x4_wmt
+prototype unsigned int vp9_sad4x4 "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int  ref_stride, unsigned int max_sad"
+specialize vp9_sad4x4 mmx sse2
+vp9_sad4x4_sse2=vp9_sad4x4_wmt
 
-prototype unsigned int vp8_variance_halfpixvar16x16_h "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int ref_stride, unsigned int *sse"
-specialize vp8_variance_halfpixvar16x16_h mmx sse2
-vp8_variance_halfpixvar16x16_h_sse2=vp8_variance_halfpixvar16x16_h_wmt
+prototype unsigned int vp9_variance_halfpixvar16x16_h "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int ref_stride, unsigned int *sse"
+specialize vp9_variance_halfpixvar16x16_h mmx sse2
+vp9_variance_halfpixvar16x16_h_sse2=vp9_variance_halfpixvar16x16_h_wmt
 
-prototype unsigned int vp8_variance_halfpixvar16x16_v "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int ref_stride, unsigned int *sse"
-specialize vp8_variance_halfpixvar16x16_v mmx sse2
-vp8_variance_halfpixvar16x16_v_sse2=vp8_variance_halfpixvar16x16_v_wmt
+prototype unsigned int vp9_variance_halfpixvar16x16_v "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int ref_stride, unsigned int *sse"
+specialize vp9_variance_halfpixvar16x16_v mmx sse2
+vp9_variance_halfpixvar16x16_v_sse2=vp9_variance_halfpixvar16x16_v_wmt
 
-prototype unsigned int vp8_variance_halfpixvar16x16_hv "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int ref_stride, unsigned int *sse"
-specialize vp8_variance_halfpixvar16x16_hv mmx sse2
-vp8_variance_halfpixvar16x16_hv_sse2=vp8_variance_halfpixvar16x16_hv_wmt
+prototype unsigned int vp9_variance_halfpixvar16x16_hv "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int ref_stride, unsigned int *sse"
+specialize vp9_variance_halfpixvar16x16_hv mmx sse2
+vp9_variance_halfpixvar16x16_hv_sse2=vp9_variance_halfpixvar16x16_hv_wmt
 
-prototype unsigned int vp8_variance_halfpixvar32x32_h "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int ref_stride, unsigned int *sse"
-specialize vp8_variance_halfpixvar32x32_h
+prototype unsigned int vp9_variance_halfpixvar32x32_h "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int ref_stride, unsigned int *sse"
+specialize vp9_variance_halfpixvar32x32_h
 
-prototype unsigned int vp8_variance_halfpixvar32x32_v "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int ref_stride, unsigned int *sse"
-specialize vp8_variance_halfpixvar32x32_v
+prototype unsigned int vp9_variance_halfpixvar32x32_v "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int ref_stride, unsigned int *sse"
+specialize vp9_variance_halfpixvar32x32_v
 
-prototype unsigned int vp8_variance_halfpixvar32x32_hv "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int ref_stride, unsigned int *sse"
-specialize vp8_variance_halfpixvar32x32_hv
+prototype unsigned int vp9_variance_halfpixvar32x32_hv "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int ref_stride, unsigned int *sse"
+specialize vp9_variance_halfpixvar32x32_hv
 
-prototype void vp8_sad32x32x3 "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int  ref_stride, unsigned int *sad_array"
-specialize vp8_sad32x32x3
+prototype void vp9_sad32x32x3 "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int  ref_stride, unsigned int *sad_array"
+specialize vp9_sad32x32x3
 
-prototype void vp8_sad16x16x3 "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int  ref_stride, unsigned int *sad_array"
-specialize vp8_sad16x16x3 sse3 ssse3
+prototype void vp9_sad16x16x3 "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int  ref_stride, unsigned int *sad_array"
+specialize vp9_sad16x16x3 sse3 ssse3
 
-prototype void vp8_sad16x8x3 "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int  ref_stride, unsigned int *sad_array"
-specialize vp8_sad16x8x3 sse3 ssse3
+prototype void vp9_sad16x8x3 "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int  ref_stride, unsigned int *sad_array"
+specialize vp9_sad16x8x3 sse3 ssse3
 
-prototype void vp8_sad8x16x3 "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int  ref_stride, unsigned int *sad_array"
-specialize vp8_sad8x16x3 sse3
+prototype void vp9_sad8x16x3 "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int  ref_stride, unsigned int *sad_array"
+specialize vp9_sad8x16x3 sse3
 
-prototype void vp8_sad8x8x3 "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int  ref_stride, unsigned int *sad_array"
-specialize vp8_sad8x8x3 sse3
+prototype void vp9_sad8x8x3 "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int  ref_stride, unsigned int *sad_array"
+specialize vp9_sad8x8x3 sse3
 
-prototype void vp8_sad4x4x3 "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int  ref_stride, unsigned int *sad_array"
-specialize vp8_sad4x4x3 sse3
+prototype void vp9_sad4x4x3 "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int  ref_stride, unsigned int *sad_array"
+specialize vp9_sad4x4x3 sse3
 
-prototype void vp8_sad32x32x8 "const unsigned char *src_ptr, int  src_stride, const unsigned char *ref_ptr, int  ref_stride, unsigned short *sad_array"
-specialize vp8_sad32x32x8
+prototype void vp9_sad32x32x8 "const unsigned char *src_ptr, int  src_stride, const unsigned char *ref_ptr, int  ref_stride, unsigned short *sad_array"
+specialize vp9_sad32x32x8
 
-prototype void vp8_sad16x16x8 "const unsigned char *src_ptr, int  src_stride, const unsigned char *ref_ptr, int  ref_stride, unsigned short *sad_array"
-specialize vp8_sad16x16x8 sse4
+prototype void vp9_sad16x16x8 "const unsigned char *src_ptr, int  src_stride, const unsigned char *ref_ptr, int  ref_stride, unsigned short *sad_array"
+specialize vp9_sad16x16x8 sse4
 
-prototype void vp8_sad16x8x8 "const unsigned char *src_ptr, int  src_stride, const unsigned char *ref_ptr, int  ref_stride, unsigned short *sad_array"
-specialize vp8_sad16x8x8 sse4
+prototype void vp9_sad16x8x8 "const unsigned char *src_ptr, int  src_stride, const unsigned char *ref_ptr, int  ref_stride, unsigned short *sad_array"
+specialize vp9_sad16x8x8 sse4
 
-prototype void vp8_sad8x16x8 "const unsigned char *src_ptr, int  src_stride, const unsigned char *ref_ptr, int  ref_stride, unsigned short *sad_array"
-specialize vp8_sad8x16x8 sse4
+prototype void vp9_sad8x16x8 "const unsigned char *src_ptr, int  src_stride, const unsigned char *ref_ptr, int  ref_stride, unsigned short *sad_array"
+specialize vp9_sad8x16x8 sse4
 
-prototype void vp8_sad8x8x8 "const unsigned char *src_ptr, int  src_stride, const unsigned char *ref_ptr, int  ref_stride, unsigned short *sad_array"
-specialize vp8_sad8x8x8 sse4
+prototype void vp9_sad8x8x8 "const unsigned char *src_ptr, int  src_stride, const unsigned char *ref_ptr, int  ref_stride, unsigned short *sad_array"
+specialize vp9_sad8x8x8 sse4
 
-prototype void vp8_sad4x4x8 "const unsigned char *src_ptr, int  src_stride, const unsigned char *ref_ptr, int  ref_stride, unsigned short *sad_array"
-specialize vp8_sad4x4x8 sse4
+prototype void vp9_sad4x4x8 "const unsigned char *src_ptr, int  src_stride, const unsigned char *ref_ptr, int  ref_stride, unsigned short *sad_array"
+specialize vp9_sad4x4x8 sse4
 
-prototype void vp8_sad32x32x4d "const unsigned char *src_ptr, int  src_stride, unsigned char *ref_ptr[], int  ref_stride, unsigned int *sad_array"
-specialize vp8_sad32x32x4d
+prototype void vp9_sad32x32x4d "const unsigned char *src_ptr, int  src_stride, unsigned char *ref_ptr[], int  ref_stride, unsigned int *sad_array"
+specialize vp9_sad32x32x4d
 
-prototype void vp8_sad16x16x4d "const unsigned char *src_ptr, int  src_stride, unsigned char *ref_ptr[], int  ref_stride, unsigned int *sad_array"
-specialize vp8_sad16x16x4d sse3
+prototype void vp9_sad16x16x4d "const unsigned char *src_ptr, int  src_stride, unsigned char *ref_ptr[], int  ref_stride, unsigned int *sad_array"
+specialize vp9_sad16x16x4d sse3
 
-prototype void vp8_sad16x8x4d "const unsigned char *src_ptr, int  src_stride, unsigned char *ref_ptr[], int  ref_stride, unsigned int *sad_array"
-specialize vp8_sad16x8x4d sse3
+prototype void vp9_sad16x8x4d "const unsigned char *src_ptr, int  src_stride, unsigned char *ref_ptr[], int  ref_stride, unsigned int *sad_array"
+specialize vp9_sad16x8x4d sse3
 
-prototype void vp8_sad8x16x4d "const unsigned char *src_ptr, int  src_stride, unsigned char *ref_ptr[], int  ref_stride, unsigned int *sad_array"
-specialize vp8_sad8x16x4d sse3
+prototype void vp9_sad8x16x4d "const unsigned char *src_ptr, int  src_stride, unsigned char *ref_ptr[], int  ref_stride, unsigned int *sad_array"
+specialize vp9_sad8x16x4d sse3
 
-prototype void vp8_sad8x8x4d "const unsigned char *src_ptr, int  src_stride, unsigned char *ref_ptr[], int  ref_stride, unsigned int *sad_array"
-specialize vp8_sad8x8x4d sse3
+prototype void vp9_sad8x8x4d "const unsigned char *src_ptr, int  src_stride, unsigned char *ref_ptr[], int  ref_stride, unsigned int *sad_array"
+specialize vp9_sad8x8x4d sse3
 
-prototype void vp8_sad4x4x4d "const unsigned char *src_ptr, int  src_stride, unsigned char *ref_ptr[], int  ref_stride, unsigned int *sad_array"
-specialize vp8_sad4x4x4d sse3
+prototype void vp9_sad4x4x4d "const unsigned char *src_ptr, int  src_stride, unsigned char *ref_ptr[], int  ref_stride, unsigned int *sad_array"
+specialize vp9_sad4x4x4d sse3
 
 #
 # Block copy
 #
 case $arch in
     x86*)
-    prototype void vp8_copy32xn "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int ref_stride, int n"
-    specialize vp8_copy32xn sse2 sse3
+    prototype void vp9_copy32xn "const unsigned char *src_ptr, int source_stride, const unsigned char *ref_ptr, int ref_stride, int n"
+    specialize vp9_copy32xn sse2 sse3
     ;;
 esac
 
-prototype unsigned int vp8_sub_pixel_mse16x16 "const unsigned char  *src_ptr, int  src_pixels_per_line, int  xoffset, int  yoffset, const unsigned char *dst_ptr, int dst_pixels_per_line, unsigned int *sse"
-specialize vp8_sub_pixel_mse16x16 sse2 mmx
-vp8_sub_pixel_mse16x16_sse2=vp8_sub_pixel_mse16x16_wmt
+prototype unsigned int vp9_sub_pixel_mse16x16 "const unsigned char  *src_ptr, int  src_pixels_per_line, int  xoffset, int  yoffset, const unsigned char *dst_ptr, int dst_pixels_per_line, unsigned int *sse"
+specialize vp9_sub_pixel_mse16x16 sse2 mmx
+vp9_sub_pixel_mse16x16_sse2=vp9_sub_pixel_mse16x16_wmt
 
-prototype unsigned int vp8_mse16x16 "const unsigned char *src_ptr, int  source_stride, const unsigned char *ref_ptr, int  recon_stride, unsigned int *sse"
-specialize vp8_mse16x16 mmx sse2
-vp8_mse16x16_sse2=vp8_mse16x16_wmt
+prototype unsigned int vp9_mse16x16 "const unsigned char *src_ptr, int  source_stride, const unsigned char *ref_ptr, int  recon_stride, unsigned int *sse"
+specialize vp9_mse16x16 mmx sse2
+vp9_mse16x16_sse2=vp9_mse16x16_wmt
 
-prototype unsigned int vp8_sub_pixel_mse32x32 "const unsigned char  *src_ptr, int  source_stride, int  xoffset, int  yoffset, const unsigned char *ref_ptr, int Refstride, unsigned int *sse"
-specialize vp8_sub_pixel_mse32x32
+prototype unsigned int vp9_sub_pixel_mse32x32 "const unsigned char  *src_ptr, int  source_stride, int  xoffset, int  yoffset, const unsigned char *ref_ptr, int Refstride, unsigned int *sse"
+specialize vp9_sub_pixel_mse32x32
 
-prototype unsigned int vp8_get_mb_ss "const short *"
-specialize vp8_get_mb_ss mmx sse2
+prototype unsigned int vp9_get_mb_ss "const short *"
+specialize vp9_get_mb_ss mmx sse2
 # ENCODEMB INVOKE
-prototype int vp8_mbblock_error "struct macroblock *mb, int dc"
-specialize vp8_mbblock_error mmx sse2
-vp8_mbblock_error_sse2=vp8_mbblock_error_xmm
+prototype int vp9_mbblock_error "struct macroblock *mb, int dc"
+specialize vp9_mbblock_error mmx sse2
+vp9_mbblock_error_sse2=vp9_mbblock_error_xmm
 
-prototype int vp8_block_error "short *coeff, short *dqcoeff, int block_size"
-specialize vp8_block_error mmx sse2
-vp8_block_error_sse2=vp8_block_error_xmm
+prototype int vp9_block_error "short *coeff, short *dqcoeff, int block_size"
+specialize vp9_block_error mmx sse2
+vp9_block_error_sse2=vp9_block_error_xmm
 
-prototype void vp8_subtract_b "struct block *be, struct blockd *bd, int pitch"
-specialize vp8_subtract_b mmx sse2
+prototype void vp9_subtract_b "struct block *be, struct blockd *bd, int pitch"
+specialize vp9_subtract_b mmx sse2
 
-prototype int vp8_mbuverror "struct macroblock *mb"
-specialize vp8_mbuverror mmx sse2
-vp8_mbuverror_sse2=vp8_mbuverror_xmm
+prototype int vp9_mbuverror "struct macroblock *mb"
+specialize vp9_mbuverror mmx sse2
+vp9_mbuverror_sse2=vp9_mbuverror_xmm
 
-prototype void vp8_subtract_b "struct block *be, struct blockd *bd, int pitch"
-specialize vp8_subtract_b mmx sse2
+prototype void vp9_subtract_b "struct block *be, struct blockd *bd, int pitch"
+specialize vp9_subtract_b mmx sse2
 
-prototype void vp8_subtract_mby "short *diff, unsigned char *src, unsigned char *pred, int stride"
-specialize vp8_subtract_mby mmx sse2
+prototype void vp9_subtract_mby "short *diff, unsigned char *src, unsigned char *pred, int stride"
+specialize vp9_subtract_mby mmx sse2
 
-prototype void vp8_subtract_mbuv "short *diff, unsigned char *usrc, unsigned char *vsrc, unsigned char *pred, int stride"
-specialize vp8_subtract_mbuv mmx sse2
+prototype void vp9_subtract_mbuv "short *diff, unsigned char *usrc, unsigned char *vsrc, unsigned char *pred, int stride"
+specialize vp9_subtract_mbuv mmx sse2
 
 #
 # Structured Similarity (SSIM)
@@ -440,40 +440,40 @@ specialize vp8_subtract_mbuv mmx sse2
 if [ "$CONFIG_INTERNAL_STATS" = "yes" ]; then
     [ $arch = "x86_64" ] && sse2_on_x86_64=sse2
 
-    prototype void vp8_ssim_parms_8x8 "unsigned char *s, int sp, unsigned char *r, int rp, unsigned long *sum_s, unsigned long *sum_r, unsigned long *sum_sq_s, unsigned long *sum_sq_r, unsigned long *sum_sxr"
-    specialize vp8_ssim_parms_8x8 $sse2_on_x86_64
+    prototype void vp9_ssim_parms_8x8 "unsigned char *s, int sp, unsigned char *r, int rp, unsigned long *sum_s, unsigned long *sum_r, unsigned long *sum_sq_s, unsigned long *sum_sq_r, unsigned long *sum_sxr"
+    specialize vp9_ssim_parms_8x8 $sse2_on_x86_64
 
-    prototype void vp8_ssim_parms_16x16 "unsigned char *s, int sp, unsigned char *r, int rp, unsigned long *sum_s, unsigned long *sum_r, unsigned long *sum_sq_s, unsigned long *sum_sq_r, unsigned long *sum_sxr"
-    specialize vp8_ssim_parms_16x16 $sse2_on_x86_64
+    prototype void vp9_ssim_parms_16x16 "unsigned char *s, int sp, unsigned char *r, int rp, unsigned long *sum_s, unsigned long *sum_r, unsigned long *sum_sq_s, unsigned long *sum_sq_r, unsigned long *sum_sxr"
+    specialize vp9_ssim_parms_16x16 $sse2_on_x86_64
 fi
 
 # fdct functions
-prototype void vp8_short_fdct8x8 "short *InputData, short *OutputData, int pitch"
-specialize vp8_short_fdct8x8
+prototype void vp9_short_fdct8x8 "short *InputData, short *OutputData, int pitch"
+specialize vp9_short_fdct8x8
 
-prototype void vp8_short_fhaar2x2 "short *InputData, short *OutputData, int pitch"
-specialize vp8_short_fhaar2x2
+prototype void vp9_short_fhaar2x2 "short *InputData, short *OutputData, int pitch"
+specialize vp9_short_fhaar2x2
 
-prototype void vp8_short_fdct4x4 "short *InputData, short *OutputData, int pitch"
-specialize vp8_short_fdct4x4
+prototype void vp9_short_fdct4x4 "short *InputData, short *OutputData, int pitch"
+specialize vp9_short_fdct4x4
 
-prototype void vp8_short_fdct8x4 "short *InputData, short *OutputData, int pitch"
-specialize vp8_short_fdct8x4
+prototype void vp9_short_fdct8x4 "short *InputData, short *OutputData, int pitch"
+specialize vp9_short_fdct8x4
 
-prototype void vp8_short_walsh4x4 "short *InputData, short *OutputData, int pitch"
-specialize vp8_short_walsh4x4
+prototype void vp9_short_walsh4x4 "short *InputData, short *OutputData, int pitch"
+specialize vp9_short_walsh4x4
 
-prototype void vp8_short_fdct16x16 "short *InputData, short *OutputData, int pitch"
-specialize vp8_short_fdct16x16
+prototype void vp9_short_fdct16x16 "short *InputData, short *OutputData, int pitch"
+specialize vp9_short_fdct16x16
 
-prototype void vp8_short_walsh4x4_lossless "short *InputData, short *OutputData, int pitch"
-specialize vp8_short_walsh4x4_lossless
+prototype void vp9_short_walsh4x4_lossless "short *InputData, short *OutputData, int pitch"
+specialize vp9_short_walsh4x4_lossless
 
-prototype void vp8_short_walsh4x4_x8 "short *InputData, short *OutputData, int pitch"
-specialize vp8_short_walsh4x4_x8
+prototype void vp9_short_walsh4x4_x8 "short *InputData, short *OutputData, int pitch"
+specialize vp9_short_walsh4x4_x8
 
-prototype void vp8_short_walsh8x4_x8 "short *InputData, short *OutputData, int pitch"
-specialize vp8_short_walsh8x4_x8
+prototype void vp9_short_walsh8x4_x8 "short *InputData, short *OutputData, int pitch"
+specialize vp9_short_walsh8x4_x8
 
 fi
 # end encoder functions
diff --git a/vp8/encoder/arm/arm_csystemdependent.c b/vp8/encoder/arm/arm_csystemdependent.c
index 1166efd96f25f12282e047c4966bc7663ff5345f..2efeafb993cf6205291e89284d3ac1efbc7eb23f 100644
--- a/vp8/encoder/arm/arm_csystemdependent.c
+++ b/vp8/encoder/arm/arm_csystemdependent.c
@@ -14,8 +14,8 @@
 #include "vp8/encoder/variance.h"
 #include "vp8/encoder/onyx_int.h"
 
-extern void (*vp8_yv12_copy_partial_frame_ptr)(YV12_BUFFER_CONFIG *src_ybc, YV12_BUFFER_CONFIG *dst_ybc, int Fraction);
-extern void vp8_yv12_copy_partial_frame(YV12_BUFFER_CONFIG *src_ybc, YV12_BUFFER_CONFIG *dst_ybc, int Fraction);
+extern void (*vp9_yv12_copy_partial_frame_ptr)(YV12_BUFFER_CONFIG *src_ybc, YV12_BUFFER_CONFIG *dst_ybc, int Fraction);
+extern void vp9_yv12_copy_partial_frame(YV12_BUFFER_CONFIG *src_ybc, YV12_BUFFER_CONFIG *dst_ybc, int Fraction);
 extern void vpxyv12_copy_partial_frame_neon(YV12_BUFFER_CONFIG *src_ybc, YV12_BUFFER_CONFIG *dst_ybc, int Fraction);
 
 void vp8_arch_arm_encoder_init(VP8_COMP *cpi) {
@@ -29,42 +29,42 @@ void vp8_arch_arm_encoder_init(VP8_COMP *cpi) {
 
 #if HAVE_ARMV6
   if (flags & HAS_MEDIA) {
-    cpi->rtcd.variance.sad16x16              = vp8_sad16x16_armv6;
-    /*cpi->rtcd.variance.sad16x8               = vp8_sad16x8_c;
-    cpi->rtcd.variance.sad8x16               = vp8_sad8x16_c;
-    cpi->rtcd.variance.sad8x8                = vp8_sad8x8_c;
-    cpi->rtcd.variance.sad4x4                = vp8_sad4x4_c;*/
-
-    /*cpi->rtcd.variance.var4x4                = vp8_variance4x4_c;*/
-    cpi->rtcd.variance.var8x8                = vp8_variance8x8_armv6;
-    /*cpi->rtcd.variance.var8x16               = vp8_variance8x16_c;
-    cpi->rtcd.variance.var16x8               = vp8_variance16x8_c;*/
-    cpi->rtcd.variance.var16x16              = vp8_variance16x16_armv6;
-
-    /*cpi->rtcd.variance.subpixvar4x4          = vp8_sub_pixel_variance4x4_c;*/
-    cpi->rtcd.variance.subpixvar8x8          = vp8_sub_pixel_variance8x8_armv6;
-    /*cpi->rtcd.variance.subpixvar8x16         = vp8_sub_pixel_variance8x16_c;
-    cpi->rtcd.variance.subpixvar16x8         = vp8_sub_pixel_variance16x8_c;*/
-    cpi->rtcd.variance.subpixvar16x16        = vp8_sub_pixel_variance16x16_armv6;
-    cpi->rtcd.variance.halfpixvar16x16_h     = vp8_variance_halfpixvar16x16_h_armv6;
-    cpi->rtcd.variance.halfpixvar16x16_v     = vp8_variance_halfpixvar16x16_v_armv6;
-    cpi->rtcd.variance.halfpixvar16x16_hv    = vp8_variance_halfpixvar16x16_hv_armv6;
-
-    cpi->rtcd.variance.mse16x16              = vp8_mse16x16_armv6;
-    /*cpi->rtcd.variance.getmbss               = vp8_get_mb_ss_c;*/
-
-    cpi->rtcd.fdct.short4x4                  = vp8_short_fdct4x4_armv6;
-    cpi->rtcd.fdct.short8x4                  = vp8_short_fdct8x4_armv6;
-    cpi->rtcd.fdct.fast4x4                   = vp8_short_fdct4x4_armv6;
-    cpi->rtcd.fdct.fast8x4                   = vp8_short_fdct8x4_armv6;
-    cpi->rtcd.fdct.walsh_short4x4            = vp8_short_walsh4x4_armv6;
-
-    /*cpi->rtcd.encodemb.berr                  = vp8_block_error_c;
-    cpi->rtcd.encodemb.mberr                 = vp8_mbblock_error_c;
-    cpi->rtcd.encodemb.mbuverr               = vp8_mbuverror_c;*/
-    cpi->rtcd.encodemb.subb                  = vp8_subtract_b_armv6;
-    cpi->rtcd.encodemb.submby                = vp8_subtract_mby_armv6;
-    cpi->rtcd.encodemb.submbuv               = vp8_subtract_mbuv_armv6;
+    cpi->rtcd.variance.sad16x16              = vp9_sad16x16_armv6;
+    /*cpi->rtcd.variance.sad16x8               = vp9_sad16x8_c;
+    cpi->rtcd.variance.sad8x16               = vp9_sad8x16_c;
+    cpi->rtcd.variance.sad8x8                = vp9_sad8x8_c;
+    cpi->rtcd.variance.sad4x4                = vp9_sad4x4_c;*/
+
+    /*cpi->rtcd.variance.var4x4                = vp9_variance4x4_c;*/
+    cpi->rtcd.variance.var8x8                = vp9_variance8x8_armv6;
+    /*cpi->rtcd.variance.var8x16               = vp9_variance8x16_c;
+    cpi->rtcd.variance.var16x8               = vp9_variance16x8_c;*/
+    cpi->rtcd.variance.var16x16              = vp9_variance16x16_armv6;
+
+    /*cpi->rtcd.variance.subpixvar4x4          = vp9_sub_pixel_variance4x4_c;*/
+    cpi->rtcd.variance.subpixvar8x8          = vp9_sub_pixel_variance8x8_armv6;
+    /*cpi->rtcd.variance.subpixvar8x16         = vp9_sub_pixel_variance8x16_c;
+    cpi->rtcd.variance.subpixvar16x8         = vp9_sub_pixel_variance16x8_c;*/
+    cpi->rtcd.variance.subpixvar16x16        = vp9_sub_pixel_variance16x16_armv6;
+    cpi->rtcd.variance.halfpixvar16x16_h     = vp9_variance_halfpixvar16x16_h_armv6;
+    cpi->rtcd.variance.halfpixvar16x16_v     = vp9_variance_halfpixvar16x16_v_armv6;
+    cpi->rtcd.variance.halfpixvar16x16_hv    = vp9_variance_halfpixvar16x16_hv_armv6;
+
+    cpi->rtcd.variance.mse16x16              = vp9_mse16x16_armv6;
+    /*cpi->rtcd.variance.getmbss               = vp9_get_mb_ss_c;*/
+
+    cpi->rtcd.fdct.short4x4                  = vp9_short_fdct4x4_armv6;
+    cpi->rtcd.fdct.short8x4                  = vp9_short_fdct8x4_armv6;
+    cpi->rtcd.fdct.fast4x4                   = vp9_short_fdct4x4_armv6;
+    cpi->rtcd.fdct.fast8x4                   = vp9_short_fdct8x4_armv6;
+    cpi->rtcd.fdct.walsh_short4x4            = vp9_short_walsh4x4_armv6;
+
+    /*cpi->rtcd.encodemb.berr                  = vp9_block_error_c;
+    cpi->rtcd.encodemb.mberr                 = vp9_mbblock_error_c;
+    cpi->rtcd.encodemb.mbuverr               = vp9_mbuverror_c;*/
+    cpi->rtcd.encodemb.subb                  = vp9_subtract_b_armv6;
+    cpi->rtcd.encodemb.submby                = vp9_subtract_mby_armv6;
+    cpi->rtcd.encodemb.submbuv               = vp9_subtract_mbuv_armv6;
 
     /*cpi->rtcd.quantize.quantb                = vp8_regular_quantize_b;*/
     cpi->rtcd.quantize.fastquantb            = vp8_fast_quantize_b_armv6;
@@ -73,42 +73,42 @@ void vp8_arch_arm_encoder_init(VP8_COMP *cpi) {
 
 #if HAVE_ARMV7
   if (flags & HAS_NEON) {
-    cpi->rtcd.variance.sad16x16              = vp8_sad16x16_neon;
-    cpi->rtcd.variance.sad16x8               = vp8_sad16x8_neon;
-    cpi->rtcd.variance.sad8x16               = vp8_sad8x16_neon;
-    cpi->rtcd.variance.sad8x8                = vp8_sad8x8_neon;
-    cpi->rtcd.variance.sad4x4                = vp8_sad4x4_neon;
-
-    /*cpi->rtcd.variance.var4x4                = vp8_variance4x4_c;*/
-    cpi->rtcd.variance.var8x8                = vp8_variance8x8_neon;
-    cpi->rtcd.variance.var8x16               = vp8_variance8x16_neon;
-    cpi->rtcd.variance.var16x8               = vp8_variance16x8_neon;
-    cpi->rtcd.variance.var16x16              = vp8_variance16x16_neon;
-
-    /*cpi->rtcd.variance.subpixvar4x4          = vp8_sub_pixel_variance4x4_c;*/
-    cpi->rtcd.variance.subpixvar8x8          = vp8_sub_pixel_variance8x8_neon;
-    /*cpi->rtcd.variance.subpixvar8x16         = vp8_sub_pixel_variance8x16_c;
-    cpi->rtcd.variance.subpixvar16x8         = vp8_sub_pixel_variance16x8_c;*/
-    cpi->rtcd.variance.subpixvar16x16        = vp8_sub_pixel_variance16x16_neon;
-    cpi->rtcd.variance.halfpixvar16x16_h     = vp8_variance_halfpixvar16x16_h_neon;
-    cpi->rtcd.variance.halfpixvar16x16_v     = vp8_variance_halfpixvar16x16_v_neon;
-    cpi->rtcd.variance.halfpixvar16x16_hv    = vp8_variance_halfpixvar16x16_hv_neon;
-
-    cpi->rtcd.variance.mse16x16              = vp8_mse16x16_neon;
-    /*cpi->rtcd.variance.getmbss               = vp8_get_mb_ss_c;*/
-
-    cpi->rtcd.fdct.short4x4                  = vp8_short_fdct4x4_neon;
-    cpi->rtcd.fdct.short8x4                  = vp8_short_fdct8x4_neon;
-    cpi->rtcd.fdct.fast4x4                   = vp8_short_fdct4x4_neon;
-    cpi->rtcd.fdct.fast8x4                   = vp8_short_fdct8x4_neon;
-    cpi->rtcd.fdct.walsh_short4x4            = vp8_short_walsh4x4_neon;
-
-    /*cpi->rtcd.encodemb.berr                  = vp8_block_error_c;
-    cpi->rtcd.encodemb.mberr                 = vp8_mbblock_error_c;
-    cpi->rtcd.encodemb.mbuverr               = vp8_mbuverror_c;*/
-    cpi->rtcd.encodemb.subb                  = vp8_subtract_b_neon;
-    cpi->rtcd.encodemb.submby                = vp8_subtract_mby_neon;
-    cpi->rtcd.encodemb.submbuv               = vp8_subtract_mbuv_neon;
+    cpi->rtcd.variance.sad16x16              = vp9_sad16x16_neon;
+    cpi->rtcd.variance.sad16x8               = vp9_sad16x8_neon;
+    cpi->rtcd.variance.sad8x16               = vp9_sad8x16_neon;
+    cpi->rtcd.variance.sad8x8                = vp9_sad8x8_neon;
+    cpi->rtcd.variance.sad4x4                = vp9_sad4x4_neon;
+
+    /*cpi->rtcd.variance.var4x4                = vp9_variance4x4_c;*/
+    cpi->rtcd.variance.var8x8                = vp9_variance8x8_neon;
+    cpi->rtcd.variance.var8x16               = vp9_variance8x16_neon;
+    cpi->rtcd.variance.var16x8               = vp9_variance16x8_neon;
+    cpi->rtcd.variance.var16x16              = vp9_variance16x16_neon;
+
+    /*cpi->rtcd.variance.subpixvar4x4          = vp9_sub_pixel_variance4x4_c;*/
+    cpi->rtcd.variance.subpixvar8x8          = vp9_sub_pixel_variance8x8_neon;
+    /*cpi->rtcd.variance.subpixvar8x16         = vp9_sub_pixel_variance8x16_c;
+    cpi->rtcd.variance.subpixvar16x8         = vp9_sub_pixel_variance16x8_c;*/
+    cpi->rtcd.variance.subpixvar16x16        = vp9_sub_pixel_variance16x16_neon;
+    cpi->rtcd.variance.halfpixvar16x16_h     = vp9_variance_halfpixvar16x16_h_neon;
+    cpi->rtcd.variance.halfpixvar16x16_v     = vp9_variance_halfpixvar16x16_v_neon;
+    cpi->rtcd.variance.halfpixvar16x16_hv    = vp9_variance_halfpixvar16x16_hv_neon;
+
+    cpi->rtcd.variance.mse16x16              = vp9_mse16x16_neon;
+    /*cpi->rtcd.variance.getmbss               = vp9_get_mb_ss_c;*/
+
+    cpi->rtcd.fdct.short4x4                  = vp9_short_fdct4x4_neon;
+    cpi->rtcd.fdct.short8x4                  = vp9_short_fdct8x4_neon;
+    cpi->rtcd.fdct.fast4x4                   = vp9_short_fdct4x4_neon;
+    cpi->rtcd.fdct.fast8x4                   = vp9_short_fdct8x4_neon;
+    cpi->rtcd.fdct.walsh_short4x4            = vp9_short_walsh4x4_neon;
+
+    /*cpi->rtcd.encodemb.berr                  = vp9_block_error_c;
+    cpi->rtcd.encodemb.mberr                 = vp9_mbblock_error_c;
+    cpi->rtcd.encodemb.mbuverr               = vp9_mbuverror_c;*/
+    cpi->rtcd.encodemb.subb                  = vp9_subtract_b_neon;
+    cpi->rtcd.encodemb.submby                = vp9_subtract_mby_neon;
+    cpi->rtcd.encodemb.submbuv               = vp9_subtract_mbuv_neon;
 
     /*cpi->rtcd.quantize.quantb                = vp8_regular_quantize_b;
     cpi->rtcd.quantize.quantb_pair           = vp8_regular_quantize_b_pair;*/
@@ -122,7 +122,7 @@ void vp8_arch_arm_encoder_init(VP8_COMP *cpi) {
   if (flags & HAS_NEON)
 #endif
   {
-    vp8_yv12_copy_partial_frame_ptr = vpxyv12_copy_partial_frame_neon;
+    vp9_yv12_copy_partial_frame_ptr = vpxyv12_copy_partial_frame_neon;
   }
 #endif
 #endif
diff --git a/vp8/encoder/arm/dct_arm.c b/vp8/encoder/arm/dct_arm.c
index 97376de32306ddfce9f232f7fd7839a53a1abecc..3fd04f3830fd1878a87e8838f6bad321b9b6fe48 100644
--- a/vp8/encoder/arm/dct_arm.c
+++ b/vp8/encoder/arm/dct_arm.c
@@ -13,9 +13,9 @@
 
 #if HAVE_ARMV6
 
-void vp8_short_fdct8x4_armv6(short *input, short *output, int pitch) {
-  vp8_short_fdct4x4_armv6(input,   output,    pitch);
-  vp8_short_fdct4x4_armv6(input + 4, output + 16, pitch);
+void vp9_short_fdct8x4_armv6(short *input, short *output, int pitch) {
+  vp9_short_fdct4x4_armv6(input,   output,    pitch);
+  vp9_short_fdct4x4_armv6(input + 4, output + 16, pitch);
 }
 
 #endif /* HAVE_ARMV6 */
diff --git a/vp8/encoder/arm/dct_arm.h b/vp8/encoder/arm/dct_arm.h
index db553c4e076233e406a2a6478031c6ed6d64549b..83c446e7eb20205574d681c2a6be689bf3e48a5d 100644
--- a/vp8/encoder/arm/dct_arm.h
+++ b/vp8/encoder/arm/dct_arm.h
@@ -13,51 +13,51 @@
 #define DCT_ARM_H
 
 #if HAVE_ARMV6
-extern prototype_fdct(vp8_short_walsh4x4_armv6);
-extern prototype_fdct(vp8_short_fdct4x4_armv6);
-extern prototype_fdct(vp8_short_fdct8x4_armv6);
+extern prototype_fdct(vp9_short_walsh4x4_armv6);
+extern prototype_fdct(vp9_short_fdct4x4_armv6);
+extern prototype_fdct(vp9_short_fdct8x4_armv6);
 
 #if !CONFIG_RUNTIME_CPU_DETECT
 #undef  vp8_fdct_walsh_short4x4
-#define vp8_fdct_walsh_short4x4 vp8_short_walsh4x4_armv6
+#define vp8_fdct_walsh_short4x4 vp9_short_walsh4x4_armv6
 
 #undef  vp8_fdct_short4x4
-#define vp8_fdct_short4x4 vp8_short_fdct4x4_armv6
+#define vp8_fdct_short4x4 vp9_short_fdct4x4_armv6
 
 #undef  vp8_fdct_short8x4
-#define vp8_fdct_short8x4 vp8_short_fdct8x4_armv6
+#define vp8_fdct_short8x4 vp9_short_fdct8x4_armv6
 
 #undef  vp8_fdct_fast4x4
-#define vp8_fdct_fast4x4 vp8_short_fdct4x4_armv6
+#define vp8_fdct_fast4x4 vp9_short_fdct4x4_armv6
 
 #undef  vp8_fdct_fast8x4
-#define vp8_fdct_fast8x4 vp8_short_fdct8x4_armv6
+#define vp8_fdct_fast8x4 vp9_short_fdct8x4_armv6
 #endif
 
 #endif /* HAVE_ARMV6 */
 
 #if HAVE_ARMV7
-extern prototype_fdct(vp8_short_fdct4x4_neon);
-extern prototype_fdct(vp8_short_fdct8x4_neon);
+extern prototype_fdct(vp9_short_fdct4x4_neon);
+extern prototype_fdct(vp9_short_fdct8x4_neon);
 extern prototype_fdct(vp8_fast_fdct4x4_neon);
 extern prototype_fdct(vp8_fast_fdct8x4_neon);
-extern prototype_fdct(vp8_short_walsh4x4_neon);
+extern prototype_fdct(vp9_short_walsh4x4_neon);
 
 #if !CONFIG_RUNTIME_CPU_DETECT
 #undef  vp8_fdct_short4x4
-#define vp8_fdct_short4x4 vp8_short_fdct4x4_neon
+#define vp8_fdct_short4x4 vp9_short_fdct4x4_neon
 
 #undef  vp8_fdct_short8x4
-#define vp8_fdct_short8x4 vp8_short_fdct8x4_neon
+#define vp8_fdct_short8x4 vp9_short_fdct8x4_neon
 
 #undef  vp8_fdct_fast4x4
-#define vp8_fdct_fast4x4 vp8_short_fdct4x4_neon
+#define vp8_fdct_fast4x4 vp9_short_fdct4x4_neon
 
 #undef  vp8_fdct_fast8x4
-#define vp8_fdct_fast8x4 vp8_short_fdct8x4_neon
+#define vp8_fdct_fast8x4 vp9_short_fdct8x4_neon
 
 #undef  vp8_fdct_walsh_short4x4
-#define vp8_fdct_walsh_short4x4 vp8_short_walsh4x4_neon
+#define vp8_fdct_walsh_short4x4 vp9_short_walsh4x4_neon
 #endif
 
 #endif
diff --git a/vp8/encoder/arm/encodemb_arm.h b/vp8/encoder/arm/encodemb_arm.h
index 7bcc747fcca03132f012815a502afd0c74f5cb54..80bff79df01a402332584a72cee9fddb0be01048 100644
--- a/vp8/encoder/arm/encodemb_arm.h
+++ b/vp8/encoder/arm/encodemb_arm.h
@@ -13,50 +13,50 @@
 #define ENCODEMB_ARM_H
 
 #if HAVE_ARMV6
-extern prototype_subb(vp8_subtract_b_armv6);
-extern prototype_submby(vp8_subtract_mby_armv6);
-extern prototype_submbuv(vp8_subtract_mbuv_armv6);
+extern prototype_subb(vp9_subtract_b_armv6);
+extern prototype_submby(vp9_subtract_mby_armv6);
+extern prototype_submbuv(vp9_subtract_mbuv_armv6);
 
 #if !CONFIG_RUNTIME_CPU_DETECT
 #undef  vp8_encodemb_subb
-#define vp8_encodemb_subb vp8_subtract_b_armv6
+#define vp8_encodemb_subb vp9_subtract_b_armv6
 
 #undef  vp8_encodemb_submby
-#define vp8_encodemb_submby vp8_subtract_mby_armv6
+#define vp8_encodemb_submby vp9_subtract_mby_armv6
 
 #undef  vp8_encodemb_submbuv
-#define vp8_encodemb_submbuv vp8_subtract_mbuv_armv6
+#define vp8_encodemb_submbuv vp9_subtract_mbuv_armv6
 #endif
 
 #endif /* HAVE_ARMV6 */
 
 #if HAVE_ARMV7
-// extern prototype_berr(vp8_block_error_c);
-// extern prototype_mberr(vp8_mbblock_error_c);
-// extern prototype_mbuverr(vp8_mbuverror_c);
+// extern prototype_berr(vp9_block_error_c);
+// extern prototype_mberr(vp9_mbblock_error_c);
+// extern prototype_mbuverr(vp9_mbuverror_c);
 
-extern prototype_subb(vp8_subtract_b_neon);
-extern prototype_submby(vp8_subtract_mby_neon);
-extern prototype_submbuv(vp8_subtract_mbuv_neon);
+extern prototype_subb(vp9_subtract_b_neon);
+extern prototype_submby(vp9_subtract_mby_neon);
+extern prototype_submbuv(vp9_subtract_mbuv_neon);
 
 // #undef  vp8_encodemb_berr
-// #define vp8_encodemb_berr vp8_block_error_c
+// #define vp8_encodemb_berr vp9_block_error_c
 
 // #undef  vp8_encodemb_mberr
-// #define vp8_encodemb_mberr vp8_mbblock_error_c
+// #define vp8_encodemb_mberr vp9_mbblock_error_c
 
 // #undef  vp8_encodemb_mbuverr
-// #define vp8_encodemb_mbuverr vp8_mbuverror_c
+// #define vp8_encodemb_mbuverr vp9_mbuverror_c
 
 #if !CONFIG_RUNTIME_CPU_DETECT
 #undef  vp8_encodemb_subb
-#define vp8_encodemb_subb vp8_subtract_b_neon
+#define vp8_encodemb_subb vp9_subtract_b_neon
 
 #undef  vp8_encodemb_submby
-#define vp8_encodemb_submby vp8_subtract_mby_neon
+#define vp8_encodemb_submby vp9_subtract_mby_neon
 
 #undef  vp8_encodemb_submbuv
-#define vp8_encodemb_submbuv vp8_subtract_mbuv_neon
+#define vp8_encodemb_submbuv vp9_subtract_mbuv_neon
 #endif
 
 #endif
diff --git a/vp8/encoder/arm/variance_arm.c b/vp8/encoder/arm/variance_arm.c
index e11d7bbb30d4e514bbe66ec4a24e04b7beca791d..80696e489219438206007d14d553a56f040bab6b 100644
--- a/vp8/encoder/arm/variance_arm.c
+++ b/vp8/encoder/arm/variance_arm.c
@@ -17,7 +17,7 @@
 
 #if HAVE_ARMV6
 
-unsigned int vp8_sub_pixel_variance8x8_armv6
+unsigned int vp9_sub_pixel_variance8x8_armv6
 (
   const unsigned char  *src_ptr,
   int  src_pixels_per_line,
@@ -40,11 +40,11 @@ unsigned int vp8_sub_pixel_variance8x8_armv6
   vp8_filter_block2d_bil_second_pass_armv6(first_pass, second_pass,
                                            8, 8, 8, VFilter);
 
-  return vp8_variance8x8_armv6(second_pass, 8, dst_ptr,
+  return vp9_variance8x8_armv6(second_pass, 8, dst_ptr,
                                dst_pixels_per_line, sse);
 }
 
-unsigned int vp8_sub_pixel_variance16x16_armv6
+unsigned int vp9_sub_pixel_variance16x16_armv6
 (
   const unsigned char  *src_ptr,
   int  src_pixels_per_line,
@@ -60,13 +60,13 @@ unsigned int vp8_sub_pixel_variance16x16_armv6
   unsigned int var;
 
   if (xoffset == HALFNDX && yoffset == 0) {
-    var = vp8_variance_halfpixvar16x16_h_armv6(src_ptr, src_pixels_per_line,
+    var = vp9_variance_halfpixvar16x16_h_armv6(src_ptr, src_pixels_per_line,
                                                dst_ptr, dst_pixels_per_line, sse);
   } else if (xoffset == 0 && yoffset == HALFNDX) {
-    var = vp8_variance_halfpixvar16x16_v_armv6(src_ptr, src_pixels_per_line,
+    var = vp9_variance_halfpixvar16x16_v_armv6(src_ptr, src_pixels_per_line,
                                                dst_ptr, dst_pixels_per_line, sse);
   } else if (xoffset == HALFNDX && yoffset == HALFNDX) {
-    var = vp8_variance_halfpixvar16x16_hv_armv6(src_ptr, src_pixels_per_line,
+    var = vp9_variance_halfpixvar16x16_hv_armv6(src_ptr, src_pixels_per_line,
                                                 dst_ptr, dst_pixels_per_line, sse);
   } else {
     HFilter = vp8_bilinear_filters[xoffset];
@@ -78,7 +78,7 @@ unsigned int vp8_sub_pixel_variance16x16_armv6
     vp8_filter_block2d_bil_second_pass_armv6(first_pass, second_pass,
                                              16, 16, 16, VFilter);
 
-    var = vp8_variance16x16_armv6(second_pass, 16, dst_ptr,
+    var = vp9_variance16x16_armv6(second_pass, 16, dst_ptr,
                                   dst_pixels_per_line, sse);
   }
   return var;
@@ -89,7 +89,7 @@ unsigned int vp8_sub_pixel_variance16x16_armv6
 
 #if HAVE_ARMV7
 
-unsigned int vp8_sub_pixel_variance16x16_neon
+unsigned int vp9_sub_pixel_variance16x16_neon
 (
   const unsigned char  *src_ptr,
   int  src_pixels_per_line,
@@ -100,13 +100,13 @@ unsigned int vp8_sub_pixel_variance16x16_neon
   unsigned int *sse
 ) {
   if (xoffset == HALFNDX && yoffset == 0)
-    return vp8_variance_halfpixvar16x16_h_neon(src_ptr, src_pixels_per_line, dst_ptr, dst_pixels_per_line, sse);
+    return vp9_variance_halfpixvar16x16_h_neon(src_ptr, src_pixels_per_line, dst_ptr, dst_pixels_per_line, sse);
   else if (xoffset == 0 && yoffset == HALFNDX)
-    return vp8_variance_halfpixvar16x16_v_neon(src_ptr, src_pixels_per_line, dst_ptr, dst_pixels_per_line, sse);
+    return vp9_variance_halfpixvar16x16_v_neon(src_ptr, src_pixels_per_line, dst_ptr, dst_pixels_per_line, sse);
   else if (xoffset == HALFNDX && yoffset == HALFNDX)
-    return vp8_variance_halfpixvar16x16_hv_neon(src_ptr, src_pixels_per_line, dst_ptr, dst_pixels_per_line, sse);
+    return vp9_variance_halfpixvar16x16_hv_neon(src_ptr, src_pixels_per_line, dst_ptr, dst_pixels_per_line, sse);
   else
-    return vp8_sub_pixel_variance16x16_neon_func(src_ptr, src_pixels_per_line, xoffset, yoffset, dst_ptr, dst_pixels_per_line, sse);
+    return vp9_sub_pixel_variance16x16_neon_func(src_ptr, src_pixels_per_line, xoffset, yoffset, dst_ptr, dst_pixels_per_line, sse);
 }
 
 #endif
diff --git a/vp8/encoder/arm/variance_arm.h b/vp8/encoder/arm/variance_arm.h
index 8deb72d4a04e262f23cf416b8998d620eaebbe19..8ca402677dfb7f0754af0cf4eca90899482194bb 100644
--- a/vp8/encoder/arm/variance_arm.h
+++ b/vp8/encoder/arm/variance_arm.h
@@ -14,44 +14,44 @@
 
 #if HAVE_ARMV6
 
-extern prototype_sad(vp8_sad16x16_armv6);
-extern prototype_variance(vp8_variance16x16_armv6);
-extern prototype_variance(vp8_variance8x8_armv6);
-extern prototype_subpixvariance(vp8_sub_pixel_variance16x16_armv6);
-extern prototype_subpixvariance(vp8_sub_pixel_variance8x8_armv6);
-extern prototype_variance(vp8_variance_halfpixvar16x16_h_armv6);
-extern prototype_variance(vp8_variance_halfpixvar16x16_v_armv6);
-extern prototype_variance(vp8_variance_halfpixvar16x16_hv_armv6);
-extern prototype_variance(vp8_mse16x16_armv6);
+extern prototype_sad(vp9_sad16x16_armv6);
+extern prototype_variance(vp9_variance16x16_armv6);
+extern prototype_variance(vp9_variance8x8_armv6);
+extern prototype_subpixvariance(vp9_sub_pixel_variance16x16_armv6);
+extern prototype_subpixvariance(vp9_sub_pixel_variance8x8_armv6);
+extern prototype_variance(vp9_variance_halfpixvar16x16_h_armv6);
+extern prototype_variance(vp9_variance_halfpixvar16x16_v_armv6);
+extern prototype_variance(vp9_variance_halfpixvar16x16_hv_armv6);
+extern prototype_variance(vp9_mse16x16_armv6);
 
 #if !CONFIG_RUNTIME_CPU_DETECT
 
 #undef  vp8_variance_sad16x16
-#define vp8_variance_sad16x16 vp8_sad16x16_armv6
+#define vp8_variance_sad16x16 vp9_sad16x16_armv6
 
 #undef  vp8_variance_subpixvar16x16
-#define vp8_variance_subpixvar16x16 vp8_sub_pixel_variance16x16_armv6
+#define vp8_variance_subpixvar16x16 vp9_sub_pixel_variance16x16_armv6
 
 #undef  vp8_variance_subpixvar8x8
-#define vp8_variance_subpixvar8x8 vp8_sub_pixel_variance8x8_armv6
+#define vp8_variance_subpixvar8x8 vp9_sub_pixel_variance8x8_armv6
 
 #undef  vp8_variance_var16x16
-#define vp8_variance_var16x16 vp8_variance16x16_armv6
+#define vp8_variance_var16x16 vp9_variance16x16_armv6
 
 #undef  vp8_variance_mse16x16
-#define vp8_variance_mse16x16 vp8_mse16x16_armv6
+#define vp8_variance_mse16x16 vp9_mse16x16_armv6
 
 #undef  vp8_variance_var8x8
-#define vp8_variance_var8x8 vp8_variance8x8_armv6
+#define vp8_variance_var8x8 vp9_variance8x8_armv6
 
-#undef  vp8_variance_halfpixvar16x16_h
-#define vp8_variance_halfpixvar16x16_h vp8_variance_halfpixvar16x16_h_armv6
+#undef  vp9_variance_halfpixvar16x16_h
+#define vp9_variance_halfpixvar16x16_h vp9_variance_halfpixvar16x16_h_armv6
 
-#undef  vp8_variance_halfpixvar16x16_v
-#define vp8_variance_halfpixvar16x16_v vp8_variance_halfpixvar16x16_v_armv6
+#undef  vp9_variance_halfpixvar16x16_v
+#define vp9_variance_halfpixvar16x16_v vp9_variance_halfpixvar16x16_v_armv6
 
-#undef  vp8_variance_halfpixvar16x16_hv
-#define vp8_variance_halfpixvar16x16_hv vp8_variance_halfpixvar16x16_hv_armv6
+#undef  vp9_variance_halfpixvar16x16_hv
+#define vp9_variance_halfpixvar16x16_hv vp9_variance_halfpixvar16x16_hv_armv6
 
 #endif /* !CONFIG_RUNTIME_CPU_DETECT */
 
@@ -59,91 +59,91 @@ extern prototype_variance(vp8_mse16x16_armv6);
 
 
 #if HAVE_ARMV7
-extern prototype_sad(vp8_sad4x4_neon);
-extern prototype_sad(vp8_sad8x8_neon);
-extern prototype_sad(vp8_sad8x16_neon);
-extern prototype_sad(vp8_sad16x8_neon);
-extern prototype_sad(vp8_sad16x16_neon);
-
-// extern prototype_variance(vp8_variance4x4_c);
-extern prototype_variance(vp8_variance8x8_neon);
-extern prototype_variance(vp8_variance8x16_neon);
-extern prototype_variance(vp8_variance16x8_neon);
-extern prototype_variance(vp8_variance16x16_neon);
-
-// extern prototype_subpixvariance(vp8_sub_pixel_variance4x4_c);
-extern prototype_subpixvariance(vp8_sub_pixel_variance8x8_neon);
-// extern prototype_subpixvariance(vp8_sub_pixel_variance8x16_c);
-// extern prototype_subpixvariance(vp8_sub_pixel_variance16x8_c);
-extern prototype_subpixvariance(vp8_sub_pixel_variance16x16_neon);
-extern prototype_subpixvariance(vp8_sub_pixel_variance16x16_neon_func);
-extern prototype_variance(vp8_variance_halfpixvar16x16_h_neon);
-extern prototype_variance(vp8_variance_halfpixvar16x16_v_neon);
-extern prototype_variance(vp8_variance_halfpixvar16x16_hv_neon);
-
-// extern prototype_getmbss(vp8_get_mb_ss_c);
-extern prototype_variance(vp8_mse16x16_neon);
+extern prototype_sad(vp9_sad4x4_neon);
+extern prototype_sad(vp9_sad8x8_neon);
+extern prototype_sad(vp9_sad8x16_neon);
+extern prototype_sad(vp9_sad16x8_neon);
+extern prototype_sad(vp9_sad16x16_neon);
+
+// extern prototype_variance(vp9_variance4x4_c);
+extern prototype_variance(vp9_variance8x8_neon);
+extern prototype_variance(vp9_variance8x16_neon);
+extern prototype_variance(vp9_variance16x8_neon);
+extern prototype_variance(vp9_variance16x16_neon);
+
+// extern prototype_subpixvariance(vp9_sub_pixel_variance4x4_c);
+extern prototype_subpixvariance(vp9_sub_pixel_variance8x8_neon);
+// extern prototype_subpixvariance(vp9_sub_pixel_variance8x16_c);
+// extern prototype_subpixvariance(vp9_sub_pixel_variance16x8_c);
+extern prototype_subpixvariance(vp9_sub_pixel_variance16x16_neon);
+extern prototype_subpixvariance(vp9_sub_pixel_variance16x16_neon_func);
+extern prototype_variance(vp9_variance_halfpixvar16x16_h_neon);
+extern prototype_variance(vp9_variance_halfpixvar16x16_v_neon);
+extern prototype_variance(vp9_variance_halfpixvar16x16_hv_neon);
+
+// extern prototype_getmbss(vp9_get_mb_ss_c);
+extern prototype_variance(vp9_mse16x16_neon);
 
 #if !CONFIG_RUNTIME_CPU_DETECT
 #undef  vp8_variance_sad4x4
-#define vp8_variance_sad4x4 vp8_sad4x4_neon
+#define vp8_variance_sad4x4 vp9_sad4x4_neon
 
 #undef  vp8_variance_sad8x8
-#define vp8_variance_sad8x8 vp8_sad8x8_neon
+#define vp8_variance_sad8x8 vp9_sad8x8_neon
 
 #undef  vp8_variance_sad8x16
-#define vp8_variance_sad8x16 vp8_sad8x16_neon
+#define vp8_variance_sad8x16 vp9_sad8x16_neon
 
 #undef  vp8_variance_sad16x8
-#define vp8_variance_sad16x8 vp8_sad16x8_neon
+#define vp8_variance_sad16x8 vp9_sad16x8_neon
 
 #undef  vp8_variance_sad16x16
-#define vp8_variance_sad16x16 vp8_sad16x16_neon
+#define vp8_variance_sad16x16 vp9_sad16x16_neon
 
 // #undef  vp8_variance_var4x4
-// #define vp8_variance_var4x4 vp8_variance4x4_c
+// #define vp8_variance_var4x4 vp9_variance4x4_c
 
 #undef  vp8_variance_var8x8
-#define vp8_variance_var8x8 vp8_variance8x8_neon
+#define vp8_variance_var8x8 vp9_variance8x8_neon
 
 #undef  vp8_variance_var8x16
-#define vp8_variance_var8x16 vp8_variance8x16_neon
+#define vp8_variance_var8x16 vp9_variance8x16_neon
 
 #undef  vp8_variance_var16x8
-#define vp8_variance_var16x8 vp8_variance16x8_neon
+#define vp8_variance_var16x8 vp9_variance16x8_neon
 
 #undef  vp8_variance_var16x16
-#define vp8_variance_var16x16 vp8_variance16x16_neon
+#define vp8_variance_var16x16 vp9_variance16x16_neon
 
 // #undef  vp8_variance_subpixvar4x4
-// #define vp8_variance_subpixvar4x4 vp8_sub_pixel_variance4x4_c
+// #define vp8_variance_subpixvar4x4 vp9_sub_pixel_variance4x4_c
 
 #undef  vp8_variance_subpixvar8x8
-#define vp8_variance_subpixvar8x8 vp8_sub_pixel_variance8x8_neon
+#define vp8_variance_subpixvar8x8 vp9_sub_pixel_variance8x8_neon
 
 // #undef  vp8_variance_subpixvar8x16
-// #define vp8_variance_subpixvar8x16 vp8_sub_pixel_variance8x16_c
+// #define vp8_variance_subpixvar8x16 vp9_sub_pixel_variance8x16_c
 
 // #undef  vp8_variance_subpixvar16x8
-// #define vp8_variance_subpixvar16x8 vp8_sub_pixel_variance16x8_c
+// #define vp8_variance_subpixvar16x8 vp9_sub_pixel_variance16x8_c
 
 #undef  vp8_variance_subpixvar16x16
-#define vp8_variance_subpixvar16x16 vp8_sub_pixel_variance16x16_neon
+#define vp8_variance_subpixvar16x16 vp9_sub_pixel_variance16x16_neon
 
-#undef  vp8_variance_halfpixvar16x16_h
-#define vp8_variance_halfpixvar16x16_h vp8_variance_halfpixvar16x16_h_neon
+#undef  vp9_variance_halfpixvar16x16_h
+#define vp9_variance_halfpixvar16x16_h vp9_variance_halfpixvar16x16_h_neon
 
-#undef  vp8_variance_halfpixvar16x16_v
-#define vp8_variance_halfpixvar16x16_v vp8_variance_halfpixvar16x16_v_neon
+#undef  vp9_variance_halfpixvar16x16_v
+#define vp9_variance_halfpixvar16x16_v vp9_variance_halfpixvar16x16_v_neon
 
-#undef  vp8_variance_halfpixvar16x16_hv
-#define vp8_variance_halfpixvar16x16_hv vp8_variance_halfpixvar16x16_hv_neon
+#undef  vp9_variance_halfpixvar16x16_hv
+#define vp9_variance_halfpixvar16x16_hv vp9_variance_halfpixvar16x16_hv_neon
 
 // #undef  vp8_variance_getmbss
-// #define vp8_variance_getmbss vp8_get_mb_ss_c
+// #define vp8_variance_getmbss vp9_get_mb_ss_c
 
 #undef  vp8_variance_mse16x16
-#define vp8_variance_mse16x16 vp8_mse16x16_neon
+#define vp8_variance_mse16x16 vp9_mse16x16_neon
 
 #endif
 
diff --git a/vp8/encoder/bitstream.c b/vp8/encoder/bitstream.c
index 13ac1dd005bbd1f185e2c19658cc4af6e863b8a7..155453543ea3513ac371d64d0523e78e24f2fc9d 100644
--- a/vp8/encoder/bitstream.c
+++ b/vp8/encoder/bitstream.c
@@ -83,7 +83,7 @@ static int update_bits[255];
 static void compute_update_table() {
   int i;
   for (i = 0; i < 255; i++)
-    update_bits[i] = vp8_count_term_subexp(i, SUBEXP_PARAM, 255);
+    update_bits[i] = vp9_count_term_subexp(i, SUBEXP_PARAM, 255);
 }
 
 static int split_index(int i, int n, int modulus) {
@@ -109,7 +109,7 @@ static int remap_prob(int v, int m) {
 static void write_prob_diff_update(vp8_writer *const bc,
                                    vp8_prob newp, vp8_prob oldp) {
   int delp = remap_prob(newp, oldp);
-  vp8_encode_term_subexp(bc, delp, SUBEXP_PARAM, 255);
+  vp9_encode_term_subexp(bc, delp, SUBEXP_PARAM, 255);
 }
 
 static int prob_diff_update_cost(vp8_prob newp, vp8_prob oldp) {
@@ -618,12 +618,12 @@ static void write_nmv(vp8_writer *bc, const MV *mv, const int_mv *ref,
   e.row = mv->row - ref->as_mv.row;
   e.col = mv->col - ref->as_mv.col;
 
-  vp8_encode_nmv(bc, &e, &ref->as_mv, nmvc);
-  vp8_encode_nmv_fp(bc, &e, &ref->as_mv, nmvc, usehp);
+  vp9_encode_nmv(bc, &e, &ref->as_mv, nmvc);
+  vp9_encode_nmv_fp(bc, &e, &ref->as_mv, nmvc, usehp);
 }
 
 #if CONFIG_NEW_MVREF
-static int vp8_cost_mv_ref_id(vp8_prob * ref_id_probs, int mv_ref_id) {
+static int vp9_cost_mv_ref_id(vp8_prob * ref_id_probs, int mv_ref_id) {
   int cost;
 
   // Encode the index for the MV reference.
@@ -698,8 +698,8 @@ static unsigned int pick_best_mv_ref(MACROBLOCK *x,
   MACROBLOCKD *xd = &x->e_mbd;
   int max_mv = MV_MAX;
 
-  cost = vp8_cost_mv_ref_id(xd->mb_mv_ref_id_probs[ref_frame], 0) +
-         vp8_mv_bit_cost(&target_mv,
+  cost = vp9_cost_mv_ref_id(xd->mb_mv_ref_id_probs[ref_frame], 0) +
+         vp9_mv_bit_cost(&target_mv,
                          &mv_ref_list[0],
                          XMVCOST, 96,
                          xd->allow_high_precision_mv);
@@ -722,8 +722,8 @@ static unsigned int pick_best_mv_ref(MACROBLOCK *x,
       continue;
     }
 
-    cost2 = vp8_cost_mv_ref_id(xd->mb_mv_ref_id_probs[ref_frame], i) +
-            vp8_mv_bit_cost(&target_mv,
+    cost2 = vp9_cost_mv_ref_id(xd->mb_mv_ref_id_probs[ref_frame], i) +
+            vp9_mv_bit_cost(&target_mv,
                             &mv_ref_list[i],
                             XMVCOST, 96,
                             xd->allow_high_precision_mv);
@@ -1820,13 +1820,13 @@ static void decide_kf_ymode_entropy(VP8_COMP *cpi) {
   int i, j;
 
   for (i = 0; i < 8; i++) {
-    vp8_cost_tokens(mode_cost, cpi->common.kf_ymode_prob[i], vp8_kf_ymode_tree);
+    vp9_cost_tokens(mode_cost, cpi->common.kf_ymode_prob[i], vp8_kf_ymode_tree);
     cost = 0;
     for (j = 0; j < VP8_YMODES; j++) {
       cost += mode_cost[j] * cpi->ymode_count[j];
     }
 #if CONFIG_SUPERBLOCKS
-    vp8_cost_tokens(mode_cost, cpi->common.sb_kf_ymode_prob[i],
+    vp9_cost_tokens(mode_cost, cpi->common.sb_kf_ymode_prob[i],
                     vp8_sb_ymode_tree);
     for (j = 0; j < VP8_I32X32_MODES; j++) {
       cost += mode_cost[j] * cpi->sb_ymode_count[j];
@@ -1860,7 +1860,7 @@ static void segment_reference_frames(VP8_COMP *cpi) {
   }
 }
 
-void vp8_pack_bitstream(VP8_COMP *cpi, unsigned char *dest, unsigned long *size) {
+void vp9_pack_bitstream(VP8_COMP *cpi, unsigned char *dest, unsigned long *size) {
   int i, j;
   VP8_HEADER oh;
   VP8_COMMON *const pc = &cpi->common;
@@ -1883,7 +1883,7 @@ void vp8_pack_bitstream(VP8_COMP *cpi, unsigned char *dest, unsigned long *size)
 
   compute_update_table();
 
-  // vp8_kf_default_bmode_probs() is called in vp8_setup_key_frame() once for each
+  // vp8_kf_default_bmode_probs() is called in vp9_setup_key_frame() once for each
   // K frame before encode frame. pc->kf_bmode_prob doesn't get changed anywhere
   // else. No need to call it again here. --yw
   // vp8_kf_default_bmode_probs( pc->kf_bmode_prob);
@@ -1908,14 +1908,14 @@ void vp8_pack_bitstream(VP8_COMP *cpi, unsigned char *dest, unsigned long *size)
     extra_bytes_packed = 7;
     cx_data += extra_bytes_packed;
 
-    vp8_start_encode(&header_bc, cx_data);
+    vp9_start_encode(&header_bc, cx_data);
 
     // signal clr type
     vp8_write_bit(&header_bc, pc->clr_type);
     vp8_write_bit(&header_bc, pc->clamp_type);
 
   } else {
-    vp8_start_encode(&header_bc, cx_data);
+    vp9_start_encode(&header_bc, cx_data);
   }
 
   // Signal whether or not Segmentation is enabled
@@ -2272,10 +2272,10 @@ void vp8_pack_bitstream(VP8_COMP *cpi, unsigned char *dest, unsigned long *size)
     vpx_memset(xd->mb_mv_ref_id_probs, 192, sizeof(xd->mb_mv_ref_id_probs));
 #endif
 
-    vp8_write_nmvprobs(cpi, xd->allow_high_precision_mv, &header_bc);
+    vp9_write_nmvprobs(cpi, xd->allow_high_precision_mv, &header_bc);
   }
 
-  vp8_stop_encode(&header_bc);
+  vp9_stop_encode(&header_bc);
 
   oh.first_partition_length_in_bytes = header_bc.pos;
 
@@ -2292,7 +2292,7 @@ void vp8_pack_bitstream(VP8_COMP *cpi, unsigned char *dest, unsigned long *size)
   }
 
   *size = VP8_HEADER_SIZE + extra_bytes_packed + header_bc.pos;
-  vp8_start_encode(&residual_bc, cx_data + header_bc.pos);
+  vp9_start_encode(&residual_bc, cx_data + header_bc.pos);
 
   if (pc->frame_type == KEY_FRAME) {
     decide_kf_ymode_entropy(cpi);
@@ -2303,7 +2303,7 @@ void vp8_pack_bitstream(VP8_COMP *cpi, unsigned char *dest, unsigned long *size)
   }
 
 
-  vp8_stop_encode(&residual_bc);
+  vp9_stop_encode(&residual_bc);
 
   *size += residual_bc.pos;
 
diff --git a/vp8/encoder/block.h b/vp8/encoder/block.h
index 9a6483adb3022e169bae7d85249264af47f12234..1e70f16170e514f49d190fa54425c30d8d5092f0 100644
--- a/vp8/encoder/block.h
+++ b/vp8/encoder/block.h
@@ -166,13 +166,13 @@ typedef struct macroblock {
   PICK_MODE_CONTEXT sb_context[4];
 #endif
 
-  void (*vp8_short_fdct4x4)(short *input, short *output, int pitch);
-  void (*vp8_short_fdct8x4)(short *input, short *output, int pitch);
+  void (*vp9_short_fdct4x4)(short *input, short *output, int pitch);
+  void (*vp9_short_fdct8x4)(short *input, short *output, int pitch);
   void (*short_walsh4x4)(short *input, short *output, int pitch);
   void (*quantize_b_4x4)(BLOCK *b, BLOCKD *d);
   void (*quantize_b_4x4_pair)(BLOCK *b1, BLOCK *b2, BLOCKD *d0, BLOCKD *d1);
-  void (*vp8_short_fdct8x8)(short *input, short *output, int pitch);
-  void (*vp8_short_fdct16x16)(short *input, short *output, int pitch);
+  void (*vp9_short_fdct8x8)(short *input, short *output, int pitch);
+  void (*vp9_short_fdct16x16)(short *input, short *output, int pitch);
   void (*short_fhaar2x2)(short *input, short *output, int pitch);
   void (*quantize_b_16x16)(BLOCK *b, BLOCKD *d);
   void (*quantize_b_8x8)(BLOCK *b, BLOCKD *d);
diff --git a/vp8/encoder/boolhuff.c b/vp8/encoder/boolhuff.c
index 9206e55967c430b3bdd1823b7366bc97343b5cb2..ef03d8b143f7b70200ee628abe8087a95bff9563 100644
--- a/vp8/encoder/boolhuff.c
+++ b/vp8/encoder/boolhuff.c
@@ -39,7 +39,7 @@ const unsigned int vp8_prob_cost[256] = {
   22,   21,   19,   18,   16,   15,   13,   12,   10,    9,    7,    6,    4,    3,    1,   1
 };
 
-void vp8_start_encode(BOOL_CODER *br, unsigned char *source) {
+void vp9_start_encode(BOOL_CODER *br, unsigned char *source) {
 
   br->lowvalue = 0;
   br->range    = 255;
@@ -49,7 +49,7 @@ void vp8_start_encode(BOOL_CODER *br, unsigned char *source) {
   br->pos      = 0;
 }
 
-void vp8_stop_encode(BOOL_CODER *br) {
+void vp9_stop_encode(BOOL_CODER *br) {
   int i;
 
   for (i = 0; i < 32; i++)
@@ -57,7 +57,7 @@ void vp8_stop_encode(BOOL_CODER *br) {
 }
 
 
-void vp8_encode_value(BOOL_CODER *br, int data, int bits) {
+void vp9_encode_value(BOOL_CODER *br, int data, int bits) {
   int bit;
 
   for (bit = bits - 1; bit >= 0; bit--)
@@ -81,20 +81,20 @@ static int get_unsigned_bits(unsigned num_values) {
   return cat;
 }
 
-void vp8_encode_uniform(BOOL_CODER *br, int v, int n) {
+void vp9_encode_uniform(BOOL_CODER *br, int v, int n) {
   int l = get_unsigned_bits(n);
   int m;
   if (l == 0) return;
   m = (1 << l) - n;
   if (v < m)
-    vp8_encode_value(br, v, l - 1);
+    vp9_encode_value(br, v, l - 1);
   else {
-    vp8_encode_value(br, m + ((v - m) >> 1), l - 1);
-    vp8_encode_value(br, (v - m) & 1, 1);
+    vp9_encode_value(br, m + ((v - m) >> 1), l - 1);
+    vp9_encode_value(br, (v - m) & 1, 1);
   }
 }
 
-int vp8_count_uniform(int v, int n) {
+int vp9_count_uniform(int v, int n) {
   int l = get_unsigned_bits(n);
   int m;
   if (l == 0) return 0;
@@ -105,30 +105,30 @@ int vp8_count_uniform(int v, int n) {
     return l;
 }
 
-void vp8_encode_term_subexp(BOOL_CODER *br, int word, int k, int num_syms) {
+void vp9_encode_term_subexp(BOOL_CODER *br, int word, int k, int num_syms) {
   int i = 0;
   int mk = 0;
   while (1) {
     int b = (i ? k + i - 1 : k);
     int a = (1 << b);
     if (num_syms <= mk + 3 * a) {
-      vp8_encode_uniform(br, word - mk, num_syms - mk);
+      vp9_encode_uniform(br, word - mk, num_syms - mk);
       break;
     } else {
       int t = (word >= mk + a);
-      vp8_encode_value(br, t, 1);
+      vp9_encode_value(br, t, 1);
       if (t) {
         i = i + 1;
         mk += a;
       } else {
-        vp8_encode_value(br, word - mk, b);
+        vp9_encode_value(br, word - mk, b);
         break;
       }
     }
   }
 }
 
-int vp8_count_term_subexp(int word, int k, int num_syms) {
+int vp9_count_term_subexp(int word, int k, int num_syms) {
   int count = 0;
   int i = 0;
   int mk = 0;
@@ -136,7 +136,7 @@ int vp8_count_term_subexp(int word, int k, int num_syms) {
     int b = (i ? k + i - 1 : k);
     int a = (1 << b);
     if (num_syms <= mk + 3 * a) {
-      count += vp8_count_uniform(word - mk, num_syms - mk);
+      count += vp9_count_uniform(word - mk, num_syms - mk);
       break;
     } else {
       int t = (word >= mk + a);
diff --git a/vp8/encoder/boolhuff.h b/vp8/encoder/boolhuff.h
index 6c86509c6cc87cf110648a276b16657496721863..8516aa219d56953b89f2ca5aa6480a6e516308f3 100644
--- a/vp8/encoder/boolhuff.h
+++ b/vp8/encoder/boolhuff.h
@@ -34,16 +34,16 @@ typedef struct {
   unsigned long bit_counter;
 } BOOL_CODER;
 
-extern void vp8_start_encode(BOOL_CODER *bc, unsigned char *buffer);
+extern void vp9_start_encode(BOOL_CODER *bc, unsigned char *buffer);
 
-extern void vp8_encode_value(BOOL_CODER *br, int data, int bits);
-extern void vp8_stop_encode(BOOL_CODER *bc);
+extern void vp9_encode_value(BOOL_CODER *br, int data, int bits);
+extern void vp9_stop_encode(BOOL_CODER *bc);
 extern const unsigned int vp8_prob_cost[256];
 
-extern void vp8_encode_uniform(BOOL_CODER *bc, int v, int n);
-extern void vp8_encode_term_subexp(BOOL_CODER *bc, int v, int k, int n);
-extern int vp8_count_uniform(int v, int n);
-extern int vp8_count_term_subexp(int v, int k, int n);
+extern void vp9_encode_uniform(BOOL_CODER *bc, int v, int n);
+extern void vp9_encode_term_subexp(BOOL_CODER *bc, int v, int k, int n);
+extern int vp9_count_uniform(int v, int n);
+extern int vp9_count_term_subexp(int v, int k, int n);
 extern int vp9_recenter_nonneg(int v, int m);
 
 DECLARE_ALIGNED(16, extern const unsigned char, vp8_norm[256]);
diff --git a/vp8/encoder/dct.c b/vp8/encoder/dct.c
index fb23b0d102a4536be7b2eaedb21b13c6edcc332b..4a1d338f71cf472dc10e54a18519f8cd899b4f42 100644
--- a/vp8/encoder/dct.c
+++ b/vp8/encoder/dct.c
@@ -281,7 +281,7 @@ static const int xC7S1 =  3196;
 #define IN_SHIFT (FINAL_SHIFT+1)
 
 
-void vp8_short_fdct8x8_c(short *InputData, short *OutputData, int pitch) {
+void vp9_short_fdct8x8_c(short *InputData, short *OutputData, int pitch) {
   int loop;
   int short_pitch = pitch >> 1;
   int is07, is12, is34, is56;
@@ -503,7 +503,7 @@ void vp8_short_fdct8x8_c(short *InputData, short *OutputData, int pitch) {
   }
 }
 
-void vp8_short_fhaar2x2_c(short *input, short *output, int pitch) { // pitch = 8
+void vp9_short_fhaar2x2_c(short *input, short *output, int pitch) { // pitch = 8
   /* [1 1; 1 -1] orthogonal transform */
   /* use position: 0,1, 4, 8 */
   int i;
@@ -523,12 +523,12 @@ void vp8_short_fhaar2x2_c(short *input, short *output, int pitch) { // pitch = 8
 /* For test */
 #define TEST_INT 1
 #if TEST_INT
-#define vp8_fht_int_c vp8_fht_c
+#define vp9_fht_int_c vp9_fht_c
 #else
-#define vp8_fht_float_c vp8_fht_c
+#define vp9_fht_float_c vp9_fht_c
 #endif
 
-void vp8_fht_float_c(const int16_t *input, int pitch, int16_t *output,
+void vp9_fht_float_c(const int16_t *input, int pitch, int16_t *output,
                TX_TYPE tx_type, int tx_dim) {
   vp8_clear_system_state();  // Make it simd safe : __asm emms;
   {
@@ -650,7 +650,7 @@ void vp8_fht_float_c(const int16_t *input, int pitch, int16_t *output,
 #define VERTICAL_ROUNDING ((1 << (VERTICAL_SHIFT - 1)) - 1)
 #define HORIZONTAL_SHIFT 16
 #define HORIZONTAL_ROUNDING ((1 << (HORIZONTAL_SHIFT - 1)) - 1)
-void vp8_fht_int_c(const int16_t *input, int pitch, int16_t *output,
+void vp9_fht_int_c(const int16_t *input, int pitch, int16_t *output,
                    TX_TYPE tx_type, int tx_dim) {
   int i, j, k;
   int16_t imbuf[256];
@@ -728,7 +728,7 @@ void vp8_fht_int_c(const int16_t *input, int pitch, int16_t *output,
   }
 }
 
-void vp8_short_fdct4x4_c(short *input, short *output, int pitch) {
+void vp9_short_fdct4x4_c(short *input, short *output, int pitch) {
   int i;
   int a1, b1, c1, d1;
   short *ip = input;
@@ -769,13 +769,13 @@ void vp8_short_fdct4x4_c(short *input, short *output, int pitch) {
   }
 }
 
-void vp8_short_fdct8x4_c(short *input, short *output, int pitch)
+void vp9_short_fdct8x4_c(short *input, short *output, int pitch)
 {
-    vp8_short_fdct4x4_c(input,   output,    pitch);
-    vp8_short_fdct4x4_c(input + 4, output + 16, pitch);
+    vp9_short_fdct4x4_c(input,   output,    pitch);
+    vp9_short_fdct4x4_c(input + 4, output + 16, pitch);
 }
 
-void vp8_short_walsh4x4_c(short *input, short *output, int pitch) {
+void vp9_short_walsh4x4_c(short *input, short *output, int pitch) {
   int i;
   int a1, b1, c1, d1;
   short *ip = input;
@@ -816,7 +816,7 @@ void vp8_short_walsh4x4_c(short *input, short *output, int pitch) {
 }
 
 #if CONFIG_LOSSLESS
-void vp8_short_walsh4x4_lossless_c(short *input, short *output, int pitch) {
+void vp9_short_walsh4x4_lossless_c(short *input, short *output, int pitch) {
   int i;
   int a1, b1, c1, d1;
   short *ip = input;
@@ -856,7 +856,7 @@ void vp8_short_walsh4x4_lossless_c(short *input, short *output, int pitch) {
   }
 }
 
-void vp8_short_walsh4x4_x8_c(short *input, short *output, int pitch) {
+void vp9_short_walsh4x4_x8_c(short *input, short *output, int pitch) {
   int i;
   int a1, b1, c1, d1;
   short *ip = input;
@@ -896,9 +896,9 @@ void vp8_short_walsh4x4_x8_c(short *input, short *output, int pitch) {
   }
 }
 
-void vp8_short_walsh8x4_x8_c(short *input, short *output, int pitch) {
-  vp8_short_walsh4x4_x8_c(input,   output,    pitch);
-  vp8_short_walsh4x4_x8_c(input + 4, output + 16, pitch);
+void vp9_short_walsh8x4_x8_c(short *input, short *output, int pitch) {
+  vp9_short_walsh4x4_x8_c(input,   output,    pitch);
+  vp9_short_walsh4x4_x8_c(input + 4, output + 16, pitch);
 }
 #endif
 
@@ -1078,7 +1078,7 @@ static void dct16x16_1d(double input[16], double output[16]) {
   vp8_clear_system_state(); // Make it simd safe : __asm emms;
 }
 
-void vp8_short_fdct16x16_c(short *input, short *out, int pitch) {
+void vp9_short_fdct16x16_c(short *input, short *out, int pitch) {
   vp8_clear_system_state(); // Make it simd safe : __asm emms;
   {
     int shortpitch = pitch >> 1;
diff --git a/vp8/encoder/encodeframe.c b/vp8/encoder/encodeframe.c
index cae689cf5a5a9a52375b7ace83761765f31350c6..a3a97e8967729c8fcffb2173a01082716309cba4 100644
--- a/vp8/encoder/encodeframe.c
+++ b/vp8/encoder/encodeframe.c
@@ -54,29 +54,29 @@ int enc_debug = 0;
 int mb_row_debug, mb_col_debug;
 #endif
 
-extern void vp8cx_initialize_me_consts(VP8_COMP *cpi, int QIndex);
-extern void vp8_auto_select_speed(VP8_COMP *cpi);
+extern void vp9cx_initialize_me_consts(VP8_COMP *cpi, int QIndex);
+extern void vp9_auto_select_speed(VP8_COMP *cpi);
 extern void vp8cx_init_mbrthread_data(VP8_COMP *cpi,
                                       MACROBLOCK *x,
                                       MB_ROW_COMP *mbr_ei,
                                       int mb_row,
                                       int count);
-int64_t vp8_rd_pick_inter_mode_sb(VP8_COMP *cpi, MACROBLOCK *x,
+int64_t vp9_rd_pick_inter_mode_sb(VP8_COMP *cpi, MACROBLOCK *x,
                               int recon_yoffset, int recon_uvoffset,
                               int *returnrate, int *returndistortion);
-extern void vp8cx_pick_mode_inter_macroblock(VP8_COMP *cpi, MACROBLOCK *x,
+extern void vp9cx_pick_mode_inter_macroblock(VP8_COMP *cpi, MACROBLOCK *x,
                                             int recon_yoffset,
                                             int recon_uvoffset, int *r, int *d);
-void vp8_build_block_offsets(MACROBLOCK *x);
-void vp8_setup_block_ptrs(MACROBLOCK *x);
-void vp8cx_encode_inter_macroblock(VP8_COMP *cpi, MACROBLOCK *x, TOKENEXTRA **t,
+void vp9_build_block_offsets(MACROBLOCK *x);
+void vp9_setup_block_ptrs(MACROBLOCK *x);
+void vp9cx_encode_inter_macroblock(VP8_COMP *cpi, MACROBLOCK *x, TOKENEXTRA **t,
                                    int recon_yoffset, int recon_uvoffset,
                                    int output_enabled);
-void vp8cx_encode_inter_superblock(VP8_COMP *cpi, MACROBLOCK *x, TOKENEXTRA **t,
+void vp9cx_encode_inter_superblock(VP8_COMP *cpi, MACROBLOCK *x, TOKENEXTRA **t,
                                    int recon_yoffset, int recon_uvoffset, int mb_col, int mb_row);
-void vp8cx_encode_intra_macro_block(VP8_COMP *cpi, MACROBLOCK *x,
+void vp9cx_encode_intra_macro_block(VP8_COMP *cpi, MACROBLOCK *x,
                                     TOKENEXTRA **t, int output_enabled);
-void vp8cx_encode_intra_super_block(VP8_COMP *cpi,
+void vp9cx_encode_intra_super_block(VP8_COMP *cpi,
                                     MACROBLOCK *x,
                                     TOKENEXTRA **t, int mb_col);
 static void adjust_act_zbin(VP8_COMP *cpi, MACROBLOCK *x);
@@ -96,7 +96,7 @@ unsigned int b_modes[B_MODE_COUNT];
 /* activity_avg must be positive, or flat regions could get a zero weight
  *  (infinite lambda), which confounds analysis.
  * This also avoids the need for divide by zero checks in
- *  vp8_activity_masking().
+ *  vp9_activity_masking().
  */
 #define VP8_ACTIVITY_AVG_MIN (64)
 
@@ -121,7 +121,7 @@ static unsigned int tt_activity_measure(VP8_COMP *cpi, MACROBLOCK *x) {
    *  lambda using a non-linear combination (e.g., the smallest, or second
    *  smallest, etc.).
    */
-  act = vp8_variance16x16(x->src.y_buffer, x->src.y_stride, VP8_VAR_OFFS, 0,
+  act = vp9_variance16x16(x->src.y_buffer, x->src.y_stride, VP8_VAR_OFFS, 0,
                           &sse);
   act = act << 4;
 
@@ -135,7 +135,7 @@ static unsigned int tt_activity_measure(VP8_COMP *cpi, MACROBLOCK *x) {
 // Stub for alternative experimental activity measures.
 static unsigned int alt_activity_measure(VP8_COMP *cpi,
                                          MACROBLOCK *x, int use_dc_pred) {
-  return vp8_encode_intra(cpi, x, use_dc_pred);
+  return vp9_encode_intra(cpi, x, use_dc_pred);
 }
 
 
@@ -347,7 +347,7 @@ static void build_activity_map(VP8_COMP *cpi) {
 }
 
 // Macroblock activity masking
-void vp8_activity_masking(VP8_COMP *cpi, MACROBLOCK *x) {
+void vp9_activity_masking(VP8_COMP *cpi, MACROBLOCK *x) {
 #if USE_ACT_INDEX
   x->rdmult += *(x->mb_activity_ptr) * (x->rdmult >> 2);
   x->errorperbit = x->rdmult * 100 / (110 * x->rddiv);
@@ -593,7 +593,7 @@ static void pick_mb_modes(VP8_COMP *cpi,
     x->rdmult = cpi->RDMULT;
 
     if (cpi->oxcf.tuning == VP8_TUNE_SSIM)
-      vp8_activity_masking(cpi, x);
+      vp9_activity_masking(cpi, x);
 
     // Is segmentation enabled
     if (xd->segmentation_enabled) {
@@ -605,7 +605,7 @@ static void pick_mb_modes(VP8_COMP *cpi,
       if (mbmi->segment_id > 3)
         mbmi->segment_id = 0;
 
-      vp8cx_mb_init_quantizer(cpi, x);
+      vp9cx_mb_init_quantizer(cpi, x);
     } else
       // Set to Segment 0 by default
       mbmi->segment_id = 0;
@@ -624,12 +624,12 @@ static void pick_mb_modes(VP8_COMP *cpi,
     // as a predictor for MBs that follow in the SB
     if (cm->frame_type == KEY_FRAME) {
       int r, d;
-      vp8_rd_pick_intra_mode(cpi, x, &r, &d);
+      vp9_rd_pick_intra_mode(cpi, x, &r, &d);
       *totalrate += r;
       *totaldist += d;
 
       // Dummy encode, do not do the tokenization
-      vp8cx_encode_intra_macro_block(cpi, x, tp, 0);
+      vp9cx_encode_intra_macro_block(cpi, x, tp, 0);
       // Note the encoder may have changed the segment_id
 
       // Save the coding context
@@ -650,13 +650,13 @@ static void pick_mb_modes(VP8_COMP *cpi,
         cpi->seg0_progress = (((mb_col & ~1) * 2 + (mb_row & ~1) * cm->mb_cols + i) << 16) / cm->MBs;
       }
 
-      vp8cx_pick_mode_inter_macroblock(cpi, x, recon_yoffset,
+      vp9cx_pick_mode_inter_macroblock(cpi, x, recon_yoffset,
                                        recon_uvoffset, &r, &d);
       *totalrate += r;
       *totaldist += d;
 
       // Dummy encode, do not do the tokenization
-      vp8cx_encode_inter_macroblock(cpi, x, tp,
+      vp9cx_encode_inter_macroblock(cpi, x, tp,
                                     recon_yoffset, recon_uvoffset, 0);
 
       seg_id = mbmi->segment_id;
@@ -777,7 +777,7 @@ static void pick_sb_modes (VP8_COMP *cpi,
   x->rddiv = cpi->RDDIV;
   x->rdmult = cpi->RDMULT;
   if(cpi->oxcf.tuning == VP8_TUNE_SSIM)
-    vp8_activity_masking(cpi, x);
+    vp9_activity_masking(cpi, x);
   /* Is segmentation enabled */
   if (xd->segmentation_enabled)
   {
@@ -797,7 +797,7 @@ static void pick_sb_modes (VP8_COMP *cpi,
     if (xd->mode_info_context->mbmi.segment_id > 3)
       xd->mode_info_context->mbmi.segment_id = 0;
 
-    vp8cx_mb_init_quantizer(cpi, x);
+    vp9cx_mb_init_quantizer(cpi, x);
   }
   else
     /* Set to Segment 0 by default */
@@ -811,7 +811,7 @@ static void pick_sb_modes (VP8_COMP *cpi,
    * as a predictor for MBs that follow in the SB */
   if (cm->frame_type == KEY_FRAME)
   {
-    vp8_rd_pick_intra_mode_sb(cpi, x,
+    vp9_rd_pick_intra_mode_sb(cpi, x,
                               totalrate,
                               totaldist);
 
@@ -832,7 +832,7 @@ static void pick_sb_modes (VP8_COMP *cpi,
         (((mb_col & ~1) * 2 + (mb_row & ~1) * cm->mb_cols) << 16) / cm->MBs;
     }
 
-    vp8_rd_pick_inter_mode_sb(cpi, x,
+    vp9_rd_pick_inter_mode_sb(cpi, x,
                               recon_yoffset,
                               recon_uvoffset,
                               totalrate,
@@ -968,11 +968,11 @@ static void encode_sb(VP8_COMP *cpi,
     vp8_copy_mem16x16(x->src.y_buffer, x->src.y_stride, x->thismb, 16);
 
     if (cpi->oxcf.tuning == VP8_TUNE_SSIM)
-      vp8_activity_masking(cpi, x);
+      vp9_activity_masking(cpi, x);
 
     // Is segmentation enabled
     if (xd->segmentation_enabled) {
-      vp8cx_mb_init_quantizer(cpi, x);
+      vp9cx_mb_init_quantizer(cpi, x);
     }
 
     x->active_ptr = cpi->active_map + map_index;
@@ -987,10 +987,10 @@ static void encode_sb(VP8_COMP *cpi,
     if (cm->frame_type == KEY_FRAME) {
 #if CONFIG_SUPERBLOCKS
       if (xd->mode_info_context->mbmi.encoded_as_sb)
-        vp8cx_encode_intra_super_block(cpi, x, tp, mb_col);
+        vp9cx_encode_intra_super_block(cpi, x, tp, mb_col);
       else
 #endif
-        vp8cx_encode_intra_macro_block(cpi, x, tp, 1);
+        vp9cx_encode_intra_macro_block(cpi, x, tp, 1);
         // Note the encoder may have changed the segment_id
 
 #ifdef MODE_STATS
@@ -1013,10 +1013,10 @@ static void encode_sb(VP8_COMP *cpi,
 
 #if CONFIG_SUPERBLOCKS
       if (xd->mode_info_context->mbmi.encoded_as_sb)
-        vp8cx_encode_inter_superblock(cpi, x, tp, recon_yoffset, recon_uvoffset, mb_col, mb_row);
+        vp9cx_encode_inter_superblock(cpi, x, tp, recon_yoffset, recon_uvoffset, mb_col, mb_row);
       else
 #endif
-        vp8cx_encode_inter_macroblock(cpi, x, tp,
+        vp9cx_encode_inter_macroblock(cpi, x, tp,
                                       recon_yoffset, recon_uvoffset, 1);
         // Note the encoder may have changed the segment_id
 
@@ -1261,11 +1261,11 @@ static void init_encode_frame_mb_context(VP8_COMP *cpi) {
   // set up frame for intra coded blocks
   vp8_setup_intra_recon(&cm->yv12_fb[cm->new_fb_idx]);
 
-  vp8_build_block_offsets(x);
+  vp9_build_block_offsets(x);
 
   vp8_setup_block_dptrs(&x->e_mbd);
 
-  vp8_setup_block_ptrs(x);
+  vp9_setup_block_ptrs(x);
 
   xd->mode_info_context->mbmi.mode = DC_PRED;
   xd->mode_info_context->mbmi.uv_mode = DC_PRED;
@@ -1360,10 +1360,10 @@ static void encode_frame_internal(VP8_COMP *cpi) {
   vp8_zero(cpi->coef_counts_16x16);
   vp8_zero(cpi->hybrid_coef_counts_16x16);
 
-  vp8cx_frame_init_quantizer(cpi);
+  vp9cx_frame_init_quantizer(cpi);
 
-  vp8_initialize_rd_consts(cpi, cm->base_qindex + cm->y1dc_delta_q);
-  vp8cx_initialize_me_consts(cpi, cm->base_qindex);
+  vp9_initialize_rd_consts(cpi, cm->base_qindex + cm->y1dc_delta_q);
+  vp9cx_initialize_me_consts(cpi, cm->base_qindex);
 
   if (cpi->oxcf.tuning == VP8_TUNE_SSIM) {
     // Initialize encode frame context.
@@ -1487,7 +1487,7 @@ static void reset_skip_txfm_size(VP8_COMP *cpi, TX_SIZE txfm_max) {
   }
 }
 
-void vp8_encode_frame(VP8_COMP *cpi) {
+void vp9_encode_frame(VP8_COMP *cpi) {
   if (cpi->sf.RD) {
     int i, frame_type, pred_type;
     TXFM_MODE txfm_type;
@@ -1640,7 +1640,7 @@ void vp8_encode_frame(VP8_COMP *cpi) {
 
 }
 
-void vp8_setup_block_ptrs(MACROBLOCK *x) {
+void vp9_setup_block_ptrs(MACROBLOCK *x) {
   int r, c;
   int i;
 
@@ -1671,7 +1671,7 @@ void vp8_setup_block_ptrs(MACROBLOCK *x) {
   }
 }
 
-void vp8_build_block_offsets(MACROBLOCK *x) {
+void vp9_build_block_offsets(MACROBLOCK *x) {
   int block = 0;
   int br, bc;
 
@@ -1836,7 +1836,7 @@ static void update_sb_skip_coeff_state(VP8_COMP *cpi,
     if (skip[n]) {
       x->e_mbd.above_context = &ta[n];
       x->e_mbd.left_context  = &tl[n];
-      vp8_stuff_mb(cpi, &x->e_mbd, tp, 0);
+      vp9_stuff_mb(cpi, &x->e_mbd, tp, 0);
     } else {
       if (n_tokens[n]) {
         memcpy(*tp, tokens[n], sizeof(*t[0]) * n_tokens[n]);
@@ -1846,7 +1846,7 @@ static void update_sb_skip_coeff_state(VP8_COMP *cpi,
   }
 }
 
-void vp8cx_encode_intra_super_block(VP8_COMP *cpi,
+void vp9cx_encode_intra_super_block(VP8_COMP *cpi,
                                     MACROBLOCK *x,
                                     TOKENEXTRA **t,
                                     int mb_col) {
@@ -1870,7 +1870,7 @@ void vp8cx_encode_intra_super_block(VP8_COMP *cpi,
 
   if ((cpi->oxcf.tuning == VP8_TUNE_SSIM) && output_enabled) {
     adjust_act_zbin(cpi, x);
-    vp8_update_zbin_extra(cpi, x);
+    vp9_update_zbin_extra(cpi, x);
   }
 
   vp8_build_intra_predictors_sby_s(&x->e_mbd);
@@ -1883,23 +1883,23 @@ void vp8cx_encode_intra_super_block(VP8_COMP *cpi,
     xd->above_context = cm->above_context + mb_col + (n & 1);
     xd->left_context = cm->left_context + (n >> 1);
 
-    vp8_subtract_mby_s_c(x->src_diff,
+    vp9_subtract_mby_s_c(x->src_diff,
                          src + x_idx * 16 + y_idx * 16 * src_y_stride,
                          src_y_stride,
                          dst + x_idx * 16 + y_idx * 16 * dst_y_stride,
                          dst_y_stride);
-    vp8_subtract_mbuv_s_c(x->src_diff,
+    vp9_subtract_mbuv_s_c(x->src_diff,
                           usrc + x_idx * 8 + y_idx * 8 * src_uv_stride,
                           vsrc + x_idx * 8 + y_idx * 8 * src_uv_stride,
                           src_uv_stride,
                           udst + x_idx * 8 + y_idx * 8 * dst_uv_stride,
                           vdst + x_idx * 8 + y_idx * 8 * dst_uv_stride,
                           dst_uv_stride);
-    vp8_transform_mb_8x8(x);
-    vp8_quantize_mb_8x8(x);
+    vp9_transform_mb_8x8(x);
+    vp9_quantize_mb_8x8(x);
     if (x->optimize) {
-      vp8_optimize_mby_8x8(x, rtcd);
-      vp8_optimize_mbuv_8x8(x, rtcd);
+      vp9_optimize_mby_8x8(x, rtcd);
+      vp9_optimize_mbuv_8x8(x, rtcd);
     }
     vp8_inverse_transform_mb_8x8(IF_RTCD(&rtcd->common->idct), &x->e_mbd);
     vp8_recon_mby_s_c(&x->e_mbd, dst + x_idx * 16 + y_idx * 16 * dst_y_stride);
@@ -1912,7 +1912,7 @@ void vp8cx_encode_intra_super_block(VP8_COMP *cpi,
       memcpy(&tl[n], xd->left_context, sizeof(tl[n]));
       tp[n] = *t;
       xd->mode_info_context = mi + x_idx + y_idx * cm->mode_info_stride;
-      vp8_tokenize_mb(cpi, &x->e_mbd, t, 0);
+      vp9_tokenize_mb(cpi, &x->e_mbd, t, 0);
       skip[n] = xd->mode_info_context->mbmi.mb_skip_coeff;
     }
   }
@@ -1926,26 +1926,26 @@ void vp8cx_encode_intra_super_block(VP8_COMP *cpi,
 }
 #endif /* CONFIG_SUPERBLOCKS */
 
-void vp8cx_encode_intra_macro_block(VP8_COMP *cpi,
+void vp9cx_encode_intra_macro_block(VP8_COMP *cpi,
                                     MACROBLOCK *x,
                                     TOKENEXTRA **t,
                                     int output_enabled) {
   MB_MODE_INFO * mbmi = &x->e_mbd.mode_info_context->mbmi;
   if ((cpi->oxcf.tuning == VP8_TUNE_SSIM) && output_enabled) {
     adjust_act_zbin(cpi, x);
-    vp8_update_zbin_extra(cpi, x);
+    vp9_update_zbin_extra(cpi, x);
   }
   if (mbmi->mode == I8X8_PRED) {
-    vp8_encode_intra8x8mby(IF_RTCD(&cpi->rtcd), x);
-    vp8_encode_intra8x8mbuv(IF_RTCD(&cpi->rtcd), x);
+    vp9_encode_intra8x8mby(IF_RTCD(&cpi->rtcd), x);
+    vp9_encode_intra8x8mbuv(IF_RTCD(&cpi->rtcd), x);
   } else if (mbmi->mode == B_PRED) {
-    vp8_encode_intra4x4mby(IF_RTCD(&cpi->rtcd), x);
+    vp9_encode_intra4x4mby(IF_RTCD(&cpi->rtcd), x);
   } else {
-    vp8_encode_intra16x16mby(IF_RTCD(&cpi->rtcd), x);
+    vp9_encode_intra16x16mby(IF_RTCD(&cpi->rtcd), x);
   }
 
   if (mbmi->mode != I8X8_PRED) {
-    vp8_encode_intra16x16mbuv(IF_RTCD(&cpi->rtcd), x);
+    vp9_encode_intra16x16mbuv(IF_RTCD(&cpi->rtcd), x);
   }
 
   if (output_enabled) {
@@ -1953,7 +1953,7 @@ void vp8cx_encode_intra_macro_block(VP8_COMP *cpi,
 
     // Tokenize
     sum_intra_stats(cpi, x);
-    vp8_tokenize_mb(cpi, &x->e_mbd, t, 0);
+    vp9_tokenize_mb(cpi, &x->e_mbd, t, 0);
 
     if (cpi->common.txfm_mode == TX_MODE_SELECT &&
         !((cpi->common.mb_no_coeff_skip && mbmi->mb_skip_coeff) ||
@@ -1975,16 +1975,16 @@ void vp8cx_encode_intra_macro_block(VP8_COMP *cpi,
   }
 #if CONFIG_NEWBESTREFMV
   else
-    vp8_tokenize_mb(cpi, &x->e_mbd, t, 1);
+    vp9_tokenize_mb(cpi, &x->e_mbd, t, 1);
 #endif
 }
 #ifdef SPEEDSTATS
 extern int cnt_pm;
 #endif
 
-extern void vp8_fix_contexts(MACROBLOCKD *xd);
+extern void vp9_fix_contexts(MACROBLOCKD *xd);
 
-void vp8cx_encode_inter_macroblock (VP8_COMP *cpi, MACROBLOCK *x,
+void vp9cx_encode_inter_macroblock (VP8_COMP *cpi, MACROBLOCK *x,
                                     TOKENEXTRA **t, int recon_yoffset,
                                     int recon_uvoffset, int output_enabled) {
   VP8_COMMON *cm = &cpi->common;
@@ -2023,7 +2023,7 @@ void vp8cx_encode_inter_macroblock (VP8_COMP *cpi, MACROBLOCK *x,
       }
     }
 
-    vp8_update_zbin_extra(cpi, x);
+    vp9_update_zbin_extra(cpi, x);
   }
 
   seg_ref_active = vp9_segfeature_active(xd, *segment_id, SEG_LVL_REF_FRAME);
@@ -2036,14 +2036,14 @@ void vp8cx_encode_inter_macroblock (VP8_COMP *cpi, MACROBLOCK *x,
 
   if (mbmi->ref_frame == INTRA_FRAME) {
     if (mbmi->mode == B_PRED) {
-      vp8_encode_intra16x16mbuv(IF_RTCD(&cpi->rtcd), x);
-      vp8_encode_intra4x4mby(IF_RTCD(&cpi->rtcd), x);
+      vp9_encode_intra16x16mbuv(IF_RTCD(&cpi->rtcd), x);
+      vp9_encode_intra4x4mby(IF_RTCD(&cpi->rtcd), x);
     } else if (mbmi->mode == I8X8_PRED) {
-      vp8_encode_intra8x8mby(IF_RTCD(&cpi->rtcd), x);
-      vp8_encode_intra8x8mbuv(IF_RTCD(&cpi->rtcd), x);
+      vp9_encode_intra8x8mby(IF_RTCD(&cpi->rtcd), x);
+      vp9_encode_intra8x8mbuv(IF_RTCD(&cpi->rtcd), x);
     } else {
-      vp8_encode_intra16x16mbuv(IF_RTCD(&cpi->rtcd), x);
-      vp8_encode_intra16x16mby(IF_RTCD(&cpi->rtcd), x);
+      vp9_encode_intra16x16mbuv(IF_RTCD(&cpi->rtcd), x);
+      vp9_encode_intra16x16mby(IF_RTCD(&cpi->rtcd), x);
     }
 
     if (output_enabled)
@@ -2081,7 +2081,7 @@ void vp8cx_encode_inter_macroblock (VP8_COMP *cpi, MACROBLOCK *x,
     }
 
     if (!x->skip) {
-      vp8_encode_inter16x16(IF_RTCD(&cpi->rtcd), x);
+      vp9_encode_inter16x16(IF_RTCD(&cpi->rtcd), x);
 
       // Clear mb_skip_coeff if mb_no_coeff_skip is not set
       if (!cpi->common.mb_no_coeff_skip)
@@ -2114,7 +2114,7 @@ void vp8cx_encode_inter_macroblock (VP8_COMP *cpi, MACROBLOCK *x,
     }
 #endif
 
-    vp8_tokenize_mb(cpi, xd, t, !output_enabled);
+    vp9_tokenize_mb(cpi, xd, t, !output_enabled);
 
 #ifdef ENC_DEBUG
     if (enc_debug) {
@@ -2132,9 +2132,9 @@ void vp8cx_encode_inter_macroblock (VP8_COMP *cpi, MACROBLOCK *x,
       mbmi->mb_skip_coeff = 1;
       if (output_enabled)
         cpi->skip_true_count[mb_skip_context]++;
-      vp8_fix_contexts(xd);
+      vp9_fix_contexts(xd);
     } else {
-      vp8_stuff_mb(cpi, xd, t, !output_enabled);
+      vp9_stuff_mb(cpi, xd, t, !output_enabled);
       mbmi->mb_skip_coeff = 0;
       if (output_enabled)
         cpi->skip_false_count[mb_skip_context]++;
@@ -2170,7 +2170,7 @@ void vp8cx_encode_inter_macroblock (VP8_COMP *cpi, MACROBLOCK *x,
 }
 
 #if CONFIG_SUPERBLOCKS
-void vp8cx_encode_inter_superblock(VP8_COMP *cpi, MACROBLOCK *x, TOKENEXTRA **t,
+void vp9cx_encode_inter_superblock(VP8_COMP *cpi, MACROBLOCK *x, TOKENEXTRA **t,
                                    int recon_yoffset, int recon_uvoffset, int mb_col, int mb_row) {
   const int output_enabled = 1;
   VP8_COMMON *cm = &cpi->common;
@@ -2218,7 +2218,7 @@ void vp8cx_encode_inter_superblock(VP8_COMP *cpi, MACROBLOCK *x, TOKENEXTRA **t,
       }
     }
 
-    vp8_update_zbin_extra(cpi, x);
+    vp9_update_zbin_extra(cpi, x);
   }
 
   seg_ref_active = vp9_segfeature_active(xd, segment_id, SEG_LVL_REF_FRAME);
@@ -2274,23 +2274,23 @@ void vp8cx_encode_inter_superblock(VP8_COMP *cpi, MACROBLOCK *x, TOKENEXTRA **t,
   for (n = 0; n < 4; n++) {
     int x_idx = n & 1, y_idx = n >> 1;
 
-    vp8_subtract_mby_s_c(x->src_diff,
+    vp9_subtract_mby_s_c(x->src_diff,
                          src + x_idx * 16 + y_idx * 16 * src_y_stride,
                          src_y_stride,
                          dst + x_idx * 16 + y_idx * 16 * dst_y_stride,
                          dst_y_stride);
-    vp8_subtract_mbuv_s_c(x->src_diff,
+    vp9_subtract_mbuv_s_c(x->src_diff,
                           usrc + x_idx * 8 + y_idx * 8 * src_uv_stride,
                           vsrc + x_idx * 8 + y_idx * 8 * src_uv_stride,
                           src_uv_stride,
                           udst + x_idx * 8 + y_idx * 8 * dst_uv_stride,
                           vdst + x_idx * 8 + y_idx * 8 * dst_uv_stride,
                           dst_uv_stride);
-    vp8_transform_mb_8x8(x);
-    vp8_quantize_mb_8x8(x);
+    vp9_transform_mb_8x8(x);
+    vp9_quantize_mb_8x8(x);
     if (x->optimize) {
-      vp8_optimize_mby_8x8(x, rtcd);
-      vp8_optimize_mbuv_8x8(x, rtcd);
+      vp9_optimize_mby_8x8(x, rtcd);
+      vp9_optimize_mbuv_8x8(x, rtcd);
     }
     vp8_inverse_transform_mb_8x8(IF_RTCD(&rtcd->common->idct), &x->e_mbd);
     vp8_recon_mby_s_c( &x->e_mbd,
@@ -2307,7 +2307,7 @@ void vp8cx_encode_inter_superblock(VP8_COMP *cpi, MACROBLOCK *x, TOKENEXTRA **t,
         memcpy(&tl[n], xd->left_context, sizeof(tl[n]));
         tp[n] = *t;
         xd->mode_info_context = mi + x_idx + y_idx * cm->mode_info_stride;
-        vp8_tokenize_mb(cpi, &x->e_mbd, t, 0);
+        vp9_tokenize_mb(cpi, &x->e_mbd, t, 0);
         skip[n] = xd->mode_info_context->mbmi.mb_skip_coeff;
       }
     } else {
@@ -2324,9 +2324,9 @@ void vp8cx_encode_inter_superblock(VP8_COMP *cpi, MACROBLOCK *x, TOKENEXTRA **t,
         memcpy(&tl[n], xd->left_context, sizeof(tl[n]));
         tp[n] = *t;
         cpi->skip_true_count[mb_skip_context]++;
-        vp8_fix_contexts(xd);
+        vp9_fix_contexts(xd);
       } else {
-        vp8_stuff_mb(cpi, xd, t, 0);
+        vp9_stuff_mb(cpi, xd, t, 0);
         xd->mode_info_context->mbmi.mb_skip_coeff = 0;
         cpi->skip_false_count[mb_skip_context]++;
       }
diff --git a/vp8/encoder/encodeintra.c b/vp8/encoder/encodeintra.c
index d23c2305be67a6acc0b7d01f01b45d665e0645e8..597d218d0d240ce54acc34c6bf009caecc8d81c2 100644
--- a/vp8/encoder/encodeintra.c
+++ b/vp8/encoder/encodeintra.c
@@ -25,7 +25,7 @@
 #define IF_RTCD(x) NULL
 #endif
 
-int vp8_encode_intra(VP8_COMP *cpi, MACROBLOCK *x, int use_16x16_pred) {
+int vp9_encode_intra(VP8_COMP *cpi, MACROBLOCK *x, int use_16x16_pred) {
   int i;
   int intra_pred_var = 0;
   MB_MODE_INFO * mbmi = &x->e_mbd.mode_info_context->mbmi;
@@ -39,20 +39,20 @@ int vp8_encode_intra(VP8_COMP *cpi, MACROBLOCK *x, int use_16x16_pred) {
     mbmi->uv_mode = DC_PRED;
     mbmi->ref_frame = INTRA_FRAME;
 
-    vp8_encode_intra16x16mby(IF_RTCD(&cpi->rtcd), x);
+    vp9_encode_intra16x16mby(IF_RTCD(&cpi->rtcd), x);
   } else {
     for (i = 0; i < 16; i++) {
       x->e_mbd.block[i].bmi.as_mode.first = B_DC_PRED;
-      vp8_encode_intra4x4block(IF_RTCD(&cpi->rtcd), x, i);
+      vp9_encode_intra4x4block(IF_RTCD(&cpi->rtcd), x, i);
     }
   }
 
-  intra_pred_var = vp8_get_mb_ss(x->src_diff);
+  intra_pred_var = vp9_get_mb_ss(x->src_diff);
 
   return intra_pred_var;
 }
 
-void vp8_encode_intra4x4block(const VP8_ENCODER_RTCD *rtcd,
+void vp9_encode_intra4x4block(const VP8_ENCODER_RTCD *rtcd,
                               MACROBLOCK *x, int ib) {
   BLOCKD *b = &x->e_mbd.block[ib];
   BLOCK *be = &x->block[ib];
@@ -69,15 +69,15 @@ void vp8_encode_intra4x4block(const VP8_ENCODER_RTCD *rtcd,
   }
 #endif
 
-  vp8_subtract_b(be, b, 16);
+  vp9_subtract_b(be, b, 16);
 
   tx_type = get_tx_type(&x->e_mbd, b);
   if (tx_type != DCT_DCT) {
-    vp8_fht_c(be->src_diff, 32, be->coeff, tx_type, 4);
-    vp8_ht_quantize_b_4x4(be, b, tx_type);
+    vp9_fht_c(be->src_diff, 32, be->coeff, tx_type, 4);
+    vp9_ht_quantize_b_4x4(be, b, tx_type);
     vp8_ihtllm_c(b->dqcoeff, b->diff, 32, tx_type, 4);
   } else {
-    x->vp8_short_fdct4x4(be->src_diff, be->coeff, 32) ;
+    x->vp9_short_fdct4x4(be->src_diff, be->coeff, 32) ;
     x->quantize_b_4x4(be, b) ;
     vp8_inverse_transform_b_4x4(IF_RTCD(&rtcd->common->idct), b, 32) ;
   }
@@ -85,15 +85,15 @@ void vp8_encode_intra4x4block(const VP8_ENCODER_RTCD *rtcd,
   vp8_recon_b(b->predictor, b->diff, *(b->base_dst) + b->dst, b->dst_stride);
 }
 
-void vp8_encode_intra4x4mby(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *mb) {
+void vp9_encode_intra4x4mby(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *mb) {
   int i;
 
   for (i = 0; i < 16; i++)
-    vp8_encode_intra4x4block(rtcd, mb, i);
+    vp9_encode_intra4x4block(rtcd, mb, i);
   return;
 }
 
-void vp8_encode_intra16x16mby(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x) {
+void vp9_encode_intra16x16mby(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x) {
   MACROBLOCKD *xd = &x->e_mbd;
   BLOCK *b = &x->block[0];
   TX_SIZE tx_size = xd->mode_info_context->mbmi.txfm_size;
@@ -108,42 +108,42 @@ void vp8_encode_intra16x16mby(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x) {
     vp8_build_comp_intra_predictors_mby(xd);
 #endif
 
-  vp8_subtract_mby(x->src_diff, *(b->base_src), xd->predictor, b->src_stride);
+  vp9_subtract_mby(x->src_diff, *(b->base_src), xd->predictor, b->src_stride);
 
   if (tx_size == TX_16X16) {
     BLOCKD  *bd = &xd->block[0];
     tx_type = get_tx_type(xd, bd);
     if (tx_type != DCT_DCT) {
-      vp8_fht_c(b->src_diff, 32, b->coeff, tx_type, 16);
-      vp8_quantize_mby_16x16(x);
+      vp9_fht_c(b->src_diff, 32, b->coeff, tx_type, 16);
+      vp9_quantize_mby_16x16(x);
       if (x->optimize)
-        vp8_optimize_mby_16x16(x, rtcd);
+        vp9_optimize_mby_16x16(x, rtcd);
       vp8_ihtllm_c(bd->dqcoeff, bd->diff, 32, tx_type, 16);
     } else {
-      vp8_transform_mby_16x16(x);
-      vp8_quantize_mby_16x16(x);
+      vp9_transform_mby_16x16(x);
+      vp9_quantize_mby_16x16(x);
       if (x->optimize)
-        vp8_optimize_mby_16x16(x, rtcd);
+        vp9_optimize_mby_16x16(x, rtcd);
       vp8_inverse_transform_mby_16x16(IF_RTCD(&rtcd->common->idct), xd);
     }
   } else if (tx_size == TX_8X8) {
-    vp8_transform_mby_8x8(x);
-    vp8_quantize_mby_8x8(x);
+    vp9_transform_mby_8x8(x);
+    vp9_quantize_mby_8x8(x);
     if (x->optimize)
-      vp8_optimize_mby_8x8(x, rtcd);
+      vp9_optimize_mby_8x8(x, rtcd);
     vp8_inverse_transform_mby_8x8(IF_RTCD(&rtcd->common->idct), xd);
   } else {
-    vp8_transform_mby_4x4(x);
-    vp8_quantize_mby_4x4(x);
+    vp9_transform_mby_4x4(x);
+    vp9_quantize_mby_4x4(x);
     if (x->optimize)
-      vp8_optimize_mby_4x4(x, rtcd);
+      vp9_optimize_mby_4x4(x, rtcd);
     vp8_inverse_transform_mby_4x4(IF_RTCD(&rtcd->common->idct), xd);
   }
 
   vp8_recon_mby(xd);
 }
 
-void vp8_encode_intra16x16mbuv(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x) {
+void vp9_encode_intra16x16mbuv(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x) {
   MACROBLOCKD *xd = &x->e_mbd;
   TX_SIZE tx_size = xd->mode_info_context->mbmi.txfm_size;
 
@@ -157,27 +157,27 @@ void vp8_encode_intra16x16mbuv(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x) {
   }
 #endif
 
-  vp8_subtract_mbuv(x->src_diff, x->src.u_buffer, x->src.v_buffer,
+  vp9_subtract_mbuv(x->src_diff, x->src.u_buffer, x->src.v_buffer,
                     xd->predictor, x->src.uv_stride);
 
   if (tx_size == TX_4X4) {
-    vp8_transform_mbuv_4x4(x);
-    vp8_quantize_mbuv_4x4(x);
+    vp9_transform_mbuv_4x4(x);
+    vp9_quantize_mbuv_4x4(x);
     if (x->optimize)
-      vp8_optimize_mbuv_4x4(x, rtcd);
+      vp9_optimize_mbuv_4x4(x, rtcd);
     vp8_inverse_transform_mbuv_4x4(IF_RTCD(&rtcd->common->idct), xd);
   } else /* 16x16 or 8x8 */ {
-    vp8_transform_mbuv_8x8(x);
-    vp8_quantize_mbuv_8x8(x);
+    vp9_transform_mbuv_8x8(x);
+    vp9_quantize_mbuv_8x8(x);
     if (x->optimize)
-      vp8_optimize_mbuv_8x8(x, rtcd);
+      vp9_optimize_mbuv_8x8(x, rtcd);
     vp8_inverse_transform_mbuv_8x8(IF_RTCD(&rtcd->common->idct), xd);
   }
 
   vp8_recon_intra_mbuv(xd);
 }
 
-void vp8_encode_intra8x8(const VP8_ENCODER_RTCD *rtcd,
+void vp9_encode_intra8x8(const VP8_ENCODER_RTCD *rtcd,
                          MACROBLOCK *x, int ib) {
   MACROBLOCKD *xd = &x->e_mbd;
   BLOCKD *b = &xd->block[ib];
@@ -201,17 +201,17 @@ void vp8_encode_intra8x8(const VP8_ENCODER_RTCD *rtcd,
     int idx = (ib & 0x02) ? (ib + 2) : ib;
 
     // generate residual blocks
-    vp8_subtract_4b_c(be, b, 16);
+    vp9_subtract_4b_c(be, b, 16);
 
     tx_type = get_tx_type(xd, xd->block + idx);
     if (tx_type != DCT_DCT) {
-      vp8_fht_c(be->src_diff, 32, (x->block + idx)->coeff,
+      vp9_fht_c(be->src_diff, 32, (x->block + idx)->coeff,
                 tx_type, 8);
       x->quantize_b_8x8(x->block + idx, xd->block + idx);
       vp8_ihtllm_c(xd->block[idx].dqcoeff, xd->block[ib].diff, 32,
                    tx_type, 8);
     } else {
-      x->vp8_short_fdct8x8(be->src_diff, (x->block + idx)->coeff, 32);
+      x->vp9_short_fdct8x8(be->src_diff, (x->block + idx)->coeff, 32);
       x->quantize_b_8x8(x->block + idx, xd->block + idx);
       vp8_idct_idct8(xd->block[idx].dqcoeff, xd->block[ib].diff, 32);
     }
@@ -219,8 +219,8 @@ void vp8_encode_intra8x8(const VP8_ENCODER_RTCD *rtcd,
     for (i = 0; i < 4; i++) {
       b = &xd->block[ib + iblock[i]];
       be = &x->block[ib + iblock[i]];
-      vp8_subtract_b(be, b, 16);
-      x->vp8_short_fdct4x4(be->src_diff, be->coeff, 32);
+      vp9_subtract_b(be, b, 16);
+      x->vp9_short_fdct4x4(be->src_diff, be->coeff, 32);
       x->quantize_b_4x4(be, b);
       vp8_inverse_transform_b_4x4(IF_RTCD(&rtcd->common->idct), b, 32);
     }
@@ -234,16 +234,16 @@ void vp8_encode_intra8x8(const VP8_ENCODER_RTCD *rtcd,
   }
 }
 
-void vp8_encode_intra8x8mby(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x) {
+void vp9_encode_intra8x8mby(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x) {
   int i, ib;
 
   for (i = 0; i < 4; i++) {
     ib = vp8_i8x8_block[i];
-    vp8_encode_intra8x8(rtcd, x, ib);
+    vp9_encode_intra8x8(rtcd, x, ib);
   }
 }
 
-void vp8_encode_intra_uv4x4(const VP8_ENCODER_RTCD *rtcd,
+void vp9_encode_intra_uv4x4(const VP8_ENCODER_RTCD *rtcd,
                             MACROBLOCK *x, int ib,
                             int mode, int second) {
   BLOCKD *b = &x->e_mbd.block[ib];
@@ -259,16 +259,16 @@ void vp8_encode_intra_uv4x4(const VP8_ENCODER_RTCD *rtcd,
   }
 #endif
 
-  vp8_subtract_b(be, b, 8);
+  vp9_subtract_b(be, b, 8);
 
-  x->vp8_short_fdct4x4(be->src_diff, be->coeff, 16);
+  x->vp9_short_fdct4x4(be->src_diff, be->coeff, 16);
   x->quantize_b_4x4(be, b);
   vp8_inverse_transform_b_4x4(IF_RTCD(&rtcd->common->idct), b, 16);
 
   vp8_recon_uv_b_c(b->predictor,b->diff, *(b->base_dst) + b->dst, b->dst_stride);
 }
 
-void vp8_encode_intra8x8mbuv(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x) {
+void vp9_encode_intra8x8mbuv(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x) {
   int i, ib, mode, second;
   BLOCKD *b;
 
@@ -282,8 +282,8 @@ void vp8_encode_intra8x8mbuv(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x) {
     second = -1;
 #endif
     /*u */
-    vp8_encode_intra_uv4x4(rtcd, x, i + 16, mode, second);
+    vp9_encode_intra_uv4x4(rtcd, x, i + 16, mode, second);
     /*v */
-    vp8_encode_intra_uv4x4(rtcd, x, i + 20, mode, second);
+    vp9_encode_intra_uv4x4(rtcd, x, i + 20, mode, second);
   }
 }
diff --git a/vp8/encoder/encodeintra.h b/vp8/encoder/encodeintra.h
index 7d122938f17f8053793f1276d800d45167b3d7b7..9f2d4a286835769ae323abc996d144bd7880ae92 100644
--- a/vp8/encoder/encodeintra.h
+++ b/vp8/encoder/encodeintra.h
@@ -13,15 +13,15 @@
 #define _ENCODEINTRA_H_
 #include "onyx_int.h"
 
-int vp8_encode_intra(VP8_COMP *cpi, MACROBLOCK *x, int use_16x16_pred);
-void vp8_encode_intra16x16mby(const VP8_ENCODER_RTCD *, MACROBLOCK *x);
-void vp8_encode_intra16x16mbuv(const VP8_ENCODER_RTCD *, MACROBLOCK *x);
-void vp8_encode_intra4x4mby(const VP8_ENCODER_RTCD *, MACROBLOCK *mb);
-void vp8_encode_intra4x4block(const VP8_ENCODER_RTCD *rtcd,
+int vp9_encode_intra(VP8_COMP *cpi, MACROBLOCK *x, int use_16x16_pred);
+void vp9_encode_intra16x16mby(const VP8_ENCODER_RTCD *, MACROBLOCK *x);
+void vp9_encode_intra16x16mbuv(const VP8_ENCODER_RTCD *, MACROBLOCK *x);
+void vp9_encode_intra4x4mby(const VP8_ENCODER_RTCD *, MACROBLOCK *mb);
+void vp9_encode_intra4x4block(const VP8_ENCODER_RTCD *rtcd,
                               MACROBLOCK *x, int ib);
-void vp8_encode_intra8x8mby(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x);
-void vp8_encode_intra8x8mbuv(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x);
-void vp8_encode_intra8x8(const VP8_ENCODER_RTCD *rtcd,
+void vp9_encode_intra8x8mby(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x);
+void vp9_encode_intra8x8mbuv(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x);
+void vp9_encode_intra8x8(const VP8_ENCODER_RTCD *rtcd,
                          MACROBLOCK *x, int ib);
 
 #endif
diff --git a/vp8/encoder/encodemb.c b/vp8/encoder/encodemb.c
index 64512be6ffe9f2f0e762e36852c883608f4d3d88..b16f0557d110fc4d5aac59a7b08a802486fed0f6 100644
--- a/vp8/encoder/encodemb.c
+++ b/vp8/encoder/encodemb.c
@@ -30,7 +30,7 @@
 extern int enc_debug;
 #endif
 
-void vp8_subtract_b_c(BLOCK *be, BLOCKD *bd, int pitch) {
+void vp9_subtract_b_c(BLOCK *be, BLOCKD *bd, int pitch) {
   unsigned char *src_ptr = (*(be->base_src) + be->src);
   short *diff_ptr = be->src_diff;
   unsigned char *pred_ptr = bd->predictor;
@@ -49,7 +49,7 @@ void vp8_subtract_b_c(BLOCK *be, BLOCKD *bd, int pitch) {
   }
 }
 
-void vp8_subtract_4b_c(BLOCK *be, BLOCKD *bd, int pitch) {
+void vp9_subtract_4b_c(BLOCK *be, BLOCKD *bd, int pitch) {
   unsigned char *src_ptr = (*(be->base_src) + be->src);
   short *diff_ptr = be->src_diff;
   unsigned char *pred_ptr = bd->predictor;
@@ -66,7 +66,7 @@ void vp8_subtract_4b_c(BLOCK *be, BLOCKD *bd, int pitch) {
   }
 }
 
-void vp8_subtract_mbuv_s_c(short *diff, const unsigned char *usrc,
+void vp9_subtract_mbuv_s_c(short *diff, const unsigned char *usrc,
                            const unsigned char *vsrc, int src_stride,
                            const unsigned char *upred,
                            const unsigned char *vpred, int dst_stride) {
@@ -95,15 +95,15 @@ void vp8_subtract_mbuv_s_c(short *diff, const unsigned char *usrc,
   }
 }
 
-void vp8_subtract_mbuv_c(short *diff, unsigned char *usrc,
+void vp9_subtract_mbuv_c(short *diff, unsigned char *usrc,
                          unsigned char *vsrc, unsigned char *pred, int stride) {
   unsigned char *upred = pred + 256;
   unsigned char *vpred = pred + 320;
 
-  vp8_subtract_mbuv_s_c(diff, usrc, vsrc, stride, upred, vpred, 8);
+  vp9_subtract_mbuv_s_c(diff, usrc, vsrc, stride, upred, vpred, 8);
 }
 
-void vp8_subtract_mby_s_c(short *diff, const unsigned char *src, int src_stride,
+void vp9_subtract_mby_s_c(short *diff, const unsigned char *src, int src_stride,
                           const unsigned char *pred, int dst_stride) {
   int r, c;
 
@@ -118,17 +118,17 @@ void vp8_subtract_mby_s_c(short *diff, const unsigned char *src, int src_stride,
   }
 }
 
-void vp8_subtract_mby_c(short *diff, unsigned char *src,
+void vp9_subtract_mby_c(short *diff, unsigned char *src,
                         unsigned char *pred, int stride) {
-  vp8_subtract_mby_s_c(diff, src, stride, pred, 16);
+  vp9_subtract_mby_s_c(diff, src, stride, pred, 16);
 }
 
 static void vp8_subtract_mb(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x) {
   BLOCK *b = &x->block[0];
 
-  vp8_subtract_mby(x->src_diff, *(b->base_src), x->e_mbd.predictor,
+  vp9_subtract_mby(x->src_diff, *(b->base_src), x->e_mbd.predictor,
                    b->src_stride);
-  vp8_subtract_mbuv(x->src_diff, x->src.u_buffer, x->src.v_buffer,
+  vp9_subtract_mbuv(x->src_diff, x->src.u_buffer, x->src.v_buffer,
                     x->e_mbd.predictor, x->src.uv_stride);
 }
 
@@ -141,11 +141,11 @@ static void build_dcblock_4x4(MACROBLOCK *x) {
   }
 }
 
-void vp8_transform_mby_4x4(MACROBLOCK *x) {
+void vp9_transform_mby_4x4(MACROBLOCK *x) {
   int i;
 
   for (i = 0; i < 16; i += 2) {
-    x->vp8_short_fdct8x4(&x->block[i].src_diff[0],
+    x->vp9_short_fdct8x4(&x->block[i].src_diff[0],
                          &x->block[i].coeff[0], 32);
   }
 
@@ -159,21 +159,21 @@ void vp8_transform_mby_4x4(MACROBLOCK *x) {
   }
 }
 
-void vp8_transform_mbuv_4x4(MACROBLOCK *x) {
+void vp9_transform_mbuv_4x4(MACROBLOCK *x) {
   int i;
 
   for (i = 16; i < 24; i += 2) {
-    x->vp8_short_fdct8x4(&x->block[i].src_diff[0],
+    x->vp9_short_fdct8x4(&x->block[i].src_diff[0],
                          &x->block[i].coeff[0], 16);
   }
 }
 
 static void transform_mb_4x4(MACROBLOCK *x) {
-  vp8_transform_mby_4x4(x);
-  vp8_transform_mbuv_4x4(x);
+  vp9_transform_mby_4x4(x);
+  vp9_transform_mbuv_4x4(x);
 }
 
-void vp8_build_dcblock_8x8(MACROBLOCK *x) {
+void vp9_build_dcblock_8x8(MACROBLOCK *x) {
   int16_t *src_diff_ptr = x->block[24].src_diff;
   int i;
 
@@ -186,21 +186,21 @@ void vp8_build_dcblock_8x8(MACROBLOCK *x) {
   src_diff_ptr[8] = x->coeff[12 * 16];
 }
 
-void vp8_transform_mby_8x8(MACROBLOCK *x) {
+void vp9_transform_mby_8x8(MACROBLOCK *x) {
   int i;
 
   for (i = 0; i < 9; i += 8) {
-    x->vp8_short_fdct8x8(&x->block[i].src_diff[0],
+    x->vp9_short_fdct8x8(&x->block[i].src_diff[0],
                          &x->block[i].coeff[0], 32);
   }
   for (i = 2; i < 11; i += 8) {
-    x->vp8_short_fdct8x8(&x->block[i].src_diff[0],
+    x->vp9_short_fdct8x8(&x->block[i].src_diff[0],
                          &x->block[i + 2].coeff[0], 32);
   }
 
   if (x->e_mbd.mode_info_context->mbmi.mode != SPLITMV) {
     // build dc block from 2x2 y dc values
-    vp8_build_dcblock_8x8(x);
+    vp9_build_dcblock_8x8(x);
 
     // do 2nd order transform on the dc block
     x->short_fhaar2x2(&x->block[24].src_diff[0],
@@ -208,29 +208,29 @@ void vp8_transform_mby_8x8(MACROBLOCK *x) {
   }
 }
 
-void vp8_transform_mbuv_8x8(MACROBLOCK *x) {
+void vp9_transform_mbuv_8x8(MACROBLOCK *x) {
   int i;
 
   for (i = 16; i < 24; i += 4) {
-    x->vp8_short_fdct8x8(&x->block[i].src_diff[0],
+    x->vp9_short_fdct8x8(&x->block[i].src_diff[0],
                          &x->block[i].coeff[0], 16);
   }
 }
 
-void vp8_transform_mb_8x8(MACROBLOCK *x) {
-  vp8_transform_mby_8x8(x);
-  vp8_transform_mbuv_8x8(x);
+void vp9_transform_mb_8x8(MACROBLOCK *x) {
+  vp9_transform_mby_8x8(x);
+  vp9_transform_mbuv_8x8(x);
 }
 
-void vp8_transform_mby_16x16(MACROBLOCK *x) {
+void vp9_transform_mby_16x16(MACROBLOCK *x) {
   vp8_clear_system_state();
-  x->vp8_short_fdct16x16(&x->block[0].src_diff[0],
+  x->vp9_short_fdct16x16(&x->block[0].src_diff[0],
                          &x->block[0].coeff[0], 32);
 }
 
-void vp8_transform_mb_16x16(MACROBLOCK *x) {
-  vp8_transform_mby_16x16(x);
-  vp8_transform_mbuv_8x8(x);
+void vp9_transform_mb_16x16(MACROBLOCK *x) {
+  vp9_transform_mby_16x16(x);
+  vp9_transform_mbuv_8x8(x);
 }
 
 #define RDTRUNC(RM,DM,R,D) ( (128+(R)*(RM)) & 0xFF )
@@ -571,7 +571,7 @@ static void check_reset_8x8_2nd_coeffs(MACROBLOCKD *xd,
   }
 }
 
-void vp8_optimize_mby_4x4(MACROBLOCK *x, const VP8_ENCODER_RTCD *rtcd) {
+void vp9_optimize_mby_4x4(MACROBLOCK *x, const VP8_ENCODER_RTCD *rtcd) {
   int b;
   PLANE_TYPE type;
   int has_2nd_order;
@@ -606,7 +606,7 @@ void vp8_optimize_mby_4x4(MACROBLOCK *x, const VP8_ENCODER_RTCD *rtcd) {
   }
 }
 
-void vp8_optimize_mbuv_4x4(MACROBLOCK *x, const VP8_ENCODER_RTCD *rtcd) {
+void vp9_optimize_mbuv_4x4(MACROBLOCK *x, const VP8_ENCODER_RTCD *rtcd) {
   int b;
   ENTROPY_CONTEXT_PLANES t_above, t_left;
   ENTROPY_CONTEXT *ta;
@@ -628,11 +628,11 @@ void vp8_optimize_mbuv_4x4(MACROBLOCK *x, const VP8_ENCODER_RTCD *rtcd) {
 }
 
 static void optimize_mb_4x4(MACROBLOCK *x, const VP8_ENCODER_RTCD *rtcd) {
-  vp8_optimize_mby_4x4(x, rtcd);
-  vp8_optimize_mbuv_4x4(x, rtcd);
+  vp9_optimize_mby_4x4(x, rtcd);
+  vp9_optimize_mbuv_4x4(x, rtcd);
 }
 
-void vp8_optimize_mby_8x8(MACROBLOCK *x, const VP8_ENCODER_RTCD *rtcd) {
+void vp9_optimize_mby_8x8(MACROBLOCK *x, const VP8_ENCODER_RTCD *rtcd) {
   int b;
   PLANE_TYPE type;
   ENTROPY_CONTEXT_PLANES t_above, t_left;
@@ -665,7 +665,7 @@ void vp8_optimize_mby_8x8(MACROBLOCK *x, const VP8_ENCODER_RTCD *rtcd) {
   }
 }
 
-void vp8_optimize_mbuv_8x8(MACROBLOCK *x, const VP8_ENCODER_RTCD *rtcd) {
+void vp9_optimize_mbuv_8x8(MACROBLOCK *x, const VP8_ENCODER_RTCD *rtcd) {
   int b;
   ENTROPY_CONTEXT_PLANES t_above, t_left;
   ENTROPY_CONTEXT *ta;
@@ -690,8 +690,8 @@ void vp8_optimize_mbuv_8x8(MACROBLOCK *x, const VP8_ENCODER_RTCD *rtcd) {
 }
 
 static void optimize_mb_8x8(MACROBLOCK *x, const VP8_ENCODER_RTCD *rtcd) {
-  vp8_optimize_mby_8x8(x, rtcd);
-  vp8_optimize_mbuv_8x8(x, rtcd);
+  vp9_optimize_mby_8x8(x, rtcd);
+  vp9_optimize_mbuv_8x8(x, rtcd);
 }
 
 static void optimize_b_16x16(MACROBLOCK *mb, int i, PLANE_TYPE type,
@@ -868,7 +868,7 @@ static void optimize_b_16x16(MACROBLOCK *mb, int i, PLANE_TYPE type,
   *a = *l = (d->eob != !type);
 }
 
-void vp8_optimize_mby_16x16(MACROBLOCK *x, const VP8_ENCODER_RTCD *rtcd) {
+void vp9_optimize_mby_16x16(MACROBLOCK *x, const VP8_ENCODER_RTCD *rtcd) {
   ENTROPY_CONTEXT_PLANES t_above, t_left;
   ENTROPY_CONTEXT *ta, *tl;
 
@@ -884,11 +884,11 @@ void vp8_optimize_mby_16x16(MACROBLOCK *x, const VP8_ENCODER_RTCD *rtcd) {
 }
 
 static void optimize_mb_16x16(MACROBLOCK *x, const VP8_ENCODER_RTCD *rtcd) {
-  vp8_optimize_mby_16x16(x, rtcd);
-  vp8_optimize_mbuv_8x8(x, rtcd);
+  vp9_optimize_mby_16x16(x, rtcd);
+  vp9_optimize_mbuv_8x8(x, rtcd);
 }
 
-void vp8_encode_inter16x16(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x) {
+void vp9_encode_inter16x16(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x) {
   MACROBLOCKD *xd = &x->e_mbd;
   TX_SIZE tx_size = xd->mode_info_context->mbmi.txfm_size;
 
@@ -896,34 +896,34 @@ void vp8_encode_inter16x16(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x) {
   vp8_subtract_mb(rtcd, x);
 
   if (tx_size == TX_16X16) {
-    vp8_transform_mb_16x16(x);
-    vp8_quantize_mb_16x16(x);
+    vp9_transform_mb_16x16(x);
+    vp9_quantize_mb_16x16(x);
     if (x->optimize)
       optimize_mb_16x16(x, rtcd);
     vp8_inverse_transform_mb_16x16(IF_RTCD(&rtcd->common->idct), xd);
   } else if (tx_size == TX_8X8) {
     if (xd->mode_info_context->mbmi.mode == SPLITMV) {
       assert(xd->mode_info_context->mbmi.partitioning != PARTITIONING_4X4);
-      vp8_transform_mby_8x8(x);
-      vp8_transform_mbuv_4x4(x);
-      vp8_quantize_mby_8x8(x);
-      vp8_quantize_mbuv_4x4(x);
+      vp9_transform_mby_8x8(x);
+      vp9_transform_mbuv_4x4(x);
+      vp9_quantize_mby_8x8(x);
+      vp9_quantize_mbuv_4x4(x);
       if (x->optimize) {
-        vp8_optimize_mby_8x8(x, rtcd);
-        vp8_optimize_mbuv_4x4(x, rtcd);
+        vp9_optimize_mby_8x8(x, rtcd);
+        vp9_optimize_mbuv_4x4(x, rtcd);
       }
       vp8_inverse_transform_mby_8x8(IF_RTCD(&rtcd->common->idct), xd);
       vp8_inverse_transform_mbuv_4x4(IF_RTCD(&rtcd->common->idct), xd);
     } else {
-      vp8_transform_mb_8x8(x);
-      vp8_quantize_mb_8x8(x);
+      vp9_transform_mb_8x8(x);
+      vp9_quantize_mb_8x8(x);
       if (x->optimize)
         optimize_mb_8x8(x, rtcd);
       vp8_inverse_transform_mb_8x8(IF_RTCD(&rtcd->common->idct), xd);
     }
   } else {
     transform_mb_4x4(x);
-    vp8_quantize_mb_4x4(x);
+    vp9_quantize_mb_4x4(x);
     if (x->optimize)
       optimize_mb_4x4(x, rtcd);
     vp8_inverse_transform_mb_4x4(IF_RTCD(&rtcd->common->idct), xd);
@@ -933,7 +933,7 @@ void vp8_encode_inter16x16(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x) {
 }
 
 /* this function is used by first pass only */
-void vp8_encode_inter16x16y(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x) {
+void vp9_encode_inter16x16y(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x) {
   MACROBLOCKD *xd = &x->e_mbd;
   BLOCK *b = &x->block[0];
 
@@ -944,10 +944,10 @@ void vp8_encode_inter16x16y(const VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x) {
 
   vp8_build_1st_inter16x16_predictors_mby(xd, xd->predictor, 16, 0);
 
-  vp8_subtract_mby(x->src_diff, *(b->base_src), xd->predictor, b->src_stride);
+  vp9_subtract_mby(x->src_diff, *(b->base_src), xd->predictor, b->src_stride);
 
-  vp8_transform_mby_4x4(x);
-  vp8_quantize_mby_4x4(x);
+  vp9_transform_mby_4x4(x);
+  vp9_quantize_mby_4x4(x);
   vp8_inverse_transform_mby_4x4(IF_RTCD(&rtcd->common->idct), xd);
 
   vp8_recon_mby(xd);
diff --git a/vp8/encoder/encodemb.h b/vp8/encoder/encodemb.h
index 6c2300e34f0fa8a2c343ccbac96fc04c4a32e5b4..c524e0531ca4cd2964e0d3892024480466d3378e 100644
--- a/vp8/encoder/encodemb.h
+++ b/vp8/encoder/encodemb.h
@@ -35,36 +35,36 @@ typedef struct {
 
 #include "onyx_int.h"
 struct VP8_ENCODER_RTCD;
-void vp8_encode_inter16x16(const struct VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x);
+void vp9_encode_inter16x16(const struct VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x);
 
 void vp8_build_dcblock(MACROBLOCK *b);
 void vp8_transform_mb_4x4(MACROBLOCK *mb);
-void vp8_transform_mbuv_4x4(MACROBLOCK *x);
-void vp8_transform_mby_4x4(MACROBLOCK *x);
+void vp9_transform_mbuv_4x4(MACROBLOCK *x);
+void vp9_transform_mby_4x4(MACROBLOCK *x);
 
-void vp8_optimize_mby_4x4(MACROBLOCK *x, const struct VP8_ENCODER_RTCD *rtcd);
-void vp8_optimize_mbuv_4x4(MACROBLOCK *x, const struct VP8_ENCODER_RTCD *rtcd);
-void vp8_encode_inter16x16y(const struct VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x);
+void vp9_optimize_mby_4x4(MACROBLOCK *x, const struct VP8_ENCODER_RTCD *rtcd);
+void vp9_optimize_mbuv_4x4(MACROBLOCK *x, const struct VP8_ENCODER_RTCD *rtcd);
+void vp9_encode_inter16x16y(const struct VP8_ENCODER_RTCD *rtcd, MACROBLOCK *x);
 
-void vp8_transform_mb_8x8(MACROBLOCK *mb);
-void vp8_transform_mby_8x8(MACROBLOCK *x);
-void vp8_transform_mbuv_8x8(MACROBLOCK *x);
-void vp8_build_dcblock_8x8(MACROBLOCK *b);
-void vp8_optimize_mby_8x8(MACROBLOCK *x, const struct VP8_ENCODER_RTCD *rtcd);
-void vp8_optimize_mbuv_8x8(MACROBLOCK *x, const struct VP8_ENCODER_RTCD *rtcd);
+void vp9_transform_mb_8x8(MACROBLOCK *mb);
+void vp9_transform_mby_8x8(MACROBLOCK *x);
+void vp9_transform_mbuv_8x8(MACROBLOCK *x);
+void vp9_build_dcblock_8x8(MACROBLOCK *b);
+void vp9_optimize_mby_8x8(MACROBLOCK *x, const struct VP8_ENCODER_RTCD *rtcd);
+void vp9_optimize_mbuv_8x8(MACROBLOCK *x, const struct VP8_ENCODER_RTCD *rtcd);
 
-void vp8_transform_mb_16x16(MACROBLOCK *mb);
-void vp8_transform_mby_16x16(MACROBLOCK *x);
-void vp8_optimize_mby_16x16(MACROBLOCK *x, const struct VP8_ENCODER_RTCD *rtcd);
+void vp9_transform_mb_16x16(MACROBLOCK *mb);
+void vp9_transform_mby_16x16(MACROBLOCK *x);
+void vp9_optimize_mby_16x16(MACROBLOCK *x, const struct VP8_ENCODER_RTCD *rtcd);
 
-void vp8_subtract_4b_c(BLOCK *be, BLOCKD *bd, int pitch);
+void vp9_subtract_4b_c(BLOCK *be, BLOCKD *bd, int pitch);
 
 #if CONFIG_SUPERBLOCKS
-void vp8_subtract_mbuv_s_c(short *diff, const unsigned char *usrc,
+void vp9_subtract_mbuv_s_c(short *diff, const unsigned char *usrc,
                            const unsigned char *vsrc, int src_stride,
                            const unsigned char *upred,
                            const unsigned char *vpred, int dst_stride);
-void vp8_subtract_mby_s_c(short *diff, const unsigned char *src,
+void vp9_subtract_mby_s_c(short *diff, const unsigned char *src,
                           int src_stride, const unsigned char *pred,
                           int dst_stride);
 #endif
diff --git a/vp8/encoder/encodemv.c b/vp8/encoder/encodemv.c
index 75dad2f9b35fe336d5a553da02335507a7402d77..ea0874c719f003fa3feae0ed7b2595de9944f2d9 100644
--- a/vp8/encoder/encodemv.c
+++ b/vp8/encoder/encodemv.c
@@ -99,16 +99,16 @@ static void build_nmv_component_cost_table(int *mvcost,
 
   sign_cost[0] = vp8_cost_zero(mvcomp->sign);
   sign_cost[1] = vp8_cost_one(mvcomp->sign);
-  vp8_cost_tokens(class_cost, mvcomp->classes, vp8_mv_class_tree);
-  vp8_cost_tokens(class0_cost, mvcomp->class0, vp8_mv_class0_tree);
+  vp9_cost_tokens(class_cost, mvcomp->classes, vp8_mv_class_tree);
+  vp9_cost_tokens(class0_cost, mvcomp->class0, vp8_mv_class0_tree);
   for (i = 0; i < MV_OFFSET_BITS; ++i) {
     bits_cost[i][0] = vp8_cost_zero(mvcomp->bits[i]);
     bits_cost[i][1] = vp8_cost_one(mvcomp->bits[i]);
   }
 
   for (i = 0; i < CLASS0_SIZE; ++i)
-    vp8_cost_tokens(class0_fp_cost[i], mvcomp->class0_fp[i], vp8_mv_fp_tree);
-  vp8_cost_tokens(fp_cost, mvcomp->fp, vp8_mv_fp_tree);
+    vp9_cost_tokens(class0_fp_cost[i], mvcomp->class0_fp[i], vp8_mv_fp_tree);
+  vp9_cost_tokens(fp_cost, mvcomp->fp, vp8_mv_fp_tree);
 
   if (usehp) {
     class0_hp_cost[0] = vp8_cost_zero(mvcomp->class0_hp);
@@ -356,7 +356,7 @@ static void add_nmvcount(nmv_context_counts* const dst,
 }
 #endif
 
-void vp8_write_nmvprobs(VP8_COMP* const cpi, int usehp, vp8_writer* const bc) {
+void vp9_write_nmvprobs(VP8_COMP* const cpi, int usehp, vp8_writer* const bc) {
   int i, j;
   nmv_context prob;
   unsigned int branch_ct_joint[MV_JOINTS - 1][2];
@@ -508,7 +508,7 @@ void vp8_write_nmvprobs(VP8_COMP* const cpi, int usehp, vp8_writer* const bc) {
   }
 }
 
-void vp8_encode_nmv(vp8_writer* const bc, const MV* const mv,
+void vp9_encode_nmv(vp8_writer* const bc, const MV* const mv,
                     const MV* const ref, const nmv_context* const mvctx) {
   MV_JOINT_TYPE j = vp8_get_mv_joint(*mv);
   vp8_write_token(bc, vp8_mv_joint_tree, mvctx->joints,
@@ -521,7 +521,7 @@ void vp8_encode_nmv(vp8_writer* const bc, const MV* const mv,
   }
 }
 
-void vp8_encode_nmv_fp(vp8_writer* const bc, const MV* const mv,
+void vp9_encode_nmv_fp(vp8_writer* const bc, const MV* const mv,
                        const MV* const ref, const nmv_context* const mvctx,
                        int usehp) {
   MV_JOINT_TYPE j = vp8_get_mv_joint(*mv);
@@ -534,14 +534,14 @@ void vp8_encode_nmv_fp(vp8_writer* const bc, const MV* const mv,
   }
 }
 
-void vp8_build_nmv_cost_table(int *mvjoint,
+void vp9_build_nmv_cost_table(int *mvjoint,
                               int *mvcost[2],
                               const nmv_context* const mvctx,
                               int usehp,
                               int mvc_flag_v,
                               int mvc_flag_h) {
   vp8_clear_system_state();
-  vp8_cost_tokens(mvjoint, mvctx->joints, vp8_mv_joint_tree);
+  vp9_cost_tokens(mvjoint, mvctx->joints, vp8_mv_joint_tree);
   if (mvc_flag_v)
     build_nmv_component_cost_table(mvcost[0], &mvctx->comps[0], usehp);
   if (mvc_flag_h)
diff --git a/vp8/encoder/encodemv.h b/vp8/encoder/encodemv.h
index 2545365804c659d36c8c69439e7648a7a7c7f83f..d15a4d3f72586de6f19da9a68c1522fd44e5f9f5 100644
--- a/vp8/encoder/encodemv.h
+++ b/vp8/encoder/encodemv.h
@@ -14,13 +14,13 @@
 
 #include "onyx_int.h"
 
-void vp8_write_nmvprobs(VP8_COMP* const, int usehp, vp8_writer* const);
-void vp8_encode_nmv(vp8_writer* const w, const MV* const mv,
+void vp9_write_nmvprobs(VP8_COMP* const, int usehp, vp8_writer* const);
+void vp9_encode_nmv(vp8_writer* const w, const MV* const mv,
                     const MV* const ref, const nmv_context* const mvctx);
-void vp8_encode_nmv_fp(vp8_writer* const w, const MV* const mv,
+void vp9_encode_nmv_fp(vp8_writer* const w, const MV* const mv,
                        const MV* const ref, const nmv_context *mvctx,
                        int usehp);
-void vp8_build_nmv_cost_table(int *mvjoint,
+void vp9_build_nmv_cost_table(int *mvjoint,
                               int *mvcost[2],
                               const nmv_context *mvctx,
                               int usehp,
diff --git a/vp8/encoder/firstpass.c b/vp8/encoder/firstpass.c
index 2a622ffef34a7902bc1306b66750f1ba80919d9e..f29e6633eb5183d0068b703ab7c005f7a5a50321 100644
--- a/vp8/encoder/firstpass.c
+++ b/vp8/encoder/firstpass.c
@@ -39,11 +39,11 @@
 #define IF_RTCD(x) NULL
 #endif
 
-extern void vp8_build_block_offsets(MACROBLOCK *x);
-extern void vp8_setup_block_ptrs(MACROBLOCK *x);
-extern void vp8cx_frame_init_quantizer(VP8_COMP *cpi);
-extern void vp8_set_mbmode_and_mvs(MACROBLOCK *x, MB_PREDICTION_MODE mb, int_mv *mv);
-extern void vp8_alloc_compressor_data(VP8_COMP *cpi);
+extern void vp9_build_block_offsets(MACROBLOCK *x);
+extern void vp9_setup_block_ptrs(MACROBLOCK *x);
+extern void vp9cx_frame_init_quantizer(VP8_COMP *cpi);
+extern void vp9_set_mbmode_and_mvs(MACROBLOCK *x, MB_PREDICTION_MODE mb, int_mv *mv);
+extern void vp9_alloc_compressor_data(VP8_COMP *cpi);
 
 #define IIFACTOR   12.5
 #define IIKFACTOR1 12.5
@@ -66,10 +66,10 @@ static int select_cq_level(int qindex) {
   int ret_val = QINDEX_RANGE - 1;
   int i;
 
-  double target_q = (vp8_convert_qindex_to_q(qindex) * 0.5847) + 1.0;
+  double target_q = (vp9_convert_qindex_to_q(qindex) * 0.5847) + 1.0;
 
   for (i = 0; i < QINDEX_RANGE; i++) {
-    if (target_q <= vp8_convert_qindex_to_q(i)) {
+    if (target_q <= vp9_convert_qindex_to_q(i)) {
       ret_val = i;
       break;
     }
@@ -341,11 +341,11 @@ static int frame_max_bits(VP8_COMP *cpi) {
   return max_bits;
 }
 
-void vp8_init_first_pass(VP8_COMP *cpi) {
+void vp9_init_first_pass(VP8_COMP *cpi) {
   zero_stats(cpi->twopass.total_stats);
 }
 
-void vp8_end_first_pass(VP8_COMP *cpi) {
+void vp9_end_first_pass(VP8_COMP *cpi) {
   output_stats(cpi, cpi->output_pkt_list, cpi->twopass.total_stats);
 }
 
@@ -364,7 +364,7 @@ static void zz_motion_search(VP8_COMP *cpi, MACROBLOCK *x, YV12_BUFFER_CONFIG *r
 
   ref_ptr = (unsigned char *)(*(d->base_pre) + d->pre);
 
-  vp8_mse16x16(src_ptr, src_stride, ref_ptr, ref_stride,
+  vp9_mse16x16(src_ptr, src_stride, ref_ptr, ref_stride,
                (unsigned int *)(best_motion_err));
 }
 
@@ -388,7 +388,7 @@ static void first_pass_motion_search(VP8_COMP *cpi, MACROBLOCK *x,
   int new_mv_mode_penalty = 256;
 
   // override the default variance function to use MSE
-  v_fn_ptr.vf = vp8_mse16x16;
+  v_fn_ptr.vf = vp9_mse16x16;
 
   // Set up pointers for this macro block recon buffer
   xd->pre.y_buffer = recon_buffer->y_buffer + recon_yoffset;
@@ -435,7 +435,7 @@ static void first_pass_motion_search(VP8_COMP *cpi, MACROBLOCK *x,
   }
 }
 
-void vp8_first_pass(VP8_COMP *cpi) {
+void vp9_first_pass(VP8_COMP *cpi) {
   int mb_row, mb_col;
   MACROBLOCK *const x = &cpi->mb;
   VP8_COMMON *const cm = &cpi->common;
@@ -477,15 +477,15 @@ void vp8_first_pass(VP8_COMP *cpi) {
 
   xd->mode_info_context = cm->mi;
 
-  vp8_build_block_offsets(x);
+  vp9_build_block_offsets(x);
 
   vp8_setup_block_dptrs(&x->e_mbd);
 
-  vp8_setup_block_ptrs(x);
+  vp9_setup_block_ptrs(x);
 
   // set up frame new frame for intra coded blocks
   vp8_setup_intra_recon(new_yv12);
-  vp8cx_frame_init_quantizer(cpi);
+  vp9cx_frame_init_quantizer(cpi);
 
   // Initialise the MV cost table to the defaults
   // if( cm->current_video_frame == 0)
@@ -493,7 +493,7 @@ void vp8_first_pass(VP8_COMP *cpi) {
   {
     int flag[2] = {1, 1};
     vp8_init_mv_probs(cm);
-    vp8_initialize_rd_consts(cpi, cm->base_qindex + cm->y1dc_delta_q);
+    vp9_initialize_rd_consts(cpi, cm->base_qindex + cm->y1dc_delta_q);
   }
 
   // for each macroblock row in image
@@ -527,7 +527,7 @@ void vp8_first_pass(VP8_COMP *cpi) {
       vp8_copy_mem16x16(x->src.y_buffer, x->src.y_stride, x->thismb, 16);
 
       // do intra 16x16 prediction
-      this_error = vp8_encode_intra(cpi, x, use_dc_pred);
+      this_error = vp9_encode_intra(cpi, x, use_dc_pred);
 
       // "intrapenalty" below deals with situations where the intra and inter error scores are very low (eg a plain black frame)
       // We do not have special cases in first pass for 0,0 and nearest etc so all inter modes carry an overhead cost estimate fot the mv.
@@ -618,9 +618,9 @@ void vp8_first_pass(VP8_COMP *cpi) {
           mv.as_mv.row <<= 3;
           mv.as_mv.col <<= 3;
           this_error = motion_error;
-          vp8_set_mbmode_and_mvs(x, NEWMV, &mv);
+          vp9_set_mbmode_and_mvs(x, NEWMV, &mv);
           xd->mode_info_context->mbmi.txfm_size = TX_4X4;
-          vp8_encode_inter16x16y(IF_RTCD(&cpi->rtcd), x);
+          vp9_encode_inter16x16y(IF_RTCD(&cpi->rtcd), x);
           sum_mvr += mv.as_mv.row;
           sum_mvr_abs += abs(mv.as_mv.row);
           sum_mvc += mv.as_mv.col;
@@ -820,11 +820,11 @@ static long long estimate_modemvcost(VP8_COMP *cpi,
   intra_cost = bitcost(av_intra);
 
   // Estimate of extra bits per mv overhead for mbs
-  // << 9 is the normalization to the (bits * 512) used in vp8_bits_per_mb
+  // << 9 is the normalization to the (bits * 512) used in vp9_bits_per_mb
   mv_cost = ((int)(fpstats->new_mv_count / fpstats->count) * 8) << 9;
 
   // Crude estimate of overhead cost from modes
-  // << 9 is the normalization to (bits * 512) used in vp8_bits_per_mb
+  // << 9 is the normalization to (bits * 512) used in vp9_bits_per_mb
   mode_cost =
     (int)((((av_pct_inter - av_pct_motion) * zz_cost) +
            (av_pct_motion * motion_cost) +
@@ -845,7 +845,7 @@ static double calc_correction_factor(double err_per_mb,
   double correction_factor;
 
   // Adjustment based on actual quantizer to power term.
-  power_term = (vp8_convert_qindex_to_q(Q) * 0.01) + pt_low;
+  power_term = (vp9_convert_qindex_to_q(Q) * 0.01) + pt_low;
   power_term = (power_term > pt_high) ? pt_high : power_term;
 
   // Adjustments to error term
@@ -875,7 +875,7 @@ static void adjust_maxq_qrange(VP8_COMP *cpi) {
   cpi->twopass.maxq_max_limit = cpi->worst_quality;
   for (i = cpi->best_quality; i <= cpi->worst_quality; i++) {
     cpi->twopass.maxq_max_limit = i;
-    if (vp8_convert_qindex_to_q(i) >= q)
+    if (vp9_convert_qindex_to_q(i) >= q)
       break;
   }
 
@@ -884,7 +884,7 @@ static void adjust_maxq_qrange(VP8_COMP *cpi) {
   cpi->twopass.maxq_min_limit = cpi->best_quality;
   for (i = cpi->worst_quality; i >= cpi->best_quality; i--) {
     cpi->twopass.maxq_min_limit = i;
-    if (vp8_convert_qindex_to_q(i) <= q)
+    if (vp9_convert_qindex_to_q(i) <= q)
       break;
   }
 }
@@ -978,7 +978,7 @@ static int estimate_max_q(VP8_COMP *cpi,
       err_correction_factor = 5.0;
 
     bits_per_mb_at_this_q =
-      vp8_bits_per_mb(INTER_FRAME, Q) + overhead_bits_per_mb;
+      vp9_bits_per_mb(INTER_FRAME, Q) + overhead_bits_per_mb;
 
     bits_per_mb_at_this_q = (int)(.5 + err_correction_factor *
                                   (double)bits_per_mb_at_this_q);
@@ -1084,7 +1084,7 @@ static int estimate_cq(VP8_COMP *cpi,
       err_correction_factor = 5.0;
 
     bits_per_mb_at_this_q =
-      vp8_bits_per_mb(INTER_FRAME, Q) + overhead_bits_per_mb;
+      vp9_bits_per_mb(INTER_FRAME, Q) + overhead_bits_per_mb;
 
     bits_per_mb_at_this_q = (int)(.5 + err_correction_factor *
                                   (double)bits_per_mb_at_this_q);
@@ -1111,9 +1111,9 @@ static int estimate_cq(VP8_COMP *cpi,
 }
 
 
-extern void vp8_new_frame_rate(VP8_COMP *cpi, double framerate);
+extern void vp9_new_frame_rate(VP8_COMP *cpi, double framerate);
 
-void vp8_init_second_pass(VP8_COMP *cpi) {
+void vp9_init_second_pass(VP8_COMP *cpi) {
   FIRSTPASS_STATS this_frame;
   FIRSTPASS_STATS *start_pos;
 
@@ -1138,7 +1138,7 @@ void vp8_init_second_pass(VP8_COMP *cpi) {
   // encoded in the second pass is a guess.  However the sum duration is not.
   // Its calculated based on the actual durations of all frames from the first
   // pass.
-  vp8_new_frame_rate(cpi, 10000000.0 * cpi->twopass.total_stats->count / cpi->twopass.total_stats->duration);
+  vp9_new_frame_rate(cpi, 10000000.0 * cpi->twopass.total_stats->count / cpi->twopass.total_stats->duration);
 
   cpi->output_frame_rate = cpi->oxcf.frame_rate;
   cpi->twopass.bits_left = (int64_t)(cpi->twopass.total_stats->duration * cpi->oxcf.target_bandwidth / 10000000.0);
@@ -1191,7 +1191,7 @@ void vp8_init_second_pass(VP8_COMP *cpi) {
   }
 }
 
-void vp8_end_second_pass(VP8_COMP *cpi) {
+void vp9_end_second_pass(VP8_COMP *cpi) {
 }
 
 // This function gives and estimate of how badly we believe
@@ -1727,7 +1727,7 @@ static void define_gf_group(VP8_COMP *cpi, FIRSTPASS_STATS *this_frame) {
     int Q = (cpi->oxcf.fixed_q < 0) ? cpi->last_q[INTER_FRAME] : cpi->oxcf.fixed_q;
     int gf_bits;
 
-    boost = (cpi->gfu_boost * vp8_gfboost_qadjust(Q)) / 100;
+    boost = (cpi->gfu_boost * vp9_gfboost_qadjust(Q)) / 100;
 
     // Set max and minimum boost and hence minimum allocation
     if (boost > ((cpi->baseline_gf_interval + 1) * 200))
@@ -1919,21 +1919,21 @@ static int adjust_active_maxq(int old_maxqi, int new_maxqi) {
   double new_q;
   double target_q;
 
-  old_q = vp8_convert_qindex_to_q(old_maxqi);
-  new_q = vp8_convert_qindex_to_q(new_maxqi);
+  old_q = vp9_convert_qindex_to_q(old_maxqi);
+  new_q = vp9_convert_qindex_to_q(new_maxqi);
 
   target_q = ((old_q * 7.0) + new_q) / 8.0;
 
   if (target_q > old_q) {
     for (i = old_maxqi; i <= new_maxqi; i++) {
-      if (vp8_convert_qindex_to_q(i) >= target_q) {
+      if (vp9_convert_qindex_to_q(i) >= target_q) {
         ret_val = i;
         break;
       }
     }
   } else {
     for (i = old_maxqi; i >= new_maxqi; i--) {
-      if (vp8_convert_qindex_to_q(i) <= target_q) {
+      if (vp9_convert_qindex_to_q(i) <= target_q) {
         ret_val = i;
         break;
       }
@@ -1943,7 +1943,7 @@ static int adjust_active_maxq(int old_maxqi, int new_maxqi) {
   return ret_val;
 }
 
-void vp8_second_pass(VP8_COMP *cpi) {
+void vp9_second_pass(VP8_COMP *cpi) {
   int tmp_q;
   int frames_left = (int)(cpi->twopass.total_stats->count - cpi->common.current_video_frame);
 
@@ -2059,7 +2059,7 @@ void vp8_second_pass(VP8_COMP *cpi) {
 
     cpi->active_worst_quality         = tmp_q;
     cpi->ni_av_qi                     = tmp_q;
-    cpi->avg_q                        = vp8_convert_qindex_to_q(tmp_q);
+    cpi->avg_q                        = vp9_convert_qindex_to_q(tmp_q);
 
     // Limit the maxq value returned subsequently.
     // This increases the risk of overspend or underspend if the initial
diff --git a/vp8/encoder/firstpass.h b/vp8/encoder/firstpass.h
index 95e1e5463c6dcb44c460d4b4f6a175252817a825..805fb0d5eb40c45516c626eb42c1dd498a3cf5e1 100644
--- a/vp8/encoder/firstpass.h
+++ b/vp8/encoder/firstpass.h
@@ -12,13 +12,13 @@
 #if !defined __INC_FIRSTPASS_H
 #define      __INC_FIRSTPASS_H
 
-extern void vp8_init_first_pass(VP8_COMP *cpi);
-extern void vp8_first_pass(VP8_COMP *cpi);
-extern void vp8_end_first_pass(VP8_COMP *cpi);
+extern void vp9_init_first_pass(VP8_COMP *cpi);
+extern void vp9_first_pass(VP8_COMP *cpi);
+extern void vp9_end_first_pass(VP8_COMP *cpi);
 
-extern void vp8_init_second_pass(VP8_COMP *cpi);
-extern void vp8_second_pass(VP8_COMP *cpi);
-extern void vp8_end_second_pass(VP8_COMP *cpi);
+extern void vp9_init_second_pass(VP8_COMP *cpi);
+extern void vp9_second_pass(VP8_COMP *cpi);
+extern void vp9_end_second_pass(VP8_COMP *cpi);
 
 extern size_t vp8_firstpass_stats_sz(unsigned int mb_count);
 #endif
diff --git a/vp8/encoder/generic/csystemdependent.c b/vp8/encoder/generic/csystemdependent.c
index dbe36249d0cbc6b20866e746e2a862c609431c51..3b8d71d7331b1d2ffb9252b8a7a13c839f36f220 100644
--- a/vp8/encoder/generic/csystemdependent.c
+++ b/vp8/encoder/generic/csystemdependent.c
@@ -14,26 +14,26 @@
 #include "vp8/encoder/onyx_int.h"
 
 
-void vp8_arch_x86_encoder_init(VP8_COMP *cpi);
+void vp9_arch_x86_encoder_init(VP8_COMP *cpi);
 void vp8_arch_arm_encoder_init(VP8_COMP *cpi);
 
-void (*vp8_yv12_copy_partial_frame_ptr)(YV12_BUFFER_CONFIG *src_ybc, YV12_BUFFER_CONFIG *dst_ybc, int Fraction);
-extern void vp8_yv12_copy_partial_frame(YV12_BUFFER_CONFIG *src_ybc, YV12_BUFFER_CONFIG *dst_ybc, int Fraction);
+void (*vp9_yv12_copy_partial_frame_ptr)(YV12_BUFFER_CONFIG *src_ybc, YV12_BUFFER_CONFIG *dst_ybc, int Fraction);
+extern void vp9_yv12_copy_partial_frame(YV12_BUFFER_CONFIG *src_ybc, YV12_BUFFER_CONFIG *dst_ybc, int Fraction);
 
-void vp8_cmachine_specific_config(VP8_COMP *cpi) {
+void vp9_cmachine_specific_config(VP8_COMP *cpi) {
 #if CONFIG_RUNTIME_CPU_DETECT
   cpi->rtcd.common                    = &cpi->common.rtcd;
 
-  cpi->rtcd.search.full_search             = vp8_full_search_sad;
-  cpi->rtcd.search.refining_search         = vp8_refining_search_sad;
-  cpi->rtcd.search.diamond_search          = vp8_diamond_search_sad;
-  cpi->rtcd.temporal.apply                 = vp8_temporal_filter_apply_c;
+  cpi->rtcd.search.full_search             = vp9_full_search_sad;
+  cpi->rtcd.search.refining_search         = vp9_refining_search_sad;
+  cpi->rtcd.search.diamond_search          = vp9_diamond_search_sad;
+  cpi->rtcd.temporal.apply                 = vp9_temporal_filter_apply_c;
 #endif
 
-  vp8_yv12_copy_partial_frame_ptr = vp8_yv12_copy_partial_frame;
+  vp9_yv12_copy_partial_frame_ptr = vp9_yv12_copy_partial_frame;
 
 #if ARCH_X86 || ARCH_X86_64
-  vp8_arch_x86_encoder_init(cpi);
+  vp9_arch_x86_encoder_init(cpi);
 #endif
 
 #if ARCH_ARM
diff --git a/vp8/encoder/lookahead.c b/vp8/encoder/lookahead.c
index ae21f9729054f9ebddede50051e12d979ce484f0..260c84ea6e728b63a8f47d119176aca178cdb623 100644
--- a/vp8/encoder/lookahead.c
+++ b/vp8/encoder/lookahead.c
@@ -40,7 +40,7 @@ pop(struct lookahead_ctx *ctx,
 
 
 void
-vp8_lookahead_destroy(struct lookahead_ctx *ctx) {
+vp9_lookahead_destroy(struct lookahead_ctx *ctx) {
   if (ctx) {
     if (ctx->buf) {
       int i;
@@ -55,7 +55,7 @@ vp8_lookahead_destroy(struct lookahead_ctx *ctx) {
 
 
 struct lookahead_ctx *
-vp8_lookahead_init(unsigned int width,
+vp9_lookahead_init(unsigned int width,
                    unsigned int height,
                    unsigned int depth) {
   struct lookahead_ctx *ctx = NULL;
@@ -85,13 +85,13 @@ vp8_lookahead_init(unsigned int width,
   }
   return ctx;
 bail:
-  vp8_lookahead_destroy(ctx);
+  vp9_lookahead_destroy(ctx);
   return NULL;
 }
 
 
 int
-vp8_lookahead_push(struct lookahead_ctx *ctx,
+vp9_lookahead_push(struct lookahead_ctx *ctx,
                    YV12_BUFFER_CONFIG   *src,
                    int64_t               ts_start,
                    int64_t               ts_end,
@@ -157,7 +157,7 @@ vp8_lookahead_push(struct lookahead_ctx *ctx,
 
 
 struct lookahead_entry *
-vp8_lookahead_pop(struct lookahead_ctx *ctx,
+vp9_lookahead_pop(struct lookahead_ctx *ctx,
                   int                   drain) {
   struct lookahead_entry *buf = NULL;
 
@@ -170,7 +170,7 @@ vp8_lookahead_pop(struct lookahead_ctx *ctx,
 
 
 struct lookahead_entry *
-vp8_lookahead_peek(struct lookahead_ctx *ctx,
+vp9_lookahead_peek(struct lookahead_ctx *ctx,
                    int                   index) {
   struct lookahead_entry *buf = NULL;
 
@@ -186,6 +186,6 @@ vp8_lookahead_peek(struct lookahead_ctx *ctx,
 
 
 unsigned int
-vp8_lookahead_depth(struct lookahead_ctx *ctx) {
+vp9_lookahead_depth(struct lookahead_ctx *ctx) {
   return ctx->sz;
 }
diff --git a/vp8/encoder/lookahead.h b/vp8/encoder/lookahead.h
index 3c7abf90185f8fbeb9fde25ba95a1891870552b5..44c31d66da1fcf3eed979f6dc6ccf72c15e4e290 100644
--- a/vp8/encoder/lookahead.h
+++ b/vp8/encoder/lookahead.h
@@ -29,7 +29,7 @@ struct lookahead_ctx;
  *
  *
  */
-struct lookahead_ctx *vp8_lookahead_init(unsigned int width,
+struct lookahead_ctx *vp9_lookahead_init(unsigned int width,
                                          unsigned int height,
                                          unsigned int depth
                                         );
@@ -38,7 +38,7 @@ struct lookahead_ctx *vp8_lookahead_init(unsigned int width,
 /**\brief Destroys the lookahead stage
  *
  */
-void vp8_lookahead_destroy(struct lookahead_ctx *ctx);
+void vp9_lookahead_destroy(struct lookahead_ctx *ctx);
 
 
 /**\brief Enqueue a source buffer
@@ -57,7 +57,7 @@ void vp8_lookahead_destroy(struct lookahead_ctx *ctx);
  * \param[in] active_map  Map that specifies which macroblock is active
  */
 int
-vp8_lookahead_push(struct lookahead_ctx *ctx,
+vp9_lookahead_push(struct lookahead_ctx *ctx,
                    YV12_BUFFER_CONFIG   *src,
                    int64_t               ts_start,
                    int64_t               ts_end,
@@ -77,7 +77,7 @@ vp8_lookahead_push(struct lookahead_ctx *ctx,
  *
  */
 struct lookahead_entry *
-vp8_lookahead_pop(struct lookahead_ctx *ctx,
+vp9_lookahead_pop(struct lookahead_ctx *ctx,
                   int                   drain);
 
 
@@ -90,7 +90,7 @@ vp8_lookahead_pop(struct lookahead_ctx *ctx,
  *
  */
 struct lookahead_entry *
-vp8_lookahead_peek(struct lookahead_ctx *ctx,
+vp9_lookahead_peek(struct lookahead_ctx *ctx,
                    int                   index);
 
 
@@ -99,7 +99,7 @@ vp8_lookahead_peek(struct lookahead_ctx *ctx,
  * \param[in] ctx       Pointer to the lookahead context
  */
 unsigned int
-vp8_lookahead_depth(struct lookahead_ctx *ctx);
+vp9_lookahead_depth(struct lookahead_ctx *ctx);
 
 
 #endif
diff --git a/vp8/encoder/mbgraph.c b/vp8/encoder/mbgraph.c
index f72eac060c57ce2b83b6ac904e947bc20a022db9..a13b66b270a2812d08beed9c4553dfb6492c1603 100644
--- a/vp8/encoder/mbgraph.c
+++ b/vp8/encoder/mbgraph.c
@@ -47,13 +47,13 @@ static unsigned int do_16x16_motion_iteration
     further_steps = 0;
   }
 
-  vp8_clamp_mv_min_max(x, ref_mv);
+  vp9_clamp_mv_min_max(x, ref_mv);
 
   ref_full.as_mv.col = ref_mv->as_mv.col >> 3;
   ref_full.as_mv.row = ref_mv->as_mv.row >> 3;
 
   /*cpi->sf.search_method == HEX*/
-  best_err = vp8_hex_search(
+  best_err = vp9_hex_search(
       x, b, d,
       &ref_full, dst_mv,
       step_param,
@@ -81,9 +81,9 @@ static unsigned int do_16x16_motion_iteration
   xd->mode_info_context->mbmi.pred_filter_enabled = 0;
 #endif
 
-  vp8_set_mbmode_and_mvs(x, NEWMV, dst_mv);
+  vp9_set_mbmode_and_mvs(x, NEWMV, dst_mv);
   vp8_build_1st_inter16x16_predictors_mby(xd, xd->predictor, 16, 0);
-  best_err = vp8_sad16x16(xd->dst.y_buffer, xd->dst.y_stride,
+  best_err = vp9_sad16x16(xd->dst.y_buffer, xd->dst.y_stride,
                           xd->predictor, 16, INT_MAX);
 
   /* restore UMV window */
@@ -128,7 +128,7 @@ static int do_16x16_motion_search
   // FIXME should really use something like near/nearest MV and/or MV prediction
   xd->pre.y_buffer = ref->y_buffer + mb_y_offset;
   xd->pre.y_stride = ref->y_stride;
-  err = vp8_sad16x16(ref->y_buffer + mb_y_offset, ref->y_stride,
+  err = vp9_sad16x16(ref->y_buffer + mb_y_offset, ref->y_stride,
                      xd->dst.y_buffer, xd->dst.y_stride, INT_MAX);
   dst_mv->as_int = 0;
 
@@ -188,7 +188,7 @@ static int do_16x16_zerozero_search
   xd->pre.y_buffer = ref->y_buffer + mb_y_offset;
   xd->pre.y_stride = ref->y_stride;
   // VARIANCE_INVOKE(&cpi->rtcd.variance, satd16x16)
-  err = vp8_sad16x16(ref->y_buffer + mb_y_offset, ref->y_stride,
+  err = vp9_sad16x16(ref->y_buffer + mb_y_offset, ref->y_stride,
                      xd->dst.y_buffer, xd->dst.y_stride, INT_MAX);
 
   dst_mv->as_int = 0;
@@ -214,7 +214,7 @@ static int find_best_16x16_intra
 
     xd->mode_info_context->mbmi.mode = mode;
     vp8_build_intra_predictors_mby(xd);
-    err = vp8_sad16x16(xd->predictor, 16, buf->y_buffer + mb_y_offset,
+    err = vp9_sad16x16(xd->predictor, 16, buf->y_buffer + mb_y_offset,
                        buf->y_stride, best_err);
     // find best
     if (err < best_err) {
@@ -428,22 +428,22 @@ static void separate_arf_mbs(VP8_COMP *cpi) {
       cpi->static_mb_pct = 0;
 
     cpi->seg0_cnt = ncnt[0];
-    vp8_enable_segmentation((VP8_PTR) cpi);
+    vp9_enable_segmentation((VP8_PTR) cpi);
   } else {
     cpi->static_mb_pct = 0;
-    vp8_disable_segmentation((VP8_PTR) cpi);
+    vp9_disable_segmentation((VP8_PTR) cpi);
   }
 
   // Free localy allocated storage
   vpx_free(arf_not_zz);
 }
 
-void vp8_update_mbgraph_stats
+void vp9_update_mbgraph_stats
 (
   VP8_COMP *cpi
 ) {
   VP8_COMMON *const cm = &cpi->common;
-  int i, n_frames = vp8_lookahead_depth(cpi->lookahead);
+  int i, n_frames = vp9_lookahead_depth(cpi->lookahead);
   YV12_BUFFER_CONFIG *golden_ref = &cm->yv12_fb[cm->gld_fb_idx];
 
   // we need to look ahead beyond where the ARF transitions into
@@ -469,7 +469,7 @@ void vp8_update_mbgraph_stats
   for (i = 0; i < n_frames; i++) {
     MBGRAPH_FRAME_STATS *frame_stats = &cpi->mbgraph_stats[i];
     struct lookahead_entry *q_cur =
-      vp8_lookahead_peek(cpi->lookahead, i);
+      vp9_lookahead_peek(cpi->lookahead, i);
 
     assert(q_cur != NULL);
 
diff --git a/vp8/encoder/mbgraph.h b/vp8/encoder/mbgraph.h
index 11ff3e2f06577fddcb225d345c91a515ae6b2ce7..757b2754d3ea9d60d8cf098bf57022520af4815b 100644
--- a/vp8/encoder/mbgraph.h
+++ b/vp8/encoder/mbgraph.h
@@ -11,6 +11,6 @@
 #ifndef __INC_MBGRAPH_H__
 #define __INC_MBGRAPH_H__ 1
 
-extern void vp8_update_mbgraph_stats(VP8_COMP *cpi);
+extern void vp9_update_mbgraph_stats(VP8_COMP *cpi);
 
 #endif /* __INC_MBGRAPH_H__ */
diff --git a/vp8/encoder/mcomp.c b/vp8/encoder/mcomp.c
index 210887491cbe860069652cc518889843d7c46788..ad9d966fb75883e5eb5ab4ce516b912f94757154 100644
--- a/vp8/encoder/mcomp.c
+++ b/vp8/encoder/mcomp.c
@@ -23,7 +23,7 @@ static int mv_ref_ct [31] [4] [2];
 static int mv_mode_cts [4] [2];
 #endif
 
-void vp8_clamp_mv_min_max(MACROBLOCK *x, int_mv *ref_mv) {
+void vp9_clamp_mv_min_max(MACROBLOCK *x, int_mv *ref_mv) {
   int col_min = (ref_mv->as_mv.col >> 3) - MAX_FULL_PEL_VAL +
       ((ref_mv->as_mv.col & 7) ? 1 : 0);
   int row_min = (ref_mv->as_mv.row >> 3) - MAX_FULL_PEL_VAL +
@@ -42,7 +42,7 @@ void vp8_clamp_mv_min_max(MACROBLOCK *x, int_mv *ref_mv) {
     x->mv_row_max = row_max;
 }
 
-int vp8_mv_bit_cost(int_mv *mv, int_mv *ref, DEC_MVCOSTS,
+int vp9_mv_bit_cost(int_mv *mv, int_mv *ref, DEC_MVCOSTS,
                     int Weight, int ishp) {
   MV v;
   v.row = (mv->as_mv.row - ref->as_mv.row);
@@ -79,7 +79,7 @@ static int mvsad_err_cost(int_mv *mv, int_mv *ref, DEC_MVSADCOSTS,
   return 0;
 }
 
-void vp8_init_dsmotion_compensation(MACROBLOCK *x, int stride) {
+void vp9_init_dsmotion_compensation(MACROBLOCK *x, int stride) {
   int Len;
   int search_site_count = 0;
 
@@ -125,7 +125,7 @@ void vp8_init_dsmotion_compensation(MACROBLOCK *x, int stride) {
   x->searches_per_step = 4;
 }
 
-void vp8_init3smotion_compensation(MACROBLOCK *x, int stride) {
+void vp9_init3smotion_compensation(MACROBLOCK *x, int stride) {
   int Len;
   int search_site_count = 0;
 
@@ -244,7 +244,7 @@ void vp8_init3smotion_compensation(MACROBLOCK *x, int stride) {
 #define MIN(x,y) (((x)<(y))?(x):(y))
 #define MAX(x,y) (((x)>(y))?(x):(y))
 
-int vp8_find_best_sub_pixel_step_iteratively(MACROBLOCK *x, BLOCK *b, BLOCKD *d,
+int vp9_find_best_sub_pixel_step_iteratively(MACROBLOCK *x, BLOCK *b, BLOCKD *d,
                                              int_mv *bestmv, int_mv *ref_mv,
                                              int error_per_bit,
                                              const vp8_variance_fn_ptr_t *vfp,
@@ -443,7 +443,7 @@ int vp8_find_best_sub_pixel_step_iteratively(MACROBLOCK *x, BLOCK *b, BLOCKD *d,
 #undef MIN
 #undef MAX
 
-int vp8_find_best_sub_pixel_step(MACROBLOCK *x, BLOCK *b, BLOCKD *d,
+int vp9_find_best_sub_pixel_step(MACROBLOCK *x, BLOCK *b, BLOCKD *d,
                                  int_mv *bestmv, int_mv *ref_mv,
                                  int error_per_bit,
                                  const vp8_variance_fn_ptr_t *vfp,
@@ -923,7 +923,7 @@ int vp8_find_best_sub_pixel_step(MACROBLOCK *x, BLOCK *b, BLOCKD *d,
 
 #undef SP
 
-int vp8_find_best_half_pixel_step(MACROBLOCK *x, BLOCK *b, BLOCKD *d,
+int vp9_find_best_half_pixel_step(MACROBLOCK *x, BLOCK *b, BLOCKD *d,
                                   int_mv *bestmv, int_mv *ref_mv,
                                   int error_per_bit,
                                   const vp8_variance_fn_ptr_t *vfp,
@@ -1099,7 +1099,7 @@ static const MV next_chkpts[6][3] = {
   {{ -1, 2}, { -2, 0}, { -1, -2}}
 };
 
-int vp8_hex_search
+int vp9_hex_search
 (
   MACROBLOCK *x,
   BLOCK *b,
@@ -1255,7 +1255,7 @@ cal_neighbors:
 #undef CHECK_POINT
 #undef CHECK_BETTER
 
-int vp8_diamond_search_sad(MACROBLOCK *x, BLOCK *b, BLOCKD *d,
+int vp9_diamond_search_sad(MACROBLOCK *x, BLOCK *b, BLOCKD *d,
                            int_mv *ref_mv, int_mv *best_mv,
                            int search_param, int sad_per_bit, int *num00,
                            vp8_variance_fn_ptr_t *fn_ptr, DEC_MVCOSTS,
@@ -1364,7 +1364,7 @@ int vp8_diamond_search_sad(MACROBLOCK *x, BLOCK *b, BLOCKD *d,
                   xd->allow_high_precision_mv);
 }
 
-int vp8_diamond_search_sadx4(MACROBLOCK *x, BLOCK *b, BLOCKD *d,
+int vp9_diamond_search_sadx4(MACROBLOCK *x, BLOCK *b, BLOCKD *d,
                              int_mv *ref_mv, int_mv *best_mv, int search_param,
                              int sad_per_bit, int *num00,
                              vp8_variance_fn_ptr_t *fn_ptr,
@@ -1512,7 +1512,7 @@ int vp8_diamond_search_sadx4(MACROBLOCK *x, BLOCK *b, BLOCKD *d,
 /* do_refine: If last step (1-away) of n-step search doesn't pick the center
               point as the best match, we will do a final 1-away diamond
               refining search  */
-int vp8_full_pixel_diamond(VP8_COMP *cpi, MACROBLOCK *x, BLOCK *b,
+int vp9_full_pixel_diamond(VP8_COMP *cpi, MACROBLOCK *x, BLOCK *b,
                            BLOCKD *d, int_mv *mvp_full, int step_param,
                            int sadpb, int further_steps,
                            int do_refine, vp8_variance_fn_ptr_t *fn_ptr,
@@ -1568,7 +1568,7 @@ int vp8_full_pixel_diamond(VP8_COMP *cpi, MACROBLOCK *x, BLOCK *b,
   return bestsme;
 }
 
-int vp8_full_search_sad(MACROBLOCK *x, BLOCK *b, BLOCKD *d, int_mv *ref_mv,
+int vp9_full_search_sad(MACROBLOCK *x, BLOCK *b, BLOCKD *d, int_mv *ref_mv,
                         int sad_per_bit, int distance,
                         vp8_variance_fn_ptr_t *fn_ptr, DEC_MVCOSTS,
                         int_mv *center_mv) {
@@ -1662,7 +1662,7 @@ int vp8_full_search_sad(MACROBLOCK *x, BLOCK *b, BLOCKD *d, int_mv *ref_mv,
     return INT_MAX;
 }
 
-int vp8_full_search_sadx3(MACROBLOCK *x, BLOCK *b, BLOCKD *d, int_mv *ref_mv,
+int vp9_full_search_sadx3(MACROBLOCK *x, BLOCK *b, BLOCKD *d, int_mv *ref_mv,
                           int sad_per_bit, int distance,
                           vp8_variance_fn_ptr_t *fn_ptr, DEC_MVCOSTS,
                           int_mv *center_mv) {
@@ -1789,7 +1789,7 @@ int vp8_full_search_sadx3(MACROBLOCK *x, BLOCK *b, BLOCKD *d, int_mv *ref_mv,
     return INT_MAX;
 }
 
-int vp8_full_search_sadx8(MACROBLOCK *x, BLOCK *b, BLOCKD *d, int_mv *ref_mv,
+int vp9_full_search_sadx8(MACROBLOCK *x, BLOCK *b, BLOCKD *d, int_mv *ref_mv,
                           int sad_per_bit, int distance,
                           vp8_variance_fn_ptr_t *fn_ptr,
                           DEC_MVCOSTS,
@@ -1943,7 +1943,7 @@ int vp8_full_search_sadx8(MACROBLOCK *x, BLOCK *b, BLOCKD *d, int_mv *ref_mv,
     return INT_MAX;
 }
 
-int vp8_refining_search_sad(MACROBLOCK *x, BLOCK *b, BLOCKD *d, int_mv *ref_mv,
+int vp9_refining_search_sad(MACROBLOCK *x, BLOCK *b, BLOCKD *d, int_mv *ref_mv,
                             int error_per_bit, int search_range,
                             vp8_variance_fn_ptr_t *fn_ptr, DEC_MVCOSTS,
                             int_mv *center_mv) {
@@ -2019,7 +2019,7 @@ int vp8_refining_search_sad(MACROBLOCK *x, BLOCK *b, BLOCKD *d, int_mv *ref_mv,
     return INT_MAX;
 }
 
-int vp8_refining_search_sadx4(MACROBLOCK *x, BLOCK *b, BLOCKD *d,
+int vp9_refining_search_sadx4(MACROBLOCK *x, BLOCK *b, BLOCKD *d,
                               int_mv *ref_mv, int error_per_bit,
                               int search_range, vp8_variance_fn_ptr_t *fn_ptr,
                               DEC_MVCOSTS, int_mv *center_mv) {
diff --git a/vp8/encoder/mcomp.h b/vp8/encoder/mcomp.h
index f0910692789730c03ccd7008ebfb74c3722b0d47..1c6e322e045db74a52f9b78de29617a07cd398b0 100644
--- a/vp8/encoder/mcomp.h
+++ b/vp8/encoder/mcomp.h
@@ -32,20 +32,20 @@ extern void accum_mv_refs(MB_PREDICTION_MODE, const int near_mv_ref_cts[4]);
 #define MAX_FULL_PEL_VAL ((1 << (MAX_MVSEARCH_STEPS)) - 1)      // Max full pel mv specified in 1 pel units
 #define MAX_FIRST_STEP (1 << (MAX_MVSEARCH_STEPS-1))            // Maximum size of the first step in full pel units
 
-extern void vp8_clamp_mv_min_max(MACROBLOCK *x, int_mv *ref_mv);
-extern int vp8_mv_bit_cost(int_mv *mv, int_mv *ref, DEC_MVCOSTS,
+extern void vp9_clamp_mv_min_max(MACROBLOCK *x, int_mv *ref_mv);
+extern int vp9_mv_bit_cost(int_mv *mv, int_mv *ref, DEC_MVCOSTS,
                            int Weight, int ishp);
-extern void vp8_init_dsmotion_compensation(MACROBLOCK *x, int stride);
-extern void vp8_init3smotion_compensation(MACROBLOCK *x,  int stride);
+extern void vp9_init_dsmotion_compensation(MACROBLOCK *x, int stride);
+extern void vp9_init3smotion_compensation(MACROBLOCK *x,  int stride);
 // Runs sequence of diamond searches in smaller steps for RD
 struct VP8_COMP;
-int vp8_full_pixel_diamond(struct VP8_COMP *cpi, MACROBLOCK *x, BLOCK *b,
+int vp9_full_pixel_diamond(struct VP8_COMP *cpi, MACROBLOCK *x, BLOCK *b,
                            BLOCKD *d, int_mv *mvp_full, int step_param,
                            int sadpb, int further_steps, int do_refine,
                            vp8_variance_fn_ptr_t *fn_ptr,
                            int_mv *ref_mv, int_mv *dst_mv);
 
-extern int vp8_hex_search
+extern int vp9_hex_search
 (
   MACROBLOCK *x,
   BLOCK *b,
@@ -64,9 +64,9 @@ typedef int (fractional_mv_step_fp)
 (MACROBLOCK *x, BLOCK *b, BLOCKD *d, int_mv *bestmv, int_mv *ref_mv,
  int error_per_bit, const vp8_variance_fn_ptr_t *vfp, DEC_MVCOSTS,
  int *distortion, unsigned int *sse);
-extern fractional_mv_step_fp vp8_find_best_sub_pixel_step_iteratively;
-extern fractional_mv_step_fp vp8_find_best_sub_pixel_step;
-extern fractional_mv_step_fp vp8_find_best_half_pixel_step;
+extern fractional_mv_step_fp vp9_find_best_sub_pixel_step_iteratively;
+extern fractional_mv_step_fp vp9_find_best_sub_pixel_step;
+extern fractional_mv_step_fp vp9_find_best_half_pixel_step;
 extern fractional_mv_step_fp vp8_skip_fractional_mv_step;
 
 #define prototype_full_search_sad(sym)\
@@ -118,30 +118,30 @@ extern fractional_mv_step_fp vp8_skip_fractional_mv_step;
 #endif
 
 typedef prototype_full_search_sad(*vp8_full_search_fn_t);
-extern prototype_full_search_sad(vp8_full_search_sad);
-extern prototype_full_search_sad(vp8_full_search_sadx3);
-extern prototype_full_search_sad(vp8_full_search_sadx8);
+extern prototype_full_search_sad(vp9_full_search_sad);
+extern prototype_full_search_sad(vp9_full_search_sadx3);
+extern prototype_full_search_sad(vp9_full_search_sadx8);
 
 typedef prototype_refining_search_sad(*vp8_refining_search_fn_t);
-extern prototype_refining_search_sad(vp8_refining_search_sad);
-extern prototype_refining_search_sad(vp8_refining_search_sadx4);
+extern prototype_refining_search_sad(vp9_refining_search_sad);
+extern prototype_refining_search_sad(vp9_refining_search_sadx4);
 
 typedef prototype_diamond_search_sad(*vp8_diamond_search_fn_t);
-extern prototype_diamond_search_sad(vp8_diamond_search_sad);
-extern prototype_diamond_search_sad(vp8_diamond_search_sadx4);
+extern prototype_diamond_search_sad(vp9_diamond_search_sad);
+extern prototype_diamond_search_sad(vp9_diamond_search_sadx4);
 
 #ifndef vp8_search_full_search
-#define vp8_search_full_search vp8_full_search_sad
+#define vp8_search_full_search vp9_full_search_sad
 #endif
 extern prototype_full_search_sad(vp8_search_full_search);
 
 #ifndef vp8_search_refining_search
-#define vp8_search_refining_search vp8_refining_search_sad
+#define vp8_search_refining_search vp9_refining_search_sad
 #endif
 extern prototype_refining_search_sad(vp8_search_refining_search);
 
 #ifndef vp8_search_diamond_search
-#define vp8_search_diamond_search vp8_diamond_search_sad
+#define vp8_search_diamond_search vp9_diamond_search_sad
 #endif
 extern prototype_diamond_search_sad(vp8_search_diamond_search);
 
diff --git a/vp8/encoder/modecosts.c b/vp8/encoder/modecosts.c
index 23b9973c34f27b64e1d78160d1c808b729219916..f944261bc7f30e5ac5805eec24eaffb46b5f9f16 100644
--- a/vp8/encoder/modecosts.c
+++ b/vp8/encoder/modecosts.c
@@ -15,7 +15,7 @@
 #include "vp8/common/entropymode.h"
 
 
-void vp8_init_mode_costs(VP8_COMP *c) {
+void vp9_init_mode_costs(VP8_COMP *c) {
   VP8_COMMON *x = &c->common;
   {
     const vp8_tree_p T = vp8_bmode_tree;
@@ -26,30 +26,30 @@ void vp8_init_mode_costs(VP8_COMP *c) {
       int j = 0;
 
       do {
-        vp8_cost_tokens((int *)c->mb.bmode_costs[i][j], x->kf_bmode_prob[i][j], T);
+        vp9_cost_tokens((int *)c->mb.bmode_costs[i][j], x->kf_bmode_prob[i][j], T);
       } while (++j < VP8_BINTRAMODES);
     } while (++i < VP8_BINTRAMODES);
 
-    vp8_cost_tokens((int *)c->mb.inter_bmode_costs, x->fc.bmode_prob, T);
+    vp9_cost_tokens((int *)c->mb.inter_bmode_costs, x->fc.bmode_prob, T);
   }
-  vp8_cost_tokens((int *)c->mb.inter_bmode_costs,
+  vp9_cost_tokens((int *)c->mb.inter_bmode_costs,
                   x->fc.sub_mv_ref_prob[0], vp8_sub_mv_ref_tree);
 
-  vp8_cost_tokens(c->mb.mbmode_cost[1], x->fc.ymode_prob, vp8_ymode_tree);
-  vp8_cost_tokens(c->mb.mbmode_cost[0],
+  vp9_cost_tokens(c->mb.mbmode_cost[1], x->fc.ymode_prob, vp8_ymode_tree);
+  vp9_cost_tokens(c->mb.mbmode_cost[0],
                   x->kf_ymode_prob[c->common.kf_ymode_probs_index],
                   vp8_kf_ymode_tree);
-  vp8_cost_tokens(c->mb.intra_uv_mode_cost[1],
+  vp9_cost_tokens(c->mb.intra_uv_mode_cost[1],
                   x->fc.uv_mode_prob[VP8_YMODES - 1], vp8_uv_mode_tree);
-  vp8_cost_tokens(c->mb.intra_uv_mode_cost[0],
+  vp9_cost_tokens(c->mb.intra_uv_mode_cost[0],
                   x->kf_uv_mode_prob[VP8_YMODES - 1], vp8_uv_mode_tree);
-  vp8_cost_tokens(c->mb.i8x8_mode_costs,
+  vp9_cost_tokens(c->mb.i8x8_mode_costs,
                   x->fc.i8x8_mode_prob, vp8_i8x8_mode_tree);
 
   {
     int i;
     for (i = 0; i <= VP8_SWITCHABLE_FILTERS; ++i)
-      vp8_cost_tokens((int *)c->mb.switchable_interp_costs[i],
+      vp9_cost_tokens((int *)c->mb.switchable_interp_costs[i],
                       x->fc.switchable_interp_prob[i],
                       vp8_switchable_interp_tree);
   }
diff --git a/vp8/encoder/modecosts.h b/vp8/encoder/modecosts.h
index 99ef119d5ddc8a7927c0ba24f191ad0d426e172f..88dc3fc79466225f1a6fc918734a46279d9ad95b 100644
--- a/vp8/encoder/modecosts.h
+++ b/vp8/encoder/modecosts.h
@@ -12,6 +12,6 @@
 #ifndef __INC_MODECOSTS_H
 #define __INC_MODECOSTS_H
 
-void vp8_init_mode_costs(VP8_COMP *x);
+void vp9_init_mode_costs(VP8_COMP *x);
 
 #endif
diff --git a/vp8/encoder/onyx_if.c b/vp8/encoder/onyx_if.c
index 4e7123cd33908515d7bb3ce7f96c6b4b814de2e1..4667b06ed6d46ec836d886edd34b49607c266ae8 100644
--- a/vp8/encoder/onyx_if.c
+++ b/vp8/encoder/onyx_if.c
@@ -60,11 +60,11 @@
 #define RTCD(x) NULL
 #endif
 
-extern void vp8cx_pick_filter_level_fast(YV12_BUFFER_CONFIG *sd, VP8_COMP *cpi);
-extern void vp8cx_set_alt_lf_level(VP8_COMP *cpi, int filt_val);
-extern void vp8cx_pick_filter_level(YV12_BUFFER_CONFIG *sd, VP8_COMP *cpi);
+extern void vp9cx_pick_filter_level_fast(YV12_BUFFER_CONFIG *sd, VP8_COMP *cpi);
+extern void vp9cx_set_alt_lf_level(VP8_COMP *cpi, int filt_val);
+extern void vp9cx_pick_filter_level(YV12_BUFFER_CONFIG *sd, VP8_COMP *cpi);
 
-extern void vp8_cmachine_specific_config(VP8_COMP *cpi);
+extern void vp9_cmachine_specific_config(VP8_COMP *cpi);
 extern void vp8_deblock_frame(YV12_BUFFER_CONFIG *source, YV12_BUFFER_CONFIG *post, int filt_lvl, int low_var_thresh, int flag);
 extern void print_parms(VP8_CONFIG *ocf, char *filenam);
 extern unsigned int vp8_get_processor_freq();
@@ -76,9 +76,9 @@ extern void vp8_yv12_copy_frame_func_neon(YV12_BUFFER_CONFIG *src_ybc, YV12_BUFF
 extern void vp8_yv12_copy_src_frame_func_neon(YV12_BUFFER_CONFIG *src_ybc, YV12_BUFFER_CONFIG *dst_ybc);
 #endif
 
-int vp8_calc_ss_err(YV12_BUFFER_CONFIG *source, YV12_BUFFER_CONFIG *dest);
+int vp9_calc_ss_err(YV12_BUFFER_CONFIG *source, YV12_BUFFER_CONFIG *dest);
 
-extern void vp8_temporal_filter_prepare_c(VP8_COMP *cpi, int distance);
+extern void vp9_temporal_filter_prepare_c(VP8_COMP *cpi, int distance);
 
 static void set_default_lf_deltas(VP8_COMP *cpi);
 
@@ -101,12 +101,12 @@ extern const int vp8_gf_interval_table[101];
 #if CONFIG_INTERNAL_STATS
 #include "math.h"
 
-extern double vp8_calc_ssim(YV12_BUFFER_CONFIG *source,
+extern double vp9_calc_ssim(YV12_BUFFER_CONFIG *source,
                             YV12_BUFFER_CONFIG *dest, int lumamask,
                             double *weight);
 
 
-extern double vp8_calc_ssimg(YV12_BUFFER_CONFIG *source,
+extern double vp9_calc_ssimg(YV12_BUFFER_CONFIG *source,
                              YV12_BUFFER_CONFIG *dest, double *ssim_y,
                              double *ssim_u, double *ssim_v);
 
@@ -166,7 +166,7 @@ extern unsigned int inter_uv_modes[VP8_UV_MODES];
 extern unsigned int inter_b_modes[B_MODE_COUNT];
 #endif
 
-extern void vp8cx_init_quantizer(VP8_COMP *cpi);
+extern void vp9cx_init_quantizer(VP8_COMP *cpi);
 
 int vp8cx_base_skip_false_prob[QINDEX_RANGE][3];
 
@@ -196,8 +196,8 @@ static int calculate_minq_index(double maxq,
     minqtarget = maxq;
 
   for (i = 0; i < QINDEX_RANGE; i++) {
-    thisq = vp8_convert_qindex_to_q(i);
-    if (minqtarget <= vp8_convert_qindex_to_q(i))
+    thisq = vp9_convert_qindex_to_q(i);
+    if (minqtarget <= vp9_convert_qindex_to_q(i))
       return i;
   }
   return QINDEX_RANGE - 1;
@@ -208,7 +208,7 @@ static void init_minq_luts(void) {
   double maxq;
 
   for (i = 0; i < QINDEX_RANGE; i++) {
-    maxq = vp8_convert_qindex_to_q(i);
+    maxq = vp9_convert_qindex_to_q(i);
 
 
     kf_low_motion_minq[i] = calculate_minq_index(maxq,
@@ -246,7 +246,7 @@ static void init_base_skip_probs(void) {
   int skip_prob, t;
 
   for (i = 0; i < QINDEX_RANGE; i++) {
-    q = vp8_convert_qindex_to_q(i);
+    q = vp9_convert_qindex_to_q(i);
 
     // Exponential decay caluclation of baseline skip prob with clamping
     // Based on crude best fit of old table.
@@ -306,15 +306,15 @@ static void update_base_skip_probs(VP8_COMP *cpi) {
 
 }
 
-void vp8_initialize() {
+void vp9_initialize() {
   static int init_done = 0;
 
   if (!init_done) {
     vp8_scale_machine_specific_config();
     vp8_initialize_common();
-    vp8_tokenize_initialize();
+    vp9_tokenize_initialize();
     vp8_init_quant_tables();
-    vp8_init_me_luts();
+    vp9_init_me_luts();
     init_minq_luts();
     init_base_skip_probs();
     init_done = 1;
@@ -381,7 +381,7 @@ static void dealloc_compressor_data(VP8_COMP *cpi) {
 #if VP8_TEMPORAL_ALT_REF
   vp8_yv12_de_alloc_frame_buffer(&cpi->alt_ref_buffer);
 #endif
-  vp8_lookahead_destroy(cpi->lookahead);
+  vp9_lookahead_destroy(cpi->lookahead);
 
   vpx_free(cpi->tok);
   cpi->tok = 0;
@@ -420,14 +420,14 @@ static int compute_qdelta(VP8_COMP *cpi, double qstart, double qtarget) {
   // Convert the average q value to an index.
   for (i = cpi->best_quality; i < cpi->worst_quality; i++) {
     start_index = i;
-    if (vp8_convert_qindex_to_q(i) >= qstart)
+    if (vp9_convert_qindex_to_q(i) >= qstart)
       break;
   }
 
   // Convert the q target to an index
   for (i = cpi->best_quality; i < cpi->worst_quality; i++) {
     target_index = i;
-    if (vp8_convert_qindex_to_q(i) >= qtarget)
+    if (vp9_convert_qindex_to_q(i) >= qtarget)
       break;
   }
 
@@ -450,7 +450,7 @@ static void init_seg_features(VP8_COMP *cpi) {
     cpi->static_mb_pct = 0;
 
     // Disable segmentation
-    vp8_disable_segmentation((VP8_PTR)cpi);
+    vp9_disable_segmentation((VP8_PTR)cpi);
 
     // Clear down the segment features.
     vp9_clearall_segfeatures(xd);
@@ -465,12 +465,12 @@ static void init_seg_features(VP8_COMP *cpi) {
     cpi->static_mb_pct = 0;
 
     // Disable segmentation and individual segment features by default
-    vp8_disable_segmentation((VP8_PTR)cpi);
+    vp9_disable_segmentation((VP8_PTR)cpi);
     vp9_clearall_segfeatures(xd);
 
     // Scan frames from current to arf frame.
     // This function re-enables segmentation if appropriate.
-    vp8_update_mbgraph_stats(cpi);
+    vp9_update_mbgraph_stats(cpi);
 
     // If segmentation was enabled set those features needed for the
     // arf itself.
@@ -526,7 +526,7 @@ static void init_seg_features(VP8_COMP *cpi) {
       // Disable segmentation and clear down features if alt ref
       // is not active for this group
       else {
-        vp8_disable_segmentation((VP8_PTR)cpi);
+        vp9_disable_segmentation((VP8_PTR)cpi);
 
         vpx_memset(cpi->segmentation_map, 0,
                    (cm->mb_rows * cm->mb_cols));
@@ -660,7 +660,7 @@ static void set_default_lf_deltas(VP8_COMP *cpi) {
   cpi->mb.e_mbd.mode_lf_deltas[3] = 4;               // Split mv
 }
 
-void vp8_set_speed_features(VP8_COMP *cpi) {
+void vp9_set_speed_features(VP8_COMP *cpi) {
   SPEED_FEATURES *sf = &cpi->sf;
   int Mode = cpi->compressor_speed;
   int Speed = cpi->Speed;
@@ -1172,48 +1172,48 @@ void vp8_set_speed_features(VP8_COMP *cpi) {
   }
 
   if (cpi->sf.search_method == NSTEP) {
-    vp8_init3smotion_compensation(&cpi->mb, cm->yv12_fb[cm->lst_fb_idx].y_stride);
+    vp9_init3smotion_compensation(&cpi->mb, cm->yv12_fb[cm->lst_fb_idx].y_stride);
   } else if (cpi->sf.search_method == DIAMOND) {
-    vp8_init_dsmotion_compensation(&cpi->mb, cm->yv12_fb[cm->lst_fb_idx].y_stride);
+    vp9_init_dsmotion_compensation(&cpi->mb, cm->yv12_fb[cm->lst_fb_idx].y_stride);
   }
 
-  cpi->mb.vp8_short_fdct16x16 = vp8_short_fdct16x16;
-  cpi->mb.vp8_short_fdct8x8 = vp8_short_fdct8x8;
-  cpi->mb.vp8_short_fdct8x4 = vp8_short_fdct8x4;
-  cpi->mb.vp8_short_fdct4x4 = vp8_short_fdct4x4;
-  cpi->mb.short_walsh4x4 = vp8_short_walsh4x4;
-  cpi->mb.short_fhaar2x2 = vp8_short_fhaar2x2;
+  cpi->mb.vp9_short_fdct16x16 = vp9_short_fdct16x16;
+  cpi->mb.vp9_short_fdct8x8 = vp9_short_fdct8x8;
+  cpi->mb.vp9_short_fdct8x4 = vp9_short_fdct8x4;
+  cpi->mb.vp9_short_fdct4x4 = vp9_short_fdct4x4;
+  cpi->mb.short_walsh4x4 = vp9_short_walsh4x4;
+  cpi->mb.short_fhaar2x2 = vp9_short_fhaar2x2;
 
 #if CONFIG_LOSSLESS
   if (cpi->oxcf.lossless) {
-    cpi->mb.vp8_short_fdct8x4 = vp8_short_walsh8x4_x8;
-    cpi->mb.vp8_short_fdct4x4 = vp8_short_walsh4x4_x8;
-    cpi->mb.short_walsh4x4 = vp8_short_walsh4x4;
-    cpi->mb.short_fhaar2x2 = vp8_short_fhaar2x2;
-    cpi->mb.short_walsh4x4 = vp8_short_walsh4x4_lossless;
+    cpi->mb.vp9_short_fdct8x4 = vp9_short_walsh8x4_x8;
+    cpi->mb.vp9_short_fdct4x4 = vp9_short_walsh4x4_x8;
+    cpi->mb.short_walsh4x4 = vp9_short_walsh4x4;
+    cpi->mb.short_fhaar2x2 = vp9_short_fhaar2x2;
+    cpi->mb.short_walsh4x4 = vp9_short_walsh4x4_lossless;
   }
 #endif
 
 
 
-  cpi->mb.quantize_b_4x4      = vp8_regular_quantize_b_4x4;
-  cpi->mb.quantize_b_4x4_pair = vp8_regular_quantize_b_4x4_pair;
-  cpi->mb.quantize_b_8x8      = vp8_regular_quantize_b_8x8;
-  cpi->mb.quantize_b_16x16    = vp8_regular_quantize_b_16x16;
-  cpi->mb.quantize_b_2x2      = vp8_regular_quantize_b_2x2;
+  cpi->mb.quantize_b_4x4      = vp9_regular_quantize_b_4x4;
+  cpi->mb.quantize_b_4x4_pair = vp9_regular_quantize_b_4x4_pair;
+  cpi->mb.quantize_b_8x8      = vp9_regular_quantize_b_8x8;
+  cpi->mb.quantize_b_16x16    = vp9_regular_quantize_b_16x16;
+  cpi->mb.quantize_b_2x2      = vp9_regular_quantize_b_2x2;
 
-  vp8cx_init_quantizer(cpi);
+  vp9cx_init_quantizer(cpi);
 
 #if CONFIG_RUNTIME_CPU_DETECT
   cpi->mb.e_mbd.rtcd = &cpi->common.rtcd;
 #endif
 
   if (cpi->sf.iterative_sub_pixel == 1) {
-    cpi->find_fractional_mv_step = vp8_find_best_sub_pixel_step_iteratively;
+    cpi->find_fractional_mv_step = vp9_find_best_sub_pixel_step_iteratively;
   } else if (cpi->sf.quarter_pixel_search) {
-    cpi->find_fractional_mv_step = vp8_find_best_sub_pixel_step;
+    cpi->find_fractional_mv_step = vp9_find_best_sub_pixel_step;
   } else if (cpi->sf.half_pixel_search) {
-    cpi->find_fractional_mv_step = vp8_find_best_half_pixel_step;
+    cpi->find_fractional_mv_step = vp9_find_best_half_pixel_step;
   }
 
   if (cpi->sf.optimize_coefficients == 1 && cpi->pass != 1)
@@ -1229,7 +1229,7 @@ static void alloc_raw_frame_buffers(VP8_COMP *cpi) {
   int width = (cpi->oxcf.Width + 15) & ~15;
   int height = (cpi->oxcf.Height + 15) & ~15;
 
-  cpi->lookahead = vp8_lookahead_init(cpi->oxcf.Width, cpi->oxcf.Height,
+  cpi->lookahead = vp9_lookahead_init(cpi->oxcf.Width, cpi->oxcf.Height,
                                       cpi->oxcf.lag_in_frames);
   if (!cpi->lookahead)
     vpx_internal_error(&cpi->common.error, VPX_CODEC_MEM_ERROR,
@@ -1259,7 +1259,7 @@ static int vp8_alloc_partition_data(VP8_COMP *cpi) {
   return 0;
 }
 
-void vp8_alloc_compressor_data(VP8_COMP *cpi) {
+void vp9_alloc_compressor_data(VP8_COMP *cpi) {
   VP8_COMMON *cm = &cpi->common;
 
   int width = cm->Width;
@@ -1363,7 +1363,7 @@ static const int q_trans[] = {
   224, 228, 232, 236, 240, 244, 249, 255,
 };
 
-int vp8_reverse_trans(int x) {
+int vp9_reverse_trans(int x) {
   int i;
 
   for (i = 0; i < 64; i++)
@@ -1372,7 +1372,7 @@ int vp8_reverse_trans(int x) {
 
   return 63;
 };
-void vp8_new_frame_rate(VP8_COMP *cpi, double framerate) {
+void vp9_new_frame_rate(VP8_COMP *cpi, double framerate) {
   if (framerate < .1)
     framerate = 30;
 
@@ -1430,7 +1430,7 @@ static void init_config(VP8_PTR ptr, VP8_CONFIG *oxcf) {
   vp8_setup_version(cm);
 
   // change includes all joint functionality
-  vp8_change_config(ptr, oxcf);
+  vp9_change_config(ptr, oxcf);
 
   // Initialize active best and worst q and average q values.
   cpi->active_worst_quality         = cpi->oxcf.worst_allowed_q;
@@ -1464,7 +1464,7 @@ static void init_config(VP8_PTR ptr, VP8_CONFIG *oxcf) {
 }
 
 
-void vp8_change_config(VP8_PTR ptr, VP8_CONFIG *oxcf) {
+void vp9_change_config(VP8_PTR ptr, VP8_CONFIG *oxcf) {
   VP8_COMP *cpi = (VP8_COMP *)(ptr);
   VP8_COMMON *cm = &cpi->common;
 
@@ -1576,7 +1576,7 @@ void vp8_change_config(VP8_PTR ptr, VP8_CONFIG *oxcf) {
               cpi->oxcf.target_bandwidth, 1000);
 
   // Set up frame rate and related parameters rate control values.
-  vp8_new_frame_rate(cpi, cpi->oxcf.frame_rate);
+  vp9_new_frame_rate(cpi, cpi->oxcf.frame_rate);
 
   // Set absolute upper and lower quality limits
   cpi->worst_quality               = cpi->oxcf.worst_allowed_q;
@@ -1639,7 +1639,7 @@ void vp8_change_config(VP8_PTR ptr, VP8_CONFIG *oxcf) {
       cm->yv12_fb[cm->lst_fb_idx].y_height ||
       cm->yv12_fb[cm->lst_fb_idx].y_width == 0) {
     alloc_raw_frame_buffers(cpi);
-    vp8_alloc_compressor_data(cpi);
+    vp9_alloc_compressor_data(cpi);
   }
 
   if (cpi->oxcf.fixed_q >= 0) {
@@ -1710,7 +1710,7 @@ static void cal_nmvsadcosts_hp(int *mvsadcost[2]) {
   } while (++i <= MV_MAX);
 }
 
-VP8_PTR vp8_create_compressor(VP8_CONFIG *oxcf) {
+VP8_PTR vp9_create_compressor(VP8_CONFIG *oxcf) {
   int i;
   volatile union {
     VP8_COMP *cpi;
@@ -1733,7 +1733,7 @@ VP8_PTR vp8_create_compressor(VP8_CONFIG *oxcf) {
     VP8_PTR ptr = ctx.ptr;
 
     ctx.cpi->common.error.setjmp = 0;
-    vp8_remove_compressor(&ptr);
+    vp9_remove_compressor(&ptr);
     return 0;
   }
 
@@ -1742,7 +1742,7 @@ VP8_PTR vp8_create_compressor(VP8_CONFIG *oxcf) {
   CHECK_MEM_ERROR(cpi->mb.ss, vpx_calloc(sizeof(search_site), (MAX_MVSEARCH_STEPS * 8) + 1));
 
   vp8_create_common(&cpi->common);
-  vp8_cmachine_specific_config(cpi);
+  vp9_cmachine_specific_config(cpi);
 
   init_config((VP8_PTR)cpi, oxcf);
 
@@ -1883,7 +1883,7 @@ VP8_PTR vp8_create_compressor(VP8_CONFIG *oxcf) {
   cpi->ni_tot_qi                    = 0;
   cpi->ni_frames                   = 0;
   cpi->tot_q = 0.0;
-  cpi->avg_q = vp8_convert_qindex_to_q(cpi->oxcf.worst_allowed_q);
+  cpi->avg_q = vp9_convert_qindex_to_q(cpi->oxcf.worst_allowed_q);
   cpi->total_byte_count             = 0;
 
   cpi->rate_correction_factor         = 1.0;
@@ -1923,7 +1923,7 @@ VP8_PTR vp8_create_compressor(VP8_CONFIG *oxcf) {
   cpi->output_pkt_list = oxcf->output_pkt_list;
 
   if (cpi->pass == 1) {
-    vp8_init_first_pass(cpi);
+    vp9_init_first_pass(cpi);
   } else if (cpi->pass == 2) {
     size_t packet_sz = sizeof(FIRSTPASS_STATS);
     int packets = oxcf->two_pass_stats_in.sz / packet_sz;
@@ -1932,10 +1932,10 @@ VP8_PTR vp8_create_compressor(VP8_CONFIG *oxcf) {
     cpi->twopass.stats_in = cpi->twopass.stats_in_start;
     cpi->twopass.stats_in_end = (void *)((char *)cpi->twopass.stats_in
                                          + (packets - 1) * packet_sz);
-    vp8_init_second_pass(cpi);
+    vp9_init_second_pass(cpi);
   }
 
-  vp8_set_speed_features(cpi);
+  vp9_set_speed_features(cpi);
 
   // Set starting values of RD threshold multipliers (128 = *1)
   for (i = 0; i < MAX_MODES; i++) {
@@ -1959,35 +1959,35 @@ VP8_PTR vp8_create_compressor(VP8_CONFIG *oxcf) {
 
 
 #if CONFIG_SUPERBLOCKS
-  BFP(BLOCK_32X32, vp8_sad32x32, vp8_variance32x32, vp8_sub_pixel_variance32x32,
-      vp8_variance_halfpixvar32x32_h, vp8_variance_halfpixvar32x32_v,
-      vp8_variance_halfpixvar32x32_hv, vp8_sad32x32x3, vp8_sad32x32x8,
-      vp8_sad32x32x4d)
+  BFP(BLOCK_32X32, vp9_sad32x32, vp9_variance32x32, vp9_sub_pixel_variance32x32,
+      vp9_variance_halfpixvar32x32_h, vp9_variance_halfpixvar32x32_v,
+      vp9_variance_halfpixvar32x32_hv, vp9_sad32x32x3, vp9_sad32x32x8,
+      vp9_sad32x32x4d)
 #endif
 
-  BFP(BLOCK_16X16, vp8_sad16x16, vp8_variance16x16, vp8_sub_pixel_variance16x16,
-       vp8_variance_halfpixvar16x16_h, vp8_variance_halfpixvar16x16_v,
-       vp8_variance_halfpixvar16x16_hv, vp8_sad16x16x3, vp8_sad16x16x8,
-       vp8_sad16x16x4d)
+  BFP(BLOCK_16X16, vp9_sad16x16, vp9_variance16x16, vp9_sub_pixel_variance16x16,
+       vp9_variance_halfpixvar16x16_h, vp9_variance_halfpixvar16x16_v,
+       vp9_variance_halfpixvar16x16_hv, vp9_sad16x16x3, vp9_sad16x16x8,
+       vp9_sad16x16x4d)
 
-  BFP(BLOCK_16X8, vp8_sad16x8, vp8_variance16x8, vp8_sub_pixel_variance16x8,
-      NULL, NULL, NULL, vp8_sad16x8x3, vp8_sad16x8x8, vp8_sad16x8x4d)
+  BFP(BLOCK_16X8, vp9_sad16x8, vp9_variance16x8, vp9_sub_pixel_variance16x8,
+      NULL, NULL, NULL, vp9_sad16x8x3, vp9_sad16x8x8, vp9_sad16x8x4d)
 
-  BFP(BLOCK_8X16, vp8_sad8x16, vp8_variance8x16, vp8_sub_pixel_variance8x16,
-      NULL, NULL, NULL, vp8_sad8x16x3, vp8_sad8x16x8, vp8_sad8x16x4d)
+  BFP(BLOCK_8X16, vp9_sad8x16, vp9_variance8x16, vp9_sub_pixel_variance8x16,
+      NULL, NULL, NULL, vp9_sad8x16x3, vp9_sad8x16x8, vp9_sad8x16x4d)
 
-  BFP(BLOCK_8X8, vp8_sad8x8, vp8_variance8x8, vp8_sub_pixel_variance8x8,
-      NULL, NULL, NULL, vp8_sad8x8x3, vp8_sad8x8x8, vp8_sad8x8x4d)
+  BFP(BLOCK_8X8, vp9_sad8x8, vp9_variance8x8, vp9_sub_pixel_variance8x8,
+      NULL, NULL, NULL, vp9_sad8x8x3, vp9_sad8x8x8, vp9_sad8x8x4d)
 
-  BFP(BLOCK_4X4, vp8_sad4x4, vp8_variance4x4, vp8_sub_pixel_variance4x4,
-      NULL, NULL, NULL, vp8_sad4x4x3, vp8_sad4x4x8, vp8_sad4x4x4d)
+  BFP(BLOCK_4X4, vp9_sad4x4, vp9_variance4x4, vp9_sub_pixel_variance4x4,
+      NULL, NULL, NULL, vp9_sad4x4x3, vp9_sad4x4x8, vp9_sad4x4x4d)
 
 #if ARCH_X86 || ARCH_X86_64
-  cpi->fn_ptr[BLOCK_16X16].copymem  = vp8_copy32xn;
-  cpi->fn_ptr[BLOCK_16X8].copymem   = vp8_copy32xn;
-  cpi->fn_ptr[BLOCK_8X16].copymem   = vp8_copy32xn;
-  cpi->fn_ptr[BLOCK_8X8].copymem    = vp8_copy32xn;
-  cpi->fn_ptr[BLOCK_4X4].copymem    = vp8_copy32xn;
+  cpi->fn_ptr[BLOCK_16X16].copymem  = vp9_copy32xn;
+  cpi->fn_ptr[BLOCK_16X8].copymem   = vp9_copy32xn;
+  cpi->fn_ptr[BLOCK_8X16].copymem   = vp9_copy32xn;
+  cpi->fn_ptr[BLOCK_8X8].copymem    = vp9_copy32xn;
+  cpi->fn_ptr[BLOCK_4X4].copymem    = vp9_copy32xn;
 #endif
 
   cpi->full_search_sad = SEARCH_INVOKE(&cpi->rtcd.search, full_search);
@@ -1997,9 +1997,9 @@ VP8_PTR vp8_create_compressor(VP8_CONFIG *oxcf) {
   // make sure frame 1 is okay
   cpi->error_bins[0] = cpi->common.MBs;
 
-  // vp8cx_init_quantizer() is first called here. Add check in vp8cx_frame_init_quantizer() so that vp8cx_init_quantizer is only called later
-  // when needed. This will avoid unnecessary calls of vp8cx_init_quantizer() for every frame.
-  vp8cx_init_quantizer(cpi);
+  // vp9cx_init_quantizer() is first called here. Add check in vp9cx_frame_init_quantizer() so that vp9cx_init_quantizer is only called later
+  // when needed. This will avoid unnecessary calls of vp9cx_init_quantizer() for every frame.
+  vp9cx_init_quantizer(cpi);
 
   vp8_loop_filter_init(cm);
 
@@ -2010,7 +2010,7 @@ VP8_PTR vp8_create_compressor(VP8_CONFIG *oxcf) {
   return (VP8_PTR) cpi;
 }
 
-void vp8_remove_compressor(VP8_PTR *ptr) {
+void vp9_remove_compressor(VP8_PTR *ptr) {
   VP8_COMP *cpi = (VP8_COMP *)(*ptr);
   int i;
 
@@ -2019,7 +2019,7 @@ void vp8_remove_compressor(VP8_PTR *ptr) {
 
   if (cpi && (cpi->common.current_video_frame > 0)) {
     if (cpi->pass == 2) {
-      vp8_end_second_pass(cpi);
+      vp9_end_second_pass(cpi);
     }
 
 #ifdef ENTROPY_STATS
@@ -2051,8 +2051,8 @@ void vp8_remove_compressor(VP8_PTR *ptr) {
       if (cpi->b_calculate_psnr) {
         YV12_BUFFER_CONFIG *lst_yv12 = &cpi->common.yv12_fb[cpi->common.lst_fb_idx];
         double samples = 3.0 / 2 * cpi->count * lst_yv12->y_width * lst_yv12->y_height;
-        double total_psnr = vp8_mse2psnr(samples, 255.0, cpi->total_sq_error);
-        double total_psnr2 = vp8_mse2psnr(samples, 255.0, cpi->total_sq_error2);
+        double total_psnr = vp9_mse2psnr(samples, 255.0, cpi->total_sq_error);
+        double total_psnr2 = vp9_mse2psnr(samples, 255.0, cpi->total_sq_error2);
         double total_ssim = 100 * pow(cpi->summed_quality / cpi->summed_weights, 8.0);
 
         fprintf(f, "Bitrate\tAVGPsnr\tGLBPsnr\tAVPsnrP\tGLPsnrP\tVPXSSIM\t  Time(ms)\n");
@@ -2270,7 +2270,7 @@ static uint64_t calc_plane_error(unsigned char *orig, int orig_stride,
     for (col = 0; col + 16 <= cols; col += 16) {
       unsigned int sse;
 
-      vp8_mse16x16(orig + col, orig_stride, recon + col, recon_stride, &sse);
+      vp9_mse16x16(orig + col, orig_stride, recon + col, recon_stride, &sse);
       total_sse += sse;
     }
 
@@ -2348,14 +2348,14 @@ static void generate_psnr_packet(VP8_COMP *cpi) {
   pkt.data.psnr.samples[3] = width * height;
 
   for (i = 0; i < 4; i++)
-    pkt.data.psnr.psnr[i] = vp8_mse2psnr(pkt.data.psnr.samples[i], 255.0,
+    pkt.data.psnr.psnr[i] = vp9_mse2psnr(pkt.data.psnr.samples[i], 255.0,
                                          pkt.data.psnr.sse[i]);
 
   vpx_codec_pkt_list_add(cpi->output_pkt_list, &pkt);
 }
 
 
-int vp8_use_as_reference(VP8_PTR ptr, int ref_frame_flags) {
+int vp9_use_as_reference(VP8_PTR ptr, int ref_frame_flags) {
   VP8_COMP *cpi = (VP8_COMP *)(ptr);
 
   if (ref_frame_flags > 7)
@@ -2364,7 +2364,7 @@ int vp8_use_as_reference(VP8_PTR ptr, int ref_frame_flags) {
   cpi->ref_frame_flags = ref_frame_flags;
   return 0;
 }
-int vp8_update_reference(VP8_PTR ptr, int ref_frame_flags) {
+int vp9_update_reference(VP8_PTR ptr, int ref_frame_flags) {
   VP8_COMP *cpi = (VP8_COMP *)(ptr);
 
   if (ref_frame_flags > 7)
@@ -2386,7 +2386,7 @@ int vp8_update_reference(VP8_PTR ptr, int ref_frame_flags) {
   return 0;
 }
 
-int vp8_get_reference(VP8_PTR ptr, VP8_REFFRAME ref_frame_flag, YV12_BUFFER_CONFIG *sd) {
+int vp9_get_reference(VP8_PTR ptr, VP8_REFFRAME ref_frame_flag, YV12_BUFFER_CONFIG *sd) {
   VP8_COMP *cpi = (VP8_COMP *)(ptr);
   VP8_COMMON *cm = &cpi->common;
   int ref_fb_idx;
@@ -2404,7 +2404,7 @@ int vp8_get_reference(VP8_PTR ptr, VP8_REFFRAME ref_frame_flag, YV12_BUFFER_CONF
 
   return 0;
 }
-int vp8_set_reference(VP8_PTR ptr, VP8_REFFRAME ref_frame_flag, YV12_BUFFER_CONFIG *sd) {
+int vp9_set_reference(VP8_PTR ptr, VP8_REFFRAME ref_frame_flag, YV12_BUFFER_CONFIG *sd) {
   VP8_COMP *cpi = (VP8_COMP *)(ptr);
   VP8_COMMON *cm = &cpi->common;
 
@@ -2423,7 +2423,7 @@ int vp8_set_reference(VP8_PTR ptr, VP8_REFFRAME ref_frame_flag, YV12_BUFFER_CONF
 
   return 0;
 }
-int vp8_update_entropy(VP8_PTR comp, int update) {
+int vp9_update_entropy(VP8_PTR comp, int update) {
   VP8_COMP *cpi = (VP8_COMP *) comp;
   VP8_COMMON *cm = &cpi->common;
   cm->refresh_entropy_probs = update;
@@ -2574,7 +2574,7 @@ static int find_fp_qindex() {
   int i;
 
   for (i = 0; i < QINDEX_RANGE; i++) {
-    if (vp8_convert_qindex_to_q(i) >= 30.0) {
+    if (vp9_convert_qindex_to_q(i) >= 30.0) {
       break;
     }
   }
@@ -2591,8 +2591,8 @@ static void Pass1Encode(VP8_COMP *cpi, unsigned long *size, unsigned char *dest,
   (void) frame_flags;
 
 
-  vp8_set_quantizer(cpi, find_fp_qindex());
-  vp8_first_pass(cpi);
+  vp9_set_quantizer(cpi, find_fp_qindex());
+  vp9_first_pass(cpi);
 }
 
 #define WRITE_RECON_BUFFER 0
@@ -2786,17 +2786,17 @@ static void loopfilter_frame(VP8_COMP *cpi, VP8_COMMON *cm) {
 
     vpx_usec_timer_start(&timer);
     if (cpi->sf.auto_filter == 0)
-      vp8cx_pick_filter_level_fast(cpi->Source, cpi);
+      vp9cx_pick_filter_level_fast(cpi->Source, cpi);
 
     else
-      vp8cx_pick_filter_level(cpi->Source, cpi);
+      vp9cx_pick_filter_level(cpi->Source, cpi);
 
     vpx_usec_timer_mark(&timer);
     cpi->time_pick_lpf += vpx_usec_timer_elapsed(&timer);
   }
 
   if (cm->filter_level > 0) {
-    vp8cx_set_alt_lf_level(cpi, cm->filter_level);
+    vp9cx_set_alt_lf_level(cpi, cm->filter_level);
     vp8_loop_filter_frame(cm, &cpi->mb.e_mbd);
   }
 
@@ -2981,7 +2981,7 @@ static void encode_frame_to_data_rate
   init_seg_features(cpi);
 
   // Decide how big to make the frame
-  vp8_pick_frame_size(cpi);
+  vp9_pick_frame_size(cpi);
 
   vp8_clear_system_state();
 
@@ -3098,7 +3098,7 @@ static void encode_frame_to_data_rate
     Q = cpi->last_boosted_qindex;
   } else {
     // Determine initial Q to try
-    Q = vp8_regulate_q(cpi, cpi->this_frame_target);
+    Q = vp9_regulate_q(cpi, cpi->this_frame_target);
   }
   last_zbin_oq = cpi->zbin_over_quant;
 
@@ -3110,7 +3110,7 @@ static void encode_frame_to_data_rate
   else
     zbin_oq_high = ZBIN_OQ_MAX;
 
-  vp8_compute_frame_size_bounds(cpi, &frame_under_shoot_limit, &frame_over_shoot_limit);
+  vp9_compute_frame_size_bounds(cpi, &frame_under_shoot_limit, &frame_over_shoot_limit);
 
   // Limit Q range for the adaptive loop.
   bottom_index = cpi->active_best_quality;
@@ -3195,7 +3195,7 @@ static void encode_frame_to_data_rate
   do {
     vp8_clear_system_state();  // __asm emms;
 
-    vp8_set_quantizer(cpi, Q);
+    vp9_set_quantizer(cpi, Q);
     this_q = Q;
 
     if (loop_count == 0) {
@@ -3245,14 +3245,14 @@ static void encode_frame_to_data_rate
 
       // Set up entropy depending on frame type.
       if (cm->frame_type == KEY_FRAME)
-        vp8_setup_key_frame(cpi);
+        vp9_setup_key_frame(cpi);
       else
-        vp8_setup_inter_frame(cpi);
+        vp9_setup_inter_frame(cpi);
     }
 
     // transform / motion compensation build reconstruction frame
 
-    vp8_encode_frame(cpi);
+    vp9_encode_frame(cpi);
 
     // Update the skip mb flag probabilities based on the distribution
     // seen in the last encoder iteration.
@@ -3270,11 +3270,11 @@ static void encode_frame_to_data_rate
     // Dummy pack of the bitstream using up to date stats to get an
     // accurate estimate of output frame size to determine if we need
     // to recode.
-    vp8_save_coding_context(cpi);
+    vp9_save_coding_context(cpi);
     cpi->dummy_packing = 1;
-    vp8_pack_bitstream(cpi, dest, size);
+    vp9_pack_bitstream(cpi, dest, size);
     cpi->projected_frame_size = (*size) << 3;
-    vp8_restore_coding_context(cpi);
+    vp9_restore_coding_context(cpi);
 
     if (frame_over_shoot_limit == 0)
       frame_over_shoot_limit = 1;
@@ -3283,7 +3283,7 @@ static void encode_frame_to_data_rate
     // Special case handling for forced key frames
     if ((cm->frame_type == KEY_FRAME) && cpi->this_key_frame_forced) {
       int last_q = Q;
-      int kf_err = vp8_calc_ss_err(cpi->Source,
+      int kf_err = vp9_calc_ss_err(cpi->Source,
                                    &cm->yv12_fb[cm->new_fb_idx]);
 
       int high_err_target = cpi->ambient_err;
@@ -3347,7 +3347,7 @@ static void encode_frame_to_data_rate
         if (undershoot_seen || (loop_count > 1)) {
           // Update rate_correction_factor unless cpi->active_worst_quality has changed.
           if (!active_worst_qchanged)
-            vp8_update_rate_correction_factors(cpi, 1);
+            vp9_update_rate_correction_factors(cpi, 1);
 
           Q = (q_high + q_low + 1) / 2;
 
@@ -3361,13 +3361,13 @@ static void encode_frame_to_data_rate
         } else {
           // Update rate_correction_factor unless cpi->active_worst_quality has changed.
           if (!active_worst_qchanged)
-            vp8_update_rate_correction_factors(cpi, 0);
+            vp9_update_rate_correction_factors(cpi, 0);
 
-          Q = vp8_regulate_q(cpi, cpi->this_frame_target);
+          Q = vp9_regulate_q(cpi, cpi->this_frame_target);
 
           while (((Q < q_low) || (cpi->zbin_over_quant < zbin_oq_low)) && (Retries < 10)) {
-            vp8_update_rate_correction_factors(cpi, 0);
-            Q = vp8_regulate_q(cpi, cpi->this_frame_target);
+            vp9_update_rate_correction_factors(cpi, 0);
+            Q = vp9_regulate_q(cpi, cpi->this_frame_target);
             Retries++;
           }
         }
@@ -3384,7 +3384,7 @@ static void encode_frame_to_data_rate
         if (overshoot_seen || (loop_count > 1)) {
           // Update rate_correction_factor unless cpi->active_worst_quality has changed.
           if (!active_worst_qchanged)
-            vp8_update_rate_correction_factors(cpi, 1);
+            vp9_update_rate_correction_factors(cpi, 1);
 
           Q = (q_high + q_low) / 2;
 
@@ -3396,9 +3396,9 @@ static void encode_frame_to_data_rate
         } else {
           // Update rate_correction_factor unless cpi->active_worst_quality has changed.
           if (!active_worst_qchanged)
-            vp8_update_rate_correction_factors(cpi, 0);
+            vp9_update_rate_correction_factors(cpi, 0);
 
-          Q = vp8_regulate_q(cpi, cpi->this_frame_target);
+          Q = vp9_regulate_q(cpi, cpi->this_frame_target);
 
           // Special case reset for qlow for constrained quality.
           // This should only trigger where there is very substantial
@@ -3410,8 +3410,8 @@ static void encode_frame_to_data_rate
           }
 
           while (((Q > q_high) || (cpi->zbin_over_quant > zbin_oq_high)) && (Retries < 10)) {
-            vp8_update_rate_correction_factors(cpi, 0);
-            Q = vp8_regulate_q(cpi, cpi->this_frame_target);
+            vp9_update_rate_correction_factors(cpi, 0);
+            Q = vp9_regulate_q(cpi, cpi->this_frame_target);
             Retries++;
           }
         }
@@ -3472,7 +3472,7 @@ static void encode_frame_to_data_rate
 
     if (Loop == FALSE && cm->frame_type != KEY_FRAME && sf->search_best_filter) {
       if (mcomp_filter_index < mcomp_filters) {
-        INT64 err = vp8_calc_ss_err(cpi->Source,
+        INT64 err = vp9_calc_ss_err(cpi->Source,
                                     &cm->yv12_fb[cm->new_fb_idx]);
         INT64 rate = cpi->projected_frame_size << 8;
         mcomp_filter_cost[mcomp_filter_index] =
@@ -3534,7 +3534,7 @@ static void encode_frame_to_data_rate
   // fixed interval. Note the reconstruction error if it is the frame before
   // the force key frame
   if (cpi->next_key_frame_forced && (cpi->twopass.frames_to_key == 0)) {
-    cpi->ambient_err = vp8_calc_ss_err(cpi->Source,
+    cpi->ambient_err = vp9_calc_ss_err(cpi->Source,
                                        &cm->yv12_fb[cm->new_fb_idx]);
   }
 
@@ -3564,7 +3564,7 @@ static void encode_frame_to_data_rate
   // Update the GF useage maps.
   // This is done after completing the compression of a frame when all modes
   // etc. are finalized but before loop filter
-  vp8_update_gf_useage_maps(cpi, cm, &cpi->mb);
+  vp9_update_gf_useage_maps(cpi, cm, &cpi->mb);
 
   if (cm->frame_type == KEY_FRAME)
     cm->refresh_last_frame = 1;
@@ -3593,7 +3593,7 @@ static void encode_frame_to_data_rate
 
   // build the bitstream
   cpi->dummy_packing = 0;
-  vp8_pack_bitstream(cpi, dest, size);
+  vp9_pack_bitstream(cpi, dest, size);
 
   if (cpi->mb.e_mbd.update_mb_segmentation_map) {
     update_reference_segmentation_map(cpi);
@@ -3641,7 +3641,7 @@ static void encode_frame_to_data_rate
   cpi->projected_frame_size = (*size) << 3;
 
   if (!active_worst_qchanged)
-    vp8_update_rate_correction_factors(cpi, 2);
+    vp9_update_rate_correction_factors(cpi, 2);
 
   cpi->last_q[cm->frame_type] = cm->base_qindex;
 
@@ -3659,7 +3659,7 @@ static void encode_frame_to_data_rate
   }
 
   if (cm->frame_type == KEY_FRAME) {
-    vp8_adjust_key_frame_context(cpi);
+    vp9_adjust_key_frame_context(cpi);
   }
 
   // Keep a record of ambient average Q.
@@ -3669,7 +3669,7 @@ static void encode_frame_to_data_rate
   // Keep a record from which we can calculate the average Q excluding GF updates and key frames
   if ((cm->frame_type != KEY_FRAME) && !cm->refresh_golden_frame && !cm->refresh_alt_ref_frame) {
     cpi->ni_frames++;
-    cpi->tot_q += vp8_convert_qindex_to_q(Q);
+    cpi->tot_q += vp9_convert_qindex_to_q(Q);
     cpi->avg_q = cpi->tot_q / (double)cpi->ni_frames;
 
     // Calculate the average Q for normal inter frames (not key or GFU
@@ -3740,7 +3740,7 @@ static void encode_frame_to_data_rate
 
     vp8_clear_system_state();  // __asm emms;
 
-    recon_err = vp8_calc_ss_err(cpi->Source,
+    recon_err = vp9_calc_ss_err(cpi->Source,
                                 &cm->yv12_fb[cm->new_fb_idx]);
 
     if (cpi->twopass.total_left_stats->coded_error != 0.0)
@@ -3754,13 +3754,13 @@ static void encode_frame_to_data_rate
               (int)cpi->total_target_vs_actual,
               (cpi->oxcf.starting_buffer_level - cpi->bits_off_target),
               (int)cpi->total_actual_bits,
-              vp8_convert_qindex_to_q(cm->base_qindex),
+              vp9_convert_qindex_to_q(cm->base_qindex),
               (double)vp8_dc_quant(cm->base_qindex, 0) / 4.0,
-              vp8_convert_qindex_to_q(cpi->active_best_quality),
-              vp8_convert_qindex_to_q(cpi->active_worst_quality),
+              vp9_convert_qindex_to_q(cpi->active_best_quality),
+              vp9_convert_qindex_to_q(cpi->active_worst_quality),
               cpi->avg_q,
-              vp8_convert_qindex_to_q(cpi->ni_av_qi),
-              vp8_convert_qindex_to_q(cpi->cq_target_quality),
+              vp9_convert_qindex_to_q(cpi->ni_av_qi),
+              vp9_convert_qindex_to_q(cpi->cq_target_quality),
               cpi->zbin_over_quant,
               // cpi->avg_frame_qindex, cpi->zbin_over_quant,
               cm->refresh_golden_frame, cm->refresh_alt_ref_frame,
@@ -3784,13 +3784,13 @@ static void encode_frame_to_data_rate
               (int)cpi->total_target_vs_actual,
               (cpi->oxcf.starting_buffer_level - cpi->bits_off_target),
               (int)cpi->total_actual_bits,
-              vp8_convert_qindex_to_q(cm->base_qindex),
+              vp9_convert_qindex_to_q(cm->base_qindex),
               (double)vp8_dc_quant(cm->base_qindex, 0) / 4.0,
-              vp8_convert_qindex_to_q(cpi->active_best_quality),
-              vp8_convert_qindex_to_q(cpi->active_worst_quality),
+              vp9_convert_qindex_to_q(cpi->active_best_quality),
+              vp9_convert_qindex_to_q(cpi->active_worst_quality),
               cpi->avg_q,
-              vp8_convert_qindex_to_q(cpi->ni_av_qi),
-              vp8_convert_qindex_to_q(cpi->cq_target_quality),
+              vp9_convert_qindex_to_q(cpi->ni_av_qi),
+              vp9_convert_qindex_to_q(cpi->cq_target_quality),
               cpi->zbin_over_quant,
               // cpi->avg_frame_qindex, cpi->zbin_over_quant,
               cm->refresh_golden_frame, cm->refresh_alt_ref_frame,
@@ -3929,7 +3929,7 @@ static void encode_frame_to_data_rate
 static void Pass2Encode(VP8_COMP *cpi, unsigned long *size, unsigned char *dest, unsigned int *frame_flags) {
 
   if (!cpi->common.refresh_alt_ref_frame)
-    vp8_second_pass(cpi);
+    vp9_second_pass(cpi);
 
   encode_frame_to_data_rate(cpi, size, dest, frame_flags);
   cpi->twopass.bits_left -= 8 * *size;
@@ -3953,7 +3953,7 @@ extern void vp8_pop_neon(int64_t *store);
 #endif
 
 
-int vp8_receive_raw_frame(VP8_PTR ptr, unsigned int frame_flags, YV12_BUFFER_CONFIG *sd, int64_t time_stamp, int64_t end_time) {
+int vp9_receive_raw_frame(VP8_PTR ptr, unsigned int frame_flags, YV12_BUFFER_CONFIG *sd, int64_t time_stamp, int64_t end_time) {
 #if HAVE_ARMV7
   int64_t store_reg[8];
 #endif
@@ -3972,7 +3972,7 @@ int vp8_receive_raw_frame(VP8_PTR ptr, unsigned int frame_flags, YV12_BUFFER_CON
 #endif
 
   vpx_usec_timer_start(&timer);
-  if (vp8_lookahead_push(cpi->lookahead, sd, time_stamp, end_time,
+  if (vp9_lookahead_push(cpi->lookahead, sd, time_stamp, end_time,
                          frame_flags, cpi->active_map_enabled ? cpi->active_map : NULL))
     res = -1;
   cm->clr_type = sd->clrtype;
@@ -4005,7 +4005,7 @@ static int frame_is_reference(const VP8_COMP *cpi) {
 }
 
 
-int vp8_get_compressed_data(VP8_PTR ptr, unsigned int *frame_flags, unsigned long *size, unsigned char *dest, int64_t *time_stamp, int64_t *time_end, int flush) {
+int vp9_get_compressed_data(VP8_PTR ptr, unsigned int *frame_flags, unsigned long *size, unsigned char *dest, int64_t *time_stamp, int64_t *time_end, int flush) {
 #if HAVE_ARMV7
   int64_t store_reg[8];
 #endif
@@ -4034,11 +4034,11 @@ int vp8_get_compressed_data(VP8_PTR ptr, unsigned int *frame_flags, unsigned lon
   // Should we code an alternate reference frame
   if (cpi->oxcf.play_alternate &&
       cpi->source_alt_ref_pending) {
-    if ((cpi->source = vp8_lookahead_peek(cpi->lookahead,
+    if ((cpi->source = vp9_lookahead_peek(cpi->lookahead,
                                           cpi->frames_till_gf_update_due))) {
       cpi->alt_ref_source = cpi->source;
       if (cpi->oxcf.arnr_max_frames > 0) {
-        vp8_temporal_filter_prepare_c(cpi,
+        vp9_temporal_filter_prepare_c(cpi,
                                       cpi->frames_till_gf_update_due);
         force_src_buffer = &cpi->alt_ref_buffer;
       }
@@ -4053,7 +4053,7 @@ int vp8_get_compressed_data(VP8_PTR ptr, unsigned int *frame_flags, unsigned lon
   }
 
   if (!cpi->source) {
-    if ((cpi->source = vp8_lookahead_pop(cpi->lookahead, flush))) {
+    if ((cpi->source = vp9_lookahead_pop(cpi->lookahead, flush))) {
       cm->show_frame = 1;
 
       cpi->is_src_frame_alt_ref = cpi->alt_ref_source
@@ -4073,7 +4073,7 @@ int vp8_get_compressed_data(VP8_PTR ptr, unsigned int *frame_flags, unsigned lon
   } else {
     *size = 0;
     if (flush && cpi->pass == 1 && !cpi->twopass.first_pass_done) {
-      vp8_end_first_pass(cpi);    /* get last stats packet */
+      vp9_end_first_pass(cpi);    /* get last stats packet */
       cpi->twopass.first_pass_done = 1;
     }
 
@@ -4114,7 +4114,7 @@ int vp8_get_compressed_data(VP8_PTR ptr, unsigned int *frame_flags, unsigned lon
 
     if (this_duration) {
       if (step)
-        vp8_new_frame_rate(cpi, 10000000.0 / this_duration);
+        vp9_new_frame_rate(cpi, 10000000.0 / this_duration);
       else {
         double avg_duration, interval;
 
@@ -4130,7 +4130,7 @@ int vp8_get_compressed_data(VP8_PTR ptr, unsigned int *frame_flags, unsigned lon
         avg_duration *= (interval - avg_duration + this_duration);
         avg_duration /= interval;
 
-        vp8_new_frame_rate(cpi, 10000000.0 / avg_duration);
+        vp9_new_frame_rate(cpi, 10000000.0 / avg_duration);
       }
     }
 
@@ -4240,11 +4240,11 @@ int vp8_get_compressed_data(VP8_PTR ptr, unsigned int *frame_flags, unsigned lon
 
         sq_error = ye + ue + ve;
 
-        frame_psnr = vp8_mse2psnr(t_samples, 255.0, sq_error);
+        frame_psnr = vp9_mse2psnr(t_samples, 255.0, sq_error);
 
-        cpi->total_y += vp8_mse2psnr(y_samples, 255.0, ye);
-        cpi->total_u += vp8_mse2psnr(uv_samples, 255.0, ue);
-        cpi->total_v += vp8_mse2psnr(uv_samples, 255.0, ve);
+        cpi->total_y += vp9_mse2psnr(y_samples, 255.0, ye);
+        cpi->total_u += vp9_mse2psnr(uv_samples, 255.0, ue);
+        cpi->total_v += vp9_mse2psnr(uv_samples, 255.0, ve);
         cpi->total_sq_error += sq_error;
         cpi->total  += frame_psnr;
         {
@@ -4269,15 +4269,15 @@ int vp8_get_compressed_data(VP8_PTR ptr, unsigned int *frame_flags, unsigned lon
 
           sq_error = ye + ue + ve;
 
-          frame_psnr2 = vp8_mse2psnr(t_samples, 255.0, sq_error);
+          frame_psnr2 = vp9_mse2psnr(t_samples, 255.0, sq_error);
 
-          cpi->totalp_y += vp8_mse2psnr(y_samples, 255.0, ye);
-          cpi->totalp_u += vp8_mse2psnr(uv_samples, 255.0, ue);
-          cpi->totalp_v += vp8_mse2psnr(uv_samples, 255.0, ve);
+          cpi->totalp_y += vp9_mse2psnr(y_samples, 255.0, ye);
+          cpi->totalp_u += vp9_mse2psnr(uv_samples, 255.0, ue);
+          cpi->totalp_v += vp9_mse2psnr(uv_samples, 255.0, ve);
           cpi->total_sq_error2 += sq_error;
           cpi->totalp  += frame_psnr2;
 
-          frame_ssim2 = vp8_calc_ssim(cpi->Source,
+          frame_ssim2 = vp9_calc_ssim(cpi->Source,
                                       &cm->post_proc_buffer, 1, &weight);
 
           cpi->summed_quality += frame_ssim2 * weight;
@@ -4296,7 +4296,7 @@ int vp8_get_compressed_data(VP8_PTR ptr, unsigned int *frame_flags, unsigned lon
 
       if (cpi->b_calculate_ssimg) {
         double y, u, v, frame_all;
-        frame_all =  vp8_calc_ssimg(cpi->Source, cm->frame_to_show,
+        frame_all =  vp9_calc_ssimg(cpi->Source, cm->frame_to_show,
                                     &y, &u, &v);
         cpi->total_ssimg_y += y;
         cpi->total_ssimg_u += u;
@@ -4321,7 +4321,7 @@ int vp8_get_compressed_data(VP8_PTR ptr, unsigned int *frame_flags, unsigned lon
   return 0;
 }
 
-int vp8_get_preview_raw_frame(VP8_PTR comp, YV12_BUFFER_CONFIG *dest, vp8_ppflags_t *flags) {
+int vp9_get_preview_raw_frame(VP8_PTR comp, YV12_BUFFER_CONFIG *dest, vp8_ppflags_t *flags) {
   VP8_COMP *cpi = (VP8_COMP *) comp;
 
   if (cpi->common.refresh_alt_ref_frame)
@@ -4348,7 +4348,7 @@ int vp8_get_preview_raw_frame(VP8_PTR comp, YV12_BUFFER_CONFIG *dest, vp8_ppflag
   }
 }
 
-int vp8_set_roimap(VP8_PTR comp, unsigned char *map, unsigned int rows, unsigned int cols, int delta_q[4], int delta_lf[4], unsigned int threshold[4]) {
+int vp9_set_roimap(VP8_PTR comp, unsigned char *map, unsigned int rows, unsigned int cols, int delta_q[4], int delta_lf[4], unsigned int threshold[4]) {
   VP8_COMP *cpi = (VP8_COMP *) comp;
   signed char feature_data[SEG_LVL_MAX][MAX_MB_SEGMENTS];
   MACROBLOCKD *xd = &cpi->mb.e_mbd;
@@ -4358,15 +4358,15 @@ int vp8_set_roimap(VP8_PTR comp, unsigned char *map, unsigned int rows, unsigned
     return -1;
 
   if (!map) {
-    vp8_disable_segmentation((VP8_PTR)cpi);
+    vp9_disable_segmentation((VP8_PTR)cpi);
     return 0;
   }
 
   // Set the segmentation Map
-  vp8_set_segmentation_map((VP8_PTR)cpi, map);
+  vp9_set_segmentation_map((VP8_PTR)cpi, map);
 
   // Activate segmentation.
-  vp8_enable_segmentation((VP8_PTR)cpi);
+  vp9_enable_segmentation((VP8_PTR)cpi);
 
   // Set up the quant segment data
   feature_data[SEG_LVL_ALT_Q][0] = delta_q[0];
@@ -4400,12 +4400,12 @@ int vp8_set_roimap(VP8_PTR comp, unsigned char *map, unsigned int rows, unsigned
 
   // Initialise the feature data structure
   // SEGMENT_DELTADATA    0, SEGMENT_ABSDATA      1
-  vp8_set_segment_data((VP8_PTR)cpi, &feature_data[0][0], SEGMENT_DELTADATA);
+  vp9_set_segment_data((VP8_PTR)cpi, &feature_data[0][0], SEGMENT_DELTADATA);
 
   return 0;
 }
 
-int vp8_set_active_map(VP8_PTR comp, unsigned char *map, unsigned int rows, unsigned int cols) {
+int vp9_set_active_map(VP8_PTR comp, unsigned char *map, unsigned int rows, unsigned int cols) {
   VP8_COMP *cpi = (VP8_COMP *) comp;
 
   if (rows == cpi->common.mb_rows && cols == cpi->common.mb_cols) {
@@ -4422,7 +4422,7 @@ int vp8_set_active_map(VP8_PTR comp, unsigned char *map, unsigned int rows, unsi
   }
 }
 
-int vp8_set_internal_size(VP8_PTR comp, VPX_SCALING horiz_mode, VPX_SCALING vert_mode) {
+int vp9_set_internal_size(VP8_PTR comp, VPX_SCALING horiz_mode, VPX_SCALING vert_mode) {
   VP8_COMP *cpi = (VP8_COMP *) comp;
 
   if (horiz_mode <= ONETWO)
@@ -4440,7 +4440,7 @@ int vp8_set_internal_size(VP8_PTR comp, VPX_SCALING horiz_mode, VPX_SCALING vert
 
 
 
-int vp8_calc_ss_err(YV12_BUFFER_CONFIG *source, YV12_BUFFER_CONFIG *dest) {
+int vp9_calc_ss_err(YV12_BUFFER_CONFIG *source, YV12_BUFFER_CONFIG *dest) {
   int i, j;
   int Total = 0;
 
@@ -4451,7 +4451,7 @@ int vp8_calc_ss_err(YV12_BUFFER_CONFIG *source, YV12_BUFFER_CONFIG *dest) {
   for (i = 0; i < source->y_height; i += 16) {
     for (j = 0; j < source->y_width; j += 16) {
       unsigned int sse;
-      Total += vp8_mse16x16(src + j, source->y_stride, dst + j, dest->y_stride,
+      Total += vp9_mse16x16(src + j, source->y_stride, dst + j, dest->y_stride,
                             &sse);
     }
 
@@ -4463,7 +4463,7 @@ int vp8_calc_ss_err(YV12_BUFFER_CONFIG *source, YV12_BUFFER_CONFIG *dest) {
 }
 
 
-int vp8_get_quantizer(VP8_PTR c) {
+int vp9_get_quantizer(VP8_PTR c) {
   VP8_COMP   *cpi = (VP8_COMP *) c;
   return cpi->common.base_qindex;
 }
diff --git a/vp8/encoder/onyx_int.h b/vp8/encoder/onyx_int.h
index 75a767b197e4dd32b186643166d71d0baf97f732..c95600c79f6ea3b8cecc52c13d11c57503ef9cd1 100644
--- a/vp8/encoder/onyx_int.h
+++ b/vp8/encoder/onyx_int.h
@@ -759,18 +759,18 @@ typedef struct VP8_COMP {
 
 void control_data_rate(VP8_COMP *cpi);
 
-void vp8_encode_frame(VP8_COMP *cpi);
+void vp9_encode_frame(VP8_COMP *cpi);
 
-void vp8_pack_bitstream(VP8_COMP *cpi, unsigned char *dest, unsigned long *size);
+void vp9_pack_bitstream(VP8_COMP *cpi, unsigned char *dest, unsigned long *size);
 
-void vp8_activity_masking(VP8_COMP *cpi, MACROBLOCK *x);
+void vp9_activity_masking(VP8_COMP *cpi, MACROBLOCK *x);
 
 int rd_cost_intra_mb(MACROBLOCKD *x);
 
-void vp8_tokenize_mb(VP8_COMP *, MACROBLOCKD *, TOKENEXTRA **, int dry_run);
-void vp8_stuff_mb(VP8_COMP *cpi, MACROBLOCKD *xd, TOKENEXTRA **t, int dry_run);
+void vp9_tokenize_mb(VP8_COMP *, MACROBLOCKD *, TOKENEXTRA **, int dry_run);
+void vp9_stuff_mb(VP8_COMP *cpi, MACROBLOCKD *xd, TOKENEXTRA **t, int dry_run);
 
-void vp8_set_speed_features(VP8_COMP *cpi);
+void vp9_set_speed_features(VP8_COMP *cpi);
 
 #if CONFIG_DEBUG
 #define CHECK_MEM_ERROR(lval,expr) do {\
diff --git a/vp8/encoder/picklpf.c b/vp8/encoder/picklpf.c
index 9e773dd0e73751dce34066acbfa6e7f0861f7f28..f8fd6e74e3b0d8eeec84425508c0278342bc1ef6 100644
--- a/vp8/encoder/picklpf.c
+++ b/vp8/encoder/picklpf.c
@@ -21,7 +21,7 @@
 #include "vpx_ports/arm.h"
 #endif
 
-extern int vp8_calc_ss_err(YV12_BUFFER_CONFIG *source,
+extern int vp9_calc_ss_err(YV12_BUFFER_CONFIG *source,
                            YV12_BUFFER_CONFIG *dest);
 #if HAVE_ARMV7
 extern void vp8_yv12_copy_frame_yonly_no_extend_frame_borders_neon(YV12_BUFFER_CONFIG *src_ybc, YV12_BUFFER_CONFIG *dst_ybc);
@@ -34,7 +34,7 @@ extern void vp8_yv12_copy_frame_yonly_no_extend_frame_borders_neon(YV12_BUFFER_C
 #endif
 
 extern void
-(*vp8_yv12_copy_partial_frame_ptr)(YV12_BUFFER_CONFIG *src_ybc,
+(*vp9_yv12_copy_partial_frame_ptr)(YV12_BUFFER_CONFIG *src_ybc,
                                    YV12_BUFFER_CONFIG *dst_ybc,
                                    int Fraction);
 
@@ -47,7 +47,7 @@ extern void vp8_loop_filter_frame_segment
 );
 
 void
-vp8_yv12_copy_partial_frame(YV12_BUFFER_CONFIG *src_ybc, YV12_BUFFER_CONFIG *dst_ybc, int Fraction) {
+vp9_yv12_copy_partial_frame(YV12_BUFFER_CONFIG *src_ybc, YV12_BUFFER_CONFIG *dst_ybc, int Fraction) {
   unsigned char *src_y, *dst_y;
   int yheight;
   int ystride;
@@ -98,7 +98,7 @@ static int vp8_calc_partial_ssl_err(YV12_BUFFER_CONFIG *source,
   for (i = 0; i < linestocopy; i += 16) {
     for (j = 0; j < source->y_width; j += 16) {
       unsigned int sse;
-      Total += vp8_mse16x16(src + j, source->y_stride, dst + j, dest->y_stride,
+      Total += vp9_mse16x16(src + j, source->y_stride, dst + j, dest->y_stride,
                             &sse);
     }
 
@@ -112,7 +112,7 @@ static int vp8_calc_partial_ssl_err(YV12_BUFFER_CONFIG *source,
 // Enforce a minimum filter level based upon baseline Q
 static int get_min_filter_level(VP8_COMP *cpi, int base_qindex) {
   int min_filter_level;
-  /*int q = (int) vp8_convert_qindex_to_q(base_qindex);
+  /*int q = (int) vp9_convert_qindex_to_q(base_qindex);
 
   if (cpi->source_alt_ref_active && cpi->common.refresh_golden_frame && !cpi->common.refresh_alt_ref_frame)
       min_filter_level = 0;
@@ -146,7 +146,7 @@ static int get_max_filter_level(VP8_COMP *cpi, int base_qindex) {
   return max_filter_level;
 }
 
-void vp8cx_pick_filter_level_fast(YV12_BUFFER_CONFIG *sd, VP8_COMP *cpi) {
+void vp9cx_pick_filter_level_fast(YV12_BUFFER_CONFIG *sd, VP8_COMP *cpi) {
   VP8_COMMON *cm = &cpi->common;
 
   int best_err = 0;
@@ -157,7 +157,7 @@ void vp8cx_pick_filter_level_fast(YV12_BUFFER_CONFIG *sd, VP8_COMP *cpi) {
   int best_filt_val = cm->filter_level;
 
   //  Make a copy of the unfiltered / processed recon buffer
-  vp8_yv12_copy_partial_frame_ptr(cm->frame_to_show, &cpi->last_frame_uf, 3);
+  vp9_yv12_copy_partial_frame_ptr(cm->frame_to_show, &cpi->last_frame_uf, 3);
 
   if (cm->frame_type == KEY_FRAME)
     cm->sharpness_level = 0;
@@ -184,7 +184,7 @@ void vp8cx_pick_filter_level_fast(YV12_BUFFER_CONFIG *sd, VP8_COMP *cpi) {
   best_err = vp8_calc_partial_ssl_err(sd, cm->frame_to_show, 3);
 
   //  Re-instate the unfiltered frame
-  vp8_yv12_copy_partial_frame_ptr(&cpi->last_frame_uf, cm->frame_to_show, 3);
+  vp9_yv12_copy_partial_frame_ptr(&cpi->last_frame_uf, cm->frame_to_show, 3);
 
   filt_val -= (1 + ((filt_val > 10) ? 1 : 0));
 
@@ -197,7 +197,7 @@ void vp8cx_pick_filter_level_fast(YV12_BUFFER_CONFIG *sd, VP8_COMP *cpi) {
     filt_err = vp8_calc_partial_ssl_err(sd, cm->frame_to_show, 3);
 
     //  Re-instate the unfiltered frame
-    vp8_yv12_copy_partial_frame_ptr(&cpi->last_frame_uf, cm->frame_to_show, 3);
+    vp9_yv12_copy_partial_frame_ptr(&cpi->last_frame_uf, cm->frame_to_show, 3);
 
 
     // Update the best case record or exit loop.
@@ -226,7 +226,7 @@ void vp8cx_pick_filter_level_fast(YV12_BUFFER_CONFIG *sd, VP8_COMP *cpi) {
       filt_err = vp8_calc_partial_ssl_err(sd, cm->frame_to_show, 3);
 
       //  Re-instate the unfiltered frame
-      vp8_yv12_copy_partial_frame_ptr(&cpi->last_frame_uf, cm->frame_to_show, 3);
+      vp9_yv12_copy_partial_frame_ptr(&cpi->last_frame_uf, cm->frame_to_show, 3);
 
       // Update the best case record or exit loop.
       if (filt_err < best_err) {
@@ -252,10 +252,10 @@ void vp8cx_pick_filter_level_fast(YV12_BUFFER_CONFIG *sd, VP8_COMP *cpi) {
 }
 
 // Stub function for now Alt LF not used
-void vp8cx_set_alt_lf_level(VP8_COMP *cpi, int filt_val) {
+void vp9cx_set_alt_lf_level(VP8_COMP *cpi, int filt_val) {
 }
 
-void vp8cx_pick_filter_level(YV12_BUFFER_CONFIG *sd, VP8_COMP *cpi) {
+void vp9cx_pick_filter_level(YV12_BUFFER_CONFIG *sd, VP8_COMP *cpi) {
   VP8_COMMON *cm = &cpi->common;
 
   int best_err = 0;
@@ -307,10 +307,10 @@ void vp8cx_pick_filter_level(YV12_BUFFER_CONFIG *sd, VP8_COMP *cpi) {
   filter_step = (filt_mid < 16) ? 4 : filt_mid / 4;
 
   // Get baseline error score
-  vp8cx_set_alt_lf_level(cpi, filt_mid);
+  vp9cx_set_alt_lf_level(cpi, filt_mid);
   vp8_loop_filter_frame_yonly(cm, &cpi->mb.e_mbd, filt_mid);
 
-  best_err = vp8_calc_ss_err(sd, cm->frame_to_show);
+  best_err = vp9_calc_ss_err(sd, cm->frame_to_show);
   filt_best = filt_mid;
 
   //  Re-instate the unfiltered frame
@@ -347,10 +347,10 @@ void vp8cx_pick_filter_level(YV12_BUFFER_CONFIG *sd, VP8_COMP *cpi) {
 
     if ((filt_direction <= 0) && (filt_low != filt_mid)) {
       // Get Low filter error score
-      vp8cx_set_alt_lf_level(cpi, filt_low);
+      vp9cx_set_alt_lf_level(cpi, filt_low);
       vp8_loop_filter_frame_yonly(cm, &cpi->mb.e_mbd, filt_low);
 
-      filt_err = vp8_calc_ss_err(sd, cm->frame_to_show);
+      filt_err = vp9_calc_ss_err(sd, cm->frame_to_show);
 
       //  Re-instate the unfiltered frame
 #if HAVE_ARMV7
@@ -382,10 +382,10 @@ void vp8cx_pick_filter_level(YV12_BUFFER_CONFIG *sd, VP8_COMP *cpi) {
 
     // Now look at filt_high
     if ((filt_direction >= 0) && (filt_high != filt_mid)) {
-      vp8cx_set_alt_lf_level(cpi, filt_high);
+      vp9cx_set_alt_lf_level(cpi, filt_high);
       vp8_loop_filter_frame_yonly(cm, &cpi->mb.e_mbd, filt_high);
 
-      filt_err = vp8_calc_ss_err(sd, cm->frame_to_show);
+      filt_err = vp9_calc_ss_err(sd, cm->frame_to_show);
 
       //  Re-instate the unfiltered frame
 #if HAVE_ARMV7
diff --git a/vp8/encoder/ppc/csystemdependent.c b/vp8/encoder/ppc/csystemdependent.c
index 96c9a0384e9a5ece15608979960607da9fade0fe..edecb20334fa70b489fda0104bc98f9b1f5f6266 100644
--- a/vp8/encoder/ppc/csystemdependent.c
+++ b/vp8/encoder/ppc/csystemdependent.c
@@ -12,53 +12,53 @@
 #include "vp8/encoder/variance.h"
 #include "vp8/encoder/onyx_int.h"
 
-SADFunction *vp8_sad16x16;
-SADFunction *vp8_sad16x8;
-SADFunction *vp8_sad8x16;
-SADFunction *vp8_sad8x8;
-SADFunction *vp8_sad4x4;
-
-variance_function *vp8_variance4x4;
-variance_function *vp8_variance8x8;
-variance_function *vp8_variance8x16;
-variance_function *vp8_variance16x8;
-variance_function *vp8_variance16x16;
-
-variance_function *vp8_mse16x16;
-
-sub_pixel_variance_function *vp8_sub_pixel_variance4x4;
-sub_pixel_variance_function *vp8_sub_pixel_variance8x8;
-sub_pixel_variance_function *vp8_sub_pixel_variance8x16;
-sub_pixel_variance_function *vp8_sub_pixel_variance16x8;
-sub_pixel_variance_function *vp8_sub_pixel_variance16x16;
-
-int (*vp8_block_error)(short *coeff, short *dqcoeff);
-int (*vp8_mbblock_error)(MACROBLOCK *mb, int dc);
-
-int (*vp8_mbuverror)(MACROBLOCK *mb);
-unsigned int (*vp8_get_mb_ss)(short *);
-void (*vp8_short_fdct4x4)(short *input, short *output, int pitch);
-void (*vp8_short_fdct8x4)(short *input, short *output, int pitch);
+SADFunction *vp9_sad16x16;
+SADFunction *vp9_sad16x8;
+SADFunction *vp9_sad8x16;
+SADFunction *vp9_sad8x8;
+SADFunction *vp9_sad4x4;
+
+variance_function *vp9_variance4x4;
+variance_function *vp9_variance8x8;
+variance_function *vp9_variance8x16;
+variance_function *vp9_variance16x8;
+variance_function *vp9_variance16x16;
+
+variance_function *vp9_mse16x16;
+
+sub_pixel_variance_function *vp9_sub_pixel_variance4x4;
+sub_pixel_variance_function *vp9_sub_pixel_variance8x8;
+sub_pixel_variance_function *vp9_sub_pixel_variance8x16;
+sub_pixel_variance_function *vp9_sub_pixel_variance16x8;
+sub_pixel_variance_function *vp9_sub_pixel_variance16x16;
+
+int (*vp9_block_error)(short *coeff, short *dqcoeff);
+int (*vp9_mbblock_error)(MACROBLOCK *mb, int dc);
+
+int (*vp9_mbuverror)(MACROBLOCK *mb);
+unsigned int (*vp9_get_mb_ss)(short *);
+void (*vp9_short_fdct4x4)(short *input, short *output, int pitch);
+void (*vp9_short_fdct8x4)(short *input, short *output, int pitch);
 void (*vp8_fast_fdct4x4)(short *input, short *output, int pitch);
 void (*vp8_fast_fdct8x4)(short *input, short *output, int pitch);
 void (*short_walsh4x4)(short *input, short *output, int pitch);
 
-void (*vp8_subtract_b)(BLOCK *be, BLOCKD *bd, int pitch);
-void (*vp8_subtract_mby)(short *diff, unsigned char *src, unsigned char *pred, int stride);
-void (*vp8_subtract_mbuv)(short *diff, unsigned char *usrc, unsigned char *vsrc, unsigned char *pred, int stride);
+void (*vp9_subtract_b)(BLOCK *be, BLOCKD *bd, int pitch);
+void (*vp9_subtract_mby)(short *diff, unsigned char *src, unsigned char *pred, int stride);
+void (*vp9_subtract_mbuv)(short *diff, unsigned char *usrc, unsigned char *vsrc, unsigned char *pred, int stride);
 void (*vp8_fast_quantize_b)(BLOCK *b, BLOCKD *d);
 
 // c imports
 extern int block_error_c(short *coeff, short *dqcoeff);
-extern int vp8_mbblock_error_c(MACROBLOCK *mb, int dc);
+extern int vp9_mbblock_error_c(MACROBLOCK *mb, int dc);
 
-extern int vp8_mbuverror_c(MACROBLOCK *mb);
+extern int vp9_mbuverror_c(MACROBLOCK *mb);
 extern unsigned int vp8_get8x8var_c(unsigned char *src_ptr, int  source_stride, unsigned char *ref_ptr, int  recon_stride, unsigned int *SSE, int *Sum);
 extern void short_fdct4x4_c(short *input, short *output, int pitch);
 extern void short_fdct8x4_c(short *input, short *output, int pitch);
-extern void vp8_short_walsh4x4_c(short *input, short *output, int pitch);
+extern void vp9_short_walsh4x4_c(short *input, short *output, int pitch);
 
-extern void vp8_subtract_b_c(BLOCK *be, BLOCKD *bd, int pitch);
+extern void vp9_subtract_b_c(BLOCK *be, BLOCKD *bd, int pitch);
 extern void subtract_mby_c(short *diff, unsigned char *src, unsigned char *pred, int stride);
 extern void subtract_mbuv_c(short *diff, unsigned char *usrc, unsigned char *vsrc, unsigned char *pred, int stride);
 extern void vp8_fast_quantize_b_c(BLOCK *b, BLOCKD *d);
@@ -82,74 +82,74 @@ extern sub_pixel_variance_function sub_pixel_variance8x16_c;
 extern sub_pixel_variance_function sub_pixel_variance16x8_c;
 extern sub_pixel_variance_function sub_pixel_variance16x16_c;
 
-extern unsigned int vp8_get_mb_ss_c(short *);
+extern unsigned int vp9_get_mb_ss_c(short *);
 
 // ppc
-extern int vp8_block_error_ppc(short *coeff, short *dqcoeff);
+extern int vp9_block_error_ppc(short *coeff, short *dqcoeff);
 
-extern void vp8_short_fdct4x4_ppc(short *input, short *output, int pitch);
-extern void vp8_short_fdct8x4_ppc(short *input, short *output, int pitch);
+extern void vp9_short_fdct4x4_ppc(short *input, short *output, int pitch);
+extern void vp9_short_fdct8x4_ppc(short *input, short *output, int pitch);
 
-extern void vp8_subtract_mby_ppc(short *diff, unsigned char *src, unsigned char *pred, int stride);
-extern void vp8_subtract_mbuv_ppc(short *diff, unsigned char *usrc, unsigned char *vsrc, unsigned char *pred, int stride);
+extern void vp9_subtract_mby_ppc(short *diff, unsigned char *src, unsigned char *pred, int stride);
+extern void vp9_subtract_mbuv_ppc(short *diff, unsigned char *usrc, unsigned char *vsrc, unsigned char *pred, int stride);
 
-extern SADFunction vp8_sad16x16_ppc;
-extern SADFunction vp8_sad16x8_ppc;
-extern SADFunction vp8_sad8x16_ppc;
-extern SADFunction vp8_sad8x8_ppc;
-extern SADFunction vp8_sad4x4_ppc;
+extern SADFunction vp9_sad16x16_ppc;
+extern SADFunction vp9_sad16x8_ppc;
+extern SADFunction vp9_sad8x16_ppc;
+extern SADFunction vp9_sad8x8_ppc;
+extern SADFunction vp9_sad4x4_ppc;
 
-extern variance_function vp8_variance16x16_ppc;
-extern variance_function vp8_variance8x16_ppc;
-extern variance_function vp8_variance16x8_ppc;
-extern variance_function vp8_variance8x8_ppc;
-extern variance_function vp8_variance4x4_ppc;
-extern variance_function vp8_mse16x16_ppc;
+extern variance_function vp9_variance16x16_ppc;
+extern variance_function vp9_variance8x16_ppc;
+extern variance_function vp9_variance16x8_ppc;
+extern variance_function vp9_variance8x8_ppc;
+extern variance_function vp9_variance4x4_ppc;
+extern variance_function vp9_mse16x16_ppc;
 
-extern sub_pixel_variance_function vp8_sub_pixel_variance4x4_ppc;
-extern sub_pixel_variance_function vp8_sub_pixel_variance8x8_ppc;
-extern sub_pixel_variance_function vp8_sub_pixel_variance8x16_ppc;
-extern sub_pixel_variance_function vp8_sub_pixel_variance16x8_ppc;
-extern sub_pixel_variance_function vp8_sub_pixel_variance16x16_ppc;
+extern sub_pixel_variance_function vp9_sub_pixel_variance4x4_ppc;
+extern sub_pixel_variance_function vp9_sub_pixel_variance8x8_ppc;
+extern sub_pixel_variance_function vp9_sub_pixel_variance8x16_ppc;
+extern sub_pixel_variance_function vp9_sub_pixel_variance16x8_ppc;
+extern sub_pixel_variance_function vp9_sub_pixel_variance16x16_ppc;
 
 extern unsigned int vp8_get8x8var_ppc(unsigned char *src_ptr, int  source_stride, unsigned char *ref_ptr, int  recon_stride, unsigned int *SSE, int *Sum);
 extern unsigned int vp8_get16x16var_ppc(unsigned char *src_ptr, int  source_stride, unsigned char *ref_ptr, int  recon_stride, unsigned int *SSE, int *Sum);
 
-void vp8_cmachine_specific_config(void) {
+void vp9_cmachine_specific_config(void) {
   // Pure C:
-  vp8_mbuverror               = vp8_mbuverror_c;
+  vp9_mbuverror               = vp9_mbuverror_c;
   vp8_fast_quantize_b           = vp8_fast_quantize_b_c;
-  vp8_short_fdct4x4            = vp8_short_fdct4x4_ppc;
-  vp8_short_fdct8x4            = vp8_short_fdct8x4_ppc;
-  vp8_fast_fdct4x4             = vp8_short_fdct4x4_ppc;
-  vp8_fast_fdct8x4             = vp8_short_fdct8x4_ppc;
-  short_walsh4x4               = vp8_short_walsh4x4_c;
-
-  vp8_variance4x4             = vp8_variance4x4_ppc;
-  vp8_variance8x8             = vp8_variance8x8_ppc;
-  vp8_variance8x16            = vp8_variance8x16_ppc;
-  vp8_variance16x8            = vp8_variance16x8_ppc;
-  vp8_variance16x16           = vp8_variance16x16_ppc;
-  vp8_mse16x16                = vp8_mse16x16_ppc;
-
-  vp8_sub_pixel_variance4x4     = vp8_sub_pixel_variance4x4_ppc;
-  vp8_sub_pixel_variance8x8     = vp8_sub_pixel_variance8x8_ppc;
-  vp8_sub_pixel_variance8x16    = vp8_sub_pixel_variance8x16_ppc;
-  vp8_sub_pixel_variance16x8    = vp8_sub_pixel_variance16x8_ppc;
-  vp8_sub_pixel_variance16x16   = vp8_sub_pixel_variance16x16_ppc;
-
-  vp8_get_mb_ss                 = vp8_get_mb_ss_c;
-
-  vp8_sad16x16                = vp8_sad16x16_ppc;
-  vp8_sad16x8                 = vp8_sad16x8_ppc;
-  vp8_sad8x16                 = vp8_sad8x16_ppc;
-  vp8_sad8x8                  = vp8_sad8x8_ppc;
-  vp8_sad4x4                  = vp8_sad4x4_ppc;
-
-  vp8_block_error              = vp8_block_error_ppc;
-  vp8_mbblock_error            = vp8_mbblock_error_c;
-
-  vp8_subtract_b               = vp8_subtract_b_c;
-  vp8_subtract_mby             = vp8_subtract_mby_ppc;
-  vp8_subtract_mbuv            = vp8_subtract_mbuv_ppc;
+  vp9_short_fdct4x4            = vp9_short_fdct4x4_ppc;
+  vp9_short_fdct8x4            = vp9_short_fdct8x4_ppc;
+  vp8_fast_fdct4x4             = vp9_short_fdct4x4_ppc;
+  vp8_fast_fdct8x4             = vp9_short_fdct8x4_ppc;
+  short_walsh4x4               = vp9_short_walsh4x4_c;
+
+  vp9_variance4x4             = vp9_variance4x4_ppc;
+  vp9_variance8x8             = vp9_variance8x8_ppc;
+  vp9_variance8x16            = vp9_variance8x16_ppc;
+  vp9_variance16x8            = vp9_variance16x8_ppc;
+  vp9_variance16x16           = vp9_variance16x16_ppc;
+  vp9_mse16x16                = vp9_mse16x16_ppc;
+
+  vp9_sub_pixel_variance4x4     = vp9_sub_pixel_variance4x4_ppc;
+  vp9_sub_pixel_variance8x8     = vp9_sub_pixel_variance8x8_ppc;
+  vp9_sub_pixel_variance8x16    = vp9_sub_pixel_variance8x16_ppc;
+  vp9_sub_pixel_variance16x8    = vp9_sub_pixel_variance16x8_ppc;
+  vp9_sub_pixel_variance16x16   = vp9_sub_pixel_variance16x16_ppc;
+
+  vp9_get_mb_ss                 = vp9_get_mb_ss_c;
+
+  vp9_sad16x16                = vp9_sad16x16_ppc;
+  vp9_sad16x8                 = vp9_sad16x8_ppc;
+  vp9_sad8x16                 = vp9_sad8x16_ppc;
+  vp9_sad8x8                  = vp9_sad8x8_ppc;
+  vp9_sad4x4                  = vp9_sad4x4_ppc;
+
+  vp9_block_error              = vp9_block_error_ppc;
+  vp9_mbblock_error            = vp9_mbblock_error_c;
+
+  vp9_subtract_b               = vp9_subtract_b_c;
+  vp9_subtract_mby             = vp9_subtract_mby_ppc;
+  vp9_subtract_mbuv            = vp9_subtract_mbuv_ppc;
 }
diff --git a/vp8/encoder/psnr.c b/vp8/encoder/psnr.c
index 5aa5587ee70bf7c410106342e354af65fb8d8ed1..b447871e058dcb2822186ff0ff63315d1948965f 100644
--- a/vp8/encoder/psnr.c
+++ b/vp8/encoder/psnr.c
@@ -15,7 +15,7 @@
 
 #define MAX_PSNR 100
 
-double vp8_mse2psnr(double Samples, double Peak, double Mse) {
+double vp9_mse2psnr(double Samples, double Peak, double Mse) {
   double psnr;
 
   if ((double)Mse > 0.0)
diff --git a/vp8/encoder/psnr.h b/vp8/encoder/psnr.h
index 7f6269abe4c805a2487032229da2323b8b787253..c25bea75014acaab22847a8218af6625232c9919 100644
--- a/vp8/encoder/psnr.h
+++ b/vp8/encoder/psnr.h
@@ -12,6 +12,6 @@
 #ifndef __INC_PSNR_H
 #define __INC_PSNR_H
 
-extern double vp8_mse2psnr(double Samples, double Peak, double Mse);
+extern double vp9_mse2psnr(double Samples, double Peak, double Mse);
 
 #endif
diff --git a/vp8/encoder/quantize.c b/vp8/encoder/quantize.c
index 600a515e0ef0e4b19fedba688dab828420037ff4..c06557f7f2593a8be2940f4f8d38d70704f43752 100644
--- a/vp8/encoder/quantize.c
+++ b/vp8/encoder/quantize.c
@@ -21,7 +21,7 @@
 extern int enc_debug;
 #endif
 
-void vp8_ht_quantize_b_4x4(BLOCK *b, BLOCKD *d, TX_TYPE tx_type) {
+void vp9_ht_quantize_b_4x4(BLOCK *b, BLOCKD *d, TX_TYPE tx_type) {
   int i, rc, eob;
   int zbin;
   int x, y, z, sz;
@@ -85,7 +85,7 @@ void vp8_ht_quantize_b_4x4(BLOCK *b, BLOCKD *d, TX_TYPE tx_type) {
   d->eob = eob + 1;
 }
 
-void vp8_regular_quantize_b_4x4(BLOCK *b, BLOCKD *d) {
+void vp9_regular_quantize_b_4x4(BLOCK *b, BLOCKD *d) {
   int i, rc, eob;
   int zbin;
   int x, y, z, sz;
@@ -134,7 +134,7 @@ void vp8_regular_quantize_b_4x4(BLOCK *b, BLOCKD *d) {
   d->eob = eob + 1;
 }
 
-void vp8_quantize_mby_4x4_c(MACROBLOCK *x) {
+void vp9_quantize_mby_4x4_c(MACROBLOCK *x) {
   int i;
   int has_2nd_order = x->e_mbd.mode_info_context->mbmi.mode != SPLITMV;
 
@@ -145,19 +145,19 @@ void vp8_quantize_mby_4x4_c(MACROBLOCK *x) {
     x->quantize_b_4x4(&x->block[24], &x->e_mbd.block[24]);
 }
 
-void vp8_quantize_mbuv_4x4_c(MACROBLOCK *x) {
+void vp9_quantize_mbuv_4x4_c(MACROBLOCK *x) {
   int i;
 
   for (i = 16; i < 24; i++)
     x->quantize_b_4x4(&x->block[i], &x->e_mbd.block[i]);
 }
 
-void vp8_quantize_mb_4x4_c(MACROBLOCK *x) {
-  vp8_quantize_mby_4x4_c(x);
-  vp8_quantize_mbuv_4x4_c(x);
+void vp9_quantize_mb_4x4_c(MACROBLOCK *x) {
+  vp9_quantize_mby_4x4_c(x);
+  vp9_quantize_mbuv_4x4_c(x);
 }
 
-void vp8_regular_quantize_b_2x2(BLOCK *b, BLOCKD *d) {
+void vp9_regular_quantize_b_2x2(BLOCK *b, BLOCKD *d) {
   int i, rc, eob;
   int zbin;
   int x, y, z, sz;
@@ -207,7 +207,7 @@ void vp8_regular_quantize_b_2x2(BLOCK *b, BLOCKD *d) {
   d->eob = eob + 1;
 }
 
-void vp8_regular_quantize_b_8x8(BLOCK *b, BLOCKD *d) {
+void vp9_regular_quantize_b_8x8(BLOCK *b, BLOCKD *d) {
   int i, rc, eob;
   int zbin;
   int x, y, z, sz;
@@ -255,7 +255,7 @@ void vp8_regular_quantize_b_8x8(BLOCK *b, BLOCKD *d) {
   d->eob = eob + 1;
 }
 
-void vp8_quantize_mby_8x8(MACROBLOCK *x) {
+void vp9_quantize_mby_8x8(MACROBLOCK *x) {
   int i;
   int has_2nd_order = x->e_mbd.mode_info_context->mbmi.mode != SPLITMV;
 
@@ -270,7 +270,7 @@ void vp8_quantize_mby_8x8(MACROBLOCK *x) {
     x->quantize_b_2x2(&x->block[24], &x->e_mbd.block[24]);
 }
 
-void vp8_quantize_mbuv_8x8(MACROBLOCK *x) {
+void vp9_quantize_mbuv_8x8(MACROBLOCK *x) {
   int i;
 
   for (i = 16; i < 24; i ++)
@@ -279,12 +279,12 @@ void vp8_quantize_mbuv_8x8(MACROBLOCK *x) {
     x->quantize_b_8x8(&x->block[i], &x->e_mbd.block[i]);
 }
 
-void vp8_quantize_mb_8x8(MACROBLOCK *x) {
-  vp8_quantize_mby_8x8(x);
-  vp8_quantize_mbuv_8x8(x);
+void vp9_quantize_mb_8x8(MACROBLOCK *x) {
+  vp9_quantize_mby_8x8(x);
+  vp9_quantize_mbuv_8x8(x);
 }
 
-void vp8_quantize_mby_16x16(MACROBLOCK *x) {
+void vp9_quantize_mby_16x16(MACROBLOCK *x) {
   int i;
 
   for (i = 0; i < 16; i++)
@@ -293,12 +293,12 @@ void vp8_quantize_mby_16x16(MACROBLOCK *x) {
   x->quantize_b_16x16(&x->block[0], &x->e_mbd.block[0]);
 }
 
-void vp8_quantize_mb_16x16(MACROBLOCK *x) {
-  vp8_quantize_mby_16x16(x);
-  vp8_quantize_mbuv_8x8(x);
+void vp9_quantize_mb_16x16(MACROBLOCK *x) {
+  vp9_quantize_mby_16x16(x);
+  vp9_quantize_mbuv_8x8(x);
 }
 
-void vp8_regular_quantize_b_16x16(BLOCK *b, BLOCKD *d) {
+void vp9_regular_quantize_b_16x16(BLOCK *b, BLOCKD *d) {
   int i, rc, eob;
   int zbin;
   int x, y, z, sz;
@@ -349,9 +349,9 @@ void vp8_regular_quantize_b_16x16(BLOCK *b, BLOCKD *d) {
  * these two C functions if corresponding optimized routine is not available.
  * NEON optimized version implements currently the fast quantization for pair
  * of blocks. */
-void vp8_regular_quantize_b_4x4_pair(BLOCK *b1, BLOCK *b2, BLOCKD *d1, BLOCKD *d2) {
-  vp8_regular_quantize_b_4x4(b1, d1);
-  vp8_regular_quantize_b_4x4(b2, d2);
+void vp9_regular_quantize_b_4x4_pair(BLOCK *b1, BLOCK *b2, BLOCKD *d1, BLOCKD *d2) {
+  vp9_regular_quantize_b_4x4(b1, d1);
+  vp9_regular_quantize_b_4x4(b2, d2);
 }
 
 static void invert_quant(short *quant,
@@ -366,7 +366,7 @@ static void invert_quant(short *quant,
   *shift = l;
 }
 
-void vp8cx_init_quantizer(VP8_COMP *cpi) {
+void vp9cx_init_quantizer(VP8_COMP *cpi) {
   int i;
   int quant_val;
   int Q;
@@ -530,7 +530,7 @@ void vp8cx_init_quantizer(VP8_COMP *cpi) {
   }
 }
 
-void vp8cx_mb_init_quantizer(VP8_COMP *cpi, MACROBLOCK *x) {
+void vp9cx_mb_init_quantizer(VP8_COMP *cpi, MACROBLOCK *x) {
   int i;
   int QIndex;
   MACROBLOCKD *xd = &x->e_mbd;
@@ -650,11 +650,11 @@ void vp8cx_mb_init_quantizer(VP8_COMP *cpi, MACROBLOCK *x) {
     x->block[24].eob_max_offset_8x8 = 4;
   }
 
-  /* save this macroblock QIndex for vp8_update_zbin_extra() */
+  /* save this macroblock QIndex for vp9_update_zbin_extra() */
   x->e_mbd.q_index = QIndex;
 }
 
-void vp8_update_zbin_extra(VP8_COMP *cpi, MACROBLOCK *x) {
+void vp9_update_zbin_extra(VP8_COMP *cpi, MACROBLOCK *x) {
   int i;
   int QIndex = x->e_mbd.q_index;
   int zbin_extra;
@@ -687,15 +687,15 @@ void vp8_update_zbin_extra(VP8_COMP *cpi, MACROBLOCK *x) {
   x->block[24].zbin_extra = (short)zbin_extra;
 }
 
-void vp8cx_frame_init_quantizer(VP8_COMP *cpi) {
+void vp9cx_frame_init_quantizer(VP8_COMP *cpi) {
   // Clear Zbin mode boost for default case
   cpi->zbin_mode_boost = 0;
 
   // MB level quantizer setup
-  vp8cx_mb_init_quantizer(cpi, &cpi->mb);
+  vp9cx_mb_init_quantizer(cpi, &cpi->mb);
 }
 
-void vp8_set_quantizer(struct VP8_COMP *cpi, int Q) {
+void vp9_set_quantizer(struct VP8_COMP *cpi, int Q) {
   VP8_COMMON *cm = &cpi->common;
 
   cm->base_qindex = Q;
@@ -711,5 +711,5 @@ void vp8_set_quantizer(struct VP8_COMP *cpi, int Q) {
   // quantizer has to be reinitialized if any delta_q changes.
   // As there are not any here for now this is inactive code.
   // if(update)
-  //    vp8cx_init_quantizer(cpi);
+  //    vp9cx_init_quantizer(cpi);
 }
diff --git a/vp8/encoder/quantize.h b/vp8/encoder/quantize.h
index e39433fc3f81c3d8b6f6bfa5b05bfedd4a175d0e..8fe27b8ee94599b25f88e9a9856c23fc50cd81af 100644
--- a/vp8/encoder/quantize.h
+++ b/vp8/encoder/quantize.h
@@ -32,61 +32,61 @@
 
 #define prototype_quantize_block_type(sym) \
   void (sym)(BLOCK *b, BLOCKD *d, TX_TYPE type)
-extern prototype_quantize_block_type(vp8_ht_quantize_b_4x4);
+extern prototype_quantize_block_type(vp9_ht_quantize_b_4x4);
 
 #ifndef vp8_quantize_quantb_4x4
-#define vp8_quantize_quantb_4x4 vp8_regular_quantize_b_4x4
+#define vp8_quantize_quantb_4x4 vp9_regular_quantize_b_4x4
 #endif
 extern prototype_quantize_block(vp8_quantize_quantb_4x4);
 
 #ifndef vp8_quantize_quantb_4x4_pair
-#define vp8_quantize_quantb_4x4_pair vp8_regular_quantize_b_4x4_pair
+#define vp8_quantize_quantb_4x4_pair vp9_regular_quantize_b_4x4_pair
 #endif
 extern prototype_quantize_block_pair(vp8_quantize_quantb_4x4_pair);
 
 #ifndef vp8_quantize_quantb_8x8
-#define vp8_quantize_quantb_8x8 vp8_regular_quantize_b_8x8
+#define vp8_quantize_quantb_8x8 vp9_regular_quantize_b_8x8
 #endif
 extern prototype_quantize_block(vp8_quantize_quantb_8x8);
 
 #ifndef vp8_quantize_quantb_16x16
-#define vp8_quantize_quantb_16x16 vp8_regular_quantize_b_16x16
+#define vp8_quantize_quantb_16x16 vp9_regular_quantize_b_16x16
 #endif
 extern prototype_quantize_block(vp8_quantize_quantb_16x16);
 
 #ifndef vp8_quantize_quantb_2x2
-#define vp8_quantize_quantb_2x2 vp8_regular_quantize_b_2x2
+#define vp8_quantize_quantb_2x2 vp9_regular_quantize_b_2x2
 #endif
 extern prototype_quantize_block(vp8_quantize_quantb_2x2);
 
-#ifndef vp8_quantize_mb_4x4
-#define vp8_quantize_mb_4x4 vp8_quantize_mb_4x4_c
+#ifndef vp9_quantize_mb_4x4
+#define vp9_quantize_mb_4x4 vp9_quantize_mb_4x4_c
 #endif
-extern prototype_quantize_mb(vp8_quantize_mb_4x4);
-void vp8_quantize_mb_8x8(MACROBLOCK *x);
+extern prototype_quantize_mb(vp9_quantize_mb_4x4);
+void vp9_quantize_mb_8x8(MACROBLOCK *x);
 
-#ifndef vp8_quantize_mbuv_4x4
-#define vp8_quantize_mbuv_4x4 vp8_quantize_mbuv_4x4_c
+#ifndef vp9_quantize_mbuv_4x4
+#define vp9_quantize_mbuv_4x4 vp9_quantize_mbuv_4x4_c
 #endif
-extern prototype_quantize_mb(vp8_quantize_mbuv_4x4);
+extern prototype_quantize_mb(vp9_quantize_mbuv_4x4);
 
-#ifndef vp8_quantize_mby_4x4
-#define vp8_quantize_mby_4x4 vp8_quantize_mby_4x4_c
+#ifndef vp9_quantize_mby_4x4
+#define vp9_quantize_mby_4x4 vp9_quantize_mby_4x4_c
 #endif
-extern prototype_quantize_mb(vp8_quantize_mby_4x4);
+extern prototype_quantize_mb(vp9_quantize_mby_4x4);
 
-extern prototype_quantize_mb(vp8_quantize_mby_8x8);
-extern prototype_quantize_mb(vp8_quantize_mbuv_8x8);
+extern prototype_quantize_mb(vp9_quantize_mby_8x8);
+extern prototype_quantize_mb(vp9_quantize_mbuv_8x8);
 
-void vp8_quantize_mb_16x16(MACROBLOCK *x);
+void vp9_quantize_mb_16x16(MACROBLOCK *x);
 extern prototype_quantize_block(vp8_quantize_quantb_16x16);
-extern prototype_quantize_mb(vp8_quantize_mby_16x16);
+extern prototype_quantize_mb(vp9_quantize_mby_16x16);
 
 struct VP8_COMP;
-extern void vp8_set_quantizer(struct VP8_COMP *cpi, int Q);
-extern void vp8cx_frame_init_quantizer(struct VP8_COMP *cpi);
-extern void vp8_update_zbin_extra(struct VP8_COMP *cpi, MACROBLOCK *x);
-extern void vp8cx_mb_init_quantizer(struct VP8_COMP *cpi, MACROBLOCK *x);
-extern void vp8cx_init_quantizer(struct VP8_COMP *cpi);
+extern void vp9_set_quantizer(struct VP8_COMP *cpi, int Q);
+extern void vp9cx_frame_init_quantizer(struct VP8_COMP *cpi);
+extern void vp9_update_zbin_extra(struct VP8_COMP *cpi, MACROBLOCK *x);
+extern void vp9cx_mb_init_quantizer(struct VP8_COMP *cpi, MACROBLOCK *x);
+extern void vp9cx_init_quantizer(struct VP8_COMP *cpi);
 
 #endif
diff --git a/vp8/encoder/ratectrl.c b/vp8/encoder/ratectrl.c
index 61fdf291ec8c60c84e752956ad03ead92acb2b75..eea2a9083e4d6611ead52e2f94c58f868eed3e93 100644
--- a/vp8/encoder/ratectrl.c
+++ b/vp8/encoder/ratectrl.c
@@ -88,16 +88,16 @@ static const unsigned int prior_key_frame_weight[KEY_FRAME_CONTEXT] = { 1, 2, 3,
 // These functions use formulaic calculations to make playing with the
 // quantizer tables easier. If necessary they can be replaced by lookup
 // tables if and when things settle down in the experimental bitstream
-double vp8_convert_qindex_to_q(int qindex) {
+double vp9_convert_qindex_to_q(int qindex) {
   // Convert the index to a real Q value (scaled down to match old Q values)
   return (double)vp8_ac_yquant(qindex) / 4.0;
 }
 
-int vp8_gfboost_qadjust(int qindex) {
+int vp9_gfboost_qadjust(int qindex) {
   int retval;
   double q;
 
-  q = vp8_convert_qindex_to_q(qindex);
+  q = vp9_convert_qindex_to_q(qindex);
   retval = (int)((0.00000828 * q * q * q) +
                  (-0.0055 * q * q) +
                  (1.32 * q) + 79.3);
@@ -108,28 +108,28 @@ static int kfboost_qadjust(int qindex) {
   int retval;
   double q;
 
-  q = vp8_convert_qindex_to_q(qindex);
+  q = vp9_convert_qindex_to_q(qindex);
   retval = (int)((0.00000973 * q * q * q) +
                  (-0.00613 * q * q) +
                  (1.316 * q) + 121.2);
   return retval;
 }
 
-int vp8_bits_per_mb(FRAME_TYPE frame_type, int qindex) {
+int vp9_bits_per_mb(FRAME_TYPE frame_type, int qindex) {
   if (frame_type == KEY_FRAME)
-    return (int)(4500000 / vp8_convert_qindex_to_q(qindex));
+    return (int)(4500000 / vp9_convert_qindex_to_q(qindex));
   else
-    return (int)(2850000 / vp8_convert_qindex_to_q(qindex));
+    return (int)(2850000 / vp9_convert_qindex_to_q(qindex));
 }
 
 
-void vp8_save_coding_context(VP8_COMP *cpi) {
+void vp9_save_coding_context(VP8_COMP *cpi) {
   CODING_CONTEXT *const cc = &cpi->coding_context;
   VP8_COMMON *cm = &cpi->common;
   MACROBLOCKD *xd = &cpi->mb.e_mbd;
 
   // Stores a snapshot of key state variables which can subsequently be
-  // restored with a call to vp8_restore_coding_context. These functions are
+  // restored with a call to vp9_restore_coding_context. These functions are
   // intended for use in a re-code loop in vp8_compress_frame where the
   // quantizer value is adjusted between loop iterations.
 
@@ -180,13 +180,13 @@ void vp8_save_coding_context(VP8_COMP *cpi) {
   vp8_copy(cc->switchable_interp_prob, cm->fc.switchable_interp_prob);
 }
 
-void vp8_restore_coding_context(VP8_COMP *cpi) {
+void vp9_restore_coding_context(VP8_COMP *cpi) {
   CODING_CONTEXT *const cc = &cpi->coding_context;
   VP8_COMMON *cm = &cpi->common;
   MACROBLOCKD *xd = &cpi->mb.e_mbd;
 
   // Restore key state variables to the snapshot state stored in the
-  // previous call to vp8_save_coding_context.
+  // previous call to vp9_save_coding_context.
 
   cm->fc.nmvc = cc->nmvc;
   vp8_copy(cpi->mb.nmvjointcost, cc->nmvjointcost);
@@ -237,7 +237,7 @@ void vp8_restore_coding_context(VP8_COMP *cpi) {
 }
 
 
-void vp8_setup_key_frame(VP8_COMP *cpi) {
+void vp9_setup_key_frame(VP8_COMP *cpi) {
   VP8_COMMON *cm = &cpi->common;
   // Setup for Key frame:
   vp8_default_coef_probs(& cpi->common);
@@ -269,7 +269,7 @@ void vp8_setup_key_frame(VP8_COMP *cpi) {
   vp9_update_mode_info_in_image(cm, cm->mi);
 }
 
-void vp8_setup_inter_frame(VP8_COMP *cpi) {
+void vp9_setup_inter_frame(VP8_COMP *cpi) {
   if (cpi->common.refresh_alt_ref_frame) {
     vpx_memcpy(&cpi->common.fc,
                &cpi->common.lfc_a,
@@ -290,7 +290,7 @@ void vp8_setup_inter_frame(VP8_COMP *cpi) {
 
 static int estimate_bits_at_q(int frame_kind, int Q, int MBs,
                               double correction_factor) {
-  int Bpm = (int)(.5 + correction_factor * vp8_bits_per_mb(frame_kind, Q));
+  int Bpm = (int)(.5 + correction_factor * vp9_bits_per_mb(frame_kind, Q));
 
   /* Attempt to retain reasonable accuracy without overflow. The cutoff is
    * chosen such that the maximum product of Bpm and MBs fits 31 bits. The
@@ -408,7 +408,7 @@ static void calc_pframe_target_size(VP8_COMP *cpi) {
 }
 
 
-void vp8_update_rate_correction_factors(VP8_COMP *cpi, int damp_var) {
+void vp9_update_rate_correction_factors(VP8_COMP *cpi, int damp_var) {
   int    Q = cpi->common.base_qindex;
   int    correction_factor = 100;
   double rate_correction_factor;
@@ -432,7 +432,7 @@ void vp8_update_rate_correction_factors(VP8_COMP *cpi, int damp_var) {
   // Stay in double to avoid int overflow when values are large
   projected_size_based_on_q =
     (int)(((.5 + rate_correction_factor *
-            vp8_bits_per_mb(cpi->common.frame_type, Q)) *
+            vp9_bits_per_mb(cpi->common.frame_type, Q)) *
            cpi->common.MBs) / (1 << BPER_MB_NORMBITS));
 
   // Make some allowance for cpi->zbin_over_quant
@@ -504,7 +504,7 @@ void vp8_update_rate_correction_factors(VP8_COMP *cpi, int damp_var) {
 }
 
 
-int vp8_regulate_q(VP8_COMP *cpi, int target_bits_per_frame) {
+int vp9_regulate_q(VP8_COMP *cpi, int target_bits_per_frame) {
   int Q = cpi->active_worst_quality;
 
   int i;
@@ -537,7 +537,7 @@ int vp8_regulate_q(VP8_COMP *cpi, int target_bits_per_frame) {
   do {
     bits_per_mb_at_this_q =
       (int)(.5 + correction_factor *
-            vp8_bits_per_mb(cpi->common.frame_type, i));
+            vp9_bits_per_mb(cpi->common.frame_type, i));
 
     if (bits_per_mb_at_this_q <= target_bits_per_mb) {
       if ((target_bits_per_mb - bits_per_mb_at_this_q) <= last_error)
@@ -641,7 +641,7 @@ static int estimate_keyframe_frequency(VP8_COMP *cpi) {
 }
 
 
-void vp8_adjust_key_frame_context(VP8_COMP *cpi) {
+void vp9_adjust_key_frame_context(VP8_COMP *cpi) {
   // Clear down mmx registers to allow floating point in what follows
   vp8_clear_system_state();
 
@@ -650,7 +650,7 @@ void vp8_adjust_key_frame_context(VP8_COMP *cpi) {
 }
 
 
-void vp8_compute_frame_size_bounds(VP8_COMP *cpi, int *frame_under_shoot_limit, int *frame_over_shoot_limit) {
+void vp9_compute_frame_size_bounds(VP8_COMP *cpi, int *frame_under_shoot_limit, int *frame_over_shoot_limit) {
   // Set-up bounds on acceptable frame size:
   if (cpi->oxcf.fixed_q >= 0) {
     // Fixed Q scenario: frame size never outranges target (there is no target!)
@@ -688,7 +688,7 @@ void vp8_compute_frame_size_bounds(VP8_COMP *cpi, int *frame_under_shoot_limit,
 
 
 // return of 0 means drop frame
-int vp8_pick_frame_size(VP8_COMP *cpi) {
+int vp9_pick_frame_size(VP8_COMP *cpi) {
   VP8_COMMON *cm = &cpi->common;
 
   if (cm->frame_type == KEY_FRAME)
diff --git a/vp8/encoder/ratectrl.h b/vp8/encoder/ratectrl.h
index ac1a76f4ccf2089be1b51f8071a62496ec3dc71b..bd4981377335dfd8580aba46f73ccc06fb589b48 100644
--- a/vp8/encoder/ratectrl.h
+++ b/vp8/encoder/ratectrl.h
@@ -15,21 +15,21 @@
 
 #define FRAME_OVERHEAD_BITS 200
 
-extern void vp8_save_coding_context(VP8_COMP *cpi);
-extern void vp8_restore_coding_context(VP8_COMP *cpi);
+extern void vp9_save_coding_context(VP8_COMP *cpi);
+extern void vp9_restore_coding_context(VP8_COMP *cpi);
 
-extern void vp8_setup_key_frame(VP8_COMP *cpi);
-extern void vp8_update_rate_correction_factors(VP8_COMP *cpi, int damp_var);
-extern int vp8_regulate_q(VP8_COMP *cpi, int target_bits_per_frame);
-extern void vp8_adjust_key_frame_context(VP8_COMP *cpi);
-extern void vp8_compute_frame_size_bounds(VP8_COMP *cpi, int *frame_under_shoot_limit, int *frame_over_shoot_limit);
+extern void vp9_setup_key_frame(VP8_COMP *cpi);
+extern void vp9_update_rate_correction_factors(VP8_COMP *cpi, int damp_var);
+extern int vp9_regulate_q(VP8_COMP *cpi, int target_bits_per_frame);
+extern void vp9_adjust_key_frame_context(VP8_COMP *cpi);
+extern void vp9_compute_frame_size_bounds(VP8_COMP *cpi, int *frame_under_shoot_limit, int *frame_over_shoot_limit);
 
 // return of 0 means drop frame
-extern int vp8_pick_frame_size(VP8_COMP *cpi);
+extern int vp9_pick_frame_size(VP8_COMP *cpi);
 
-extern double vp8_convert_qindex_to_q(int qindex);
-extern int vp8_gfboost_qadjust(int qindex);
-extern int vp8_bits_per_mb(FRAME_TYPE frame_type, int qindex);
-void vp8_setup_inter_frame(VP8_COMP *cpi);
+extern double vp9_convert_qindex_to_q(int qindex);
+extern int vp9_gfboost_qadjust(int qindex);
+extern int vp9_bits_per_mb(FRAME_TYPE frame_type, int qindex);
+void vp9_setup_inter_frame(VP8_COMP *cpi);
 
 #endif
diff --git a/vp8/encoder/rdopt.c b/vp8/encoder/rdopt.c
index 8207574dd370309743e2fc82e02396ed89e7b413..8c9bc6369c1dd790da8684504f435d8dbacb2f2f 100644
--- a/vp8/encoder/rdopt.c
+++ b/vp8/encoder/rdopt.c
@@ -52,8 +52,8 @@
 #define IF_RTCD(x)  NULL
 #endif
 
-extern void vp8cx_mb_init_quantizer(VP8_COMP *cpi, MACROBLOCK *x);
-extern void vp8_update_zbin_extra(VP8_COMP *cpi, MACROBLOCK *x);
+extern void vp9cx_mb_init_quantizer(VP8_COMP *cpi, MACROBLOCK *x);
+extern void vp9_update_zbin_extra(VP8_COMP *cpi, MACROBLOCK *x);
 
 #define MAXF(a,b)            (((a) > (b)) ? (a) : (b))
 
@@ -226,11 +226,11 @@ static void fill_token_costs(
     for (j = 0; j < COEF_BANDS; j++)
       for (k = 0; k < PREV_COEF_CONTEXTS; k++) {
         if (k == 0 && ((j > 0 && i > 0) || (j > 1 && i == 0)))
-          vp8_cost_tokens_skip((int *)(c [i][j][k]),
+          vp9_cost_tokens_skip((int *)(c [i][j][k]),
                                p [i][j][k],
                                vp8_coef_tree);
         else
-          vp8_cost_tokens((int *)(c [i][j][k]),
+          vp9_cost_tokens((int *)(c [i][j][k]),
                           p [i][j][k],
                           vp8_coef_tree);
       }
@@ -249,7 +249,7 @@ static int rd_iifactor [ 32 ] =  {    4,   4,   3,   2,   1,   0,   0,   0,
 static int sad_per_bit16lut[QINDEX_RANGE];
 static int sad_per_bit4lut[QINDEX_RANGE];
 
-void vp8_init_me_luts() {
+void vp9_init_me_luts() {
   int i;
 
   // Initialize the sad lut tables using a formulaic calculation for now
@@ -257,8 +257,8 @@ void vp8_init_me_luts() {
   // to the quantizer tables.
   for (i = 0; i < QINDEX_RANGE; i++) {
     sad_per_bit16lut[i] =
-      (int)((0.0418 * vp8_convert_qindex_to_q(i)) + 2.4107);
-    sad_per_bit4lut[i] = (int)((0.063 * vp8_convert_qindex_to_q(i)) + 2.742);
+      (int)((0.0418 * vp9_convert_qindex_to_q(i)) + 2.4107);
+    sad_per_bit4lut[i] = (int)((0.063 * vp9_convert_qindex_to_q(i)) + 2.742);
   }
 }
 
@@ -269,13 +269,13 @@ static int compute_rd_mult(int qindex) {
   return (11 * q * q) >> 6;
 }
 
-void vp8cx_initialize_me_consts(VP8_COMP *cpi, int QIndex) {
+void vp9cx_initialize_me_consts(VP8_COMP *cpi, int QIndex) {
   cpi->mb.sadperbit16 =  sad_per_bit16lut[QIndex];
   cpi->mb.sadperbit4  =  sad_per_bit4lut[QIndex];
 }
 
 
-void vp8_initialize_rd_consts(VP8_COMP *cpi, int QIndex) {
+void vp9_initialize_rd_consts(VP8_COMP *cpi, int QIndex) {
   int q, i;
 
   vp8_clear_system_state();  // __asm emms;
@@ -312,7 +312,7 @@ void vp8_initialize_rd_consts(VP8_COMP *cpi, int QIndex) {
   cpi->mb.errorperbit = (cpi->RDMULT / 110);
   cpi->mb.errorperbit += (cpi->mb.errorperbit == 0);
 
-  vp8_set_speed_features(cpi);
+  vp9_set_speed_features(cpi);
 
   q = (int)pow(vp8_dc_quant(QIndex, 0) >> 2, 1.25);
   q = q << 2;
@@ -380,11 +380,11 @@ void vp8_initialize_rd_consts(VP8_COMP *cpi, int QIndex) {
 
   /*rough estimate for costing*/
   cpi->common.kf_ymode_probs_index = cpi->common.base_qindex >> 4;
-  vp8_init_mode_costs(cpi);
+  vp9_init_mode_costs(cpi);
 
   if (cpi->common.frame_type != KEY_FRAME)
   {
-    vp8_build_nmv_cost_table(
+    vp9_build_nmv_cost_table(
         cpi->mb.nmvjointcost,
         cpi->mb.e_mbd.allow_high_precision_mv ?
         cpi->mb.nmvcost_hp : cpi->mb.nmvcost,
@@ -393,7 +393,7 @@ void vp8_initialize_rd_consts(VP8_COMP *cpi, int QIndex) {
   }
 }
 
-void vp8_auto_select_speed(VP8_COMP *cpi) {
+void vp9_auto_select_speed(VP8_COMP *cpi) {
   int milliseconds_for_compress = (int)(1000000 / cpi->oxcf.frame_rate);
 
   milliseconds_for_compress = milliseconds_for_compress * (16 - cpi->oxcf.cpu_used) / 16;
@@ -446,7 +446,7 @@ void vp8_auto_select_speed(VP8_COMP *cpi) {
   }
 }
 
-int vp8_block_error_c(short *coeff, short *dqcoeff, int block_size) {
+int vp9_block_error_c(short *coeff, short *dqcoeff, int block_size) {
   int i, error = 0;
 
   for (i = 0; i < block_size; i++) {
@@ -457,7 +457,7 @@ int vp8_block_error_c(short *coeff, short *dqcoeff, int block_size) {
   return error;
 }
 
-int vp8_mbblock_error_c(MACROBLOCK *mb, int dc) {
+int vp9_mbblock_error_c(MACROBLOCK *mb, int dc) {
   BLOCK  *be;
   BLOCKD *bd;
   int i, j;
@@ -480,7 +480,7 @@ int vp8_mbblock_error_c(MACROBLOCK *mb, int dc) {
   return error;
 }
 
-int vp8_mbuverror_c(MACROBLOCK *mb) {
+int vp9_mbuverror_c(MACROBLOCK *mb) {
   BLOCK  *be;
   BLOCKD *bd;
 
@@ -490,13 +490,13 @@ int vp8_mbuverror_c(MACROBLOCK *mb) {
     be = &mb->block[i];
     bd = &mb->e_mbd.block[i];
 
-    error += vp8_block_error_c(be->coeff, bd->dqcoeff, 16);
+    error += vp9_block_error_c(be->coeff, bd->dqcoeff, 16);
   }
 
   return error;
 }
 
-int vp8_uvsse(MACROBLOCK *x) {
+int vp9_uvsse(MACROBLOCK *x) {
   unsigned char *uptr, *vptr;
   unsigned char *upred_ptr = (*(x->block[16].base_src) + x->block[16].src);
   unsigned char *vpred_ptr = (*(x->block[20].base_src) + x->block[20].src);
@@ -527,14 +527,14 @@ int vp8_uvsse(MACROBLOCK *x) {
   vptr = x->e_mbd.pre.v_buffer + offset;
 
   if ((mv_row | mv_col) & 7) {
-    vp8_sub_pixel_variance8x8(uptr, pre_stride, (mv_col & 7) << 1,
+    vp9_sub_pixel_variance8x8(uptr, pre_stride, (mv_col & 7) << 1,
                               (mv_row & 7) << 1, upred_ptr, uv_stride, &sse2);
-    vp8_sub_pixel_variance8x8(vptr, pre_stride, (mv_col & 7) << 1,
+    vp9_sub_pixel_variance8x8(vptr, pre_stride, (mv_col & 7) << 1,
                               (mv_row & 7) << 1, vpred_ptr, uv_stride, &sse1);
     sse2 += sse1;
   } else {
-    vp8_variance8x8(uptr, pre_stride, upred_ptr, uv_stride, &sse2);
-    vp8_variance8x8(vptr, pre_stride, vpred_ptr, uv_stride, &sse1);
+    vp9_variance8x8(uptr, pre_stride, upred_ptr, uv_stride, &sse2);
+    vp9_variance8x8(vptr, pre_stride, vpred_ptr, uv_stride, &sse1);
     sse2 += sse1;
   }
   return sse2;
@@ -709,12 +709,12 @@ static void macro_block_yrd_4x4(MACROBLOCK *mb,
   BLOCK *beptr;
   int d;
 
-  vp8_subtract_mby(mb->src_diff, *(mb->block[0].base_src), xd->predictor,
+  vp9_subtract_mby(mb->src_diff, *(mb->block[0].base_src), xd->predictor,
                    mb->block[0].src_stride);
 
   // Fdct and building the 2nd order block
   for (beptr = mb->block; beptr < mb->block + 16; beptr += 2) {
-    mb->vp8_short_fdct8x4(beptr->src_diff, beptr->coeff, 32);
+    mb->vp9_short_fdct8x4(beptr->src_diff, beptr->coeff, 32);
     *Y2DCPtr++ = beptr->coeff[0];
     *Y2DCPtr++ = beptr->coeff[16];
   }
@@ -731,9 +731,9 @@ static void macro_block_yrd_4x4(MACROBLOCK *mb,
   mb->quantize_b_4x4(mb_y2, x_y2);
 
   // Distortion
-  d = vp8_mbblock_error(mb, 1);
+  d = vp9_mbblock_error(mb, 1);
 
-  d += vp8_block_error(mb_y2->coeff, x_y2->dqcoeff, 16);
+  d += vp9_block_error(mb_y2->coeff, x_y2->dqcoeff, 16);
 
   *Distortion = (d >> 2);
   // rate
@@ -780,11 +780,11 @@ static void macro_block_yrd_8x8(MACROBLOCK *mb,
   BLOCKD *const x_y2  = xd->block + 24;
   int d;
 
-  vp8_subtract_mby(mb->src_diff, *(mb->block[0].base_src), xd->predictor,
+  vp9_subtract_mby(mb->src_diff, *(mb->block[0].base_src), xd->predictor,
                    mb->block[0].src_stride);
 
-  vp8_transform_mby_8x8(mb);
-  vp8_quantize_mby_8x8(mb);
+  vp9_transform_mby_8x8(mb);
+  vp9_quantize_mby_8x8(mb);
 
   /* remove 1st order dc to properly combine 1st/2nd order distortion */
   mb->coeff[0] = 0;
@@ -796,8 +796,8 @@ static void macro_block_yrd_8x8(MACROBLOCK *mb,
   xd->dqcoeff[128] = 0;
   xd->dqcoeff[192] = 0;
 
-  d = vp8_mbblock_error(mb, 0);
-  d += vp8_block_error(mb_y2->coeff, x_y2->dqcoeff, 16);
+  d = vp9_mbblock_error(mb, 0);
+  d += vp9_block_error(mb_y2->coeff, x_y2->dqcoeff, 16);
 
   *Distortion = (d >> 2);
   // rate
@@ -829,23 +829,23 @@ static void macro_block_yrd_16x16(MACROBLOCK *mb, int *Rate, int *Distortion,
   BLOCK  *be = &mb->block[0];
   TX_TYPE tx_type;
 
-  vp8_subtract_mby(mb->src_diff, *(mb->block[0].base_src), mb->e_mbd.predictor,
+  vp9_subtract_mby(mb->src_diff, *(mb->block[0].base_src), mb->e_mbd.predictor,
                    mb->block[0].src_stride);
 
   tx_type = get_tx_type_16x16(xd, b);
   if (tx_type != DCT_DCT) {
-    vp8_fht_c(be->src_diff, 32, be->coeff, tx_type, 16);
+    vp9_fht_c(be->src_diff, 32, be->coeff, tx_type, 16);
   } else
-    vp8_transform_mby_16x16(mb);
+    vp9_transform_mby_16x16(mb);
 
-  vp8_quantize_mby_16x16(mb);
+  vp9_quantize_mby_16x16(mb);
   // TODO(jingning) is it possible to quickly determine whether to force
   //                trailing coefficients to be zero, instead of running trellis
   //                optimization in the rate-distortion optimization loop?
   if (mb->e_mbd.mode_info_context->mbmi.mode < I8X8_PRED)
-    vp8_optimize_mby_16x16(mb, rtcd);
+    vp9_optimize_mby_16x16(mb, rtcd);
 
-  d = vp8_mbblock_error(mb, 0);
+  d = vp9_mbblock_error(mb, 0);
 
   *Distortion = (d >> 2);
   // rate
@@ -1001,13 +1001,13 @@ static void super_block_yrd_8x8(MACROBLOCK *x,
   for (n = 0; n < 4; n++) {
     int x_idx = n & 1, y_idx = n >> 1;
 
-    vp8_subtract_mby_s_c(x->src_diff,
+    vp9_subtract_mby_s_c(x->src_diff,
                          src + x_idx * 16 + y_idx * 16 * src_y_stride,
                          src_y_stride,
                          dst + x_idx * 16 + y_idx * 16 * dst_y_stride,
                          dst_y_stride);
-    vp8_transform_mby_8x8(x);
-    vp8_quantize_mby_8x8(x);
+    vp9_transform_mby_8x8(x);
+    vp9_quantize_mby_8x8(x);
 
     /* remove 1st order dc to properly combine 1st/2nd order distortion */
     x->coeff[  0] = 0;
@@ -1019,8 +1019,8 @@ static void super_block_yrd_8x8(MACROBLOCK *x,
     xd->dqcoeff[128] = 0;
     xd->dqcoeff[192] = 0;
 
-    d += vp8_mbblock_error(x, 0);
-    d += vp8_block_error(by2->coeff, bdy2->dqcoeff, 16);
+    d += vp9_mbblock_error(x, 0);
+    d += vp9_block_error(by2->coeff, bdy2->dqcoeff, 16);
     xd->above_context = ta + x_idx;
     xd->left_context = tl + y_idx;
     r += vp8_rdcost_mby_8x8(x, 0);
@@ -1111,15 +1111,15 @@ static int64_t rd_pick_intra4x4block(VP8_COMP *cpi, MACROBLOCK *x, BLOCK *be,
         rate += bmode_costs[mode2];
       }
 #endif
-      vp8_subtract_b(be, b, 16);
+      vp9_subtract_b(be, b, 16);
 
       b->bmi.as_mode.first = mode;
       tx_type = get_tx_type_4x4(xd, b);
       if (tx_type != DCT_DCT) {
-        vp8_fht_c(be->src_diff, 32, be->coeff, tx_type, 4);
-        vp8_ht_quantize_b_4x4(be, b, tx_type);
+        vp9_fht_c(be->src_diff, 32, be->coeff, tx_type, 4);
+        vp9_ht_quantize_b_4x4(be, b, tx_type);
       } else {
-        x->vp8_short_fdct4x4(be->src_diff, be->coeff, 32);
+        x->vp9_short_fdct4x4(be->src_diff, be->coeff, 32);
         x->quantize_b_4x4(be, b);
       }
 
@@ -1128,7 +1128,7 @@ static int64_t rd_pick_intra4x4block(VP8_COMP *cpi, MACROBLOCK *x, BLOCK *be,
 
       ratey = cost_coeffs(x, b, PLANE_TYPE_Y_WITH_DC, &tempa, &templ, TX_4X4);
       rate += ratey;
-      distortion = vp8_block_error(be->coeff, b->dqcoeff, 16) >> 2;
+      distortion = vp9_block_error(be->coeff, b->dqcoeff, 16) >> 2;
 
       this_rd = RDCOST(x->rdmult, x->rddiv, rate, distortion);
 
@@ -1436,18 +1436,18 @@ static int64_t rd_pick_intra8x8block(VP8_COMP *cpi, MACROBLOCK *x, int ib,
       }
 #endif
 
-      vp8_subtract_4b_c(be, b, 16);
+      vp9_subtract_4b_c(be, b, 16);
 
       if (xd->mode_info_context->mbmi.txfm_size == TX_8X8) {
         TX_TYPE tx_type = get_tx_type_8x8(xd, b);
         if (tx_type != DCT_DCT)
-          vp8_fht_c(be->src_diff, 32, (x->block + idx)->coeff, tx_type, 8);
+          vp9_fht_c(be->src_diff, 32, (x->block + idx)->coeff, tx_type, 8);
         else
-          x->vp8_short_fdct8x8(be->src_diff, (x->block + idx)->coeff, 32);
+          x->vp9_short_fdct8x8(be->src_diff, (x->block + idx)->coeff, 32);
         x->quantize_b_8x8(x->block + idx, xd->block + idx);
 
         // compute quantization mse of 8x8 block
-        distortion = vp8_block_error_c((x->block + idx)->coeff,
+        distortion = vp9_block_error_c((x->block + idx)->coeff,
                                        (xd->block + idx)->dqcoeff, 64);
         ta0 = a[vp8_block2above_8x8[idx]];
         tl0 = l[vp8_block2left_8x8[idx]];
@@ -1459,21 +1459,21 @@ static int64_t rd_pick_intra8x8block(VP8_COMP *cpi, MACROBLOCK *x, int ib,
         ta1 = ta0;
         tl1 = tl0;
       } else {
-        x->vp8_short_fdct8x4(be->src_diff, be->coeff, 32);
-        x->vp8_short_fdct8x4((be + 4)->src_diff, (be + 4)->coeff, 32);
+        x->vp9_short_fdct8x4(be->src_diff, be->coeff, 32);
+        x->vp9_short_fdct8x4((be + 4)->src_diff, (be + 4)->coeff, 32);
 
         x->quantize_b_4x4_pair(x->block + ib, x->block + ib + 1,
                                xd->block + ib, xd->block + ib + 1);
         x->quantize_b_4x4_pair(x->block + ib + 4, x->block + ib + 5,
                                xd->block + ib + 4, xd->block + ib + 5);
 
-        distortion = vp8_block_error_c((x->block + ib)->coeff,
+        distortion = vp9_block_error_c((x->block + ib)->coeff,
                                        (xd->block + ib)->dqcoeff, 16);
-        distortion += vp8_block_error_c((x->block + ib + 1)->coeff,
+        distortion += vp9_block_error_c((x->block + ib + 1)->coeff,
                                         (xd->block + ib + 1)->dqcoeff, 16);
-        distortion += vp8_block_error_c((x->block + ib + 4)->coeff,
+        distortion += vp9_block_error_c((x->block + ib + 4)->coeff,
                                         (xd->block + ib + 4)->dqcoeff, 16);
-        distortion += vp8_block_error_c((x->block + ib + 5)->coeff,
+        distortion += vp9_block_error_c((x->block + ib + 5)->coeff,
                                         (xd->block + ib + 5)->dqcoeff, 16);
 
         ta0 = a[vp8_block2above[ib]];
@@ -1518,7 +1518,7 @@ static int64_t rd_pick_intra8x8block(VP8_COMP *cpi, MACROBLOCK *x, int ib,
 #if CONFIG_COMP_INTRA_PRED
   b->bmi.as_mode.second = (*best_second_mode);
 #endif
-  vp8_encode_intra8x8(IF_RTCD(&cpi->rtcd), x, ib);
+  vp9_encode_intra8x8(IF_RTCD(&cpi->rtcd), x, ib);
 
   if (xd->mode_info_context->mbmi.txfm_size == TX_8X8) {
     a[vp8_block2above_8x8[idx]]     = besta0;
@@ -1610,14 +1610,14 @@ static int rd_cost_mbuv(MACROBLOCK *mb) {
 
 static int64_t rd_inter16x16_uv(VP8_COMP *cpi, MACROBLOCK *x, int *rate,
                                 int *distortion, int fullpixel, int *skip) {
-  vp8_subtract_mbuv(x->src_diff, x->src.u_buffer, x->src.v_buffer,
+  vp9_subtract_mbuv(x->src_diff, x->src.u_buffer, x->src.v_buffer,
                     x->e_mbd.predictor, x->src.uv_stride);
 
-  vp8_transform_mbuv_4x4(x);
-  vp8_quantize_mbuv_4x4(x);
+  vp9_transform_mbuv_4x4(x);
+  vp9_quantize_mbuv_4x4(x);
 
   *rate       = rd_cost_mbuv(x);
-  *distortion = vp8_mbuverror(x) / 4;
+  *distortion = vp9_mbuverror(x) / 4;
   *skip       = vp9_mbuv_is_skippable_4x4(&x->e_mbd);
 
   return RDCOST(x->rdmult, x->rddiv, *rate, *distortion);
@@ -1668,7 +1668,7 @@ static int64_t rd_inter32x32_uv_8x8(VP8_COMP *cpi, MACROBLOCK *x, int *rate,
   for (n = 0; n < 4; n++) {
     int x_idx = n & 1, y_idx = n >> 1;
 
-    vp8_subtract_mbuv_s_c(x->src_diff,
+    vp9_subtract_mbuv_s_c(x->src_diff,
                           usrc + x_idx * 8 + y_idx * 8 * src_uv_stride,
                           vsrc + x_idx * 8 + y_idx * 8 * src_uv_stride,
                           src_uv_stride,
@@ -1676,13 +1676,13 @@ static int64_t rd_inter32x32_uv_8x8(VP8_COMP *cpi, MACROBLOCK *x, int *rate,
                           vdst + x_idx * 8 + y_idx * 8 * dst_uv_stride,
                           dst_uv_stride);
 
-    vp8_transform_mbuv_8x8(x);
-    vp8_quantize_mbuv_8x8(x);
+    vp9_transform_mbuv_8x8(x);
+    vp9_quantize_mbuv_8x8(x);
 
     xd->above_context = ta + x_idx;
     xd->left_context = tl + y_idx;
     r += rd_cost_mbuv_8x8(x, 0);
-    d += vp8_mbuverror(x) / 4;
+    d += vp9_mbuverror(x) / 4;
     skippable = skippable && vp9_mbuv_is_skippable_8x8(xd);
   }
 
@@ -1700,14 +1700,14 @@ static int64_t rd_inter32x32_uv_8x8(VP8_COMP *cpi, MACROBLOCK *x, int *rate,
 
 static int64_t rd_inter16x16_uv_8x8(VP8_COMP *cpi, MACROBLOCK *x, int *rate,
                                     int *distortion, int fullpixel, int *skip) {
-  vp8_subtract_mbuv(x->src_diff, x->src.u_buffer, x->src.v_buffer,
+  vp9_subtract_mbuv(x->src_diff, x->src.u_buffer, x->src.v_buffer,
                     x->e_mbd.predictor, x->src.uv_stride);
 
-  vp8_transform_mbuv_8x8(x);
-  vp8_quantize_mbuv_8x8(x);
+  vp9_transform_mbuv_8x8(x);
+  vp9_quantize_mbuv_8x8(x);
 
   *rate       = rd_cost_mbuv_8x8(x, 1);
-  *distortion = vp8_mbuverror(x) / 4;
+  *distortion = vp9_mbuverror(x) / 4;
   *skip       = vp9_mbuv_is_skippable_8x8(&x->e_mbd);
 
   return RDCOST(x->rdmult, x->rddiv, *rate, *distortion);
@@ -1717,14 +1717,14 @@ static int64_t rd_inter16x16_uv_8x8(VP8_COMP *cpi, MACROBLOCK *x, int *rate,
 static int64_t rd_inter4x4_uv(VP8_COMP *cpi, MACROBLOCK *x, int *rate,
                               int *distortion, int *skippable, int fullpixel) {
   vp8_build_inter4x4_predictors_mbuv(&x->e_mbd);
-  vp8_subtract_mbuv(x->src_diff, x->src.u_buffer, x->src.v_buffer,
+  vp9_subtract_mbuv(x->src_diff, x->src.u_buffer, x->src.v_buffer,
                     x->e_mbd.predictor, x->src.uv_stride);
 
-  vp8_transform_mbuv_4x4(x);
-  vp8_quantize_mbuv_4x4(x);
+  vp9_transform_mbuv_4x4(x);
+  vp9_quantize_mbuv_4x4(x);
 
   *rate       = rd_cost_mbuv(x);
-  *distortion = vp8_mbuverror(x) / 4;
+  *distortion = vp9_mbuverror(x) / 4;
   *skippable  = vp9_mbuv_is_skippable_4x4(&x->e_mbd);
 
   return RDCOST(x->rdmult, x->rddiv, *rate, *distortion);
@@ -1769,16 +1769,16 @@ static void rd_pick_intra_mbuv_mode(VP8_COMP *cpi,
       }
 #endif
 
-      vp8_subtract_mbuv(x->src_diff, x->src.u_buffer, x->src.v_buffer,
+      vp9_subtract_mbuv(x->src_diff, x->src.u_buffer, x->src.v_buffer,
                         x->e_mbd.predictor, x->src.uv_stride);
-      vp8_transform_mbuv_4x4(x);
-      vp8_quantize_mbuv_4x4(x);
+      vp9_transform_mbuv_4x4(x);
+      vp9_quantize_mbuv_4x4(x);
 
       rate_to = rd_cost_mbuv(x);
       rate = rate_to
              + x->intra_uv_mode_cost[x->e_mbd.frame_type][mbmi->uv_mode];
 
-      distortion = vp8_mbuverror(x) / 4;
+      distortion = vp9_mbuverror(x) / 4;
 
       this_rd = RDCOST(x->rdmult, x->rddiv, rate, distortion);
 
@@ -1827,16 +1827,16 @@ static void rd_pick_intra_mbuv_mode_8x8(VP8_COMP *cpi,
 
     mbmi->uv_mode = mode;
     vp8_build_intra_predictors_mbuv(&x->e_mbd);
-    vp8_subtract_mbuv(x->src_diff, x->src.u_buffer, x->src.v_buffer,
+    vp9_subtract_mbuv(x->src_diff, x->src.u_buffer, x->src.v_buffer,
                       x->e_mbd.predictor, x->src.uv_stride);
-    vp8_transform_mbuv_8x8(x);
+    vp9_transform_mbuv_8x8(x);
 
-    vp8_quantize_mbuv_8x8(x);
+    vp9_quantize_mbuv_8x8(x);
 
     rate_to = rd_cost_mbuv_8x8(x, 1);
     rate = rate_to + x->intra_uv_mode_cost[x->e_mbd.frame_type][mbmi->uv_mode];
 
-    distortion = vp8_mbuverror(x) / 4;
+    distortion = vp9_mbuverror(x) / 4;
     this_rd = RDCOST(x->rdmult, x->rddiv, rate, distortion);
 
     if (this_rd < best_rd) {
@@ -1875,18 +1875,18 @@ static void super_block_uvrd_8x8(MACROBLOCK *x,
   for (n = 0; n < 4; n++) {
     int x_idx = n & 1, y_idx = n >> 1;
 
-    vp8_subtract_mbuv_s_c(x->src_diff,
+    vp9_subtract_mbuv_s_c(x->src_diff,
                           usrc + x_idx * 8 + y_idx * 8 * src_uv_stride,
                           vsrc + x_idx * 8 + y_idx * 8 * src_uv_stride,
                           src_uv_stride,
                           udst + x_idx * 8 + y_idx * 8 * dst_uv_stride,
                           vdst + x_idx * 8 + y_idx * 8 * dst_uv_stride,
                           dst_uv_stride);
-    vp8_transform_mbuv_8x8(x);
-    vp8_quantize_mbuv_8x8(x);
+    vp9_transform_mbuv_8x8(x);
+    vp9_quantize_mbuv_8x8(x);
     s &= vp9_mbuv_is_skippable_8x8(xd);
 
-    d += vp8_mbuverror(x) >> 2;
+    d += vp9_mbuverror(x) >> 2;
     xd->above_context = ta + x_idx;
     xd->left_context = tl + y_idx;
     r += rd_cost_mbuv_8x8(x, 0);
@@ -1943,7 +1943,7 @@ static int64_t rd_pick_intra_sbuv_mode(VP8_COMP *cpi,
 }
 #endif
 
-int vp8_cost_mv_ref(VP8_COMP *cpi,
+int vp9_cost_mv_ref(VP8_COMP *cpi,
                     MB_PREDICTION_MODE m,
                     const int near_mv_ref_ct[4]) {
   MACROBLOCKD *xd = &cpi->mb.e_mbd;
@@ -1966,7 +1966,7 @@ int vp8_cost_mv_ref(VP8_COMP *cpi,
     return 0;
 }
 
-void vp8_set_mbmode_and_mvs(MACROBLOCK *x, MB_PREDICTION_MODE mb, int_mv *mv) {
+void vp9_set_mbmode_and_mvs(MACROBLOCK *x, MB_PREDICTION_MODE mb, int_mv *mv) {
   x->e_mbd.mode_info_context->mbmi.mode = mb;
   x->e_mbd.mode_info_context->mbmi.mv[0].as_int = mv->as_int;
 }
@@ -2014,10 +2014,10 @@ static int labels2mode(
               seg_mvs[mbmi->second_ref_frame - 1].as_int;
           }
 
-          thismvcost  = vp8_mv_bit_cost(this_mv, best_ref_mv, MVCOSTS,
+          thismvcost  = vp9_mv_bit_cost(this_mv, best_ref_mv, MVCOSTS,
                                         102, xd->allow_high_precision_mv);
           if (mbmi->second_ref_frame) {
-            thismvcost += vp8_mv_bit_cost(this_second_mv, second_best_ref_mv,
+            thismvcost += vp9_mv_bit_cost(this_second_mv, second_best_ref_mv,
                                           MVCOSTS, 102,
                                           xd->allow_high_precision_mv);
           }
@@ -2096,10 +2096,10 @@ static int64_t encode_inter_mb_segment(MACROBLOCK *x,
       vp8_build_inter_predictors_b(bd, 16, xd->subpixel_predict);
       if (xd->mode_info_context->mbmi.second_ref_frame)
         vp8_build_2nd_inter_predictors_b(bd, 16, xd->subpixel_predict_avg);
-      vp8_subtract_b(be, bd, 16);
-      x->vp8_short_fdct4x4(be->src_diff, be->coeff, 32);
+      vp9_subtract_b(be, bd, 16);
+      x->vp9_short_fdct4x4(be->src_diff, be->coeff, 32);
       x->quantize_b_4x4(be, bd);
-      thisdistortion = vp8_block_error(be->coeff, bd->dqcoeff, 16);
+      thisdistortion = vp9_block_error(be->coeff, bd->dqcoeff, 16);
       *distortion += thisdistortion;
       *labelyrate += cost_coeffs(x, bd, PLANE_TYPE_Y_WITH_DC,
                                  ta + vp8_block2above[i],
@@ -2146,13 +2146,13 @@ static int64_t encode_inter_mb_segment_8x8(MACROBLOCK *x,
       vp8_build_inter_predictors4b(xd, bd, 16);
       if (xd->mode_info_context->mbmi.second_ref_frame)
         vp8_build_2nd_inter_predictors4b(xd, bd, 16);
-      vp8_subtract_4b_c(be, bd, 16);
+      vp9_subtract_4b_c(be, bd, 16);
 
       if (xd->mode_info_context->mbmi.txfm_size == TX_4X4) {
         if (otherrd) {
-          x->vp8_short_fdct8x8(be->src_diff, be2->coeff, 32);
+          x->vp9_short_fdct8x8(be->src_diff, be2->coeff, 32);
           x->quantize_b_8x8(be2, bd2);
-          thisdistortion = vp8_block_error_c(be2->coeff, bd2->dqcoeff, 64);
+          thisdistortion = vp9_block_error_c(be2->coeff, bd2->dqcoeff, 64);
           otherdist += thisdistortion;
           othercost += cost_coeffs(x, bd2, PLANE_TYPE_Y_WITH_DC,
                                      tacp + vp8_block2above_8x8[idx],
@@ -2161,9 +2161,9 @@ static int64_t encode_inter_mb_segment_8x8(MACROBLOCK *x,
         for (j = 0; j < 4; j += 2) {
           bd = &xd->block[ib + iblock[j]];
           be = &x->block[ib + iblock[j]];
-          x->vp8_short_fdct8x4(be->src_diff, be->coeff, 32);
+          x->vp9_short_fdct8x4(be->src_diff, be->coeff, 32);
           x->quantize_b_4x4_pair(be, be + 1, bd, bd + 1);
-          thisdistortion = vp8_block_error_c(be->coeff, bd->dqcoeff, 32);
+          thisdistortion = vp9_block_error_c(be->coeff, bd->dqcoeff, 32);
           *distortion += thisdistortion;
           *labelyrate += cost_coeffs(x, bd, PLANE_TYPE_Y_WITH_DC,
                                      ta + vp8_block2above[ib + iblock[j]],
@@ -2179,9 +2179,9 @@ static int64_t encode_inter_mb_segment_8x8(MACROBLOCK *x,
           for (j = 0; j < 4; j += 2) {
             BLOCKD *bd3 = &xd->block[ib + iblock[j]];
             BLOCK *be3 = &x->block[ib + iblock[j]];
-            x->vp8_short_fdct8x4(be3->src_diff, be3->coeff, 32);
+            x->vp9_short_fdct8x4(be3->src_diff, be3->coeff, 32);
             x->quantize_b_4x4_pair(be3, be3 + 1, bd3, bd3 + 1);
-            thisdistortion = vp8_block_error_c(be3->coeff, bd3->dqcoeff, 32);
+            thisdistortion = vp9_block_error_c(be3->coeff, bd3->dqcoeff, 32);
             otherdist += thisdistortion;
             othercost += cost_coeffs(x, bd3, PLANE_TYPE_Y_WITH_DC,
                                      tacp + vp8_block2above[ib + iblock[j]],
@@ -2193,9 +2193,9 @@ static int64_t encode_inter_mb_segment_8x8(MACROBLOCK *x,
                                      TX_4X4);
           }
         }
-        x->vp8_short_fdct8x8(be->src_diff, be2->coeff, 32);
+        x->vp9_short_fdct8x8(be->src_diff, be2->coeff, 32);
         x->quantize_b_8x8(be2, bd2);
-        thisdistortion = vp8_block_error_c(be2->coeff, bd2->dqcoeff, 64);
+        thisdistortion = vp9_block_error_c(be2->coeff, bd2->dqcoeff, 64);
         *distortion += thisdistortion;
         *labelyrate += cost_coeffs(x, bd2, PLANE_TYPE_Y_WITH_DC,
                                    ta + vp8_block2above_8x8[idx],
@@ -2296,7 +2296,7 @@ static void rd_check_segment_txsize(VP8_COMP *cpi, MACROBLOCK *x,
   // Segmentation method overheads
   rate = vp8_cost_token(vp8_mbsplit_tree, vp8_mbsplit_probs,
                         vp8_mbsplit_encodings + segmentation);
-  rate += vp8_cost_mv_ref(cpi, SPLITMV, bsi->mdcounts);
+  rate += vp9_cost_mv_ref(cpi, SPLITMV, bsi->mdcounts);
   this_segment_rd += RDCOST(x->rdmult, x->rddiv, rate, 0);
   br += rate;
   other_segment_rd = this_segment_rd;
@@ -2371,7 +2371,7 @@ static void rd_check_segment_txsize(VP8_COMP *cpi, MACROBLOCK *x,
           c = &x->block[n];
           e = &x->e_mbd.block[n];
 
-          bestsme = vp8_full_pixel_diamond(cpi, x, c, e, &mvp_full, step_param,
+          bestsme = vp9_full_pixel_diamond(cpi, x, c, e, &mvp_full, step_param,
                                            sadpb, further_steps, 0, v_fn_ptr,
                                            bsi->ref_mv, &mode_mv[NEW4X4]);
 
@@ -2679,7 +2679,7 @@ static int rd_pick_best_mbsegmentation(VP8_COMP *cpi, MACROBLOCK *x,
       int tmp_row_min = x->mv_row_min;
       int tmp_row_max = x->mv_row_max;
 
-      vp8_clamp_mv_min_max(x, best_ref_mv);
+      vp9_clamp_mv_min_max(x, best_ref_mv);
 
       /* Get 8x8 result */
       bsi.sv_mvp[0].as_int = bsi.mvs[0].as_int;
@@ -2817,7 +2817,7 @@ static void insertsortsad(int arr[], int idx[], int len) {
 }
 
 // The improved MV prediction
-void vp8_mv_pred(VP8_COMP *cpi, MACROBLOCKD *xd, const MODE_INFO *here,
+void vp9_mv_pred(VP8_COMP *cpi, MACROBLOCKD *xd, const MODE_INFO *here,
                  int_mv *mvp, int refframe, int *ref_frame_sign_bias,
                  int *sr, int near_sadidx[]) {
   const MODE_INFO *above = here - xd->mode_info_stride;
@@ -3313,11 +3313,11 @@ static int64_t handle_inter_mode(VP8_COMP *cpi, MACROBLOCK *x,
         if (frame_mv[NEWMV][refs[0]].as_int == INVALID_MV ||
             frame_mv[NEWMV][refs[1]].as_int == INVALID_MV)
           return INT64_MAX;
-        *rate2 += vp8_mv_bit_cost(&frame_mv[NEWMV][refs[0]],
+        *rate2 += vp9_mv_bit_cost(&frame_mv[NEWMV][refs[0]],
                                   &frame_best_ref_mv[refs[0]],
                                   XMVCOST, 96,
                                   x->e_mbd.allow_high_precision_mv);
-        *rate2 += vp8_mv_bit_cost(&frame_mv[NEWMV][refs[1]],
+        *rate2 += vp9_mv_bit_cost(&frame_mv[NEWMV][refs[1]],
                                   &frame_best_ref_mv[refs[1]],
                                   XMVCOST, 96,
                                   x->e_mbd.allow_high_precision_mv);
@@ -3334,14 +3334,14 @@ static int64_t handle_inter_mode(VP8_COMP *cpi, MACROBLOCK *x,
         int tmp_row_min = x->mv_row_min;
         int tmp_row_max = x->mv_row_max;
 
-        vp8_clamp_mv_min_max(x, &frame_best_ref_mv[refs[0]]);
+        vp9_clamp_mv_min_max(x, &frame_best_ref_mv[refs[0]]);
 
         if (!*saddone) {
           cal_sad(cpi, xd, x, recon_yoffset, &near_sadidx[0], block_size);
           *saddone = 1;
         }
 
-        vp8_mv_pred(cpi, &x->e_mbd, x->e_mbd.mode_info_context, &mvp,
+        vp9_mv_pred(cpi, &x->e_mbd, x->e_mbd.mode_info_context, &mvp,
                     mbmi->ref_frame, cpi->common.ref_frame_sign_bias,
                     &sr, &near_sadidx[0]);
 
@@ -3354,7 +3354,7 @@ static int64_t handle_inter_mode(VP8_COMP *cpi, MACROBLOCK *x,
         // Further step/diamond searches as necessary
         further_steps = (cpi->sf.max_step_search_steps - 1) - step_param;
 
-        bestsme = vp8_full_pixel_diamond(cpi, x, b, d, &mvp_full, step_param,
+        bestsme = vp9_full_pixel_diamond(cpi, x, b, d, &mvp_full, step_param,
                                          sadpb, further_steps, 1,
                                          &cpi->fn_ptr[block_size],
                                          &frame_best_ref_mv[refs[0]], &tmp_mv);
@@ -3377,7 +3377,7 @@ static int64_t handle_inter_mode(VP8_COMP *cpi, MACROBLOCK *x,
         frame_mv[NEWMV][refs[0]].as_int = d->bmi.as_mv.first.as_int;
 
         // Add the new motion vector cost to our rolling cost variable
-        *rate2 += vp8_mv_bit_cost(&tmp_mv, &frame_best_ref_mv[refs[0]],
+        *rate2 += vp9_mv_bit_cost(&tmp_mv, &frame_best_ref_mv[refs[0]],
                                   XMVCOST, 96, xd->allow_high_precision_mv);
       }
       break;
@@ -3419,7 +3419,7 @@ static int64_t handle_inter_mode(VP8_COMP *cpi, MACROBLOCK *x,
    * if the first is known */
   *compmode_cost = vp8_cost_bit(vp9_get_pred_prob(cm, xd, PRED_COMP),
                                 is_comp_pred);
-  *rate2 += vp8_cost_mv_ref(cpi, this_mode, mdcounts);
+  *rate2 += vp9_cost_mv_ref(cpi, this_mode, mdcounts);
 
   if (block_size == BLOCK_16X16) {
     vp8_build_1st_inter16x16_predictors_mby(xd, xd->predictor, 16, 0);
@@ -3447,11 +3447,11 @@ static int64_t handle_inter_mode(VP8_COMP *cpi, MACROBLOCK *x,
       threshold = x->encode_breakout;
 
     if (block_size == BLOCK_16X16) {
-      var = vp8_variance16x16(*(b->base_src), b->src_stride,
+      var = vp9_variance16x16(*(b->base_src), b->src_stride,
                               xd->predictor, 16, &sse);
     } else {
 #if CONFIG_SUPERBLOCKS
-      var = vp8_variance32x32(*(b->base_src), b->src_stride,
+      var = vp9_variance32x32(*(b->base_src), b->src_stride,
                               xd->dst.y_buffer, xd->dst.y_stride, &sse);
 #endif
     }
@@ -3466,12 +3466,12 @@ static int64_t handle_inter_mode(VP8_COMP *cpi, MACROBLOCK *x,
         int sse2;
 
         if (block_size == BLOCK_16X16) {
-          sse2 = vp8_uvsse(x);
+          sse2 = vp9_uvsse(x);
         } else {
           unsigned int sse2u, sse2v;
-          var = vp8_variance16x16(x->src.u_buffer, x->src.uv_stride,
+          var = vp9_variance16x16(x->src.u_buffer, x->src.uv_stride,
                                   xd->dst.u_buffer, xd->dst.uv_stride, &sse2u);
-          var = vp8_variance16x16(x->src.v_buffer, x->src.uv_stride,
+          var = vp9_variance16x16(x->src.v_buffer, x->src.uv_stride,
                                   xd->dst.v_buffer, xd->dst.uv_stride, &sse2v);
           sse2 = sse2u + sse2v;
         }
@@ -3530,7 +3530,7 @@ static int64_t handle_inter_mode(VP8_COMP *cpi, MACROBLOCK *x,
   return this_rd;  // if 0, this will be re-calculated by caller
 }
 
-void vp8_rd_pick_inter_mode(VP8_COMP *cpi, MACROBLOCK *x, int recon_yoffset, int recon_uvoffset,
+void vp9_rd_pick_inter_mode(VP8_COMP *cpi, MACROBLOCK *x, int recon_yoffset, int recon_uvoffset,
                             int *returnrate, int *returndistortion, int64_t *returnintra) {
   VP8_COMMON *cm = &cpi->common;
   MACROBLOCKD *xd = &x->e_mbd;
@@ -3632,7 +3632,7 @@ void vp8_rd_pick_inter_mode(VP8_COMP *cpi, MACROBLOCK *x, int recon_yoffset, int
 
   /* Initialize zbin mode boost for uv costing */
   cpi->zbin_mode_boost = 0;
-  vp8_update_zbin_extra(cpi, x);
+  vp9_update_zbin_extra(cpi, x);
 
   rd_pick_intra_mbuv_mode(cpi, x, &uv_intra_rate,
                           &uv_intra_rate_tokenonly, &uv_intra_distortion,
@@ -3760,7 +3760,7 @@ void vp8_rd_pick_inter_mode(VP8_COMP *cpi, MACROBLOCK *x, int recon_yoffset, int
           cpi->zbin_mode_boost = MV_ZBIN_BOOST;
       }
 
-      vp8_update_zbin_extra(cpi, x);
+      vp9_update_zbin_extra(cpi, x);
     }
 
     // Intra
@@ -4277,7 +4277,7 @@ end:
 }
 
 #if CONFIG_SUPERBLOCKS
-void vp8_rd_pick_intra_mode_sb(VP8_COMP *cpi, MACROBLOCK *x,
+void vp9_rd_pick_intra_mode_sb(VP8_COMP *cpi, MACROBLOCK *x,
                                int *returnrate,
                                int *returndist) {
   VP8_COMMON *cm = &cpi->common;
@@ -4308,7 +4308,7 @@ void vp8_rd_pick_intra_mode_sb(VP8_COMP *cpi, MACROBLOCK *x,
 }
 #endif
 
-void vp8_rd_pick_intra_mode(VP8_COMP *cpi, MACROBLOCK *x,
+void vp9_rd_pick_intra_mode(VP8_COMP *cpi, MACROBLOCK *x,
                             int *returnrate, int *returndist) {
   VP8_COMMON *cm = &cpi->common;
   MACROBLOCKD *xd = &x->e_mbd;
@@ -4465,7 +4465,7 @@ void vp8_rd_pick_intra_mode(VP8_COMP *cpi, MACROBLOCK *x,
 }
 
 #if CONFIG_SUPERBLOCKS
-int64_t vp8_rd_pick_inter_mode_sb(VP8_COMP *cpi, MACROBLOCK *x,
+int64_t vp9_rd_pick_inter_mode_sb(VP8_COMP *cpi, MACROBLOCK *x,
                                   int recon_yoffset, int recon_uvoffset,
                                   int *returnrate, int *returndistortion) {
   VP8_COMMON *cm = &cpi->common;
@@ -4815,7 +4815,7 @@ int64_t vp8_rd_pick_inter_mode_sb(VP8_COMP *cpi, MACROBLOCK *x,
 }
 #endif
 
-void vp8cx_pick_mode_inter_macroblock(VP8_COMP *cpi, MACROBLOCK *x,
+void vp9cx_pick_mode_inter_macroblock(VP8_COMP *cpi, MACROBLOCK *x,
                                       int recon_yoffset,
                                       int recon_uvoffset,
                                       int *totalrate, int *totaldist) {
@@ -4835,7 +4835,7 @@ void vp8cx_pick_mode_inter_macroblock(VP8_COMP *cpi, MACROBLOCK *x,
   {
     int zbin_mode_boost_enabled = cpi->zbin_mode_boost_enabled;
 
-    vp8_rd_pick_inter_mode(cpi, x, recon_yoffset, recon_uvoffset, &rate,
+    vp9_rd_pick_inter_mode(cpi, x, recon_yoffset, recon_uvoffset, &rate,
                            &distortion, &intra_error);
 
     /* restore cpi->zbin_mode_boost_enabled */
diff --git a/vp8/encoder/rdopt.h b/vp8/encoder/rdopt.h
index 0495037eb646775fe5fa0006a71b4fab8a52db37..67fd6612bbd809b1f17d76274101d54d0c1b3159 100644
--- a/vp8/encoder/rdopt.h
+++ b/vp8/encoder/rdopt.h
@@ -15,13 +15,13 @@
 #define RDCOST(RM,DM,R,D) ( ((128+((int64_t)R)*(RM)) >> 8) + ((int64_t)DM)*(D) )
 #define RDCOST_8x8(RM,DM,R,D) ( ((128+((int64_t)R)*(RM)) >> 8) + ((int64_t)DM)*(D) )
 
-extern void vp8_initialize_rd_consts(VP8_COMP *cpi, int Qvalue);
-extern void vp8_rd_pick_inter_mode(VP8_COMP *cpi, MACROBLOCK *x, int recon_yoffset, int recon_uvoffset,
+extern void vp9_initialize_rd_consts(VP8_COMP *cpi, int Qvalue);
+extern void vp9_rd_pick_inter_mode(VP8_COMP *cpi, MACROBLOCK *x, int recon_yoffset, int recon_uvoffset,
                                    int *returnrate, int *returndistortion, int64_t *returnintra);
-extern void vp8_rd_pick_intra_mode(VP8_COMP *cpi, MACROBLOCK *x, int *r, int *d);
-extern void vp8_rd_pick_intra_mode_sb(VP8_COMP *cpi, MACROBLOCK *x, int *r, int *d);
+extern void vp9_rd_pick_intra_mode(VP8_COMP *cpi, MACROBLOCK *x, int *r, int *d);
+extern void vp9_rd_pick_intra_mode_sb(VP8_COMP *cpi, MACROBLOCK *x, int *r, int *d);
 
-extern void vp8_mv_pred
+extern void vp9_mv_pred
 (
   VP8_COMP *cpi,
   MACROBLOCKD *xd,
@@ -32,6 +32,6 @@ extern void vp8_mv_pred
   int *sr,
   int near_sadidx[]
 );
-extern void vp8_init_me_luts();
-extern void vp8_set_mbmode_and_mvs(MACROBLOCK *x, MB_PREDICTION_MODE mb, int_mv *mv);
+extern void vp9_init_me_luts();
+extern void vp9_set_mbmode_and_mvs(MACROBLOCK *x, MB_PREDICTION_MODE mb, int_mv *mv);
 #endif
diff --git a/vp8/encoder/sad_c.c b/vp8/encoder/sad_c.c
index f15e687c62f40f77d35297da7d17fae4cf3d1a9b..fdada23879b58f988cd66a86a39bc4073ce4118a 100644
--- a/vp8/encoder/sad_c.c
+++ b/vp8/encoder/sad_c.c
@@ -14,7 +14,7 @@
 #include "vpx_ports/config.h"
 #include "vpx/vpx_integer.h"
 
-unsigned int vp8_sad32x32_c(const unsigned char *src_ptr,
+unsigned int vp9_sad32x32_c(const unsigned char *src_ptr,
                             int  src_stride,
                             const unsigned char *ref_ptr,
                             int  ref_stride,
@@ -22,7 +22,7 @@ unsigned int vp8_sad32x32_c(const unsigned char *src_ptr,
   return sad_mx_n_c(src_ptr, src_stride, ref_ptr, ref_stride, 32, 32);
 }
 
-unsigned int vp8_sad16x16_c(const unsigned char *src_ptr,
+unsigned int vp9_sad16x16_c(const unsigned char *src_ptr,
                             int  src_stride,
                             const unsigned char *ref_ptr,
                             int  ref_stride,
@@ -30,7 +30,7 @@ unsigned int vp8_sad16x16_c(const unsigned char *src_ptr,
   return sad_mx_n_c(src_ptr, src_stride, ref_ptr, ref_stride, 16, 16);
 }
 
-unsigned int vp8_sad8x8_c(
+unsigned int vp9_sad8x8_c(
   const unsigned char *src_ptr,
   int  src_stride,
   const unsigned char *ref_ptr,
@@ -41,7 +41,7 @@ unsigned int vp8_sad8x8_c(
 }
 
 
-unsigned int vp8_sad16x8_c(
+unsigned int vp9_sad16x8_c(
   const unsigned char *src_ptr,
   int  src_stride,
   const unsigned char *ref_ptr,
@@ -53,7 +53,7 @@ unsigned int vp8_sad16x8_c(
 }
 
 
-unsigned int vp8_sad8x16_c(
+unsigned int vp9_sad8x16_c(
   const unsigned char *src_ptr,
   int  src_stride,
   const unsigned char *ref_ptr,
@@ -64,7 +64,7 @@ unsigned int vp8_sad8x16_c(
 }
 
 
-unsigned int vp8_sad4x4_c(
+unsigned int vp9_sad4x4_c(
   const unsigned char *src_ptr,
   int  src_stride,
   const unsigned char *ref_ptr,
@@ -74,257 +74,257 @@ unsigned int vp8_sad4x4_c(
   return sad_mx_n_c(src_ptr, src_stride, ref_ptr, ref_stride, 4, 4);
 }
 
-void vp8_sad32x32x3_c(const unsigned char *src_ptr,
+void vp9_sad32x32x3_c(const unsigned char *src_ptr,
                       int  src_stride,
                       const unsigned char *ref_ptr,
                       int  ref_stride,
                       unsigned int *sad_array
                       ) {
-  sad_array[0] = vp8_sad32x32_c(src_ptr, src_stride, ref_ptr, ref_stride, 0x7fffffff);
-  sad_array[1] = vp8_sad32x32_c(src_ptr, src_stride, ref_ptr + 1, ref_stride, 0x7fffffff);
-  sad_array[2] = vp8_sad32x32_c(src_ptr, src_stride, ref_ptr + 2, ref_stride, 0x7fffffff);
+  sad_array[0] = vp9_sad32x32_c(src_ptr, src_stride, ref_ptr, ref_stride, 0x7fffffff);
+  sad_array[1] = vp9_sad32x32_c(src_ptr, src_stride, ref_ptr + 1, ref_stride, 0x7fffffff);
+  sad_array[2] = vp9_sad32x32_c(src_ptr, src_stride, ref_ptr + 2, ref_stride, 0x7fffffff);
 }
 
-void vp8_sad32x32x8_c(const unsigned char *src_ptr,
+void vp9_sad32x32x8_c(const unsigned char *src_ptr,
                       int  src_stride,
                       const unsigned char *ref_ptr,
                       int  ref_stride,
                       unsigned short *sad_array
                       ) {
-  sad_array[0] = (unsigned short)vp8_sad32x32_c(src_ptr, src_stride, ref_ptr, ref_stride, 0x7fffffff);
-  sad_array[1] = (unsigned short)vp8_sad32x32_c(src_ptr, src_stride, ref_ptr + 1, ref_stride, 0x7fffffff);
-  sad_array[2] = (unsigned short)vp8_sad32x32_c(src_ptr, src_stride, ref_ptr + 2, ref_stride, 0x7fffffff);
-  sad_array[3] = (unsigned short)vp8_sad32x32_c(src_ptr, src_stride, ref_ptr + 3, ref_stride, 0x7fffffff);
-  sad_array[4] = (unsigned short)vp8_sad32x32_c(src_ptr, src_stride, ref_ptr + 4, ref_stride, 0x7fffffff);
-  sad_array[5] = (unsigned short)vp8_sad32x32_c(src_ptr, src_stride, ref_ptr + 5, ref_stride, 0x7fffffff);
-  sad_array[6] = (unsigned short)vp8_sad32x32_c(src_ptr, src_stride, ref_ptr + 6, ref_stride, 0x7fffffff);
-  sad_array[7] = (unsigned short)vp8_sad32x32_c(src_ptr, src_stride, ref_ptr + 7, ref_stride, 0x7fffffff);
+  sad_array[0] = (unsigned short)vp9_sad32x32_c(src_ptr, src_stride, ref_ptr, ref_stride, 0x7fffffff);
+  sad_array[1] = (unsigned short)vp9_sad32x32_c(src_ptr, src_stride, ref_ptr + 1, ref_stride, 0x7fffffff);
+  sad_array[2] = (unsigned short)vp9_sad32x32_c(src_ptr, src_stride, ref_ptr + 2, ref_stride, 0x7fffffff);
+  sad_array[3] = (unsigned short)vp9_sad32x32_c(src_ptr, src_stride, ref_ptr + 3, ref_stride, 0x7fffffff);
+  sad_array[4] = (unsigned short)vp9_sad32x32_c(src_ptr, src_stride, ref_ptr + 4, ref_stride, 0x7fffffff);
+  sad_array[5] = (unsigned short)vp9_sad32x32_c(src_ptr, src_stride, ref_ptr + 5, ref_stride, 0x7fffffff);
+  sad_array[6] = (unsigned short)vp9_sad32x32_c(src_ptr, src_stride, ref_ptr + 6, ref_stride, 0x7fffffff);
+  sad_array[7] = (unsigned short)vp9_sad32x32_c(src_ptr, src_stride, ref_ptr + 7, ref_stride, 0x7fffffff);
 }
 
-void vp8_sad16x16x3_c(
+void vp9_sad16x16x3_c(
   const unsigned char *src_ptr,
   int  src_stride,
   const unsigned char *ref_ptr,
   int  ref_stride,
   unsigned int *sad_array
 ) {
-  sad_array[0] = vp8_sad16x16_c(src_ptr, src_stride, ref_ptr, ref_stride, 0x7fffffff);
-  sad_array[1] = vp8_sad16x16_c(src_ptr, src_stride, ref_ptr + 1, ref_stride, 0x7fffffff);
-  sad_array[2] = vp8_sad16x16_c(src_ptr, src_stride, ref_ptr + 2, ref_stride, 0x7fffffff);
+  sad_array[0] = vp9_sad16x16_c(src_ptr, src_stride, ref_ptr, ref_stride, 0x7fffffff);
+  sad_array[1] = vp9_sad16x16_c(src_ptr, src_stride, ref_ptr + 1, ref_stride, 0x7fffffff);
+  sad_array[2] = vp9_sad16x16_c(src_ptr, src_stride, ref_ptr + 2, ref_stride, 0x7fffffff);
 }
 
-void vp8_sad16x16x8_c(
+void vp9_sad16x16x8_c(
   const unsigned char *src_ptr,
   int  src_stride,
   const unsigned char *ref_ptr,
   int  ref_stride,
   unsigned short *sad_array
 ) {
-  sad_array[0] = (unsigned short)vp8_sad16x16_c(src_ptr, src_stride, ref_ptr, ref_stride, 0x7fffffff);
-  sad_array[1] = (unsigned short)vp8_sad16x16_c(src_ptr, src_stride, ref_ptr + 1, ref_stride, 0x7fffffff);
-  sad_array[2] = (unsigned short)vp8_sad16x16_c(src_ptr, src_stride, ref_ptr + 2, ref_stride, 0x7fffffff);
-  sad_array[3] = (unsigned short)vp8_sad16x16_c(src_ptr, src_stride, ref_ptr + 3, ref_stride, 0x7fffffff);
-  sad_array[4] = (unsigned short)vp8_sad16x16_c(src_ptr, src_stride, ref_ptr + 4, ref_stride, 0x7fffffff);
-  sad_array[5] = (unsigned short)vp8_sad16x16_c(src_ptr, src_stride, ref_ptr + 5, ref_stride, 0x7fffffff);
-  sad_array[6] = (unsigned short)vp8_sad16x16_c(src_ptr, src_stride, ref_ptr + 6, ref_stride, 0x7fffffff);
-  sad_array[7] = (unsigned short)vp8_sad16x16_c(src_ptr, src_stride, ref_ptr + 7, ref_stride, 0x7fffffff);
+  sad_array[0] = (unsigned short)vp9_sad16x16_c(src_ptr, src_stride, ref_ptr, ref_stride, 0x7fffffff);
+  sad_array[1] = (unsigned short)vp9_sad16x16_c(src_ptr, src_stride, ref_ptr + 1, ref_stride, 0x7fffffff);
+  sad_array[2] = (unsigned short)vp9_sad16x16_c(src_ptr, src_stride, ref_ptr + 2, ref_stride, 0x7fffffff);
+  sad_array[3] = (unsigned short)vp9_sad16x16_c(src_ptr, src_stride, ref_ptr + 3, ref_stride, 0x7fffffff);
+  sad_array[4] = (unsigned short)vp9_sad16x16_c(src_ptr, src_stride, ref_ptr + 4, ref_stride, 0x7fffffff);
+  sad_array[5] = (unsigned short)vp9_sad16x16_c(src_ptr, src_stride, ref_ptr + 5, ref_stride, 0x7fffffff);
+  sad_array[6] = (unsigned short)vp9_sad16x16_c(src_ptr, src_stride, ref_ptr + 6, ref_stride, 0x7fffffff);
+  sad_array[7] = (unsigned short)vp9_sad16x16_c(src_ptr, src_stride, ref_ptr + 7, ref_stride, 0x7fffffff);
 }
 
-void vp8_sad16x8x3_c(
+void vp9_sad16x8x3_c(
   const unsigned char *src_ptr,
   int  src_stride,
   const unsigned char *ref_ptr,
   int  ref_stride,
   unsigned int *sad_array
 ) {
-  sad_array[0] = vp8_sad16x8_c(src_ptr, src_stride, ref_ptr, ref_stride, 0x7fffffff);
-  sad_array[1] = vp8_sad16x8_c(src_ptr, src_stride, ref_ptr + 1, ref_stride, 0x7fffffff);
-  sad_array[2] = vp8_sad16x8_c(src_ptr, src_stride, ref_ptr + 2, ref_stride, 0x7fffffff);
+  sad_array[0] = vp9_sad16x8_c(src_ptr, src_stride, ref_ptr, ref_stride, 0x7fffffff);
+  sad_array[1] = vp9_sad16x8_c(src_ptr, src_stride, ref_ptr + 1, ref_stride, 0x7fffffff);
+  sad_array[2] = vp9_sad16x8_c(src_ptr, src_stride, ref_ptr + 2, ref_stride, 0x7fffffff);
 }
 
-void vp8_sad16x8x8_c(
+void vp9_sad16x8x8_c(
   const unsigned char *src_ptr,
   int  src_stride,
   const unsigned char *ref_ptr,
   int  ref_stride,
   unsigned short *sad_array
 ) {
-  sad_array[0] = (unsigned short)vp8_sad16x8_c(src_ptr, src_stride, ref_ptr, ref_stride, 0x7fffffff);
-  sad_array[1] = (unsigned short)vp8_sad16x8_c(src_ptr, src_stride, ref_ptr + 1, ref_stride, 0x7fffffff);
-  sad_array[2] = (unsigned short)vp8_sad16x8_c(src_ptr, src_stride, ref_ptr + 2, ref_stride, 0x7fffffff);
-  sad_array[3] = (unsigned short)vp8_sad16x8_c(src_ptr, src_stride, ref_ptr + 3, ref_stride, 0x7fffffff);
-  sad_array[4] = (unsigned short)vp8_sad16x8_c(src_ptr, src_stride, ref_ptr + 4, ref_stride, 0x7fffffff);
-  sad_array[5] = (unsigned short)vp8_sad16x8_c(src_ptr, src_stride, ref_ptr + 5, ref_stride, 0x7fffffff);
-  sad_array[6] = (unsigned short)vp8_sad16x8_c(src_ptr, src_stride, ref_ptr + 6, ref_stride, 0x7fffffff);
-  sad_array[7] = (unsigned short)vp8_sad16x8_c(src_ptr, src_stride, ref_ptr + 7, ref_stride, 0x7fffffff);
+  sad_array[0] = (unsigned short)vp9_sad16x8_c(src_ptr, src_stride, ref_ptr, ref_stride, 0x7fffffff);
+  sad_array[1] = (unsigned short)vp9_sad16x8_c(src_ptr, src_stride, ref_ptr + 1, ref_stride, 0x7fffffff);
+  sad_array[2] = (unsigned short)vp9_sad16x8_c(src_ptr, src_stride, ref_ptr + 2, ref_stride, 0x7fffffff);
+  sad_array[3] = (unsigned short)vp9_sad16x8_c(src_ptr, src_stride, ref_ptr + 3, ref_stride, 0x7fffffff);
+  sad_array[4] = (unsigned short)vp9_sad16x8_c(src_ptr, src_stride, ref_ptr + 4, ref_stride, 0x7fffffff);
+  sad_array[5] = (unsigned short)vp9_sad16x8_c(src_ptr, src_stride, ref_ptr + 5, ref_stride, 0x7fffffff);
+  sad_array[6] = (unsigned short)vp9_sad16x8_c(src_ptr, src_stride, ref_ptr + 6, ref_stride, 0x7fffffff);
+  sad_array[7] = (unsigned short)vp9_sad16x8_c(src_ptr, src_stride, ref_ptr + 7, ref_stride, 0x7fffffff);
 }
 
-void vp8_sad8x8x3_c(
+void vp9_sad8x8x3_c(
   const unsigned char *src_ptr,
   int  src_stride,
   const unsigned char *ref_ptr,
   int  ref_stride,
   unsigned int *sad_array
 ) {
-  sad_array[0] = vp8_sad8x8_c(src_ptr, src_stride, ref_ptr, ref_stride, 0x7fffffff);
-  sad_array[1] = vp8_sad8x8_c(src_ptr, src_stride, ref_ptr + 1, ref_stride, 0x7fffffff);
-  sad_array[2] = vp8_sad8x8_c(src_ptr, src_stride, ref_ptr + 2, ref_stride, 0x7fffffff);
+  sad_array[0] = vp9_sad8x8_c(src_ptr, src_stride, ref_ptr, ref_stride, 0x7fffffff);
+  sad_array[1] = vp9_sad8x8_c(src_ptr, src_stride, ref_ptr + 1, ref_stride, 0x7fffffff);
+  sad_array[2] = vp9_sad8x8_c(src_ptr, src_stride, ref_ptr + 2, ref_stride, 0x7fffffff);
 }
 
-void vp8_sad8x8x8_c(
+void vp9_sad8x8x8_c(
   const unsigned char *src_ptr,
   int  src_stride,
   const unsigned char *ref_ptr,
   int  ref_stride,
   unsigned short *sad_array
 ) {
-  sad_array[0] = (unsigned short)vp8_sad8x8_c(src_ptr, src_stride, ref_ptr, ref_stride, 0x7fffffff);
-  sad_array[1] = (unsigned short)vp8_sad8x8_c(src_ptr, src_stride, ref_ptr + 1, ref_stride, 0x7fffffff);
-  sad_array[2] = (unsigned short)vp8_sad8x8_c(src_ptr, src_stride, ref_ptr + 2, ref_stride, 0x7fffffff);
-  sad_array[3] = (unsigned short)vp8_sad8x8_c(src_ptr, src_stride, ref_ptr + 3, ref_stride, 0x7fffffff);
-  sad_array[4] = (unsigned short)vp8_sad8x8_c(src_ptr, src_stride, ref_ptr + 4, ref_stride, 0x7fffffff);
-  sad_array[5] = (unsigned short)vp8_sad8x8_c(src_ptr, src_stride, ref_ptr + 5, ref_stride, 0x7fffffff);
-  sad_array[6] = (unsigned short)vp8_sad8x8_c(src_ptr, src_stride, ref_ptr + 6, ref_stride, 0x7fffffff);
-  sad_array[7] = (unsigned short)vp8_sad8x8_c(src_ptr, src_stride, ref_ptr + 7, ref_stride, 0x7fffffff);
+  sad_array[0] = (unsigned short)vp9_sad8x8_c(src_ptr, src_stride, ref_ptr, ref_stride, 0x7fffffff);
+  sad_array[1] = (unsigned short)vp9_sad8x8_c(src_ptr, src_stride, ref_ptr + 1, ref_stride, 0x7fffffff);
+  sad_array[2] = (unsigned short)vp9_sad8x8_c(src_ptr, src_stride, ref_ptr + 2, ref_stride, 0x7fffffff);
+  sad_array[3] = (unsigned short)vp9_sad8x8_c(src_ptr, src_stride, ref_ptr + 3, ref_stride, 0x7fffffff);
+  sad_array[4] = (unsigned short)vp9_sad8x8_c(src_ptr, src_stride, ref_ptr + 4, ref_stride, 0x7fffffff);
+  sad_array[5] = (unsigned short)vp9_sad8x8_c(src_ptr, src_stride, ref_ptr + 5, ref_stride, 0x7fffffff);
+  sad_array[6] = (unsigned short)vp9_sad8x8_c(src_ptr, src_stride, ref_ptr + 6, ref_stride, 0x7fffffff);
+  sad_array[7] = (unsigned short)vp9_sad8x8_c(src_ptr, src_stride, ref_ptr + 7, ref_stride, 0x7fffffff);
 }
 
-void vp8_sad8x16x3_c(
+void vp9_sad8x16x3_c(
   const unsigned char *src_ptr,
   int  src_stride,
   const unsigned char *ref_ptr,
   int  ref_stride,
   unsigned int *sad_array
 ) {
-  sad_array[0] = vp8_sad8x16_c(src_ptr, src_stride, ref_ptr, ref_stride, 0x7fffffff);
-  sad_array[1] = vp8_sad8x16_c(src_ptr, src_stride, ref_ptr + 1, ref_stride, 0x7fffffff);
-  sad_array[2] = vp8_sad8x16_c(src_ptr, src_stride, ref_ptr + 2, ref_stride, 0x7fffffff);
+  sad_array[0] = vp9_sad8x16_c(src_ptr, src_stride, ref_ptr, ref_stride, 0x7fffffff);
+  sad_array[1] = vp9_sad8x16_c(src_ptr, src_stride, ref_ptr + 1, ref_stride, 0x7fffffff);
+  sad_array[2] = vp9_sad8x16_c(src_ptr, src_stride, ref_ptr + 2, ref_stride, 0x7fffffff);
 }
 
-void vp8_sad8x16x8_c(
+void vp9_sad8x16x8_c(
   const unsigned char *src_ptr,
   int  src_stride,
   const unsigned char *ref_ptr,
   int  ref_stride,
   unsigned short *sad_array
 ) {
-  sad_array[0] = (unsigned short)vp8_sad8x16_c(src_ptr, src_stride, ref_ptr, ref_stride, 0x7fffffff);
-  sad_array[1] = (unsigned short)vp8_sad8x16_c(src_ptr, src_stride, ref_ptr + 1, ref_stride, 0x7fffffff);
-  sad_array[2] = (unsigned short)vp8_sad8x16_c(src_ptr, src_stride, ref_ptr + 2, ref_stride, 0x7fffffff);
-  sad_array[3] = (unsigned short)vp8_sad8x16_c(src_ptr, src_stride, ref_ptr + 3, ref_stride, 0x7fffffff);
-  sad_array[4] = (unsigned short)vp8_sad8x16_c(src_ptr, src_stride, ref_ptr + 4, ref_stride, 0x7fffffff);
-  sad_array[5] = (unsigned short)vp8_sad8x16_c(src_ptr, src_stride, ref_ptr + 5, ref_stride, 0x7fffffff);
-  sad_array[6] = (unsigned short)vp8_sad8x16_c(src_ptr, src_stride, ref_ptr + 6, ref_stride, 0x7fffffff);
-  sad_array[7] = (unsigned short)vp8_sad8x16_c(src_ptr, src_stride, ref_ptr + 7, ref_stride, 0x7fffffff);
+  sad_array[0] = (unsigned short)vp9_sad8x16_c(src_ptr, src_stride, ref_ptr, ref_stride, 0x7fffffff);
+  sad_array[1] = (unsigned short)vp9_sad8x16_c(src_ptr, src_stride, ref_ptr + 1, ref_stride, 0x7fffffff);
+  sad_array[2] = (unsigned short)vp9_sad8x16_c(src_ptr, src_stride, ref_ptr + 2, ref_stride, 0x7fffffff);
+  sad_array[3] = (unsigned short)vp9_sad8x16_c(src_ptr, src_stride, ref_ptr + 3, ref_stride, 0x7fffffff);
+  sad_array[4] = (unsigned short)vp9_sad8x16_c(src_ptr, src_stride, ref_ptr + 4, ref_stride, 0x7fffffff);
+  sad_array[5] = (unsigned short)vp9_sad8x16_c(src_ptr, src_stride, ref_ptr + 5, ref_stride, 0x7fffffff);
+  sad_array[6] = (unsigned short)vp9_sad8x16_c(src_ptr, src_stride, ref_ptr + 6, ref_stride, 0x7fffffff);
+  sad_array[7] = (unsigned short)vp9_sad8x16_c(src_ptr, src_stride, ref_ptr + 7, ref_stride, 0x7fffffff);
 }
 
-void vp8_sad4x4x3_c(
+void vp9_sad4x4x3_c(
   const unsigned char *src_ptr,
   int  src_stride,
   const unsigned char *ref_ptr,
   int  ref_stride,
   unsigned int *sad_array
 ) {
-  sad_array[0] = vp8_sad4x4_c(src_ptr, src_stride, ref_ptr, ref_stride, 0x7fffffff);
-  sad_array[1] = vp8_sad4x4_c(src_ptr, src_stride, ref_ptr + 1, ref_stride, 0x7fffffff);
-  sad_array[2] = vp8_sad4x4_c(src_ptr, src_stride, ref_ptr + 2, ref_stride, 0x7fffffff);
+  sad_array[0] = vp9_sad4x4_c(src_ptr, src_stride, ref_ptr, ref_stride, 0x7fffffff);
+  sad_array[1] = vp9_sad4x4_c(src_ptr, src_stride, ref_ptr + 1, ref_stride, 0x7fffffff);
+  sad_array[2] = vp9_sad4x4_c(src_ptr, src_stride, ref_ptr + 2, ref_stride, 0x7fffffff);
 }
 
-void vp8_sad4x4x8_c(
+void vp9_sad4x4x8_c(
   const unsigned char *src_ptr,
   int  src_stride,
   const unsigned char *ref_ptr,
   int  ref_stride,
   unsigned short *sad_array
 ) {
-  sad_array[0] = (unsigned short)vp8_sad4x4_c(src_ptr, src_stride, ref_ptr, ref_stride, 0x7fffffff);
-  sad_array[1] = (unsigned short)vp8_sad4x4_c(src_ptr, src_stride, ref_ptr + 1, ref_stride, 0x7fffffff);
-  sad_array[2] = (unsigned short)vp8_sad4x4_c(src_ptr, src_stride, ref_ptr + 2, ref_stride, 0x7fffffff);
-  sad_array[3] = (unsigned short)vp8_sad4x4_c(src_ptr, src_stride, ref_ptr + 3, ref_stride, 0x7fffffff);
-  sad_array[4] = (unsigned short)vp8_sad4x4_c(src_ptr, src_stride, ref_ptr + 4, ref_stride, 0x7fffffff);
-  sad_array[5] = (unsigned short)vp8_sad4x4_c(src_ptr, src_stride, ref_ptr + 5, ref_stride, 0x7fffffff);
-  sad_array[6] = (unsigned short)vp8_sad4x4_c(src_ptr, src_stride, ref_ptr + 6, ref_stride, 0x7fffffff);
-  sad_array[7] = (unsigned short)vp8_sad4x4_c(src_ptr, src_stride, ref_ptr + 7, ref_stride, 0x7fffffff);
+  sad_array[0] = (unsigned short)vp9_sad4x4_c(src_ptr, src_stride, ref_ptr, ref_stride, 0x7fffffff);
+  sad_array[1] = (unsigned short)vp9_sad4x4_c(src_ptr, src_stride, ref_ptr + 1, ref_stride, 0x7fffffff);
+  sad_array[2] = (unsigned short)vp9_sad4x4_c(src_ptr, src_stride, ref_ptr + 2, ref_stride, 0x7fffffff);
+  sad_array[3] = (unsigned short)vp9_sad4x4_c(src_ptr, src_stride, ref_ptr + 3, ref_stride, 0x7fffffff);
+  sad_array[4] = (unsigned short)vp9_sad4x4_c(src_ptr, src_stride, ref_ptr + 4, ref_stride, 0x7fffffff);
+  sad_array[5] = (unsigned short)vp9_sad4x4_c(src_ptr, src_stride, ref_ptr + 5, ref_stride, 0x7fffffff);
+  sad_array[6] = (unsigned short)vp9_sad4x4_c(src_ptr, src_stride, ref_ptr + 6, ref_stride, 0x7fffffff);
+  sad_array[7] = (unsigned short)vp9_sad4x4_c(src_ptr, src_stride, ref_ptr + 7, ref_stride, 0x7fffffff);
 }
 
-void vp8_sad32x32x4d_c(const unsigned char *src_ptr,
+void vp9_sad32x32x4d_c(const unsigned char *src_ptr,
                        int  src_stride,
                        unsigned char *ref_ptr[],
                        int  ref_stride,
                        unsigned int *sad_array
                        ) {
-  sad_array[0] = vp8_sad32x32_c(src_ptr, src_stride, ref_ptr[0], ref_stride, 0x7fffffff);
-  sad_array[1] = vp8_sad32x32_c(src_ptr, src_stride, ref_ptr[1], ref_stride, 0x7fffffff);
-  sad_array[2] = vp8_sad32x32_c(src_ptr, src_stride, ref_ptr[2], ref_stride, 0x7fffffff);
-  sad_array[3] = vp8_sad32x32_c(src_ptr, src_stride, ref_ptr[3], ref_stride, 0x7fffffff);
+  sad_array[0] = vp9_sad32x32_c(src_ptr, src_stride, ref_ptr[0], ref_stride, 0x7fffffff);
+  sad_array[1] = vp9_sad32x32_c(src_ptr, src_stride, ref_ptr[1], ref_stride, 0x7fffffff);
+  sad_array[2] = vp9_sad32x32_c(src_ptr, src_stride, ref_ptr[2], ref_stride, 0x7fffffff);
+  sad_array[3] = vp9_sad32x32_c(src_ptr, src_stride, ref_ptr[3], ref_stride, 0x7fffffff);
 }
 
-void vp8_sad16x16x4d_c(
+void vp9_sad16x16x4d_c(
   const unsigned char *src_ptr,
   int  src_stride,
   unsigned char *ref_ptr[],
   int  ref_stride,
   unsigned int *sad_array
 ) {
-  sad_array[0] = vp8_sad16x16_c(src_ptr, src_stride, ref_ptr[0], ref_stride, 0x7fffffff);
-  sad_array[1] = vp8_sad16x16_c(src_ptr, src_stride, ref_ptr[1], ref_stride, 0x7fffffff);
-  sad_array[2] = vp8_sad16x16_c(src_ptr, src_stride, ref_ptr[2], ref_stride, 0x7fffffff);
-  sad_array[3] = vp8_sad16x16_c(src_ptr, src_stride, ref_ptr[3], ref_stride, 0x7fffffff);
+  sad_array[0] = vp9_sad16x16_c(src_ptr, src_stride, ref_ptr[0], ref_stride, 0x7fffffff);
+  sad_array[1] = vp9_sad16x16_c(src_ptr, src_stride, ref_ptr[1], ref_stride, 0x7fffffff);
+  sad_array[2] = vp9_sad16x16_c(src_ptr, src_stride, ref_ptr[2], ref_stride, 0x7fffffff);
+  sad_array[3] = vp9_sad16x16_c(src_ptr, src_stride, ref_ptr[3], ref_stride, 0x7fffffff);
 }
 
-void vp8_sad16x8x4d_c(
+void vp9_sad16x8x4d_c(
   const unsigned char *src_ptr,
   int  src_stride,
   unsigned char *ref_ptr[],
   int  ref_stride,
   unsigned int *sad_array
 ) {
-  sad_array[0] = vp8_sad16x8_c(src_ptr, src_stride, ref_ptr[0], ref_stride, 0x7fffffff);
-  sad_array[1] = vp8_sad16x8_c(src_ptr, src_stride, ref_ptr[1], ref_stride, 0x7fffffff);
-  sad_array[2] = vp8_sad16x8_c(src_ptr, src_stride, ref_ptr[2], ref_stride, 0x7fffffff);
-  sad_array[3] = vp8_sad16x8_c(src_ptr, src_stride, ref_ptr[3], ref_stride, 0x7fffffff);
+  sad_array[0] = vp9_sad16x8_c(src_ptr, src_stride, ref_ptr[0], ref_stride, 0x7fffffff);
+  sad_array[1] = vp9_sad16x8_c(src_ptr, src_stride, ref_ptr[1], ref_stride, 0x7fffffff);
+  sad_array[2] = vp9_sad16x8_c(src_ptr, src_stride, ref_ptr[2], ref_stride, 0x7fffffff);
+  sad_array[3] = vp9_sad16x8_c(src_ptr, src_stride, ref_ptr[3], ref_stride, 0x7fffffff);
 }
 
-void vp8_sad8x8x4d_c(
+void vp9_sad8x8x4d_c(
   const unsigned char *src_ptr,
   int  src_stride,
   unsigned char *ref_ptr[],
   int  ref_stride,
   unsigned int *sad_array
 ) {
-  sad_array[0] = vp8_sad8x8_c(src_ptr, src_stride, ref_ptr[0], ref_stride, 0x7fffffff);
-  sad_array[1] = vp8_sad8x8_c(src_ptr, src_stride, ref_ptr[1], ref_stride, 0x7fffffff);
-  sad_array[2] = vp8_sad8x8_c(src_ptr, src_stride, ref_ptr[2], ref_stride, 0x7fffffff);
-  sad_array[3] = vp8_sad8x8_c(src_ptr, src_stride, ref_ptr[3], ref_stride, 0x7fffffff);
+  sad_array[0] = vp9_sad8x8_c(src_ptr, src_stride, ref_ptr[0], ref_stride, 0x7fffffff);
+  sad_array[1] = vp9_sad8x8_c(src_ptr, src_stride, ref_ptr[1], ref_stride, 0x7fffffff);
+  sad_array[2] = vp9_sad8x8_c(src_ptr, src_stride, ref_ptr[2], ref_stride, 0x7fffffff);
+  sad_array[3] = vp9_sad8x8_c(src_ptr, src_stride, ref_ptr[3], ref_stride, 0x7fffffff);
 }
 
-void vp8_sad8x16x4d_c(
+void vp9_sad8x16x4d_c(
   const unsigned char *src_ptr,
   int  src_stride,
   unsigned char *ref_ptr[],
   int  ref_stride,
   unsigned int *sad_array
 ) {
-  sad_array[0] = vp8_sad8x16_c(src_ptr, src_stride, ref_ptr[0], ref_stride, 0x7fffffff);
-  sad_array[1] = vp8_sad8x16_c(src_ptr, src_stride, ref_ptr[1], ref_stride, 0x7fffffff);
-  sad_array[2] = vp8_sad8x16_c(src_ptr, src_stride, ref_ptr[2], ref_stride, 0x7fffffff);
-  sad_array[3] = vp8_sad8x16_c(src_ptr, src_stride, ref_ptr[3], ref_stride, 0x7fffffff);
+  sad_array[0] = vp9_sad8x16_c(src_ptr, src_stride, ref_ptr[0], ref_stride, 0x7fffffff);
+  sad_array[1] = vp9_sad8x16_c(src_ptr, src_stride, ref_ptr[1], ref_stride, 0x7fffffff);
+  sad_array[2] = vp9_sad8x16_c(src_ptr, src_stride, ref_ptr[2], ref_stride, 0x7fffffff);
+  sad_array[3] = vp9_sad8x16_c(src_ptr, src_stride, ref_ptr[3], ref_stride, 0x7fffffff);
 }
 
-void vp8_sad4x4x4d_c(
+void vp9_sad4x4x4d_c(
   const unsigned char *src_ptr,
   int  src_stride,
   unsigned char *ref_ptr[],
   int  ref_stride,
   unsigned int *sad_array
 ) {
-  sad_array[0] = vp8_sad4x4_c(src_ptr, src_stride, ref_ptr[0], ref_stride, 0x7fffffff);
-  sad_array[1] = vp8_sad4x4_c(src_ptr, src_stride, ref_ptr[1], ref_stride, 0x7fffffff);
-  sad_array[2] = vp8_sad4x4_c(src_ptr, src_stride, ref_ptr[2], ref_stride, 0x7fffffff);
-  sad_array[3] = vp8_sad4x4_c(src_ptr, src_stride, ref_ptr[3], ref_stride, 0x7fffffff);
+  sad_array[0] = vp9_sad4x4_c(src_ptr, src_stride, ref_ptr[0], ref_stride, 0x7fffffff);
+  sad_array[1] = vp9_sad4x4_c(src_ptr, src_stride, ref_ptr[1], ref_stride, 0x7fffffff);
+  sad_array[2] = vp9_sad4x4_c(src_ptr, src_stride, ref_ptr[2], ref_stride, 0x7fffffff);
+  sad_array[3] = vp9_sad4x4_c(src_ptr, src_stride, ref_ptr[3], ref_stride, 0x7fffffff);
 }
 
 /* Copy 2 macroblocks to a buffer */
-void vp8_copy32xn_c(
+void vp9_copy32xn_c(
   unsigned char *src_ptr,
   int  src_stride,
   unsigned char *dst_ptr,
diff --git a/vp8/encoder/satd_c.c b/vp8/encoder/satd_c.c
index 4a76de36b5ced41c20058bbea0f2a916a1f2fa80..102aa73e59599e943103d5822468fd2d35e61a96 100644
--- a/vp8/encoder/satd_c.c
+++ b/vp8/encoder/satd_c.c
@@ -11,7 +11,7 @@
 #include <stdlib.h>
 #include "vpx_ports/mem.h"
 #include "./vpx_rtcd.h"
-unsigned int vp8_satd16x16_c(const unsigned char *src_ptr,
+unsigned int vp9_satd16x16_c(const unsigned char *src_ptr,
                              int  src_stride,
                              const unsigned char *ref_ptr,
                              int  ref_stride,
@@ -33,7 +33,7 @@ unsigned int vp8_satd16x16_c(const unsigned char *src_ptr,
   in = diff_in;
   for (r = 0; r < 16; r += 4) {
     for (c = 0; c < 16; c += 4) {
-      vp8_short_walsh4x4_c(in + c, diff_out, 32);
+      vp9_short_walsh4x4_c(in + c, diff_out, 32);
       for (i = 0; i < 16; i++)
         satd += abs(diff_out[i]);
     }
diff --git a/vp8/encoder/segmentation.c b/vp8/encoder/segmentation.c
index 4071839e8e6565c57c6efa84b1d6e487b7d29b69..1b33aa10de0ccbd0f971e81d62f52f455a5cf63c 100644
--- a/vp8/encoder/segmentation.c
+++ b/vp8/encoder/segmentation.c
@@ -14,7 +14,7 @@
 #include "segmentation.h"
 #include "vp8/common/pred_common.h"
 
-void vp8_update_gf_useage_maps(VP8_COMP *cpi, VP8_COMMON *cm, MACROBLOCK *x) {
+void vp9_update_gf_useage_maps(VP8_COMP *cpi, VP8_COMMON *cm, MACROBLOCK *x) {
   int mb_row, mb_col;
 
   MODE_INFO *this_mb_mode_info = cm->mi;
@@ -58,7 +58,7 @@ void vp8_update_gf_useage_maps(VP8_COMP *cpi, VP8_COMMON *cm, MACROBLOCK *x) {
   }
 }
 
-void vp8_enable_segmentation(VP8_PTR ptr) {
+void vp9_enable_segmentation(VP8_PTR ptr) {
   VP8_COMP *cpi = (VP8_COMP *)(ptr);
 
   // Set the appropriate feature bit
@@ -67,14 +67,14 @@ void vp8_enable_segmentation(VP8_PTR ptr) {
   cpi->mb.e_mbd.update_mb_segmentation_data = 1;
 }
 
-void vp8_disable_segmentation(VP8_PTR ptr) {
+void vp9_disable_segmentation(VP8_PTR ptr) {
   VP8_COMP *cpi = (VP8_COMP *)(ptr);
 
   // Clear the appropriate feature bit
   cpi->mb.e_mbd.segmentation_enabled = 0;
 }
 
-void vp8_set_segmentation_map(VP8_PTR ptr,
+void vp9_set_segmentation_map(VP8_PTR ptr,
                               unsigned char *segmentation_map) {
   VP8_COMP *cpi = (VP8_COMP *)(ptr);
 
@@ -87,7 +87,7 @@ void vp8_set_segmentation_map(VP8_PTR ptr,
   cpi->mb.e_mbd.update_mb_segmentation_data = 1;
 }
 
-void vp8_set_segment_data(VP8_PTR ptr,
+void vp9_set_segment_data(VP8_PTR ptr,
                           signed char *feature_data,
                           unsigned char abs_delta) {
   VP8_COMP *cpi = (VP8_COMP *)(ptr);
diff --git a/vp8/encoder/segmentation.h b/vp8/encoder/segmentation.h
index 88217dff8afdd5454a43c2482f154f0225482162..bf36505423e1023dde1fcc72a4f99c2f5c7753b2 100644
--- a/vp8/encoder/segmentation.h
+++ b/vp8/encoder/segmentation.h
@@ -16,14 +16,14 @@
 #ifndef __INC_SEGMENTATION_H__
 #define __INC_SEGMENTATION_H__ 1
 
-extern void vp8_update_gf_useage_maps(VP8_COMP *cpi, VP8_COMMON *cm, MACROBLOCK *x);
+extern void vp9_update_gf_useage_maps(VP8_COMP *cpi, VP8_COMMON *cm, MACROBLOCK *x);
 
-extern void vp8_enable_segmentation(VP8_PTR ptr);
-extern void vp8_disable_segmentation(VP8_PTR ptr);
+extern void vp9_enable_segmentation(VP8_PTR ptr);
+extern void vp9_disable_segmentation(VP8_PTR ptr);
 
 // Valid values for a segment are 0 to 3
 // Segmentation map is arrange as [Rows][Columns]
-extern void vp8_set_segmentation_map(VP8_PTR ptr, unsigned char *segmentation_map);
+extern void vp9_set_segmentation_map(VP8_PTR ptr, unsigned char *segmentation_map);
 
 // The values given for each segment can be either deltas (from the default
 // value chosen for the frame) or absolute values.
@@ -36,7 +36,7 @@ extern void vp8_set_segmentation_map(VP8_PTR ptr, unsigned char *segmentation_ma
 // abs_delta = SEGMENT_DELTADATA (deltas) abs_delta = SEGMENT_ABSDATA (use
 // the absolute values given).
 //
-extern void vp8_set_segment_data(VP8_PTR ptr, signed char *feature_data, unsigned char abs_delta);
+extern void vp9_set_segment_data(VP8_PTR ptr, signed char *feature_data, unsigned char abs_delta);
 
 extern void vp9_choose_segmap_coding_method(VP8_COMP *cpi);
 
diff --git a/vp8/encoder/ssim.c b/vp8/encoder/ssim.c
index aa3a1ca0b9dd66562208db780fa5121fbb37cf21..464dfd341ac99a1210e721e276d71e3fd657b0d7 100644
--- a/vp8/encoder/ssim.c
+++ b/vp8/encoder/ssim.c
@@ -11,7 +11,7 @@
 
 #include "onyx_int.h"
 
-void vp8_ssim_parms_16x16_c(unsigned char *s, int sp, unsigned char *r,
+void vp9_ssim_parms_16x16_c(unsigned char *s, int sp, unsigned char *r,
                             int rp, unsigned long *sum_s, unsigned long *sum_r,
                             unsigned long *sum_sq_s, unsigned long *sum_sq_r,
                             unsigned long *sum_sxr) {
@@ -26,7 +26,7 @@ void vp8_ssim_parms_16x16_c(unsigned char *s, int sp, unsigned char *r,
     }
   }
 }
-void vp8_ssim_parms_8x8_c(unsigned char *s, int sp, unsigned char *r, int rp,
+void vp9_ssim_parms_8x8_c(unsigned char *s, int sp, unsigned char *r, int rp,
                           unsigned long *sum_s, unsigned long *sum_r,
                           unsigned long *sum_sq_s, unsigned long *sum_sq_r,
                           unsigned long *sum_sxr) {
@@ -67,13 +67,13 @@ static double similarity(unsigned long sum_s, unsigned long sum_r,
 
 static double ssim_16x16(unsigned char *s, int sp, unsigned char *r, int rp) {
   unsigned long sum_s = 0, sum_r = 0, sum_sq_s = 0, sum_sq_r = 0, sum_sxr = 0;
-  vp8_ssim_parms_16x16(s, sp, r, rp, &sum_s, &sum_r, &sum_sq_s, &sum_sq_r,
+  vp9_ssim_parms_16x16(s, sp, r, rp, &sum_s, &sum_r, &sum_sq_s, &sum_sq_r,
                        &sum_sxr);
   return similarity(sum_s, sum_r, sum_sq_s, sum_sq_r, sum_sxr, 256);
 }
 static double ssim_8x8(unsigned char *s, int sp, unsigned char *r, int rp) {
   unsigned long sum_s = 0, sum_r = 0, sum_sq_s = 0, sum_sq_r = 0, sum_sxr = 0;
-  vp8_ssim_parms_8x8(s, sp, r, rp, &sum_s, &sum_r, &sum_sq_s, &sum_sq_r,
+  vp9_ssim_parms_8x8(s, sp, r, rp, &sum_s, &sum_r, &sum_sq_s, &sum_sq_r,
                      &sum_sxr);
   return similarity(sum_s, sum_r, sum_sq_s, sum_sq_r, sum_sxr, 64);
 }
@@ -81,7 +81,7 @@ static double ssim_8x8(unsigned char *s, int sp, unsigned char *r, int rp) {
 // We are using a 8x8 moving window with starting location of each 8x8 window
 // on the 4x4 pixel grid. Such arrangement allows the windows to overlap
 // block boundaries to penalize blocking artifacts.
-double vp8_ssim2(unsigned char *img1, unsigned char *img2, int stride_img1,
+double vp9_ssim2(unsigned char *img1, unsigned char *img2, int stride_img1,
                  int stride_img2, int width, int height) {
   int i, j;
   int samples = 0;
@@ -98,20 +98,20 @@ double vp8_ssim2(unsigned char *img1, unsigned char *img2, int stride_img1,
   ssim_total /= samples;
   return ssim_total;
 }
-double vp8_calc_ssim(YV12_BUFFER_CONFIG *source, YV12_BUFFER_CONFIG *dest,
+double vp9_calc_ssim(YV12_BUFFER_CONFIG *source, YV12_BUFFER_CONFIG *dest,
                      int lumamask, double *weight) {
   double a, b, c;
   double ssimv;
 
-  a = vp8_ssim2(source->y_buffer, dest->y_buffer,
+  a = vp9_ssim2(source->y_buffer, dest->y_buffer,
                 source->y_stride, dest->y_stride, source->y_width,
                 source->y_height);
 
-  b = vp8_ssim2(source->u_buffer, dest->u_buffer,
+  b = vp9_ssim2(source->u_buffer, dest->u_buffer,
                 source->uv_stride, dest->uv_stride, source->uv_width,
                 source->uv_height);
 
-  c = vp8_ssim2(source->v_buffer, dest->v_buffer,
+  c = vp9_ssim2(source->v_buffer, dest->v_buffer,
                 source->uv_stride, dest->uv_stride, source->uv_width,
                 source->uv_height);
 
@@ -122,20 +122,20 @@ double vp8_calc_ssim(YV12_BUFFER_CONFIG *source, YV12_BUFFER_CONFIG *dest,
   return ssimv;
 }
 
-double vp8_calc_ssimg(YV12_BUFFER_CONFIG *source, YV12_BUFFER_CONFIG *dest,
+double vp9_calc_ssimg(YV12_BUFFER_CONFIG *source, YV12_BUFFER_CONFIG *dest,
                       double *ssim_y, double *ssim_u, double *ssim_v) {
   double ssim_all = 0;
   double a, b, c;
 
-  a = vp8_ssim2(source->y_buffer, dest->y_buffer,
+  a = vp9_ssim2(source->y_buffer, dest->y_buffer,
                 source->y_stride, dest->y_stride, source->y_width,
                 source->y_height);
 
-  b = vp8_ssim2(source->u_buffer, dest->u_buffer,
+  b = vp9_ssim2(source->u_buffer, dest->u_buffer,
                 source->uv_stride, dest->uv_stride, source->uv_width,
                 source->uv_height);
 
-  c = vp8_ssim2(source->v_buffer, dest->v_buffer,
+  c = vp9_ssim2(source->v_buffer, dest->v_buffer,
                 source->uv_stride, dest->uv_stride, source->uv_width,
                 source->uv_height);
   *ssim_y = a;
diff --git a/vp8/encoder/temporal_filter.c b/vp8/encoder/temporal_filter.c
index e59aa14fa73a80cae9baa815cacf8a0f1f8901e3..e8bf9da9ae2e37a152223e1c1a7c3ae8f422235e 100644
--- a/vp8/encoder/temporal_filter.c
+++ b/vp8/encoder/temporal_filter.c
@@ -83,7 +83,7 @@ static void vp8_temporal_filter_predictors_mb_c
     vp8_copy_mem8x8(vptr, stride, &pred[320], 8);
   }
 }
-void vp8_temporal_filter_apply_c
+void vp9_temporal_filter_apply_c
 (
   unsigned char *frame1,
   unsigned int stride,
@@ -185,7 +185,7 @@ static int vp8_temporal_filter_find_matching_mb_c
   /*cpi->sf.search_method == HEX*/
   // TODO Check that the 16x16 vf & sdf are selected here
   // Ignore mv costing by sending NULL pointer instead of cost arrays
-  bestsme = vp8_hex_search(x, b, d, &best_ref_mv1_full, &d->bmi.as_mv.first,
+  bestsme = vp9_hex_search(x, b, d, &best_ref_mv1_full, &d->bmi.as_mv.first,
                            step_param, sadpb, &cpi->fn_ptr[BLOCK_16X16],
                            NULLMVCOST, NULLMVCOST,
                            &best_ref_mv1);
@@ -410,7 +410,7 @@ static void vp8_temporal_filter_iterate_c
   mbd->pre.v_buffer = v_buffer;
 }
 
-void vp8_temporal_filter_prepare_c
+void vp9_temporal_filter_prepare_c
 (
   VP8_COMP *cpi,
   int distance
@@ -431,7 +431,7 @@ void vp8_temporal_filter_prepare_c
   int max_frames = cpi->active_arnr_frames;
 
   num_frames_backward = distance;
-  num_frames_forward = vp8_lookahead_depth(cpi->lookahead)
+  num_frames_forward = vp9_lookahead_depth(cpi->lookahead)
                        - (num_frames_backward + 1);
 
   switch (blur_type) {
@@ -503,7 +503,7 @@ void vp8_temporal_filter_prepare_c
   vpx_memset(cpi->frames, 0, max_frames * sizeof(YV12_BUFFER_CONFIG *));
   for (frame = 0; frame < frames_to_blur; frame++) {
     int which_buffer =  start_frame - frame;
-    struct lookahead_entry *buf = vp8_lookahead_peek(cpi->lookahead,
+    struct lookahead_entry *buf = vp9_lookahead_peek(cpi->lookahead,
                                                      which_buffer);
     cpi->frames[frames_to_blur - 1 - frame] = &buf->img;
   }
diff --git a/vp8/encoder/temporal_filter.h b/vp8/encoder/temporal_filter.h
index e051c5c765b27667fadb053cb9c5407b26f64bc8..f82adb3b129b3afc7d14b144d473f86930fc7b1b 100644
--- a/vp8/encoder/temporal_filter.h
+++ b/vp8/encoder/temporal_filter.h
@@ -29,10 +29,10 @@
 #include "x86/temporal_filter_x86.h"
 #endif
 
-#ifndef vp8_temporal_filter_apply
-#define vp8_temporal_filter_apply vp8_temporal_filter_apply_c
+#ifndef vp9_temporal_filter_apply
+#define vp9_temporal_filter_apply vp9_temporal_filter_apply_c
 #endif
-extern prototype_apply(vp8_temporal_filter_apply);
+extern prototype_apply(vp9_temporal_filter_apply);
 
 typedef struct {
   prototype_apply(*apply);
diff --git a/vp8/encoder/tokenize.c b/vp8/encoder/tokenize.c
index ab2d543cbfc06385d2f89ed8fca941dcca7d5817..58037a5a74d5d1270c2458d71613dde66964d2b6 100644
--- a/vp8/encoder/tokenize.c
+++ b/vp8/encoder/tokenize.c
@@ -48,8 +48,8 @@ extern unsigned int hybrid_tree_update_hist_16x16[BLOCK_TYPES_16X16][COEF_BANDS]
                     [PREV_COEF_CONTEXTS][ENTROPY_NODES] [2];
 #endif  /* ENTROPY_STATS */
 
-void vp8_stuff_mb(VP8_COMP *cpi, MACROBLOCKD *xd, TOKENEXTRA **t, int dry_run);
-void vp8_fix_contexts(MACROBLOCKD *xd);
+void vp9_stuff_mb(VP8_COMP *cpi, MACROBLOCKD *xd, TOKENEXTRA **t, int dry_run);
+void vp9_fix_contexts(MACROBLOCKD *xd);
 
 static TOKENVALUE dct_value_tokens[DCT_MAX_VALUE * 2];
 const TOKENVALUE *vp8_dct_value_tokens_ptr;
@@ -286,7 +286,7 @@ static int mb_is_skippable_16x16(MACROBLOCKD *xd) {
   return (vp9_mby_is_skippable_16x16(xd) & vp9_mbuv_is_skippable_8x8(xd));
 }
 
-void vp8_tokenize_mb(VP8_COMP *cpi,
+void vp9_tokenize_mb(VP8_COMP *cpi,
                      MACROBLOCKD *xd,
                      TOKENEXTRA **t,
                      int dry_run) {
@@ -337,9 +337,9 @@ void vp8_tokenize_mb(VP8_COMP *cpi,
     if (!dry_run)
       cpi->skip_true_count[mb_skip_context] += skip_inc;
     if (!cpi->common.mb_no_coeff_skip) {
-      vp8_stuff_mb(cpi, xd, t, dry_run);
+      vp9_stuff_mb(cpi, xd, t, dry_run);
     } else {
-      vp8_fix_contexts(xd);
+      vp9_fix_contexts(xd);
     }
     if (dry_run)
       *t = t_backup;
@@ -663,7 +663,7 @@ void print_context_counters() {
 }
 #endif
 
-void vp8_tokenize_initialize() {
+void vp9_tokenize_initialize() {
   fill_value_tokens();
 }
 
@@ -730,7 +730,7 @@ static __inline void stuff_b(VP8_COMP *cpi,
   }
 }
 
-static void vp8_stuff_mb_8x8(VP8_COMP *cpi, MACROBLOCKD *xd,
+static void vp9_stuff_mb_8x8(VP8_COMP *cpi, MACROBLOCKD *xd,
                              TOKENEXTRA **t, int dry_run) {
   ENTROPY_CONTEXT *A = (ENTROPY_CONTEXT *)xd->above_context;
   ENTROPY_CONTEXT *L = (ENTROPY_CONTEXT *)xd->left_context;
@@ -765,7 +765,7 @@ static void vp8_stuff_mb_8x8(VP8_COMP *cpi, MACROBLOCKD *xd,
   }
 }
 
-static void vp8_stuff_mb_16x16(VP8_COMP *cpi, MACROBLOCKD *xd,
+static void vp9_stuff_mb_16x16(VP8_COMP *cpi, MACROBLOCKD *xd,
                                TOKENEXTRA **t, int dry_run) {
   ENTROPY_CONTEXT * A = (ENTROPY_CONTEXT *)xd->above_context;
   ENTROPY_CONTEXT * L = (ENTROPY_CONTEXT *)xd->left_context;
@@ -784,7 +784,7 @@ static void vp8_stuff_mb_16x16(VP8_COMP *cpi, MACROBLOCKD *xd,
   vpx_memset(&L[8], 0, sizeof(L[8]));
 }
 
-static void vp8_stuff_mb_4x4(VP8_COMP *cpi, MACROBLOCKD *xd,
+static void vp9_stuff_mb_4x4(VP8_COMP *cpi, MACROBLOCKD *xd,
                              TOKENEXTRA **t, int dry_run) {
   ENTROPY_CONTEXT *A = (ENTROPY_CONTEXT *)xd->above_context;
   ENTROPY_CONTEXT *L = (ENTROPY_CONTEXT *)xd->left_context;
@@ -811,7 +811,7 @@ static void vp8_stuff_mb_4x4(VP8_COMP *cpi, MACROBLOCKD *xd,
             L + vp8_block2left[b], TX_4X4, dry_run);
 }
 
-static void vp8_stuff_mb_8x8_4x4uv(VP8_COMP *cpi, MACROBLOCKD *xd,
+static void vp9_stuff_mb_8x8_4x4uv(VP8_COMP *cpi, MACROBLOCKD *xd,
                                    TOKENEXTRA **t, int dry_run) {
   ENTROPY_CONTEXT *A = (ENTROPY_CONTEXT *)xd->above_context;
   ENTROPY_CONTEXT *L = (ENTROPY_CONTEXT *)xd->left_context;
@@ -830,21 +830,21 @@ static void vp8_stuff_mb_8x8_4x4uv(VP8_COMP *cpi, MACROBLOCKD *xd,
             L + vp8_block2left[b], TX_4X4, dry_run);
 }
 
-void vp8_stuff_mb(VP8_COMP *cpi, MACROBLOCKD *xd, TOKENEXTRA **t, int dry_run) {
+void vp9_stuff_mb(VP8_COMP *cpi, MACROBLOCKD *xd, TOKENEXTRA **t, int dry_run) {
   TX_SIZE tx_size = xd->mode_info_context->mbmi.txfm_size;
   TOKENEXTRA * const t_backup = *t;
 
   if (tx_size == TX_16X16) {
-    vp8_stuff_mb_16x16(cpi, xd, t, dry_run);
+    vp9_stuff_mb_16x16(cpi, xd, t, dry_run);
   } else if (tx_size == TX_8X8) {
     if (xd->mode_info_context->mbmi.mode == I8X8_PRED ||
         xd->mode_info_context->mbmi.mode == SPLITMV) {
-      vp8_stuff_mb_8x8_4x4uv(cpi, xd, t, dry_run);
+      vp9_stuff_mb_8x8_4x4uv(cpi, xd, t, dry_run);
     } else {
-      vp8_stuff_mb_8x8(cpi, xd, t, dry_run);
+      vp9_stuff_mb_8x8(cpi, xd, t, dry_run);
     }
   } else {
-    vp8_stuff_mb_4x4(cpi, xd, t, dry_run);
+    vp9_stuff_mb_4x4(cpi, xd, t, dry_run);
   }
 
   if (dry_run) {
@@ -852,7 +852,7 @@ void vp8_stuff_mb(VP8_COMP *cpi, MACROBLOCKD *xd, TOKENEXTRA **t, int dry_run) {
   }
 }
 
-void vp8_fix_contexts(MACROBLOCKD *xd) {
+void vp9_fix_contexts(MACROBLOCKD *xd) {
   /* Clear entropy contexts for Y2 blocks */
   if ((xd->mode_info_context->mbmi.mode != B_PRED
       && xd->mode_info_context->mbmi.mode != I8X8_PRED
diff --git a/vp8/encoder/tokenize.h b/vp8/encoder/tokenize.h
index 46e94fb0facb35e8fad5dca14f33ce92455d4962..7e36561e0ebf81ae81c285f055ce8d67456e945b 100644
--- a/vp8/encoder/tokenize.h
+++ b/vp8/encoder/tokenize.h
@@ -15,7 +15,7 @@
 #include "vp8/common/entropy.h"
 #include "block.h"
 
-void vp8_tokenize_initialize();
+void vp9_tokenize_initialize();
 
 typedef struct {
   short Token;
diff --git a/vp8/encoder/treewriter.c b/vp8/encoder/treewriter.c
index 2dc7d8e62e2fd9e240f010bf1b2ff7b89d3a3465..3349394b1b0b0de268211666daffbd9239cc9fc2 100644
--- a/vp8/encoder/treewriter.c
+++ b/vp8/encoder/treewriter.c
@@ -30,10 +30,10 @@ static void cost(
       cost(C, T, P, j, d);
   } while (++i & 1);
 }
-void vp8_cost_tokens(int *c, const vp8_prob *p, vp8_tree t) {
+void vp9_cost_tokens(int *c, const vp8_prob *p, vp8_tree t) {
   cost(c, t, p, 0, 0);
 }
 
-void vp8_cost_tokens_skip(int *c, const vp8_prob *p, vp8_tree t) {
+void vp9_cost_tokens_skip(int *c, const vp8_prob *p, vp8_tree t) {
   cost(c, t, p, 2, 0);
 }
diff --git a/vp8/encoder/treewriter.h b/vp8/encoder/treewriter.h
index 7292072fe9eaa866b9ad1bae0f5b9e11ef70f4ac..4c34db4f1ee49076aefaf47125c432d3302ffcef 100644
--- a/vp8/encoder/treewriter.h
+++ b/vp8/encoder/treewriter.h
@@ -22,7 +22,7 @@
 typedef BOOL_CODER vp8_writer;
 
 #define vp8_write vp8_encode_bool
-#define vp8_write_literal vp8_encode_value
+#define vp8_write_literal vp9_encode_value
 #define vp8_write_bit( W, V) vp8_write( W, V, vp8_prob_half)
 
 #define vp8bc_write vp8bc_write_bool
@@ -113,10 +113,10 @@ static __inline int vp8_cost_token
 
 /* Fill array of costs for all possible token values. */
 
-void vp8_cost_tokens(
+void vp9_cost_tokens(
   int *Costs, const vp8_prob *, vp8_tree
 );
 
-void vp8_cost_tokens_skip(int *c, const vp8_prob *p, vp8_tree t);
+void vp9_cost_tokens_skip(int *c, const vp8_prob *p, vp8_tree t);
 
 #endif
diff --git a/vp8/encoder/variance.h b/vp8/encoder/variance.h
index cdeb390c3bfcdba4eab413019d2e7c92f2e025a8..8fb8db0f6894743b478b3608e62e52a77f4d8ee4 100644
--- a/vp8/encoder/variance.h
+++ b/vp8/encoder/variance.h
@@ -18,7 +18,7 @@ typedef unsigned int(*vp8_sad_fn_t)(const unsigned char *src_ptr,
                                     int ref_stride,
                                     unsigned int max_sad);
 
-typedef void (*vp8_copy32xn_fn_t)(const unsigned char *src_ptr,
+typedef void (*vp9_copy32xn_fn_t)(const unsigned char *src_ptr,
                                   int source_stride,
                                   const unsigned char *ref_ptr,
                                   int ref_stride,
@@ -78,7 +78,7 @@ typedef struct variance_vtable {
     vp8_sad_multi_fn_t      sdx3f;
     vp8_sad_multi1_fn_t     sdx8f;
     vp8_sad_multi_d_fn_t    sdx4df;
-    vp8_copy32xn_fn_t       copymem;
+    vp9_copy32xn_fn_t       copymem;
 } vp8_variance_fn_ptr_t;
 
 #endif
diff --git a/vp8/encoder/variance_c.c b/vp8/encoder/variance_c.c
index a66c80b8099b313ae3ded83cc32f139999b1fb10..3a093b7d08825c0f7143b9af61cc9b990b371cc4 100644
--- a/vp8/encoder/variance_c.c
+++ b/vp8/encoder/variance_c.c
@@ -13,7 +13,7 @@
 #include "vp8/common/filter.h"
 
 
-unsigned int vp8_get_mb_ss_c
+unsigned int vp9_get_mb_ss_c
 (
   const short *src_ptr
 ) {
@@ -56,7 +56,7 @@ static void variance(
 }
 
 #if CONFIG_SUPERBLOCKS
-unsigned int vp8_variance32x32_c(const unsigned char *src_ptr,
+unsigned int vp9_variance32x32_c(const unsigned char *src_ptr,
                                  int  source_stride,
                                  const unsigned char *ref_ptr,
                                  int  recon_stride,
@@ -70,7 +70,7 @@ unsigned int vp8_variance32x32_c(const unsigned char *src_ptr,
 }
 #endif
 
-unsigned int vp8_variance16x16_c(
+unsigned int vp9_variance16x16_c(
   const unsigned char *src_ptr,
   int  source_stride,
   const unsigned char *ref_ptr,
@@ -85,7 +85,7 @@ unsigned int vp8_variance16x16_c(
   return (var - ((avg * avg) >> 8));
 }
 
-unsigned int vp8_variance8x16_c(
+unsigned int vp9_variance8x16_c(
   const unsigned char *src_ptr,
   int  source_stride,
   const unsigned char *ref_ptr,
@@ -100,7 +100,7 @@ unsigned int vp8_variance8x16_c(
   return (var - ((avg * avg) >> 7));
 }
 
-unsigned int vp8_variance16x8_c(
+unsigned int vp9_variance16x8_c(
   const unsigned char *src_ptr,
   int  source_stride,
   const unsigned char *ref_ptr,
@@ -116,7 +116,7 @@ unsigned int vp8_variance16x8_c(
 }
 
 
-unsigned int vp8_variance8x8_c(
+unsigned int vp9_variance8x8_c(
   const unsigned char *src_ptr,
   int  source_stride,
   const unsigned char *ref_ptr,
@@ -131,7 +131,7 @@ unsigned int vp8_variance8x8_c(
   return (var - ((avg * avg) >> 6));
 }
 
-unsigned int vp8_variance4x4_c(
+unsigned int vp9_variance4x4_c(
   const unsigned char *src_ptr,
   int  source_stride,
   const unsigned char *ref_ptr,
@@ -147,7 +147,7 @@ unsigned int vp8_variance4x4_c(
 }
 
 
-unsigned int vp8_mse16x16_c(
+unsigned int vp9_mse16x16_c(
   const unsigned char *src_ptr,
   int  source_stride,
   const unsigned char *ref_ptr,
@@ -275,7 +275,7 @@ static void var_filter_block2d_bil_second_pass
 }
 
 
-unsigned int vp8_sub_pixel_variance4x4_c
+unsigned int vp9_sub_pixel_variance4x4_c
 (
   const unsigned char  *src_ptr,
   int  src_pixels_per_line,
@@ -298,11 +298,11 @@ unsigned int vp8_sub_pixel_variance4x4_c
   // Now filter Verticaly
   var_filter_block2d_bil_second_pass(FData3, temp2, 4,  4,  4,  4, VFilter);
 
-  return vp8_variance4x4_c(temp2, 4, dst_ptr, dst_pixels_per_line, sse);
+  return vp9_variance4x4_c(temp2, 4, dst_ptr, dst_pixels_per_line, sse);
 }
 
 
-unsigned int vp8_sub_pixel_variance8x8_c
+unsigned int vp9_sub_pixel_variance8x8_c
 (
   const unsigned char  *src_ptr,
   int  src_pixels_per_line,
@@ -322,10 +322,10 @@ unsigned int vp8_sub_pixel_variance8x8_c
   var_filter_block2d_bil_first_pass(src_ptr, FData3, src_pixels_per_line, 1, 9, 8, HFilter);
   var_filter_block2d_bil_second_pass(FData3, temp2, 8, 8, 8, 8, VFilter);
 
-  return vp8_variance8x8_c(temp2, 8, dst_ptr, dst_pixels_per_line, sse);
+  return vp9_variance8x8_c(temp2, 8, dst_ptr, dst_pixels_per_line, sse);
 }
 
-unsigned int vp8_sub_pixel_variance16x16_c
+unsigned int vp9_sub_pixel_variance16x16_c
 (
   const unsigned char  *src_ptr,
   int  src_pixels_per_line,
@@ -345,11 +345,11 @@ unsigned int vp8_sub_pixel_variance16x16_c
   var_filter_block2d_bil_first_pass(src_ptr, FData3, src_pixels_per_line, 1, 17, 16, HFilter);
   var_filter_block2d_bil_second_pass(FData3, temp2, 16, 16, 16, 16, VFilter);
 
-  return vp8_variance16x16_c(temp2, 16, dst_ptr, dst_pixels_per_line, sse);
+  return vp9_variance16x16_c(temp2, 16, dst_ptr, dst_pixels_per_line, sse);
 }
 
 #if CONFIG_SUPERBLOCKS
-unsigned int vp8_sub_pixel_variance32x32_c(const unsigned char  *src_ptr,
+unsigned int vp9_sub_pixel_variance32x32_c(const unsigned char  *src_ptr,
                                            int  src_pixels_per_line,
                                            int  xoffset,
                                            int  yoffset,
@@ -366,75 +366,75 @@ unsigned int vp8_sub_pixel_variance32x32_c(const unsigned char  *src_ptr,
   var_filter_block2d_bil_first_pass(src_ptr, FData3, src_pixels_per_line, 1, 33, 32, HFilter);
   var_filter_block2d_bil_second_pass(FData3, temp2, 32, 32, 32, 32, VFilter);
 
-  return vp8_variance32x32_c(temp2, 32, dst_ptr, dst_pixels_per_line, sse);
+  return vp9_variance32x32_c(temp2, 32, dst_ptr, dst_pixels_per_line, sse);
 }
 #endif
 
-unsigned int vp8_variance_halfpixvar16x16_h_c(
+unsigned int vp9_variance_halfpixvar16x16_h_c(
   const unsigned char *src_ptr,
   int  source_stride,
   const unsigned char *ref_ptr,
   int  recon_stride,
   unsigned int *sse) {
-  return vp8_sub_pixel_variance16x16_c(src_ptr, source_stride, 8, 0,
+  return vp9_sub_pixel_variance16x16_c(src_ptr, source_stride, 8, 0,
                                        ref_ptr, recon_stride, sse);
 }
 
 #if CONFIG_SUPERBLOCKS
-unsigned int vp8_variance_halfpixvar32x32_h_c(const unsigned char *src_ptr,
+unsigned int vp9_variance_halfpixvar32x32_h_c(const unsigned char *src_ptr,
                                               int  source_stride,
                                               const unsigned char *ref_ptr,
                                               int  recon_stride,
                                               unsigned int *sse) {
-  return vp8_sub_pixel_variance32x32_c(src_ptr, source_stride, 8, 0,
+  return vp9_sub_pixel_variance32x32_c(src_ptr, source_stride, 8, 0,
                                        ref_ptr, recon_stride, sse);
 }
 #endif
 
 
-unsigned int vp8_variance_halfpixvar16x16_v_c(const unsigned char *src_ptr,
+unsigned int vp9_variance_halfpixvar16x16_v_c(const unsigned char *src_ptr,
                                               int  source_stride,
                                               const unsigned char *ref_ptr,
                                               int  recon_stride,
                                               unsigned int *sse) {
-  return vp8_sub_pixel_variance16x16_c(src_ptr, source_stride, 0, 8,
+  return vp9_sub_pixel_variance16x16_c(src_ptr, source_stride, 0, 8,
                                        ref_ptr, recon_stride, sse);
 }
 
 #if CONFIG_SUPERBLOCKS
-unsigned int vp8_variance_halfpixvar32x32_v_c(
+unsigned int vp9_variance_halfpixvar32x32_v_c(
   const unsigned char *src_ptr,
   int  source_stride,
   const unsigned char *ref_ptr,
   int  recon_stride,
   unsigned int *sse) {
-  return vp8_sub_pixel_variance32x32_c(src_ptr, source_stride, 0, 8,
+  return vp9_sub_pixel_variance32x32_c(src_ptr, source_stride, 0, 8,
                                        ref_ptr, recon_stride, sse);
 }
 #endif
 
-unsigned int vp8_variance_halfpixvar16x16_hv_c(
+unsigned int vp9_variance_halfpixvar16x16_hv_c(
   const unsigned char *src_ptr,
   int  source_stride,
   const unsigned char *ref_ptr,
   int  recon_stride,
   unsigned int *sse) {
-  return vp8_sub_pixel_variance16x16_c(src_ptr, source_stride, 8, 8,
+  return vp9_sub_pixel_variance16x16_c(src_ptr, source_stride, 8, 8,
                                        ref_ptr, recon_stride, sse);
 }
 
 #if CONFIG_SUPERBLOCKS
-unsigned int vp8_variance_halfpixvar32x32_hv_c(const unsigned char *src_ptr,
+unsigned int vp9_variance_halfpixvar32x32_hv_c(const unsigned char *src_ptr,
                                                int  source_stride,
                                                const unsigned char *ref_ptr,
                                                int  recon_stride,
                                                unsigned int *sse) {
-  return vp8_sub_pixel_variance32x32_c(src_ptr, source_stride, 8, 8,
+  return vp9_sub_pixel_variance32x32_c(src_ptr, source_stride, 8, 8,
                                        ref_ptr, recon_stride, sse);
 }
 #endif
 
-unsigned int vp8_sub_pixel_mse16x16_c
+unsigned int vp9_sub_pixel_mse16x16_c
 (
   const unsigned char  *src_ptr,
   int  src_pixels_per_line,
@@ -444,24 +444,24 @@ unsigned int vp8_sub_pixel_mse16x16_c
   int dst_pixels_per_line,
   unsigned int *sse
 ) {
-  vp8_sub_pixel_variance16x16_c(src_ptr, src_pixels_per_line, xoffset, yoffset, dst_ptr, dst_pixels_per_line, sse);
+  vp9_sub_pixel_variance16x16_c(src_ptr, src_pixels_per_line, xoffset, yoffset, dst_ptr, dst_pixels_per_line, sse);
   return *sse;
 }
 
 #if CONFIG_SUPERBLOCKS
-unsigned int vp8_sub_pixel_mse32x32_c(const unsigned char  *src_ptr,
+unsigned int vp9_sub_pixel_mse32x32_c(const unsigned char  *src_ptr,
                                       int  src_pixels_per_line,
                                       int  xoffset,
                                       int  yoffset,
                                       const unsigned char *dst_ptr,
                                       int dst_pixels_per_line,
                                       unsigned int *sse) {
-  vp8_sub_pixel_variance32x32_c(src_ptr, src_pixels_per_line, xoffset, yoffset, dst_ptr, dst_pixels_per_line, sse);
+  vp9_sub_pixel_variance32x32_c(src_ptr, src_pixels_per_line, xoffset, yoffset, dst_ptr, dst_pixels_per_line, sse);
   return *sse;
 }
 #endif
 
-unsigned int vp8_sub_pixel_variance16x8_c
+unsigned int vp9_sub_pixel_variance16x8_c
 (
   const unsigned char  *src_ptr,
   int  src_pixels_per_line,
@@ -481,10 +481,10 @@ unsigned int vp8_sub_pixel_variance16x8_c
   var_filter_block2d_bil_first_pass(src_ptr, FData3, src_pixels_per_line, 1, 9, 16, HFilter);
   var_filter_block2d_bil_second_pass(FData3, temp2, 16, 16, 8, 16, VFilter);
 
-  return vp8_variance16x8_c(temp2, 16, dst_ptr, dst_pixels_per_line, sse);
+  return vp9_variance16x8_c(temp2, 16, dst_ptr, dst_pixels_per_line, sse);
 }
 
-unsigned int vp8_sub_pixel_variance8x16_c
+unsigned int vp9_sub_pixel_variance8x16_c
 (
   const unsigned char  *src_ptr,
   int  src_pixels_per_line,
@@ -506,7 +506,7 @@ unsigned int vp8_sub_pixel_variance8x16_c
   var_filter_block2d_bil_first_pass(src_ptr, FData3, src_pixels_per_line, 1, 17, 8, HFilter);
   var_filter_block2d_bil_second_pass(FData3, temp2, 8, 8, 16, 8, VFilter);
 
-  return vp8_variance8x16_c(temp2, 8, dst_ptr, dst_pixels_per_line, sse);
+  return vp9_variance8x16_c(temp2, 8, dst_ptr, dst_pixels_per_line, sse);
 }
 #if CONFIG_NEWBESTREFMV
 unsigned int vp8_variance2x16_c(
diff --git a/vp8/encoder/x86/dct_mmx.asm b/vp8/encoder/x86/dct_mmx.asm
index f07b030bde3f463fde05e8d7319325054b2c2eb0..3045466f2f72fce7e400b3fc7fb5b3b40091161b 100644
--- a/vp8/encoder/x86/dct_mmx.asm
+++ b/vp8/encoder/x86/dct_mmx.asm
@@ -11,9 +11,9 @@
 
 %include "vpx_ports/x86_abi_support.asm"
 
-;void vp8_short_fdct4x4_mmx(short *input, short *output, int pitch)
-global sym(vp8_short_fdct4x4_mmx)
-sym(vp8_short_fdct4x4_mmx):
+;void vp9_short_fdct4x4_mmx(short *input, short *output, int pitch)
+global sym(vp9_short_fdct4x4_mmx)
+sym(vp9_short_fdct4x4_mmx):
     push        rbp
     mov         rbp,        rsp
     SHADOW_ARGS_TO_STACK 3
diff --git a/vp8/encoder/x86/dct_sse2.asm b/vp8/encoder/x86/dct_sse2.asm
index 3d52a5d54f306f6e3afa86a1c52a4231ea8593eb..2821fbe352c845d3512d05453d5d22d8b1d3692a 100644
--- a/vp8/encoder/x86/dct_sse2.asm
+++ b/vp8/encoder/x86/dct_sse2.asm
@@ -60,9 +60,9 @@
     ret
 %endmacro
 
-;void vp8_short_fdct4x4_sse2(short *input, short *output, int pitch)
-global sym(vp8_short_fdct4x4_sse2)
-sym(vp8_short_fdct4x4_sse2):
+;void vp9_short_fdct4x4_sse2(short *input, short *output, int pitch)
+global sym(vp9_short_fdct4x4_sse2)
+sym(vp9_short_fdct4x4_sse2):
 
     STACK_FRAME_CREATE
 
@@ -165,9 +165,9 @@ sym(vp8_short_fdct4x4_sse2):
 
     STACK_FRAME_DESTROY
 
-;void vp8_short_fdct8x4_sse2(short *input, short *output, int pitch)
-global sym(vp8_short_fdct8x4_sse2)
-sym(vp8_short_fdct8x4_sse2):
+;void vp9_short_fdct8x4_sse2(short *input, short *output, int pitch)
+global sym(vp9_short_fdct8x4_sse2)
+sym(vp9_short_fdct8x4_sse2):
 
     STACK_FRAME_CREATE
 
diff --git a/vp8/encoder/x86/encodeopt.asm b/vp8/encoder/x86/encodeopt.asm
index 7ec7d603c99fb126863fae38b45dc3e552126e05..9e4cd1102b71d2b87f4e4c2b830c3f7c1da4e06f 100644
--- a/vp8/encoder/x86/encodeopt.asm
+++ b/vp8/encoder/x86/encodeopt.asm
@@ -11,9 +11,9 @@
 
 %include "vpx_ports/x86_abi_support.asm"
 
-;int vp8_block_error_xmm(short *coeff_ptr,  short *dcoef_ptr)
-global sym(vp8_block_error_xmm)
-sym(vp8_block_error_xmm):
+;int vp9_block_error_xmm(short *coeff_ptr,  short *dcoef_ptr)
+global sym(vp9_block_error_xmm)
+sym(vp9_block_error_xmm):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 2
@@ -59,9 +59,9 @@ sym(vp8_block_error_xmm):
     pop         rbp
     ret
 
-;int vp8_block_error_mmx(short *coeff_ptr,  short *dcoef_ptr)
-global sym(vp8_block_error_mmx)
-sym(vp8_block_error_mmx):
+;int vp9_block_error_mmx(short *coeff_ptr,  short *dcoef_ptr)
+global sym(vp9_block_error_mmx)
+sym(vp9_block_error_mmx):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 2
@@ -125,9 +125,9 @@ sym(vp8_block_error_mmx):
     ret
 
 
-;int vp8_mbblock_error_mmx_impl(short *coeff_ptr, short *dcoef_ptr, int dc);
-global sym(vp8_mbblock_error_mmx_impl)
-sym(vp8_mbblock_error_mmx_impl):
+;int vp9_mbblock_error_mmx_impl(short *coeff_ptr, short *dcoef_ptr, int dc);
+global sym(vp9_mbblock_error_mmx_impl)
+sym(vp9_mbblock_error_mmx_impl):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 3
@@ -202,9 +202,9 @@ sym(vp8_mbblock_error_mmx_impl):
     ret
 
 
-;int vp8_mbblock_error_xmm_impl(short *coeff_ptr, short *dcoef_ptr, int dc);
-global sym(vp8_mbblock_error_xmm_impl)
-sym(vp8_mbblock_error_xmm_impl):
+;int vp9_mbblock_error_xmm_impl(short *coeff_ptr, short *dcoef_ptr, int dc);
+global sym(vp9_mbblock_error_xmm_impl)
+sym(vp9_mbblock_error_xmm_impl):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 3
@@ -272,9 +272,9 @@ sym(vp8_mbblock_error_xmm_impl):
     ret
 
 
-;int vp8_mbuverror_mmx_impl(short *s_ptr, short *d_ptr);
-global sym(vp8_mbuverror_mmx_impl)
-sym(vp8_mbuverror_mmx_impl):
+;int vp9_mbuverror_mmx_impl(short *s_ptr, short *d_ptr);
+global sym(vp9_mbuverror_mmx_impl)
+sym(vp9_mbuverror_mmx_impl):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 2
@@ -329,9 +329,9 @@ sym(vp8_mbuverror_mmx_impl):
     ret
 
 
-;int vp8_mbuverror_xmm_impl(short *s_ptr, short *d_ptr);
-global sym(vp8_mbuverror_xmm_impl)
-sym(vp8_mbuverror_xmm_impl):
+;int vp9_mbuverror_xmm_impl(short *s_ptr, short *d_ptr);
+global sym(vp9_mbuverror_xmm_impl)
+sym(vp9_mbuverror_xmm_impl):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 2
diff --git a/vp8/encoder/x86/fwalsh_sse2.asm b/vp8/encoder/x86/fwalsh_sse2.asm
index 71efd5613f601ed06a256f00e7f87b8b717b6ba3..c6b18c1a188a9361b174ff7dcb30ff804589bc72 100644
--- a/vp8/encoder/x86/fwalsh_sse2.asm
+++ b/vp8/encoder/x86/fwalsh_sse2.asm
@@ -11,9 +11,9 @@
 
 %include "vpx_ports/x86_abi_support.asm"
 
-;void vp8_short_walsh4x4_sse2(short *input, short *output, int pitch)
-global sym(vp8_short_walsh4x4_sse2)
-sym(vp8_short_walsh4x4_sse2):
+;void vp9_short_walsh4x4_sse2(short *input, short *output, int pitch)
+global sym(vp9_short_walsh4x4_sse2)
+sym(vp9_short_walsh4x4_sse2):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 3
diff --git a/vp8/encoder/x86/mcomp_x86.h b/vp8/encoder/x86/mcomp_x86.h
index efe7241f2eed280db81ff4cbd8e8770f3227ffa8..ea8d1dcf162e2c8b702f2b60a7cf928ae10eea57 100644
--- a/vp8/encoder/x86/mcomp_x86.h
+++ b/vp8/encoder/x86/mcomp_x86.h
@@ -16,13 +16,13 @@
 #if !CONFIG_RUNTIME_CPU_DETECT
 
 #undef  vp8_search_full_search
-#define vp8_search_full_search vp8_full_search_sadx3
+#define vp8_search_full_search vp9_full_search_sadx3
 
 #undef  vp8_search_refining_search
-#define vp8_search_refining_search vp8_refining_search_sadx4
+#define vp8_search_refining_search vp9_refining_search_sadx4
 
 #undef  vp8_search_diamond_search
-#define vp8_search_diamond_search vp8_diamond_search_sadx4
+#define vp8_search_diamond_search vp9_diamond_search_sadx4
 
 #endif
 #endif
@@ -31,7 +31,7 @@
 #if !CONFIG_RUNTIME_CPU_DETECT
 
 #undef  vp8_search_full_search
-#define vp8_search_full_search vp8_full_search_sadx8
+#define vp8_search_full_search vp9_full_search_sadx8
 
 #endif
 #endif
diff --git a/vp8/encoder/x86/quantize_mmx.asm b/vp8/encoder/x86/quantize_mmx.asm
index f29a54ecd12832d674c62435dd9adefd73d6c48d..050119a3130ffab37f0396882538c7315082048c 100644
--- a/vp8/encoder/x86/quantize_mmx.asm
+++ b/vp8/encoder/x86/quantize_mmx.asm
@@ -11,12 +11,12 @@
 
 %include "vpx_ports/x86_abi_support.asm"
 
-;int vp8_fast_quantize_b_impl_mmx(short *coeff_ptr, short *zbin_ptr,
+;int vp9_fast_quantize_b_impl_mmx(short *coeff_ptr, short *zbin_ptr,
 ;                           short *qcoeff_ptr,short *dequant_ptr,
 ;                           short *scan_mask, short *round_ptr,
 ;                           short *quant_ptr, short *dqcoeff_ptr);
-global sym(vp8_fast_quantize_b_impl_mmx)
-sym(vp8_fast_quantize_b_impl_mmx):
+global sym(vp9_fast_quantize_b_impl_mmx)
+sym(vp9_fast_quantize_b_impl_mmx):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 8
diff --git a/vp8/encoder/x86/quantize_sse2.asm b/vp8/encoder/x86/quantize_sse2.asm
index c483933df1497d779124ad2fbf6e913e016b29fc..b12a5850a81681259e9cc76756d9410cf2d62b9e 100644
--- a/vp8/encoder/x86/quantize_sse2.asm
+++ b/vp8/encoder/x86/quantize_sse2.asm
@@ -12,12 +12,12 @@
 %include "asm_enc_offsets.asm"
 
 
-; void vp8_regular_quantize_b_sse2 | arg
+; void vp9_regular_quantize_b_sse2 | arg
 ;  (BLOCK  *b,                     |  0
 ;   BLOCKD *d)                     |  1
 
-global sym(vp8_regular_quantize_b_sse2)
-sym(vp8_regular_quantize_b_sse2):
+global sym(vp9_regular_quantize_b_sse2)
+sym(vp9_regular_quantize_b_sse2):
     push        rbp
     mov         rbp, rsp
     SAVE_XMM 7
@@ -233,12 +233,12 @@ ZIGZAG_LOOP 15
     pop         rbp
     ret
 
-; void vp8_fast_quantize_b_sse2 | arg
+; void vp9_fast_quantize_b_sse2 | arg
 ;  (BLOCK  *b,                  |  0
 ;   BLOCKD *d)                  |  1
 
-global sym(vp8_fast_quantize_b_sse2)
-sym(vp8_fast_quantize_b_sse2):
+global sym(vp9_fast_quantize_b_sse2)
+sym(vp9_fast_quantize_b_sse2):
     push        rbp
     mov         rbp, rsp
     GET_GOT     rbx
diff --git a/vp8/encoder/x86/quantize_sse4.asm b/vp8/encoder/x86/quantize_sse4.asm
index 95e1c20744f7ede82625e696a34b1927c628dcb2..76f7a9278768dd5e1e38d2176b236b229dca31f7 100644
--- a/vp8/encoder/x86/quantize_sse4.asm
+++ b/vp8/encoder/x86/quantize_sse4.asm
@@ -12,12 +12,12 @@
 %include "asm_enc_offsets.asm"
 
 
-; void vp8_regular_quantize_b_sse4 | arg
+; void vp9_regular_quantize_b_sse4 | arg
 ;  (BLOCK  *b,                     |  0
 ;   BLOCKD *d)                     |  1
 
-global sym(vp8_regular_quantize_b_sse4)
-sym(vp8_regular_quantize_b_sse4):
+global sym(vp9_regular_quantize_b_sse4)
+sym(vp9_regular_quantize_b_sse4):
 
 %if ABI_IS_32BIT
     push        rbp
diff --git a/vp8/encoder/x86/quantize_ssse3.asm b/vp8/encoder/x86/quantize_ssse3.asm
index 912007e02d7682697ce33af594954abb4e3d645e..018161e7c8ee5baa080f5b3c25b3244f90596932 100644
--- a/vp8/encoder/x86/quantize_ssse3.asm
+++ b/vp8/encoder/x86/quantize_ssse3.asm
@@ -12,13 +12,13 @@
 %include "asm_enc_offsets.asm"
 
 
-; void vp8_fast_quantize_b_ssse3 | arg
+; void vp9_fast_quantize_b_ssse3 | arg
 ;  (BLOCK  *b,                   |  0
 ;   BLOCKD *d)                   |  1
 ;
 
-global sym(vp8_fast_quantize_b_ssse3)
-sym(vp8_fast_quantize_b_ssse3):
+global sym(vp9_fast_quantize_b_ssse3)
+sym(vp9_fast_quantize_b_ssse3):
     push        rbp
     mov         rbp, rsp
     GET_GOT     rbx
diff --git a/vp8/encoder/x86/quantize_x86.h b/vp8/encoder/x86/quantize_x86.h
index 449a6bcac039b9b79a93e6ef41d7d38e0706923c..33d3a13c578e584947b640e2d687f1bbbf87d4ef 100644
--- a/vp8/encoder/x86/quantize_x86.h
+++ b/vp8/encoder/x86/quantize_x86.h
@@ -23,23 +23,23 @@
 
 
 #if HAVE_SSE2
-extern prototype_quantize_block(vp8_regular_quantize_b_sse2);
+extern prototype_quantize_block(vp9_regular_quantize_b_sse2);
 #if !CONFIG_RUNTIME_CPU_DETECT
 
 #undef vp8_quantize_quantb
-#define vp8_quantize_quantb vp8_regular_quantize_b_sse2
+#define vp8_quantize_quantb vp9_regular_quantize_b_sse2
 #endif /* !CONFIG_RUNTIME_CPU_DETECT */
 
 #endif /* HAVE_SSE2 */
 
 
 #if HAVE_SSE4_1
-extern prototype_quantize_block(vp8_regular_quantize_b_sse4);
+extern prototype_quantize_block(vp9_regular_quantize_b_sse4);
 
 #if !CONFIG_RUNTIME_CPU_DETECT
 
 #undef vp8_quantize_quantb
-#define vp8_quantize_quantb vp8_regular_quantize_b_sse4
+#define vp8_quantize_quantb vp9_regular_quantize_b_sse4
 
 #endif /* !CONFIG_RUNTIME_CPU_DETECT */
 
diff --git a/vp8/encoder/x86/sad_mmx.asm b/vp8/encoder/x86/sad_mmx.asm
index 407b3997907ca59e2a885717a0323de8ae11281d..827c58cbbd86f7cf24a0a4226ab07de3d205aa64 100644
--- a/vp8/encoder/x86/sad_mmx.asm
+++ b/vp8/encoder/x86/sad_mmx.asm
@@ -11,18 +11,18 @@
 
 %include "vpx_ports/x86_abi_support.asm"
 
-global sym(vp8_sad16x16_mmx)
-global sym(vp8_sad8x16_mmx)
-global sym(vp8_sad8x8_mmx)
-global sym(vp8_sad4x4_mmx)
-global sym(vp8_sad16x8_mmx)
+global sym(vp9_sad16x16_mmx)
+global sym(vp9_sad8x16_mmx)
+global sym(vp9_sad8x8_mmx)
+global sym(vp9_sad4x4_mmx)
+global sym(vp9_sad16x8_mmx)
 
-;unsigned int vp8_sad16x16_mmx(
+;unsigned int vp9_sad16x16_mmx(
 ;    unsigned char *src_ptr,
 ;    int  src_stride,
 ;    unsigned char *ref_ptr,
 ;    int  ref_stride)
-sym(vp8_sad16x16_mmx):
+sym(vp9_sad16x16_mmx):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 4
@@ -109,12 +109,12 @@ sym(vp8_sad16x16_mmx):
     ret
 
 
-;unsigned int vp8_sad8x16_mmx(
+;unsigned int vp9_sad8x16_mmx(
 ;    unsigned char *src_ptr,
 ;    int  src_stride,
 ;    unsigned char *ref_ptr,
 ;    int  ref_stride)
-sym(vp8_sad8x16_mmx):
+sym(vp9_sad8x16_mmx):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 4
@@ -181,12 +181,12 @@ sym(vp8_sad8x16_mmx):
     ret
 
 
-;unsigned int vp8_sad8x8_mmx(
+;unsigned int vp9_sad8x8_mmx(
 ;    unsigned char *src_ptr,
 ;    int  src_stride,
 ;    unsigned char *ref_ptr,
 ;    int  ref_stride)
-sym(vp8_sad8x8_mmx):
+sym(vp9_sad8x8_mmx):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 4
@@ -251,12 +251,12 @@ sym(vp8_sad8x8_mmx):
     ret
 
 
-;unsigned int vp8_sad4x4_mmx(
+;unsigned int vp9_sad4x4_mmx(
 ;    unsigned char *src_ptr,
 ;    int  src_stride,
 ;    unsigned char *ref_ptr,
 ;    int  ref_stride)
-sym(vp8_sad4x4_mmx):
+sym(vp9_sad4x4_mmx):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 4
@@ -340,12 +340,12 @@ sym(vp8_sad4x4_mmx):
     ret
 
 
-;unsigned int vp8_sad16x8_mmx(
+;unsigned int vp9_sad16x8_mmx(
 ;    unsigned char *src_ptr,
 ;    int  src_stride,
 ;    unsigned char *ref_ptr,
 ;    int  ref_stride)
-sym(vp8_sad16x8_mmx):
+sym(vp9_sad16x8_mmx):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 4
diff --git a/vp8/encoder/x86/sad_sse2.asm b/vp8/encoder/x86/sad_sse2.asm
index fa8e3e3f8c31b41fb1b8d1eedd1796c8f0e3c954..fe9fc4d555e3016713cb4bc2fe1297df2cbe7528 100644
--- a/vp8/encoder/x86/sad_sse2.asm
+++ b/vp8/encoder/x86/sad_sse2.asm
@@ -11,13 +11,13 @@
 
 %include "vpx_ports/x86_abi_support.asm"
 
-;unsigned int vp8_sad16x16_wmt(
+;unsigned int vp9_sad16x16_wmt(
 ;    unsigned char *src_ptr,
 ;    int  src_stride,
 ;    unsigned char *ref_ptr,
 ;    int  ref_stride)
-global sym(vp8_sad16x16_wmt)
-sym(vp8_sad16x16_wmt):
+global sym(vp9_sad16x16_wmt)
+sym(vp9_sad16x16_wmt):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 4
@@ -84,14 +84,14 @@ sym(vp8_sad16x16_wmt):
     pop         rbp
     ret
 
-;unsigned int vp8_sad8x16_wmt(
+;unsigned int vp9_sad8x16_wmt(
 ;    unsigned char *src_ptr,
 ;    int  src_stride,
 ;    unsigned char *ref_ptr,
 ;    int  ref_stride,
 ;    int  max_err)
-global sym(vp8_sad8x16_wmt)
-sym(vp8_sad8x16_wmt):
+global sym(vp9_sad8x16_wmt)
+sym(vp9_sad8x16_wmt):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 5
@@ -148,13 +148,13 @@ sym(vp8_sad8x16_wmt):
     ret
 
 
-;unsigned int vp8_sad8x8_wmt(
+;unsigned int vp9_sad8x8_wmt(
 ;    unsigned char *src_ptr,
 ;    int  src_stride,
 ;    unsigned char *ref_ptr,
 ;    int  ref_stride)
-global sym(vp8_sad8x8_wmt)
-sym(vp8_sad8x8_wmt):
+global sym(vp9_sad8x8_wmt)
+sym(vp9_sad8x8_wmt):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 5
@@ -201,13 +201,13 @@ sym(vp8_sad8x8_wmt):
     pop         rbp
     ret
 
-;unsigned int vp8_sad4x4_wmt(
+;unsigned int vp9_sad4x4_wmt(
 ;    unsigned char *src_ptr,
 ;    int  src_stride,
 ;    unsigned char *ref_ptr,
 ;    int  ref_stride)
-global sym(vp8_sad4x4_wmt)
-sym(vp8_sad4x4_wmt):
+global sym(vp9_sad4x4_wmt)
+sym(vp9_sad4x4_wmt):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 4
@@ -256,13 +256,13 @@ sym(vp8_sad4x4_wmt):
     ret
 
 
-;unsigned int vp8_sad16x8_wmt(
+;unsigned int vp9_sad16x8_wmt(
 ;    unsigned char *src_ptr,
 ;    int  src_stride,
 ;    unsigned char *ref_ptr,
 ;    int  ref_stride)
-global sym(vp8_sad16x8_wmt)
-sym(vp8_sad16x8_wmt):
+global sym(vp9_sad16x8_wmt)
+sym(vp9_sad16x8_wmt):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 5
@@ -329,14 +329,14 @@ sym(vp8_sad16x8_wmt):
     pop         rbp
     ret
 
-;void vp8_copy32xn_sse2(
+;void vp9_copy32xn_sse2(
 ;    unsigned char *src_ptr,
 ;    int  src_stride,
 ;    unsigned char *dst_ptr,
 ;    int  dst_stride,
 ;    int height);
-global sym(vp8_copy32xn_sse2)
-sym(vp8_copy32xn_sse2):
+global sym(vp9_copy32xn_sse2)
+sym(vp9_copy32xn_sse2):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 5
diff --git a/vp8/encoder/x86/sad_sse3.asm b/vp8/encoder/x86/sad_sse3.asm
index a2550974cf282bf3d48bb1f0332e07066bbd0e18..e17485e5befc0d222057516964d4e093292275ad 100644
--- a/vp8/encoder/x86/sad_sse3.asm
+++ b/vp8/encoder/x86/sad_sse3.asm
@@ -374,14 +374,14 @@
 
 %endmacro
 
-;void int vp8_sad16x16x3_sse3(
+;void int vp9_sad16x16x3_sse3(
 ;    unsigned char *src_ptr,
 ;    int  src_stride,
 ;    unsigned char *ref_ptr,
 ;    int  ref_stride,
 ;    int  *results)
-global sym(vp8_sad16x16x3_sse3)
-sym(vp8_sad16x16x3_sse3):
+global sym(vp9_sad16x16x3_sse3)
+sym(vp9_sad16x16x3_sse3):
 
     STACK_FRAME_CREATE_X3
 
@@ -416,14 +416,14 @@ sym(vp8_sad16x16x3_sse3):
 
     STACK_FRAME_DESTROY_X3
 
-;void int vp8_sad16x8x3_sse3(
+;void int vp9_sad16x8x3_sse3(
 ;    unsigned char *src_ptr,
 ;    int  src_stride,
 ;    unsigned char *ref_ptr,
 ;    int  ref_stride,
 ;    int  *results)
-global sym(vp8_sad16x8x3_sse3)
-sym(vp8_sad16x8x3_sse3):
+global sym(vp9_sad16x8x3_sse3)
+sym(vp9_sad16x8x3_sse3):
 
     STACK_FRAME_CREATE_X3
 
@@ -454,14 +454,14 @@ sym(vp8_sad16x8x3_sse3):
 
     STACK_FRAME_DESTROY_X3
 
-;void int vp8_sad8x16x3_sse3(
+;void int vp9_sad8x16x3_sse3(
 ;    unsigned char *src_ptr,
 ;    int  src_stride,
 ;    unsigned char *ref_ptr,
 ;    int  ref_stride,
 ;    int  *results)
-global sym(vp8_sad8x16x3_sse3)
-sym(vp8_sad8x16x3_sse3):
+global sym(vp9_sad8x16x3_sse3)
+sym(vp9_sad8x16x3_sse3):
 
     STACK_FRAME_CREATE_X3
 
@@ -483,14 +483,14 @@ sym(vp8_sad8x16x3_sse3):
 
     STACK_FRAME_DESTROY_X3
 
-;void int vp8_sad8x8x3_sse3(
+;void int vp9_sad8x8x3_sse3(
 ;    unsigned char *src_ptr,
 ;    int  src_stride,
 ;    unsigned char *ref_ptr,
 ;    int  ref_stride,
 ;    int  *results)
-global sym(vp8_sad8x8x3_sse3)
-sym(vp8_sad8x8x3_sse3):
+global sym(vp9_sad8x8x3_sse3)
+sym(vp9_sad8x8x3_sse3):
 
     STACK_FRAME_CREATE_X3
 
@@ -508,14 +508,14 @@ sym(vp8_sad8x8x3_sse3):
 
     STACK_FRAME_DESTROY_X3
 
-;void int vp8_sad4x4x3_sse3(
+;void int vp9_sad4x4x3_sse3(
 ;    unsigned char *src_ptr,
 ;    int  src_stride,
 ;    unsigned char *ref_ptr,
 ;    int  ref_stride,
 ;    int  *results)
-global sym(vp8_sad4x4x3_sse3)
-sym(vp8_sad4x4x3_sse3):
+global sym(vp9_sad4x4x3_sse3)
+sym(vp9_sad4x4x3_sse3):
 
     STACK_FRAME_CREATE_X3
 
@@ -582,22 +582,22 @@ sym(vp8_sad4x4x3_sse3):
 
     STACK_FRAME_DESTROY_X3
 
-;unsigned int vp8_sad16x16_sse3(
+;unsigned int vp9_sad16x16_sse3(
 ;    unsigned char *src_ptr,
 ;    int  src_stride,
 ;    unsigned char *ref_ptr,
 ;    int  ref_stride,
 ;    int  max_err)
 ;%define lddqu movdqu
-global sym(vp8_sad16x16_sse3)
-sym(vp8_sad16x16_sse3):
+global sym(vp9_sad16x16_sse3)
+sym(vp9_sad16x16_sse3):
 
     STACK_FRAME_CREATE_X3
 
         mov             end_ptr,    4
         pxor            xmm7,        xmm7
 
-.vp8_sad16x16_sse3_loop:
+.vp9_sad16x16_sse3_loop:
         movdqa          xmm0,       XMMWORD PTR [src_ptr]
         movdqu          xmm1,       XMMWORD PTR [ref_ptr]
         movdqa          xmm2,       XMMWORD PTR [src_ptr+src_stride]
@@ -627,7 +627,7 @@ sym(vp8_sad16x16_sse3):
         paddw           xmm7,        xmm6
 
         sub             end_ptr,     1
-        jne             .vp8_sad16x16_sse3_loop
+        jne             .vp9_sad16x16_sse3_loop
 
         movq            xmm0,       xmm7
         psrldq          xmm7,       8
@@ -636,14 +636,14 @@ sym(vp8_sad16x16_sse3):
 
     STACK_FRAME_DESTROY_X3
 
-;void vp8_copy32xn_sse3(
+;void vp9_copy32xn_sse3(
 ;    unsigned char *src_ptr,
 ;    int  src_stride,
 ;    unsigned char *dst_ptr,
 ;    int  dst_stride,
 ;    int height);
-global sym(vp8_copy32xn_sse3)
-sym(vp8_copy32xn_sse3):
+global sym(vp9_copy32xn_sse3)
+sym(vp9_copy32xn_sse3):
 
     STACK_FRAME_CREATE_X3
 
@@ -697,14 +697,14 @@ sym(vp8_copy32xn_sse3):
 .copy_is_done:
     STACK_FRAME_DESTROY_X3
 
-;void vp8_sad16x16x4d_sse3(
+;void vp9_sad16x16x4d_sse3(
 ;    unsigned char *src_ptr,
 ;    int  src_stride,
 ;    unsigned char *ref_ptr_base,
 ;    int  ref_stride,
 ;    int  *results)
-global sym(vp8_sad16x16x4d_sse3)
-sym(vp8_sad16x16x4d_sse3):
+global sym(vp9_sad16x16x4d_sse3)
+sym(vp9_sad16x16x4d_sse3):
 
     STACK_FRAME_CREATE_X4
 
@@ -748,14 +748,14 @@ sym(vp8_sad16x16x4d_sse3):
 
     STACK_FRAME_DESTROY_X4
 
-;void vp8_sad16x8x4d_sse3(
+;void vp9_sad16x8x4d_sse3(
 ;    unsigned char *src_ptr,
 ;    int  src_stride,
 ;    unsigned char *ref_ptr_base,
 ;    int  ref_stride,
 ;    int  *results)
-global sym(vp8_sad16x8x4d_sse3)
-sym(vp8_sad16x8x4d_sse3):
+global sym(vp9_sad16x8x4d_sse3)
+sym(vp9_sad16x8x4d_sse3):
 
     STACK_FRAME_CREATE_X4
 
@@ -795,14 +795,14 @@ sym(vp8_sad16x8x4d_sse3):
 
     STACK_FRAME_DESTROY_X4
 
-;void int vp8_sad8x16x4d_sse3(
+;void int vp9_sad8x16x4d_sse3(
 ;    unsigned char *src_ptr,
 ;    int  src_stride,
 ;    unsigned char *ref_ptr,
 ;    int  ref_stride,
 ;    int  *results)
-global sym(vp8_sad8x16x4d_sse3)
-sym(vp8_sad8x16x4d_sse3):
+global sym(vp9_sad8x16x4d_sse3)
+sym(vp9_sad8x16x4d_sse3):
 
     STACK_FRAME_CREATE_X4
 
@@ -828,14 +828,14 @@ sym(vp8_sad8x16x4d_sse3):
 
     STACK_FRAME_DESTROY_X4
 
-;void int vp8_sad8x8x4d_sse3(
+;void int vp9_sad8x8x4d_sse3(
 ;    unsigned char *src_ptr,
 ;    int  src_stride,
 ;    unsigned char *ref_ptr,
 ;    int  ref_stride,
 ;    int  *results)
-global sym(vp8_sad8x8x4d_sse3)
-sym(vp8_sad8x8x4d_sse3):
+global sym(vp9_sad8x8x4d_sse3)
+sym(vp9_sad8x8x4d_sse3):
 
     STACK_FRAME_CREATE_X4
 
@@ -857,14 +857,14 @@ sym(vp8_sad8x8x4d_sse3):
 
     STACK_FRAME_DESTROY_X4
 
-;void int vp8_sad4x4x4d_sse3(
+;void int vp9_sad4x4x4d_sse3(
 ;    unsigned char *src_ptr,
 ;    int  src_stride,
 ;    unsigned char *ref_ptr,
 ;    int  ref_stride,
 ;    int  *results)
-global sym(vp8_sad4x4x4d_sse3)
-sym(vp8_sad4x4x4d_sse3):
+global sym(vp9_sad4x4x4d_sse3)
+sym(vp9_sad4x4x4d_sse3):
 
     STACK_FRAME_CREATE_X4
 
diff --git a/vp8/encoder/x86/sad_sse4.asm b/vp8/encoder/x86/sad_sse4.asm
index 03ecec4b34c6cbb0edf4fd4acd5567f6f1d3f33f..25980d62497b3107d4be83994fe72d938316e614 100644
--- a/vp8/encoder/x86/sad_sse4.asm
+++ b/vp8/encoder/x86/sad_sse4.asm
@@ -155,14 +155,14 @@
 %endmacro
 
 
-;void vp8_sad16x16x8_sse4(
+;void vp9_sad16x16x8_sse4(
 ;    const unsigned char *src_ptr,
 ;    int  src_stride,
 ;    const unsigned char *ref_ptr,
 ;    int  ref_stride,
 ;    unsigned short *sad_array);
-global sym(vp8_sad16x16x8_sse4)
-sym(vp8_sad16x16x8_sse4):
+global sym(vp9_sad16x16x8_sse4)
+sym(vp9_sad16x16x8_sse4):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 5
@@ -196,15 +196,15 @@ sym(vp8_sad16x16x8_sse4):
     ret
 
 
-;void vp8_sad16x8x8_sse4(
+;void vp9_sad16x8x8_sse4(
 ;    const unsigned char *src_ptr,
 ;    int  src_stride,
 ;    const unsigned char *ref_ptr,
 ;    int  ref_stride,
 ;    unsigned short *sad_array
 ;);
-global sym(vp8_sad16x8x8_sse4)
-sym(vp8_sad16x8x8_sse4):
+global sym(vp9_sad16x8x8_sse4)
+sym(vp9_sad16x8x8_sse4):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 5
@@ -234,15 +234,15 @@ sym(vp8_sad16x8x8_sse4):
     ret
 
 
-;void vp8_sad8x8x8_sse4(
+;void vp9_sad8x8x8_sse4(
 ;    const unsigned char *src_ptr,
 ;    int  src_stride,
 ;    const unsigned char *ref_ptr,
 ;    int  ref_stride,
 ;    unsigned short *sad_array
 ;);
-global sym(vp8_sad8x8x8_sse4)
-sym(vp8_sad8x8x8_sse4):
+global sym(vp9_sad8x8x8_sse4)
+sym(vp9_sad8x8x8_sse4):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 5
@@ -272,15 +272,15 @@ sym(vp8_sad8x8x8_sse4):
     ret
 
 
-;void vp8_sad8x16x8_sse4(
+;void vp9_sad8x16x8_sse4(
 ;    const unsigned char *src_ptr,
 ;    int  src_stride,
 ;    const unsigned char *ref_ptr,
 ;    int  ref_stride,
 ;    unsigned short *sad_array
 ;);
-global sym(vp8_sad8x16x8_sse4)
-sym(vp8_sad8x16x8_sse4):
+global sym(vp9_sad8x16x8_sse4)
+sym(vp9_sad8x16x8_sse4):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 5
@@ -313,15 +313,15 @@ sym(vp8_sad8x16x8_sse4):
     ret
 
 
-;void vp8_sad4x4x8_c(
+;void vp9_sad4x4x8_c(
 ;    const unsigned char *src_ptr,
 ;    int  src_stride,
 ;    const unsigned char *ref_ptr,
 ;    int  ref_stride,
 ;    unsigned short *sad_array
 ;);
-global sym(vp8_sad4x4x8_sse4)
-sym(vp8_sad4x4x8_sse4):
+global sym(vp9_sad4x4x8_sse4)
+sym(vp9_sad4x4x8_sse4):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 5
diff --git a/vp8/encoder/x86/sad_ssse3.asm b/vp8/encoder/x86/sad_ssse3.asm
index 95b6c89e64832e2635b1636b11889ddb558a5d36..5623d8be46484c45428928d8b8a6f86709c9ba22 100644
--- a/vp8/encoder/x86/sad_ssse3.asm
+++ b/vp8/encoder/x86/sad_ssse3.asm
@@ -146,14 +146,14 @@
 
 %endmacro
 
-;void int vp8_sad16x16x3_ssse3(
+;void int vp9_sad16x16x3_ssse3(
 ;    unsigned char *src_ptr,
 ;    int  src_stride,
 ;    unsigned char *ref_ptr,
 ;    int  ref_stride,
 ;    int  *results)
-global sym(vp8_sad16x16x3_ssse3)
-sym(vp8_sad16x16x3_ssse3):
+global sym(vp9_sad16x16x3_ssse3)
+sym(vp9_sad16x16x3_ssse3):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 5
@@ -169,31 +169,31 @@ sym(vp8_sad16x16x3_ssse3):
         mov             rdx,        0xf
         and             rdx,        rdi
 
-        jmp .vp8_sad16x16x3_ssse3_skiptable
-.vp8_sad16x16x3_ssse3_jumptable:
-        dd .vp8_sad16x16x3_ssse3_aligned_by_0  - .vp8_sad16x16x3_ssse3_do_jump
-        dd .vp8_sad16x16x3_ssse3_aligned_by_1  - .vp8_sad16x16x3_ssse3_do_jump
-        dd .vp8_sad16x16x3_ssse3_aligned_by_2  - .vp8_sad16x16x3_ssse3_do_jump
-        dd .vp8_sad16x16x3_ssse3_aligned_by_3  - .vp8_sad16x16x3_ssse3_do_jump
-        dd .vp8_sad16x16x3_ssse3_aligned_by_4  - .vp8_sad16x16x3_ssse3_do_jump
-        dd .vp8_sad16x16x3_ssse3_aligned_by_5  - .vp8_sad16x16x3_ssse3_do_jump
-        dd .vp8_sad16x16x3_ssse3_aligned_by_6  - .vp8_sad16x16x3_ssse3_do_jump
-        dd .vp8_sad16x16x3_ssse3_aligned_by_7  - .vp8_sad16x16x3_ssse3_do_jump
-        dd .vp8_sad16x16x3_ssse3_aligned_by_8  - .vp8_sad16x16x3_ssse3_do_jump
-        dd .vp8_sad16x16x3_ssse3_aligned_by_9  - .vp8_sad16x16x3_ssse3_do_jump
-        dd .vp8_sad16x16x3_ssse3_aligned_by_10 - .vp8_sad16x16x3_ssse3_do_jump
-        dd .vp8_sad16x16x3_ssse3_aligned_by_11 - .vp8_sad16x16x3_ssse3_do_jump
-        dd .vp8_sad16x16x3_ssse3_aligned_by_12 - .vp8_sad16x16x3_ssse3_do_jump
-        dd .vp8_sad16x16x3_ssse3_aligned_by_13 - .vp8_sad16x16x3_ssse3_do_jump
-        dd .vp8_sad16x16x3_ssse3_aligned_by_14 - .vp8_sad16x16x3_ssse3_do_jump
-        dd .vp8_sad16x16x3_ssse3_aligned_by_15 - .vp8_sad16x16x3_ssse3_do_jump
-.vp8_sad16x16x3_ssse3_skiptable:
-
-        call .vp8_sad16x16x3_ssse3_do_jump
-.vp8_sad16x16x3_ssse3_do_jump:
+        jmp .vp9_sad16x16x3_ssse3_skiptable
+.vp9_sad16x16x3_ssse3_jumptable:
+        dd .vp9_sad16x16x3_ssse3_aligned_by_0  - .vp9_sad16x16x3_ssse3_do_jump
+        dd .vp9_sad16x16x3_ssse3_aligned_by_1  - .vp9_sad16x16x3_ssse3_do_jump
+        dd .vp9_sad16x16x3_ssse3_aligned_by_2  - .vp9_sad16x16x3_ssse3_do_jump
+        dd .vp9_sad16x16x3_ssse3_aligned_by_3  - .vp9_sad16x16x3_ssse3_do_jump
+        dd .vp9_sad16x16x3_ssse3_aligned_by_4  - .vp9_sad16x16x3_ssse3_do_jump
+        dd .vp9_sad16x16x3_ssse3_aligned_by_5  - .vp9_sad16x16x3_ssse3_do_jump
+        dd .vp9_sad16x16x3_ssse3_aligned_by_6  - .vp9_sad16x16x3_ssse3_do_jump
+        dd .vp9_sad16x16x3_ssse3_aligned_by_7  - .vp9_sad16x16x3_ssse3_do_jump
+        dd .vp9_sad16x16x3_ssse3_aligned_by_8  - .vp9_sad16x16x3_ssse3_do_jump
+        dd .vp9_sad16x16x3_ssse3_aligned_by_9  - .vp9_sad16x16x3_ssse3_do_jump
+        dd .vp9_sad16x16x3_ssse3_aligned_by_10 - .vp9_sad16x16x3_ssse3_do_jump
+        dd .vp9_sad16x16x3_ssse3_aligned_by_11 - .vp9_sad16x16x3_ssse3_do_jump
+        dd .vp9_sad16x16x3_ssse3_aligned_by_12 - .vp9_sad16x16x3_ssse3_do_jump
+        dd .vp9_sad16x16x3_ssse3_aligned_by_13 - .vp9_sad16x16x3_ssse3_do_jump
+        dd .vp9_sad16x16x3_ssse3_aligned_by_14 - .vp9_sad16x16x3_ssse3_do_jump
+        dd .vp9_sad16x16x3_ssse3_aligned_by_15 - .vp9_sad16x16x3_ssse3_do_jump
+.vp9_sad16x16x3_ssse3_skiptable:
+
+        call .vp9_sad16x16x3_ssse3_do_jump
+.vp9_sad16x16x3_ssse3_do_jump:
         pop             rcx                         ; get the address of do_jump
-        mov             rax,  .vp8_sad16x16x3_ssse3_jumptable - .vp8_sad16x16x3_ssse3_do_jump
-        add             rax,  rcx  ; get the absolute address of vp8_sad16x16x3_ssse3_jumptable
+        mov             rax,  .vp9_sad16x16x3_ssse3_jumptable - .vp9_sad16x16x3_ssse3_do_jump
+        add             rax,  rcx  ; get the absolute address of vp9_sad16x16x3_ssse3_jumptable
 
         movsxd          rax,  dword [rax + 4*rdx]   ; get the 32 bit offset from the jumptable
         add             rcx,        rax
@@ -203,23 +203,23 @@ sym(vp8_sad16x16x3_ssse3):
 
         jmp             rcx
 
-        PROCESS_16X16X3_OFFSET 0,  .vp8_sad16x16x3_ssse3
-        PROCESS_16X16X3_OFFSET 1,  .vp8_sad16x16x3_ssse3
-        PROCESS_16X16X3_OFFSET 2,  .vp8_sad16x16x3_ssse3
-        PROCESS_16X16X3_OFFSET 3,  .vp8_sad16x16x3_ssse3
-        PROCESS_16X16X3_OFFSET 4,  .vp8_sad16x16x3_ssse3
-        PROCESS_16X16X3_OFFSET 5,  .vp8_sad16x16x3_ssse3
-        PROCESS_16X16X3_OFFSET 6,  .vp8_sad16x16x3_ssse3
-        PROCESS_16X16X3_OFFSET 7,  .vp8_sad16x16x3_ssse3
-        PROCESS_16X16X3_OFFSET 8,  .vp8_sad16x16x3_ssse3
-        PROCESS_16X16X3_OFFSET 9,  .vp8_sad16x16x3_ssse3
-        PROCESS_16X16X3_OFFSET 10, .vp8_sad16x16x3_ssse3
-        PROCESS_16X16X3_OFFSET 11, .vp8_sad16x16x3_ssse3
-        PROCESS_16X16X3_OFFSET 12, .vp8_sad16x16x3_ssse3
-        PROCESS_16X16X3_OFFSET 13, .vp8_sad16x16x3_ssse3
-        PROCESS_16X16X3_OFFSET 14, .vp8_sad16x16x3_ssse3
-
-.vp8_sad16x16x3_ssse3_aligned_by_15:
+        PROCESS_16X16X3_OFFSET 0,  .vp9_sad16x16x3_ssse3
+        PROCESS_16X16X3_OFFSET 1,  .vp9_sad16x16x3_ssse3
+        PROCESS_16X16X3_OFFSET 2,  .vp9_sad16x16x3_ssse3
+        PROCESS_16X16X3_OFFSET 3,  .vp9_sad16x16x3_ssse3
+        PROCESS_16X16X3_OFFSET 4,  .vp9_sad16x16x3_ssse3
+        PROCESS_16X16X3_OFFSET 5,  .vp9_sad16x16x3_ssse3
+        PROCESS_16X16X3_OFFSET 6,  .vp9_sad16x16x3_ssse3
+        PROCESS_16X16X3_OFFSET 7,  .vp9_sad16x16x3_ssse3
+        PROCESS_16X16X3_OFFSET 8,  .vp9_sad16x16x3_ssse3
+        PROCESS_16X16X3_OFFSET 9,  .vp9_sad16x16x3_ssse3
+        PROCESS_16X16X3_OFFSET 10, .vp9_sad16x16x3_ssse3
+        PROCESS_16X16X3_OFFSET 11, .vp9_sad16x16x3_ssse3
+        PROCESS_16X16X3_OFFSET 12, .vp9_sad16x16x3_ssse3
+        PROCESS_16X16X3_OFFSET 13, .vp9_sad16x16x3_ssse3
+        PROCESS_16X16X3_OFFSET 14, .vp9_sad16x16x3_ssse3
+
+.vp9_sad16x16x3_ssse3_aligned_by_15:
         PROCESS_16X2X3 1
         PROCESS_16X2X3 0
         PROCESS_16X2X3 0
@@ -229,7 +229,7 @@ sym(vp8_sad16x16x3_ssse3):
         PROCESS_16X2X3 0
         PROCESS_16X2X3 0
 
-.vp8_sad16x16x3_ssse3_store_off:
+.vp9_sad16x16x3_ssse3_store_off:
         mov             rdi,        arg(4) ;Results
 
         movq            xmm0,       xmm5
@@ -259,14 +259,14 @@ sym(vp8_sad16x16x3_ssse3):
     pop         rbp
     ret
 
-;void int vp8_sad16x8x3_ssse3(
+;void int vp9_sad16x8x3_ssse3(
 ;    unsigned char *src_ptr,
 ;    int  src_stride,
 ;    unsigned char *ref_ptr,
 ;    int  ref_stride,
 ;    int  *results)
-global sym(vp8_sad16x8x3_ssse3)
-sym(vp8_sad16x8x3_ssse3):
+global sym(vp9_sad16x8x3_ssse3)
+sym(vp9_sad16x8x3_ssse3):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 5
@@ -282,31 +282,31 @@ sym(vp8_sad16x8x3_ssse3):
         mov             rdx,        0xf
         and             rdx,        rdi
 
-        jmp .vp8_sad16x8x3_ssse3_skiptable
-.vp8_sad16x8x3_ssse3_jumptable:
-        dd .vp8_sad16x8x3_ssse3_aligned_by_0  - .vp8_sad16x8x3_ssse3_do_jump
-        dd .vp8_sad16x8x3_ssse3_aligned_by_1  - .vp8_sad16x8x3_ssse3_do_jump
-        dd .vp8_sad16x8x3_ssse3_aligned_by_2  - .vp8_sad16x8x3_ssse3_do_jump
-        dd .vp8_sad16x8x3_ssse3_aligned_by_3  - .vp8_sad16x8x3_ssse3_do_jump
-        dd .vp8_sad16x8x3_ssse3_aligned_by_4  - .vp8_sad16x8x3_ssse3_do_jump
-        dd .vp8_sad16x8x3_ssse3_aligned_by_5  - .vp8_sad16x8x3_ssse3_do_jump
-        dd .vp8_sad16x8x3_ssse3_aligned_by_6  - .vp8_sad16x8x3_ssse3_do_jump
-        dd .vp8_sad16x8x3_ssse3_aligned_by_7  - .vp8_sad16x8x3_ssse3_do_jump
-        dd .vp8_sad16x8x3_ssse3_aligned_by_8  - .vp8_sad16x8x3_ssse3_do_jump
-        dd .vp8_sad16x8x3_ssse3_aligned_by_9  - .vp8_sad16x8x3_ssse3_do_jump
-        dd .vp8_sad16x8x3_ssse3_aligned_by_10 - .vp8_sad16x8x3_ssse3_do_jump
-        dd .vp8_sad16x8x3_ssse3_aligned_by_11 - .vp8_sad16x8x3_ssse3_do_jump
-        dd .vp8_sad16x8x3_ssse3_aligned_by_12 - .vp8_sad16x8x3_ssse3_do_jump
-        dd .vp8_sad16x8x3_ssse3_aligned_by_13 - .vp8_sad16x8x3_ssse3_do_jump
-        dd .vp8_sad16x8x3_ssse3_aligned_by_14 - .vp8_sad16x8x3_ssse3_do_jump
-        dd .vp8_sad16x8x3_ssse3_aligned_by_15 - .vp8_sad16x8x3_ssse3_do_jump
-.vp8_sad16x8x3_ssse3_skiptable:
-
-        call .vp8_sad16x8x3_ssse3_do_jump
-.vp8_sad16x8x3_ssse3_do_jump:
+        jmp .vp9_sad16x8x3_ssse3_skiptable
+.vp9_sad16x8x3_ssse3_jumptable:
+        dd .vp9_sad16x8x3_ssse3_aligned_by_0  - .vp9_sad16x8x3_ssse3_do_jump
+        dd .vp9_sad16x8x3_ssse3_aligned_by_1  - .vp9_sad16x8x3_ssse3_do_jump
+        dd .vp9_sad16x8x3_ssse3_aligned_by_2  - .vp9_sad16x8x3_ssse3_do_jump
+        dd .vp9_sad16x8x3_ssse3_aligned_by_3  - .vp9_sad16x8x3_ssse3_do_jump
+        dd .vp9_sad16x8x3_ssse3_aligned_by_4  - .vp9_sad16x8x3_ssse3_do_jump
+        dd .vp9_sad16x8x3_ssse3_aligned_by_5  - .vp9_sad16x8x3_ssse3_do_jump
+        dd .vp9_sad16x8x3_ssse3_aligned_by_6  - .vp9_sad16x8x3_ssse3_do_jump
+        dd .vp9_sad16x8x3_ssse3_aligned_by_7  - .vp9_sad16x8x3_ssse3_do_jump
+        dd .vp9_sad16x8x3_ssse3_aligned_by_8  - .vp9_sad16x8x3_ssse3_do_jump
+        dd .vp9_sad16x8x3_ssse3_aligned_by_9  - .vp9_sad16x8x3_ssse3_do_jump
+        dd .vp9_sad16x8x3_ssse3_aligned_by_10 - .vp9_sad16x8x3_ssse3_do_jump
+        dd .vp9_sad16x8x3_ssse3_aligned_by_11 - .vp9_sad16x8x3_ssse3_do_jump
+        dd .vp9_sad16x8x3_ssse3_aligned_by_12 - .vp9_sad16x8x3_ssse3_do_jump
+        dd .vp9_sad16x8x3_ssse3_aligned_by_13 - .vp9_sad16x8x3_ssse3_do_jump
+        dd .vp9_sad16x8x3_ssse3_aligned_by_14 - .vp9_sad16x8x3_ssse3_do_jump
+        dd .vp9_sad16x8x3_ssse3_aligned_by_15 - .vp9_sad16x8x3_ssse3_do_jump
+.vp9_sad16x8x3_ssse3_skiptable:
+
+        call .vp9_sad16x8x3_ssse3_do_jump
+.vp9_sad16x8x3_ssse3_do_jump:
         pop             rcx                         ; get the address of do_jump
-        mov             rax,  .vp8_sad16x8x3_ssse3_jumptable - .vp8_sad16x8x3_ssse3_do_jump
-        add             rax,  rcx  ; get the absolute address of vp8_sad16x8x3_ssse3_jumptable
+        mov             rax,  .vp9_sad16x8x3_ssse3_jumptable - .vp9_sad16x8x3_ssse3_do_jump
+        add             rax,  rcx  ; get the absolute address of vp9_sad16x8x3_ssse3_jumptable
 
         movsxd          rax,  dword [rax + 4*rdx]   ; get the 32 bit offset from the jumptable
         add             rcx,        rax
@@ -316,30 +316,30 @@ sym(vp8_sad16x8x3_ssse3):
 
         jmp             rcx
 
-        PROCESS_16X8X3_OFFSET 0,  .vp8_sad16x8x3_ssse3
-        PROCESS_16X8X3_OFFSET 1,  .vp8_sad16x8x3_ssse3
-        PROCESS_16X8X3_OFFSET 2,  .vp8_sad16x8x3_ssse3
-        PROCESS_16X8X3_OFFSET 3,  .vp8_sad16x8x3_ssse3
-        PROCESS_16X8X3_OFFSET 4,  .vp8_sad16x8x3_ssse3
-        PROCESS_16X8X3_OFFSET 5,  .vp8_sad16x8x3_ssse3
-        PROCESS_16X8X3_OFFSET 6,  .vp8_sad16x8x3_ssse3
-        PROCESS_16X8X3_OFFSET 7,  .vp8_sad16x8x3_ssse3
-        PROCESS_16X8X3_OFFSET 8,  .vp8_sad16x8x3_ssse3
-        PROCESS_16X8X3_OFFSET 9,  .vp8_sad16x8x3_ssse3
-        PROCESS_16X8X3_OFFSET 10, .vp8_sad16x8x3_ssse3
-        PROCESS_16X8X3_OFFSET 11, .vp8_sad16x8x3_ssse3
-        PROCESS_16X8X3_OFFSET 12, .vp8_sad16x8x3_ssse3
-        PROCESS_16X8X3_OFFSET 13, .vp8_sad16x8x3_ssse3
-        PROCESS_16X8X3_OFFSET 14, .vp8_sad16x8x3_ssse3
-
-.vp8_sad16x8x3_ssse3_aligned_by_15:
+        PROCESS_16X8X3_OFFSET 0,  .vp9_sad16x8x3_ssse3
+        PROCESS_16X8X3_OFFSET 1,  .vp9_sad16x8x3_ssse3
+        PROCESS_16X8X3_OFFSET 2,  .vp9_sad16x8x3_ssse3
+        PROCESS_16X8X3_OFFSET 3,  .vp9_sad16x8x3_ssse3
+        PROCESS_16X8X3_OFFSET 4,  .vp9_sad16x8x3_ssse3
+        PROCESS_16X8X3_OFFSET 5,  .vp9_sad16x8x3_ssse3
+        PROCESS_16X8X3_OFFSET 6,  .vp9_sad16x8x3_ssse3
+        PROCESS_16X8X3_OFFSET 7,  .vp9_sad16x8x3_ssse3
+        PROCESS_16X8X3_OFFSET 8,  .vp9_sad16x8x3_ssse3
+        PROCESS_16X8X3_OFFSET 9,  .vp9_sad16x8x3_ssse3
+        PROCESS_16X8X3_OFFSET 10, .vp9_sad16x8x3_ssse3
+        PROCESS_16X8X3_OFFSET 11, .vp9_sad16x8x3_ssse3
+        PROCESS_16X8X3_OFFSET 12, .vp9_sad16x8x3_ssse3
+        PROCESS_16X8X3_OFFSET 13, .vp9_sad16x8x3_ssse3
+        PROCESS_16X8X3_OFFSET 14, .vp9_sad16x8x3_ssse3
+
+.vp9_sad16x8x3_ssse3_aligned_by_15:
 
         PROCESS_16X2X3 1
         PROCESS_16X2X3 0
         PROCESS_16X2X3 0
         PROCESS_16X2X3 0
 
-.vp8_sad16x8x3_ssse3_store_off:
+.vp9_sad16x8x3_ssse3_store_off:
         mov             rdi,        arg(4) ;Results
 
         movq            xmm0,       xmm5
diff --git a/vp8/encoder/x86/ssim_opt.asm b/vp8/encoder/x86/ssim_opt.asm
index c6db3d1c62ccf61b754e67d6b1373f48f421fbb0..905c263a6b034f31c2fa8f8d486e7c0c876b1033 100644
--- a/vp8/encoder/x86/ssim_opt.asm
+++ b/vp8/encoder/x86/ssim_opt.asm
@@ -61,8 +61,8 @@
 ; or pavgb At this point this is just meant to be first pass for calculating
 ; all the parms needed for 16x16 ssim so we can play with dssim as distortion
 ; in mode selection code.
-global sym(vp8_ssim_parms_16x16_sse2)
-sym(vp8_ssim_parms_16x16_sse2):
+global sym(vp9_ssim_parms_16x16_sse2)
+sym(vp9_ssim_parms_16x16_sse2):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 9
@@ -151,8 +151,8 @@ sym(vp8_ssim_parms_16x16_sse2):
 ; or pavgb At this point this is just meant to be first pass for calculating
 ; all the parms needed for 16x16 ssim so we can play with dssim as distortion
 ; in mode selection code.
-global sym(vp8_ssim_parms_8x8_sse2)
-sym(vp8_ssim_parms_8x8_sse2):
+global sym(vp9_ssim_parms_8x8_sse2)
+sym(vp9_ssim_parms_8x8_sse2):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 9
diff --git a/vp8/encoder/x86/subtract_mmx.asm b/vp8/encoder/x86/subtract_mmx.asm
index 4ce16ce900f3fca67cc39f6ae2952f0378833e38..5b0e249ca69f6c6b3d278fdba907024c737bba75 100644
--- a/vp8/encoder/x86/subtract_mmx.asm
+++ b/vp8/encoder/x86/subtract_mmx.asm
@@ -11,11 +11,11 @@
 
 %include "vpx_ports/x86_abi_support.asm"
 
-;void vp8_subtract_b_mmx_impl(unsigned char *z,  int src_stride,
+;void vp9_subtract_b_mmx_impl(unsigned char *z,  int src_stride,
 ;                            short *diff, unsigned char *Predictor,
 ;                            int pitch);
-global sym(vp8_subtract_b_mmx_impl)
-sym(vp8_subtract_b_mmx_impl):
+global sym(vp9_subtract_b_mmx_impl)
+sym(vp9_subtract_b_mmx_impl):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 5
@@ -73,9 +73,9 @@ sym(vp8_subtract_b_mmx_impl):
     pop         rbp
     ret
 
-;void vp8_subtract_mby_mmx(short *diff, unsigned char *src, unsigned char *pred, int stride)
-global sym(vp8_subtract_mby_mmx)
-sym(vp8_subtract_mby_mmx):
+;void vp9_subtract_mby_mmx(short *diff, unsigned char *src, unsigned char *pred, int stride)
+global sym(vp9_subtract_mby_mmx)
+sym(vp9_subtract_mby_mmx):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 4
@@ -149,9 +149,9 @@ sym(vp8_subtract_mby_mmx):
     ret
 
 
-;void vp8_subtract_mbuv_mmx(short *diff, unsigned char *usrc, unsigned char *vsrc, unsigned char *pred, int stride)
-global sym(vp8_subtract_mbuv_mmx)
-sym(vp8_subtract_mbuv_mmx):
+;void vp9_subtract_mbuv_mmx(short *diff, unsigned char *usrc, unsigned char *vsrc, unsigned char *pred, int stride)
+global sym(vp9_subtract_mbuv_mmx)
+sym(vp9_subtract_mbuv_mmx):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 5
diff --git a/vp8/encoder/x86/subtract_sse2.asm b/vp8/encoder/x86/subtract_sse2.asm
index 3bd1ff678449b55eb85cc08358963cc48622b3ff..f84ed0697e376b37f1645d2041fb57de195aab25 100644
--- a/vp8/encoder/x86/subtract_sse2.asm
+++ b/vp8/encoder/x86/subtract_sse2.asm
@@ -11,11 +11,11 @@
 
 %include "vpx_ports/x86_abi_support.asm"
 
-;void vp8_subtract_b_sse2_impl(unsigned char *z,  int src_stride,
+;void vp9_subtract_b_sse2_impl(unsigned char *z,  int src_stride,
 ;                            short *diff, unsigned char *Predictor,
 ;                            int pitch);
-global sym(vp8_subtract_b_sse2_impl)
-sym(vp8_subtract_b_sse2_impl):
+global sym(vp9_subtract_b_sse2_impl)
+sym(vp9_subtract_b_sse2_impl):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 5
@@ -71,9 +71,9 @@ sym(vp8_subtract_b_sse2_impl):
     ret
 
 
-;void vp8_subtract_mby_sse2(short *diff, unsigned char *src, unsigned char *pred, int stride)
-global sym(vp8_subtract_mby_sse2)
-sym(vp8_subtract_mby_sse2):
+;void vp9_subtract_mby_sse2(short *diff, unsigned char *src, unsigned char *pred, int stride)
+global sym(vp9_subtract_mby_sse2)
+sym(vp9_subtract_mby_sse2):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 4
@@ -145,9 +145,9 @@ sym(vp8_subtract_mby_sse2):
     ret
 
 
-;void vp8_subtract_mbuv_sse2(short *diff, unsigned char *usrc, unsigned char *vsrc, unsigned char *pred, int stride)
-global sym(vp8_subtract_mbuv_sse2)
-sym(vp8_subtract_mbuv_sse2):
+;void vp9_subtract_mbuv_sse2(short *diff, unsigned char *usrc, unsigned char *vsrc, unsigned char *pred, int stride)
+global sym(vp9_subtract_mbuv_sse2)
+sym(vp9_subtract_mbuv_sse2):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 5
diff --git a/vp8/encoder/x86/temporal_filter_apply_sse2.asm b/vp8/encoder/x86/temporal_filter_apply_sse2.asm
index b97c694391f7e1fd4ffd78e0af5e51fbcd46e0e4..60cc80f15e0ce8104793c3f3c16f6bda36aba356 100644
--- a/vp8/encoder/x86/temporal_filter_apply_sse2.asm
+++ b/vp8/encoder/x86/temporal_filter_apply_sse2.asm
@@ -11,7 +11,7 @@
 
 %include "vpx_ports/x86_abi_support.asm"
 
-; void vp8_temporal_filter_apply_sse2 | arg
+; void vp9_temporal_filter_apply_sse2 | arg
 ;  (unsigned char  *frame1,           |  0
 ;   unsigned int    stride,           |  1
 ;   unsigned char  *frame2,           |  2
@@ -20,8 +20,8 @@
 ;   int             filter_weight,    |  5
 ;   unsigned int   *accumulator,      |  6
 ;   unsigned short *count)            |  7
-global sym(vp8_temporal_filter_apply_sse2)
-sym(vp8_temporal_filter_apply_sse2):
+global sym(vp9_temporal_filter_apply_sse2)
+sym(vp9_temporal_filter_apply_sse2):
 
     push        rbp
     mov         rbp, rsp
diff --git a/vp8/encoder/x86/temporal_filter_x86.h b/vp8/encoder/x86/temporal_filter_x86.h
index 2daa1401828b5be21f329b385cecc09a74e593cb..c72c27ba1f499c5c9b49157f5ae23b544120ca13 100644
--- a/vp8/encoder/x86/temporal_filter_x86.h
+++ b/vp8/encoder/x86/temporal_filter_x86.h
@@ -13,12 +13,12 @@
 #define __INC_VP8_TEMPORAL_FILTER_X86_H
 
 #if HAVE_SSE2
-extern prototype_apply(vp8_temporal_filter_apply_sse2);
+extern prototype_apply(vp9_temporal_filter_apply_sse2);
 
 #if !CONFIG_RUNTIME_CPU_DETECT
 
-#undef  vp8_temporal_filter_apply
-#define vp8_temporal_filter_apply vp8_temporal_filter_apply_sse2
+#undef  vp9_temporal_filter_apply
+#define vp9_temporal_filter_apply vp9_temporal_filter_apply_sse2
 
 #endif
 
diff --git a/vp8/encoder/x86/variance_impl_mmx.asm b/vp8/encoder/x86/variance_impl_mmx.asm
index 2be8bbeb3dc2152d521aae20274eae43aca5b5f8..45c30b089e8c0d57dda376ad42b1c997eb76faaa 100644
--- a/vp8/encoder/x86/variance_impl_mmx.asm
+++ b/vp8/encoder/x86/variance_impl_mmx.asm
@@ -11,9 +11,9 @@
 
 %include "vpx_ports/x86_abi_support.asm"
 
-;unsigned int vp8_get_mb_ss_mmx( short *src_ptr )
-global sym(vp8_get_mb_ss_mmx)
-sym(vp8_get_mb_ss_mmx):
+;unsigned int vp9_get_mb_ss_mmx( short *src_ptr )
+global sym(vp9_get_mb_ss_mmx)
+sym(vp9_get_mb_ss_mmx):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 7
@@ -63,7 +63,7 @@ sym(vp8_get_mb_ss_mmx):
     ret
 
 
-;unsigned int vp8_get8x8var_mmx
+;unsigned int vp9_get8x8var_mmx
 ;(
 ;    unsigned char *src_ptr,
 ;    int  source_stride,
@@ -72,8 +72,8 @@ sym(vp8_get_mb_ss_mmx):
 ;    unsigned int *SSE,
 ;    int *Sum
 ;)
-global sym(vp8_get8x8var_mmx)
-sym(vp8_get8x8var_mmx):
+global sym(vp9_get8x8var_mmx)
+sym(vp9_get8x8var_mmx):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 6
@@ -311,7 +311,7 @@ sym(vp8_get8x8var_mmx):
 
 
 ;unsigned int
-;vp8_get4x4var_mmx
+;vp9_get4x4var_mmx
 ;(
 ;    unsigned char *src_ptr,
 ;    int  source_stride,
@@ -320,8 +320,8 @@ sym(vp8_get8x8var_mmx):
 ;    unsigned int *SSE,
 ;    int *Sum
 ;)
-global sym(vp8_get4x4var_mmx)
-sym(vp8_get4x4var_mmx):
+global sym(vp9_get4x4var_mmx)
+sym(vp9_get4x4var_mmx):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 6
@@ -426,15 +426,15 @@ sym(vp8_get4x4var_mmx):
 
 
 ;unsigned int
-;vp8_get4x4sse_cs_mmx
+;vp9_get4x4sse_cs_mmx
 ;(
 ;    unsigned char *src_ptr,
 ;    int  source_stride,
 ;    unsigned char *ref_ptr,
 ;    int  recon_stride
 ;)
-global sym(vp8_get4x4sse_cs_mmx)
-sym(vp8_get4x4sse_cs_mmx):
+global sym(vp9_get4x4sse_cs_mmx)
+sym(vp9_get4x4sse_cs_mmx):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 4
@@ -511,7 +511,7 @@ sym(vp8_get4x4sse_cs_mmx):
 
 %define mmx_filter_shift            7
 
-;void vp8_filter_block2d_bil4x4_var_mmx
+;void vp9_filter_block2d_bil4x4_var_mmx
 ;(
 ;    unsigned char *ref_ptr,
 ;    int ref_pixels_per_line,
@@ -522,8 +522,8 @@ sym(vp8_get4x4sse_cs_mmx):
 ;    int *sum,
 ;    unsigned int *sumsquared
 ;)
-global sym(vp8_filter_block2d_bil4x4_var_mmx)
-sym(vp8_filter_block2d_bil4x4_var_mmx):
+global sym(vp9_filter_block2d_bil4x4_var_mmx)
+sym(vp9_filter_block2d_bil4x4_var_mmx):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 8
@@ -655,7 +655,7 @@ sym(vp8_filter_block2d_bil4x4_var_mmx):
 
 
 
-;void vp8_filter_block2d_bil_var_mmx
+;void vp9_filter_block2d_bil_var_mmx
 ;(
 ;    unsigned char *ref_ptr,
 ;    int ref_pixels_per_line,
@@ -667,8 +667,8 @@ sym(vp8_filter_block2d_bil4x4_var_mmx):
 ;    int *sum,
 ;    unsigned int *sumsquared
 ;)
-global sym(vp8_filter_block2d_bil_var_mmx)
-sym(vp8_filter_block2d_bil_var_mmx):
+global sym(vp9_filter_block2d_bil_var_mmx)
+sym(vp9_filter_block2d_bil_var_mmx):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 9
diff --git a/vp8/encoder/x86/variance_impl_sse2.asm b/vp8/encoder/x86/variance_impl_sse2.asm
index 823c2f7bf042fa602e635abdf3f3f3cd3087aa3e..349687c9e31a51ba4c2043a5cab0e18df2c5842e 100644
--- a/vp8/encoder/x86/variance_impl_sse2.asm
+++ b/vp8/encoder/x86/variance_impl_sse2.asm
@@ -13,12 +13,12 @@
 
 %define xmm_filter_shift            7
 
-;unsigned int vp8_get_mb_ss_sse2
+;unsigned int vp9_get_mb_ss_sse2
 ;(
 ;    short *src_ptr
 ;)
-global sym(vp8_get_mb_ss_sse2)
-sym(vp8_get_mb_ss_sse2):
+global sym(vp9_get_mb_ss_sse2)
+sym(vp9_get_mb_ss_sse2):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 1
@@ -71,7 +71,7 @@ sym(vp8_get_mb_ss_sse2):
     ret
 
 
-;unsigned int vp8_get16x16var_sse2
+;unsigned int vp9_get16x16var_sse2
 ;(
 ;    unsigned char   *  src_ptr,
 ;    int             source_stride,
@@ -80,8 +80,8 @@ sym(vp8_get_mb_ss_sse2):
 ;    unsigned int    *  SSE,
 ;    int             *  Sum
 ;)
-global sym(vp8_get16x16var_sse2)
-sym(vp8_get16x16var_sse2):
+global sym(vp9_get16x16var_sse2)
+sym(vp9_get16x16var_sse2):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 6
@@ -215,7 +215,7 @@ sym(vp8_get16x16var_sse2):
 
 
 
-;unsigned int vp8_get8x8var_sse2
+;unsigned int vp9_get8x8var_sse2
 ;(
 ;    unsigned char   *  src_ptr,
 ;    int             source_stride,
@@ -224,8 +224,8 @@ sym(vp8_get16x16var_sse2):
 ;    unsigned int    *  SSE,
 ;    int             *  Sum
 ;)
-global sym(vp8_get8x8var_sse2)
-sym(vp8_get8x8var_sse2):
+global sym(vp9_get8x8var_sse2)
+sym(vp9_get8x8var_sse2):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 6
@@ -400,7 +400,7 @@ sym(vp8_get8x8var_sse2):
     pop         rbp
     ret
 
-;void vp8_filter_block2d_bil_var_sse2
+;void vp9_filter_block2d_bil_var_sse2
 ;(
 ;    unsigned char *ref_ptr,
 ;    int ref_pixels_per_line,
@@ -413,8 +413,8 @@ sym(vp8_get8x8var_sse2):
 ;    unsigned int *sumsquared;;
 ;
 ;)
-global sym(vp8_filter_block2d_bil_var_sse2)
-sym(vp8_filter_block2d_bil_var_sse2):
+global sym(vp9_filter_block2d_bil_var_sse2)
+sym(vp9_filter_block2d_bil_var_sse2):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 9
@@ -680,7 +680,7 @@ filter_block2d_bil_variance:
     ret
 
 
-;void vp8_half_horiz_vert_variance8x_h_sse2
+;void vp9_half_horiz_vert_variance8x_h_sse2
 ;(
 ;    unsigned char *ref_ptr,
 ;    int ref_pixels_per_line,
@@ -690,8 +690,8 @@ filter_block2d_bil_variance:
 ;    int *sum,
 ;    unsigned int *sumsquared
 ;)
-global sym(vp8_half_horiz_vert_variance8x_h_sse2)
-sym(vp8_half_horiz_vert_variance8x_h_sse2):
+global sym(vp9_half_horiz_vert_variance8x_h_sse2)
+sym(vp9_half_horiz_vert_variance8x_h_sse2):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 7
@@ -802,7 +802,7 @@ vp8_half_horiz_vert_variance8x_h_1:
     pop         rbp
     ret
 
-;void vp8_half_horiz_vert_variance16x_h_sse2
+;void vp9_half_horiz_vert_variance16x_h_sse2
 ;(
 ;    unsigned char *ref_ptr,
 ;    int ref_pixels_per_line,
@@ -812,8 +812,8 @@ vp8_half_horiz_vert_variance8x_h_1:
 ;    int *sum,
 ;    unsigned int *sumsquared
 ;)
-global sym(vp8_half_horiz_vert_variance16x_h_sse2)
-sym(vp8_half_horiz_vert_variance16x_h_sse2):
+global sym(vp9_half_horiz_vert_variance16x_h_sse2)
+sym(vp9_half_horiz_vert_variance16x_h_sse2):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 7
@@ -918,7 +918,7 @@ vp8_half_horiz_vert_variance16x_h_1:
     ret
 
 
-;void vp8_half_vert_variance8x_h_sse2
+;void vp9_half_vert_variance8x_h_sse2
 ;(
 ;    unsigned char *ref_ptr,
 ;    int ref_pixels_per_line,
@@ -928,8 +928,8 @@ vp8_half_horiz_vert_variance16x_h_1:
 ;    int *sum,
 ;    unsigned int *sumsquared
 ;)
-global sym(vp8_half_vert_variance8x_h_sse2)
-sym(vp8_half_vert_variance8x_h_sse2):
+global sym(vp9_half_vert_variance8x_h_sse2)
+sym(vp9_half_vert_variance8x_h_sse2):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 7
@@ -1025,7 +1025,7 @@ vp8_half_vert_variance8x_h_1:
     pop         rbp
     ret
 
-;void vp8_half_vert_variance16x_h_sse2
+;void vp9_half_vert_variance16x_h_sse2
 ;(
 ;    unsigned char *ref_ptr,
 ;    int ref_pixels_per_line,
@@ -1035,8 +1035,8 @@ vp8_half_vert_variance8x_h_1:
 ;    int *sum,
 ;    unsigned int *sumsquared
 ;)
-global sym(vp8_half_vert_variance16x_h_sse2)
-sym(vp8_half_vert_variance16x_h_sse2):
+global sym(vp9_half_vert_variance16x_h_sse2)
+sym(vp9_half_vert_variance16x_h_sse2):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 7
@@ -1133,7 +1133,7 @@ vp8_half_vert_variance16x_h_1:
     ret
 
 
-;void vp8_half_horiz_variance8x_h_sse2
+;void vp9_half_horiz_variance8x_h_sse2
 ;(
 ;    unsigned char *ref_ptr,
 ;    int ref_pixels_per_line,
@@ -1143,8 +1143,8 @@ vp8_half_vert_variance16x_h_1:
 ;    int *sum,
 ;    unsigned int *sumsquared
 ;)
-global sym(vp8_half_horiz_variance8x_h_sse2)
-sym(vp8_half_horiz_variance8x_h_sse2):
+global sym(vp9_half_horiz_variance8x_h_sse2)
+sym(vp9_half_horiz_variance8x_h_sse2):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 7
@@ -1238,7 +1238,7 @@ vp8_half_horiz_variance8x_h_1:
     pop         rbp
     ret
 
-;void vp8_half_horiz_variance16x_h_sse2
+;void vp9_half_horiz_variance16x_h_sse2
 ;(
 ;    unsigned char *ref_ptr,
 ;    int ref_pixels_per_line,
@@ -1248,8 +1248,8 @@ vp8_half_horiz_variance8x_h_1:
 ;    int *sum,
 ;    unsigned int *sumsquared
 ;)
-global sym(vp8_half_horiz_variance16x_h_sse2)
-sym(vp8_half_horiz_variance16x_h_sse2):
+global sym(vp9_half_horiz_variance16x_h_sse2)
+sym(vp9_half_horiz_variance16x_h_sse2):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 7
diff --git a/vp8/encoder/x86/variance_impl_ssse3.asm b/vp8/encoder/x86/variance_impl_ssse3.asm
index 33ba51a9164fe567c358ec1dbf29ea7bb1ae2781..ca1a5719667bc55fbbc90e56e91bed94c7c486a5 100644
--- a/vp8/encoder/x86/variance_impl_ssse3.asm
+++ b/vp8/encoder/x86/variance_impl_ssse3.asm
@@ -14,7 +14,7 @@
 %define xmm_filter_shift            7
 
 
-;void vp8_filter_block2d_bil_var_ssse3
+;void vp9_filter_block2d_bil_var_ssse3
 ;(
 ;    unsigned char *ref_ptr,
 ;    int ref_pixels_per_line,
@@ -29,8 +29,8 @@
 ;)
 ;Note: The filter coefficient at offset=0 is 128. Since the second register
 ;for Pmaddubsw is signed bytes, we must calculate zero offset seperately.
-global sym(vp8_filter_block2d_bil_var_ssse3)
-sym(vp8_filter_block2d_bil_var_ssse3):
+global sym(vp9_filter_block2d_bil_var_ssse3)
+sym(vp9_filter_block2d_bil_var_ssse3):
     push        rbp
     mov         rbp, rsp
     SHADOW_ARGS_TO_STACK 9
diff --git a/vp8/encoder/x86/variance_mmx.c b/vp8/encoder/x86/variance_mmx.c
index 7e63acf9b4c932f35c7652455d047cbdf0f9e0a2..58aae066fa0b0830e96273b8d5879d977d634cd8 100644
--- a/vp8/encoder/x86/variance_mmx.c
+++ b/vp8/encoder/x86/variance_mmx.c
@@ -34,8 +34,8 @@ extern void filter_block1d_v6_mmx
   short *vp7_filter
 );
 
-extern unsigned int vp8_get_mb_ss_mmx(const short *src_ptr);
-extern unsigned int vp8_get8x8var_mmx
+extern unsigned int vp9_get_mb_ss_mmx(const short *src_ptr);
+extern unsigned int vp9_get8x8var_mmx
 (
   const unsigned char *src_ptr,
   int  source_stride,
@@ -44,7 +44,7 @@ extern unsigned int vp8_get8x8var_mmx
   unsigned int *SSE,
   int *Sum
 );
-extern unsigned int vp8_get4x4var_mmx
+extern unsigned int vp9_get4x4var_mmx
 (
   const unsigned char *src_ptr,
   int  source_stride,
@@ -53,7 +53,7 @@ extern unsigned int vp8_get4x4var_mmx
   unsigned int *SSE,
   int *Sum
 );
-extern void vp8_filter_block2d_bil4x4_var_mmx
+extern void vp9_filter_block2d_bil4x4_var_mmx
 (
   const unsigned char *ref_ptr,
   int ref_pixels_per_line,
@@ -64,7 +64,7 @@ extern void vp8_filter_block2d_bil4x4_var_mmx
   int *sum,
   unsigned int *sumsquared
 );
-extern void vp8_filter_block2d_bil_var_mmx
+extern void vp9_filter_block2d_bil_var_mmx
 (
   const unsigned char *ref_ptr,
   int ref_pixels_per_line,
@@ -78,7 +78,7 @@ extern void vp8_filter_block2d_bil_var_mmx
 );
 
 
-unsigned int vp8_variance4x4_mmx(
+unsigned int vp9_variance4x4_mmx(
   const unsigned char *src_ptr,
   int  source_stride,
   const unsigned char *ref_ptr,
@@ -87,13 +87,13 @@ unsigned int vp8_variance4x4_mmx(
   unsigned int var;
   int avg;
 
-  vp8_get4x4var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &var, &avg);
+  vp9_get4x4var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &var, &avg);
   *sse = var;
   return (var - ((avg * avg) >> 4));
 
 }
 
-unsigned int vp8_variance8x8_mmx(
+unsigned int vp9_variance8x8_mmx(
   const unsigned char *src_ptr,
   int  source_stride,
   const unsigned char *ref_ptr,
@@ -102,14 +102,14 @@ unsigned int vp8_variance8x8_mmx(
   unsigned int var;
   int avg;
 
-  vp8_get8x8var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &var, &avg);
+  vp9_get8x8var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &var, &avg);
   *sse = var;
 
   return (var - ((avg * avg) >> 6));
 
 }
 
-unsigned int vp8_mse16x16_mmx(
+unsigned int vp9_mse16x16_mmx(
   const unsigned char *src_ptr,
   int  source_stride,
   const unsigned char *ref_ptr,
@@ -119,10 +119,10 @@ unsigned int vp8_mse16x16_mmx(
   int sum0, sum1, sum2, sum3;
 
 
-  vp8_get8x8var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &sse0, &sum0);
-  vp8_get8x8var_mmx(src_ptr + 8, source_stride, ref_ptr + 8, recon_stride, &sse1, &sum1);
-  vp8_get8x8var_mmx(src_ptr + 8 * source_stride, source_stride, ref_ptr + 8 * recon_stride, recon_stride, &sse2, &sum2);
-  vp8_get8x8var_mmx(src_ptr + 8 * source_stride + 8, source_stride, ref_ptr + 8 * recon_stride + 8, recon_stride, &sse3, &sum3);
+  vp9_get8x8var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &sse0, &sum0);
+  vp9_get8x8var_mmx(src_ptr + 8, source_stride, ref_ptr + 8, recon_stride, &sse1, &sum1);
+  vp9_get8x8var_mmx(src_ptr + 8 * source_stride, source_stride, ref_ptr + 8 * recon_stride, recon_stride, &sse2, &sum2);
+  vp9_get8x8var_mmx(src_ptr + 8 * source_stride + 8, source_stride, ref_ptr + 8 * recon_stride + 8, recon_stride, &sse3, &sum3);
 
   var = sse0 + sse1 + sse2 + sse3;
   *sse = var;
@@ -130,7 +130,7 @@ unsigned int vp8_mse16x16_mmx(
 }
 
 
-unsigned int vp8_variance16x16_mmx(
+unsigned int vp9_variance16x16_mmx(
   const unsigned char *src_ptr,
   int  source_stride,
   const unsigned char *ref_ptr,
@@ -140,10 +140,10 @@ unsigned int vp8_variance16x16_mmx(
   int sum0, sum1, sum2, sum3, avg;
 
 
-  vp8_get8x8var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &sse0, &sum0);
-  vp8_get8x8var_mmx(src_ptr + 8, source_stride, ref_ptr + 8, recon_stride, &sse1, &sum1);
-  vp8_get8x8var_mmx(src_ptr + 8 * source_stride, source_stride, ref_ptr + 8 * recon_stride, recon_stride, &sse2, &sum2);
-  vp8_get8x8var_mmx(src_ptr + 8 * source_stride + 8, source_stride, ref_ptr + 8 * recon_stride + 8, recon_stride, &sse3, &sum3);
+  vp9_get8x8var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &sse0, &sum0);
+  vp9_get8x8var_mmx(src_ptr + 8, source_stride, ref_ptr + 8, recon_stride, &sse1, &sum1);
+  vp9_get8x8var_mmx(src_ptr + 8 * source_stride, source_stride, ref_ptr + 8 * recon_stride, recon_stride, &sse2, &sum2);
+  vp9_get8x8var_mmx(src_ptr + 8 * source_stride + 8, source_stride, ref_ptr + 8 * recon_stride + 8, recon_stride, &sse3, &sum3);
 
   var = sse0 + sse1 + sse2 + sse3;
   avg = sum0 + sum1 + sum2 + sum3;
@@ -151,7 +151,7 @@ unsigned int vp8_variance16x16_mmx(
   return (var - ((avg * avg) >> 8));
 }
 
-unsigned int vp8_variance16x8_mmx(
+unsigned int vp9_variance16x8_mmx(
   const unsigned char *src_ptr,
   int  source_stride,
   const unsigned char *ref_ptr,
@@ -160,8 +160,8 @@ unsigned int vp8_variance16x8_mmx(
   unsigned int sse0, sse1, var;
   int sum0, sum1, avg;
 
-  vp8_get8x8var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &sse0, &sum0);
-  vp8_get8x8var_mmx(src_ptr + 8, source_stride, ref_ptr + 8, recon_stride, &sse1, &sum1);
+  vp9_get8x8var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &sse0, &sum0);
+  vp9_get8x8var_mmx(src_ptr + 8, source_stride, ref_ptr + 8, recon_stride, &sse1, &sum1);
 
   var = sse0 + sse1;
   avg = sum0 + sum1;
@@ -171,7 +171,7 @@ unsigned int vp8_variance16x8_mmx(
 }
 
 
-unsigned int vp8_variance8x16_mmx(
+unsigned int vp9_variance8x16_mmx(
   const unsigned char *src_ptr,
   int  source_stride,
   const unsigned char *ref_ptr,
@@ -180,8 +180,8 @@ unsigned int vp8_variance8x16_mmx(
   unsigned int sse0, sse1, var;
   int sum0, sum1, avg;
 
-  vp8_get8x8var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &sse0, &sum0);
-  vp8_get8x8var_mmx(src_ptr + 8 * source_stride, source_stride, ref_ptr + 8 * recon_stride, recon_stride, &sse1, &sum1);
+  vp9_get8x8var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &sse0, &sum0);
+  vp9_get8x8var_mmx(src_ptr + 8 * source_stride, source_stride, ref_ptr + 8 * recon_stride, recon_stride, &sse1, &sum1);
 
   var = sse0 + sse1;
   avg = sum0 + sum1;
@@ -217,7 +217,7 @@ DECLARE_ALIGNED(16, const short, vp8_vp7_bilinear_filters_mmx[16][8]) = {
   {   8,  8,  8,  8, 120, 120, 120, 120 }
 };
 
-unsigned int vp8_sub_pixel_variance4x4_mmx
+unsigned int vp9_sub_pixel_variance4x4_mmx
 (
   const unsigned char  *src_ptr,
   int  src_pixels_per_line,
@@ -230,7 +230,7 @@ unsigned int vp8_sub_pixel_variance4x4_mmx
 {
   int xsum;
   unsigned int xxsum;
-  vp8_filter_block2d_bil4x4_var_mmx(
+  vp9_filter_block2d_bil4x4_var_mmx(
     src_ptr, src_pixels_per_line,
     dst_ptr, dst_pixels_per_line,
     vp8_vp7_bilinear_filters_mmx[xoffset], vp8_vp7_bilinear_filters_mmx[yoffset],
@@ -241,7 +241,7 @@ unsigned int vp8_sub_pixel_variance4x4_mmx
 }
 
 
-unsigned int vp8_sub_pixel_variance8x8_mmx
+unsigned int vp9_sub_pixel_variance8x8_mmx
 (
   const unsigned char  *src_ptr,
   int  src_pixels_per_line,
@@ -254,7 +254,7 @@ unsigned int vp8_sub_pixel_variance8x8_mmx
 
   int xsum;
   unsigned int xxsum;
-  vp8_filter_block2d_bil_var_mmx(
+  vp9_filter_block2d_bil_var_mmx(
     src_ptr, src_pixels_per_line,
     dst_ptr, dst_pixels_per_line, 8,
     vp8_vp7_bilinear_filters_mmx[xoffset], vp8_vp7_bilinear_filters_mmx[yoffset],
@@ -264,7 +264,7 @@ unsigned int vp8_sub_pixel_variance8x8_mmx
   return (xxsum - ((xsum * xsum) >> 6));
 }
 
-unsigned int vp8_sub_pixel_variance16x16_mmx
+unsigned int vp9_sub_pixel_variance16x16_mmx
 (
   const unsigned char  *src_ptr,
   int  src_pixels_per_line,
@@ -278,14 +278,14 @@ unsigned int vp8_sub_pixel_variance16x16_mmx
   int xsum0, xsum1;
   unsigned int xxsum0, xxsum1;
 
-  vp8_filter_block2d_bil_var_mmx(
+  vp9_filter_block2d_bil_var_mmx(
     src_ptr, src_pixels_per_line,
     dst_ptr, dst_pixels_per_line, 16,
     vp8_vp7_bilinear_filters_mmx[xoffset], vp8_vp7_bilinear_filters_mmx[yoffset],
     &xsum0, &xxsum0
   );
 
-  vp8_filter_block2d_bil_var_mmx(
+  vp9_filter_block2d_bil_var_mmx(
     src_ptr + 8, src_pixels_per_line,
     dst_ptr + 8, dst_pixels_per_line, 16,
     vp8_vp7_bilinear_filters_mmx[xoffset], vp8_vp7_bilinear_filters_mmx[yoffset],
@@ -301,7 +301,7 @@ unsigned int vp8_sub_pixel_variance16x16_mmx
 
 }
 
-unsigned int vp8_sub_pixel_mse16x16_mmx(
+unsigned int vp9_sub_pixel_mse16x16_mmx(
   const unsigned char  *src_ptr,
   int  src_pixels_per_line,
   int  xoffset,
@@ -310,11 +310,11 @@ unsigned int vp8_sub_pixel_mse16x16_mmx(
   int dst_pixels_per_line,
   unsigned int *sse
 ) {
-  vp8_sub_pixel_variance16x16_mmx(src_ptr, src_pixels_per_line, xoffset, yoffset, dst_ptr, dst_pixels_per_line, sse);
+  vp9_sub_pixel_variance16x16_mmx(src_ptr, src_pixels_per_line, xoffset, yoffset, dst_ptr, dst_pixels_per_line, sse);
   return *sse;
 }
 
-unsigned int vp8_sub_pixel_variance16x8_mmx
+unsigned int vp9_sub_pixel_variance16x8_mmx
 (
   const unsigned char  *src_ptr,
   int  src_pixels_per_line,
@@ -328,7 +328,7 @@ unsigned int vp8_sub_pixel_variance16x8_mmx
   unsigned int xxsum0, xxsum1;
 
 
-  vp8_filter_block2d_bil_var_mmx(
+  vp9_filter_block2d_bil_var_mmx(
     src_ptr, src_pixels_per_line,
     dst_ptr, dst_pixels_per_line, 8,
     vp8_vp7_bilinear_filters_mmx[xoffset], vp8_vp7_bilinear_filters_mmx[yoffset],
@@ -336,7 +336,7 @@ unsigned int vp8_sub_pixel_variance16x8_mmx
   );
 
 
-  vp8_filter_block2d_bil_var_mmx(
+  vp9_filter_block2d_bil_var_mmx(
     src_ptr + 8, src_pixels_per_line,
     dst_ptr + 8, dst_pixels_per_line, 8,
     vp8_vp7_bilinear_filters_mmx[xoffset], vp8_vp7_bilinear_filters_mmx[yoffset],
@@ -350,7 +350,7 @@ unsigned int vp8_sub_pixel_variance16x8_mmx
   return (xxsum0 - ((xsum0 * xsum0) >> 7));
 }
 
-unsigned int vp8_sub_pixel_variance8x16_mmx
+unsigned int vp9_sub_pixel_variance8x16_mmx
 (
   const unsigned char  *src_ptr,
   int  src_pixels_per_line,
@@ -362,7 +362,7 @@ unsigned int vp8_sub_pixel_variance8x16_mmx
 ) {
   int xsum;
   unsigned int xxsum;
-  vp8_filter_block2d_bil_var_mmx(
+  vp9_filter_block2d_bil_var_mmx(
     src_ptr, src_pixels_per_line,
     dst_ptr, dst_pixels_per_line, 16,
     vp8_vp7_bilinear_filters_mmx[xoffset], vp8_vp7_bilinear_filters_mmx[yoffset],
@@ -373,34 +373,34 @@ unsigned int vp8_sub_pixel_variance8x16_mmx
 }
 
 
-unsigned int vp8_variance_halfpixvar16x16_h_mmx(
+unsigned int vp9_variance_halfpixvar16x16_h_mmx(
   const unsigned char *src_ptr,
   int  source_stride,
   const unsigned char *ref_ptr,
   int  recon_stride,
   unsigned int *sse) {
-  return vp8_sub_pixel_variance16x16_mmx(src_ptr, source_stride, 8, 0,
+  return vp9_sub_pixel_variance16x16_mmx(src_ptr, source_stride, 8, 0,
                                          ref_ptr, recon_stride, sse);
 }
 
 
-unsigned int vp8_variance_halfpixvar16x16_v_mmx(
+unsigned int vp9_variance_halfpixvar16x16_v_mmx(
   const unsigned char *src_ptr,
   int  source_stride,
   const unsigned char *ref_ptr,
   int  recon_stride,
   unsigned int *sse) {
-  return vp8_sub_pixel_variance16x16_mmx(src_ptr, source_stride, 0, 8,
+  return vp9_sub_pixel_variance16x16_mmx(src_ptr, source_stride, 0, 8,
                                          ref_ptr, recon_stride, sse);
 }
 
 
-unsigned int vp8_variance_halfpixvar16x16_hv_mmx(
+unsigned int vp9_variance_halfpixvar16x16_hv_mmx(
   const unsigned char *src_ptr,
   int  source_stride,
   const unsigned char *ref_ptr,
   int  recon_stride,
   unsigned int *sse) {
-  return vp8_sub_pixel_variance16x16_mmx(src_ptr, source_stride, 8, 8,
+  return vp9_sub_pixel_variance16x16_mmx(src_ptr, source_stride, 8, 8,
                                          ref_ptr, recon_stride, sse);
 }
diff --git a/vp8/encoder/x86/variance_sse2.c b/vp8/encoder/x86/variance_sse2.c
index 6dbc70b80b9f1c5ce8fa0dfeab72d1df8eb5cda8..79e13005eafd2b7e6009f8930dd191b6e8a61034 100644
--- a/vp8/encoder/x86/variance_sse2.c
+++ b/vp8/encoder/x86/variance_sse2.c
@@ -20,7 +20,7 @@ extern void filter_block1d_v6_mmx(const short *src_ptr, unsigned char *output_pt
 extern void filter_block1d8_h6_sse2(const unsigned char *src_ptr, unsigned short *output_ptr, unsigned int src_pixels_per_line, unsigned int pixel_step, unsigned int output_height, unsigned int output_width, short *vp7_filter);
 extern void filter_block1d8_v6_sse2(const short *src_ptr, unsigned char *output_ptr, unsigned int pixels_per_line, unsigned int pixel_step, unsigned int output_height, unsigned int output_width, short *vp7_filter);
 
-extern void vp8_filter_block2d_bil4x4_var_mmx
+extern void vp9_filter_block2d_bil4x4_var_mmx
 (
   const unsigned char *ref_ptr,
   int ref_pixels_per_line,
@@ -32,7 +32,7 @@ extern void vp8_filter_block2d_bil4x4_var_mmx
   unsigned int *sumsquared
 );
 
-extern unsigned int vp8_get4x4var_mmx
+extern unsigned int vp9_get4x4var_mmx
 (
   const unsigned char *src_ptr,
   int  source_stride,
@@ -42,11 +42,11 @@ extern unsigned int vp8_get4x4var_mmx
   int *Sum
 );
 
-unsigned int vp8_get_mb_ss_sse2
+unsigned int vp9_get_mb_ss_sse2
 (
   const short *src_ptr
 );
-unsigned int vp8_get16x16var_sse2
+unsigned int vp9_get16x16var_sse2
 (
   const unsigned char *src_ptr,
   int source_stride,
@@ -55,7 +55,7 @@ unsigned int vp8_get16x16var_sse2
   unsigned int *SSE,
   int *Sum
 );
-unsigned int vp8_get8x8var_sse2
+unsigned int vp9_get8x8var_sse2
 (
   const unsigned char *src_ptr,
   int source_stride,
@@ -64,7 +64,7 @@ unsigned int vp8_get8x8var_sse2
   unsigned int *SSE,
   int *Sum
 );
-void vp8_filter_block2d_bil_var_sse2
+void vp9_filter_block2d_bil_var_sse2
 (
   const unsigned char *ref_ptr,
   int ref_pixels_per_line,
@@ -76,7 +76,7 @@ void vp8_filter_block2d_bil_var_sse2
   int *sum,
   unsigned int *sumsquared
 );
-void vp8_half_horiz_vert_variance8x_h_sse2
+void vp9_half_horiz_vert_variance8x_h_sse2
 (
   const unsigned char *ref_ptr,
   int ref_pixels_per_line,
@@ -86,7 +86,7 @@ void vp8_half_horiz_vert_variance8x_h_sse2
   int *sum,
   unsigned int *sumsquared
 );
-void vp8_half_horiz_vert_variance16x_h_sse2
+void vp9_half_horiz_vert_variance16x_h_sse2
 (
   const unsigned char *ref_ptr,
   int ref_pixels_per_line,
@@ -96,7 +96,7 @@ void vp8_half_horiz_vert_variance16x_h_sse2
   int *sum,
   unsigned int *sumsquared
 );
-void vp8_half_horiz_variance8x_h_sse2
+void vp9_half_horiz_variance8x_h_sse2
 (
   const unsigned char *ref_ptr,
   int ref_pixels_per_line,
@@ -106,7 +106,7 @@ void vp8_half_horiz_variance8x_h_sse2
   int *sum,
   unsigned int *sumsquared
 );
-void vp8_half_horiz_variance16x_h_sse2
+void vp9_half_horiz_variance16x_h_sse2
 (
   const unsigned char *ref_ptr,
   int ref_pixels_per_line,
@@ -116,7 +116,7 @@ void vp8_half_horiz_variance16x_h_sse2
   int *sum,
   unsigned int *sumsquared
 );
-void vp8_half_vert_variance8x_h_sse2
+void vp9_half_vert_variance8x_h_sse2
 (
   const unsigned char *ref_ptr,
   int ref_pixels_per_line,
@@ -126,7 +126,7 @@ void vp8_half_vert_variance8x_h_sse2
   int *sum,
   unsigned int *sumsquared
 );
-void vp8_half_vert_variance16x_h_sse2
+void vp9_half_vert_variance16x_h_sse2
 (
   const unsigned char *ref_ptr,
   int ref_pixels_per_line,
@@ -139,7 +139,7 @@ void vp8_half_vert_variance16x_h_sse2
 
 DECLARE_ALIGNED(16, extern short, vp8_vp7_bilinear_filters_mmx[16][8]);
 
-unsigned int vp8_variance4x4_wmt(
+unsigned int vp9_variance4x4_wmt(
   const unsigned char *src_ptr,
   int  source_stride,
   const unsigned char *ref_ptr,
@@ -148,13 +148,13 @@ unsigned int vp8_variance4x4_wmt(
   unsigned int var;
   int avg;
 
-  vp8_get4x4var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &var, &avg);
+  vp9_get4x4var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &var, &avg);
   *sse = var;
   return (var - ((avg * avg) >> 4));
 
 }
 
-unsigned int vp8_variance8x8_wmt
+unsigned int vp9_variance8x8_wmt
 (
   const unsigned char *src_ptr,
   int  source_stride,
@@ -164,14 +164,14 @@ unsigned int vp8_variance8x8_wmt
   unsigned int var;
   int avg;
 
-  vp8_get8x8var_sse2(src_ptr, source_stride, ref_ptr, recon_stride, &var, &avg);
+  vp9_get8x8var_sse2(src_ptr, source_stride, ref_ptr, recon_stride, &var, &avg);
   *sse = var;
   return (var - ((avg * avg) >> 6));
 
 }
 
 
-unsigned int vp8_variance16x16_wmt
+unsigned int vp9_variance16x16_wmt
 (
   const unsigned char *src_ptr,
   int  source_stride,
@@ -182,11 +182,11 @@ unsigned int vp8_variance16x16_wmt
   int sum0;
 
 
-  vp8_get16x16var_sse2(src_ptr, source_stride, ref_ptr, recon_stride, &sse0, &sum0);
+  vp9_get16x16var_sse2(src_ptr, source_stride, ref_ptr, recon_stride, &sse0, &sum0);
   *sse = sse0;
   return (sse0 - ((sum0 * sum0) >> 8));
 }
-unsigned int vp8_mse16x16_wmt(
+unsigned int vp9_mse16x16_wmt(
   const unsigned char *src_ptr,
   int  source_stride,
   const unsigned char *ref_ptr,
@@ -195,14 +195,14 @@ unsigned int vp8_mse16x16_wmt(
 
   unsigned int sse0;
   int sum0;
-  vp8_get16x16var_sse2(src_ptr, source_stride, ref_ptr, recon_stride, &sse0, &sum0);
+  vp9_get16x16var_sse2(src_ptr, source_stride, ref_ptr, recon_stride, &sse0, &sum0);
   *sse = sse0;
   return sse0;
 
 }
 
 
-unsigned int vp8_variance16x8_wmt
+unsigned int vp9_variance16x8_wmt
 (
   const unsigned char *src_ptr,
   int  source_stride,
@@ -212,8 +212,8 @@ unsigned int vp8_variance16x8_wmt
   unsigned int sse0, sse1, var;
   int sum0, sum1, avg;
 
-  vp8_get8x8var_sse2(src_ptr, source_stride, ref_ptr, recon_stride, &sse0, &sum0);
-  vp8_get8x8var_sse2(src_ptr + 8, source_stride, ref_ptr + 8, recon_stride, &sse1, &sum1);
+  vp9_get8x8var_sse2(src_ptr, source_stride, ref_ptr, recon_stride, &sse0, &sum0);
+  vp9_get8x8var_sse2(src_ptr + 8, source_stride, ref_ptr + 8, recon_stride, &sse1, &sum1);
 
   var = sse0 + sse1;
   avg = sum0 + sum1;
@@ -222,7 +222,7 @@ unsigned int vp8_variance16x8_wmt
 
 }
 
-unsigned int vp8_variance8x16_wmt
+unsigned int vp9_variance8x16_wmt
 (
   const unsigned char *src_ptr,
   int  source_stride,
@@ -232,8 +232,8 @@ unsigned int vp8_variance8x16_wmt
   unsigned int sse0, sse1, var;
   int sum0, sum1, avg;
 
-  vp8_get8x8var_sse2(src_ptr, source_stride, ref_ptr, recon_stride, &sse0, &sum0);
-  vp8_get8x8var_sse2(src_ptr + 8 * source_stride, source_stride, ref_ptr + 8 * recon_stride, recon_stride, &sse1, &sum1);
+  vp9_get8x8var_sse2(src_ptr, source_stride, ref_ptr, recon_stride, &sse0, &sum0);
+  vp9_get8x8var_sse2(src_ptr + 8 * source_stride, source_stride, ref_ptr + 8 * recon_stride, recon_stride, &sse1, &sum1);
 
   var = sse0 + sse1;
   avg = sum0 + sum1;
@@ -242,7 +242,7 @@ unsigned int vp8_variance8x16_wmt
 
 }
 
-unsigned int vp8_sub_pixel_variance4x4_wmt
+unsigned int vp9_sub_pixel_variance4x4_wmt
 (
   const unsigned char  *src_ptr,
   int  src_pixels_per_line,
@@ -254,7 +254,7 @@ unsigned int vp8_sub_pixel_variance4x4_wmt
 ) {
   int xsum;
   unsigned int xxsum;
-  vp8_filter_block2d_bil4x4_var_mmx(
+  vp9_filter_block2d_bil4x4_var_mmx(
     src_ptr, src_pixels_per_line,
     dst_ptr, dst_pixels_per_line,
     vp8_vp7_bilinear_filters_mmx[xoffset], vp8_vp7_bilinear_filters_mmx[yoffset],
@@ -265,7 +265,7 @@ unsigned int vp8_sub_pixel_variance4x4_wmt
 }
 
 
-unsigned int vp8_sub_pixel_variance8x8_wmt
+unsigned int vp9_sub_pixel_variance8x8_wmt
 (
   const unsigned char  *src_ptr,
   int  src_pixels_per_line,
@@ -279,22 +279,22 @@ unsigned int vp8_sub_pixel_variance8x8_wmt
   unsigned int xxsum;
 
   if (xoffset == HALFNDX && yoffset == 0) {
-    vp8_half_horiz_variance8x_h_sse2(
+    vp9_half_horiz_variance8x_h_sse2(
       src_ptr, src_pixels_per_line,
       dst_ptr, dst_pixels_per_line, 8,
       &xsum, &xxsum);
   } else if (xoffset == 0 && yoffset == HALFNDX) {
-    vp8_half_vert_variance8x_h_sse2(
+    vp9_half_vert_variance8x_h_sse2(
       src_ptr, src_pixels_per_line,
       dst_ptr, dst_pixels_per_line, 8,
       &xsum, &xxsum);
   } else if (xoffset == HALFNDX && yoffset == HALFNDX) {
-    vp8_half_horiz_vert_variance8x_h_sse2(
+    vp9_half_horiz_vert_variance8x_h_sse2(
       src_ptr, src_pixels_per_line,
       dst_ptr, dst_pixels_per_line, 8,
       &xsum, &xxsum);
   } else {
-    vp8_filter_block2d_bil_var_sse2(
+    vp9_filter_block2d_bil_var_sse2(
       src_ptr, src_pixels_per_line,
       dst_ptr, dst_pixels_per_line, 8,
       xoffset, yoffset,
@@ -305,7 +305,7 @@ unsigned int vp8_sub_pixel_variance8x8_wmt
   return (xxsum - ((xsum * xsum) >> 6));
 }
 
-unsigned int vp8_sub_pixel_variance16x16_wmt
+unsigned int vp9_sub_pixel_variance16x16_wmt
 (
   const unsigned char  *src_ptr,
   int  src_pixels_per_line,
@@ -322,29 +322,29 @@ unsigned int vp8_sub_pixel_variance16x16_wmt
   // note we could avoid these if statements if the calling function
   // just called the appropriate functions inside.
   if (xoffset == HALFNDX && yoffset == 0) {
-    vp8_half_horiz_variance16x_h_sse2(
+    vp9_half_horiz_variance16x_h_sse2(
       src_ptr, src_pixels_per_line,
       dst_ptr, dst_pixels_per_line, 16,
       &xsum0, &xxsum0);
   } else if (xoffset == 0 && yoffset == HALFNDX) {
-    vp8_half_vert_variance16x_h_sse2(
+    vp9_half_vert_variance16x_h_sse2(
       src_ptr, src_pixels_per_line,
       dst_ptr, dst_pixels_per_line, 16,
       &xsum0, &xxsum0);
   } else if (xoffset == HALFNDX && yoffset == HALFNDX) {
-    vp8_half_horiz_vert_variance16x_h_sse2(
+    vp9_half_horiz_vert_variance16x_h_sse2(
       src_ptr, src_pixels_per_line,
       dst_ptr, dst_pixels_per_line, 16,
       &xsum0, &xxsum0);
   } else {
-    vp8_filter_block2d_bil_var_sse2(
+    vp9_filter_block2d_bil_var_sse2(
       src_ptr, src_pixels_per_line,
       dst_ptr, dst_pixels_per_line, 16,
       xoffset, yoffset,
       &xsum0, &xxsum0
     );
 
-    vp8_filter_block2d_bil_var_sse2(
+    vp9_filter_block2d_bil_var_sse2(
       src_ptr + 8, src_pixels_per_line,
       dst_ptr + 8, dst_pixels_per_line, 16,
       xoffset, yoffset,
@@ -358,7 +358,7 @@ unsigned int vp8_sub_pixel_variance16x16_wmt
   return (xxsum0 - ((xsum0 * xsum0) >> 8));
 }
 
-unsigned int vp8_sub_pixel_mse16x16_wmt(
+unsigned int vp9_sub_pixel_mse16x16_wmt(
   const unsigned char  *src_ptr,
   int  src_pixels_per_line,
   int  xoffset,
@@ -367,11 +367,11 @@ unsigned int vp8_sub_pixel_mse16x16_wmt(
   int dst_pixels_per_line,
   unsigned int *sse
 ) {
-  vp8_sub_pixel_variance16x16_wmt(src_ptr, src_pixels_per_line, xoffset, yoffset, dst_ptr, dst_pixels_per_line, sse);
+  vp9_sub_pixel_variance16x16_wmt(src_ptr, src_pixels_per_line, xoffset, yoffset, dst_ptr, dst_pixels_per_line, sse);
   return *sse;
 }
 
-unsigned int vp8_sub_pixel_variance16x8_wmt
+unsigned int vp9_sub_pixel_variance16x8_wmt
 (
   const unsigned char  *src_ptr,
   int  src_pixels_per_line,
@@ -386,28 +386,28 @@ unsigned int vp8_sub_pixel_variance16x8_wmt
   unsigned int xxsum0, xxsum1;
 
   if (xoffset == HALFNDX && yoffset == 0) {
-    vp8_half_horiz_variance16x_h_sse2(
+    vp9_half_horiz_variance16x_h_sse2(
       src_ptr, src_pixels_per_line,
       dst_ptr, dst_pixels_per_line, 8,
       &xsum0, &xxsum0);
   } else if (xoffset == 0 && yoffset == HALFNDX) {
-    vp8_half_vert_variance16x_h_sse2(
+    vp9_half_vert_variance16x_h_sse2(
       src_ptr, src_pixels_per_line,
       dst_ptr, dst_pixels_per_line, 8,
       &xsum0, &xxsum0);
   } else if (xoffset == HALFNDX && yoffset == HALFNDX) {
-    vp8_half_horiz_vert_variance16x_h_sse2(
+    vp9_half_horiz_vert_variance16x_h_sse2(
       src_ptr, src_pixels_per_line,
       dst_ptr, dst_pixels_per_line, 8,
       &xsum0, &xxsum0);
   } else {
-    vp8_filter_block2d_bil_var_sse2(
+    vp9_filter_block2d_bil_var_sse2(
       src_ptr, src_pixels_per_line,
       dst_ptr, dst_pixels_per_line, 8,
       xoffset, yoffset,
       &xsum0, &xxsum0);
 
-    vp8_filter_block2d_bil_var_sse2(
+    vp9_filter_block2d_bil_var_sse2(
       src_ptr + 8, src_pixels_per_line,
       dst_ptr + 8, dst_pixels_per_line, 8,
       xoffset, yoffset,
@@ -420,7 +420,7 @@ unsigned int vp8_sub_pixel_variance16x8_wmt
   return (xxsum0 - ((xsum0 * xsum0) >> 7));
 }
 
-unsigned int vp8_sub_pixel_variance8x16_wmt
+unsigned int vp9_sub_pixel_variance8x16_wmt
 (
   const unsigned char  *src_ptr,
   int  src_pixels_per_line,
@@ -434,22 +434,22 @@ unsigned int vp8_sub_pixel_variance8x16_wmt
   unsigned int xxsum;
 
   if (xoffset == HALFNDX && yoffset == 0) {
-    vp8_half_horiz_variance8x_h_sse2(
+    vp9_half_horiz_variance8x_h_sse2(
       src_ptr, src_pixels_per_line,
       dst_ptr, dst_pixels_per_line, 16,
       &xsum, &xxsum);
   } else if (xoffset == 0 && yoffset == HALFNDX) {
-    vp8_half_vert_variance8x_h_sse2(
+    vp9_half_vert_variance8x_h_sse2(
       src_ptr, src_pixels_per_line,
       dst_ptr, dst_pixels_per_line, 16,
       &xsum, &xxsum);
   } else if (xoffset == HALFNDX && yoffset == HALFNDX) {
-    vp8_half_horiz_vert_variance8x_h_sse2(
+    vp9_half_horiz_vert_variance8x_h_sse2(
       src_ptr, src_pixels_per_line,
       dst_ptr, dst_pixels_per_line, 16,
       &xsum, &xxsum);
   } else {
-    vp8_filter_block2d_bil_var_sse2(
+    vp9_filter_block2d_bil_var_sse2(
       src_ptr, src_pixels_per_line,
       dst_ptr, dst_pixels_per_line, 16,
       xoffset, yoffset,
@@ -461,7 +461,7 @@ unsigned int vp8_sub_pixel_variance8x16_wmt
 }
 
 
-unsigned int vp8_variance_halfpixvar16x16_h_wmt(
+unsigned int vp9_variance_halfpixvar16x16_h_wmt(
   const unsigned char *src_ptr,
   int  src_pixels_per_line,
   const unsigned char *dst_ptr,
@@ -470,7 +470,7 @@ unsigned int vp8_variance_halfpixvar16x16_h_wmt(
   int xsum0;
   unsigned int xxsum0;
 
-  vp8_half_horiz_variance16x_h_sse2(
+  vp9_half_horiz_variance16x_h_sse2(
     src_ptr, src_pixels_per_line,
     dst_ptr, dst_pixels_per_line, 16,
     &xsum0, &xxsum0);
@@ -480,7 +480,7 @@ unsigned int vp8_variance_halfpixvar16x16_h_wmt(
 }
 
 
-unsigned int vp8_variance_halfpixvar16x16_v_wmt(
+unsigned int vp9_variance_halfpixvar16x16_v_wmt(
   const unsigned char *src_ptr,
   int  src_pixels_per_line,
   const unsigned char *dst_ptr,
@@ -488,7 +488,7 @@ unsigned int vp8_variance_halfpixvar16x16_v_wmt(
   unsigned int *sse) {
   int xsum0;
   unsigned int xxsum0;
-  vp8_half_vert_variance16x_h_sse2(
+  vp9_half_vert_variance16x_h_sse2(
     src_ptr, src_pixels_per_line,
     dst_ptr, dst_pixels_per_line, 16,
     &xsum0, &xxsum0);
@@ -498,7 +498,7 @@ unsigned int vp8_variance_halfpixvar16x16_v_wmt(
 }
 
 
-unsigned int vp8_variance_halfpixvar16x16_hv_wmt(
+unsigned int vp9_variance_halfpixvar16x16_hv_wmt(
   const unsigned char *src_ptr,
   int  src_pixels_per_line,
   const unsigned char *dst_ptr,
@@ -507,7 +507,7 @@ unsigned int vp8_variance_halfpixvar16x16_hv_wmt(
   int xsum0;
   unsigned int xxsum0;
 
-  vp8_half_horiz_vert_variance16x_h_sse2(
+  vp9_half_horiz_vert_variance16x_h_sse2(
     src_ptr, src_pixels_per_line,
     dst_ptr, dst_pixels_per_line, 16,
     &xsum0, &xxsum0);
diff --git a/vp8/encoder/x86/variance_ssse3.c b/vp8/encoder/x86/variance_ssse3.c
index 2e85017f142d666585d49b656afd69853d694ca9..f33c662e1740b63bc55c13bdf8ac5751592fffc3 100644
--- a/vp8/encoder/x86/variance_ssse3.c
+++ b/vp8/encoder/x86/variance_ssse3.c
@@ -15,7 +15,7 @@
 
 #define HALFNDX 8
 
-extern unsigned int vp8_get16x16var_sse2
+extern unsigned int vp9_get16x16var_sse2
 (
   const unsigned char *src_ptr,
   int source_stride,
@@ -24,7 +24,7 @@ extern unsigned int vp8_get16x16var_sse2
   unsigned int *SSE,
   int *Sum
 );
-extern void vp8_half_horiz_vert_variance16x_h_sse2
+extern void vp9_half_horiz_vert_variance16x_h_sse2
 (
   const unsigned char *ref_ptr,
   int ref_pixels_per_line,
@@ -34,7 +34,7 @@ extern void vp8_half_horiz_vert_variance16x_h_sse2
   int *sum,
   unsigned int *sumsquared
 );
-extern void vp8_half_horiz_variance16x_h_sse2
+extern void vp9_half_horiz_variance16x_h_sse2
 (
   const unsigned char *ref_ptr,
   int ref_pixels_per_line,
@@ -44,7 +44,7 @@ extern void vp8_half_horiz_variance16x_h_sse2
   int *sum,
   unsigned int *sumsquared
 );
-extern void vp8_half_vert_variance16x_h_sse2
+extern void vp9_half_vert_variance16x_h_sse2
 (
   const unsigned char *ref_ptr,
   int ref_pixels_per_line,
@@ -54,7 +54,7 @@ extern void vp8_half_vert_variance16x_h_sse2
   int *sum,
   unsigned int *sumsquared
 );
-extern void vp8_filter_block2d_bil_var_ssse3
+extern void vp9_filter_block2d_bil_var_ssse3
 (
   const unsigned char *ref_ptr,
   int ref_pixels_per_line,
@@ -67,7 +67,7 @@ extern void vp8_filter_block2d_bil_var_ssse3
   unsigned int *sumsquared
 );
 
-unsigned int vp8_sub_pixel_variance16x16_ssse3
+unsigned int vp9_sub_pixel_variance16x16_ssse3
 (
   const unsigned char  *src_ptr,
   int  src_pixels_per_line,
@@ -83,22 +83,22 @@ unsigned int vp8_sub_pixel_variance16x16_ssse3
   // note we could avoid these if statements if the calling function
   // just called the appropriate functions inside.
   if (xoffset == HALFNDX && yoffset == 0) {
-    vp8_half_horiz_variance16x_h_sse2(
+    vp9_half_horiz_variance16x_h_sse2(
       src_ptr, src_pixels_per_line,
       dst_ptr, dst_pixels_per_line, 16,
       &xsum0, &xxsum0);
   } else if (xoffset == 0 && yoffset == HALFNDX) {
-    vp8_half_vert_variance16x_h_sse2(
+    vp9_half_vert_variance16x_h_sse2(
       src_ptr, src_pixels_per_line,
       dst_ptr, dst_pixels_per_line, 16,
       &xsum0, &xxsum0);
   } else if (xoffset == HALFNDX && yoffset == HALFNDX) {
-    vp8_half_horiz_vert_variance16x_h_sse2(
+    vp9_half_horiz_vert_variance16x_h_sse2(
       src_ptr, src_pixels_per_line,
       dst_ptr, dst_pixels_per_line, 16,
       &xsum0, &xxsum0);
   } else {
-    vp8_filter_block2d_bil_var_ssse3(
+    vp9_filter_block2d_bil_var_ssse3(
       src_ptr, src_pixels_per_line,
       dst_ptr, dst_pixels_per_line, 16,
       xoffset, yoffset,
@@ -109,7 +109,7 @@ unsigned int vp8_sub_pixel_variance16x16_ssse3
   return (xxsum0 - ((xsum0 * xsum0) >> 8));
 }
 
-unsigned int vp8_sub_pixel_variance16x8_ssse3
+unsigned int vp9_sub_pixel_variance16x8_ssse3
 (
   const unsigned char  *src_ptr,
   int  src_pixels_per_line,
@@ -124,22 +124,22 @@ unsigned int vp8_sub_pixel_variance16x8_ssse3
   unsigned int xxsum0;
 
   if (xoffset == HALFNDX && yoffset == 0) {
-    vp8_half_horiz_variance16x_h_sse2(
+    vp9_half_horiz_variance16x_h_sse2(
       src_ptr, src_pixels_per_line,
       dst_ptr, dst_pixels_per_line, 8,
       &xsum0, &xxsum0);
   } else if (xoffset == 0 && yoffset == HALFNDX) {
-    vp8_half_vert_variance16x_h_sse2(
+    vp9_half_vert_variance16x_h_sse2(
       src_ptr, src_pixels_per_line,
       dst_ptr, dst_pixels_per_line, 8,
       &xsum0, &xxsum0);
   } else if (xoffset == HALFNDX && yoffset == HALFNDX) {
-    vp8_half_horiz_vert_variance16x_h_sse2(
+    vp9_half_horiz_vert_variance16x_h_sse2(
       src_ptr, src_pixels_per_line,
       dst_ptr, dst_pixels_per_line, 8,
       &xsum0, &xxsum0);
   } else {
-    vp8_filter_block2d_bil_var_ssse3(
+    vp9_filter_block2d_bil_var_ssse3(
       src_ptr, src_pixels_per_line,
       dst_ptr, dst_pixels_per_line, 8,
       xoffset, yoffset,
diff --git a/vp8/encoder/x86/x86_csystemdependent.c b/vp8/encoder/x86/x86_csystemdependent.c
index 053601939f2d58ad235d405db9af45a16e3d31de..3c39227570aec8cd0c70d451a26157bcf3230a3e 100644
--- a/vp8/encoder/x86/x86_csystemdependent.c
+++ b/vp8/encoder/x86/x86_csystemdependent.c
@@ -16,67 +16,67 @@
 
 
 #if HAVE_MMX
-void vp8_short_fdct8x4_mmx(short *input, short *output, int pitch) {
-  vp8_short_fdct4x4_mmx(input,   output,    pitch);
-  vp8_short_fdct4x4_mmx(input + 4, output + 16, pitch);
+void vp9_short_fdct8x4_mmx(short *input, short *output, int pitch) {
+  vp9_short_fdct4x4_mmx(input,   output,    pitch);
+  vp9_short_fdct4x4_mmx(input + 4, output + 16, pitch);
 }
 
-int vp8_mbblock_error_mmx_impl(short *coeff_ptr, short *dcoef_ptr, int dc);
-int vp8_mbblock_error_mmx(MACROBLOCK *mb, int dc) {
+int vp9_mbblock_error_mmx_impl(short *coeff_ptr, short *dcoef_ptr, int dc);
+int vp9_mbblock_error_mmx(MACROBLOCK *mb, int dc) {
   short *coeff_ptr =  mb->block[0].coeff;
   short *dcoef_ptr =  mb->e_mbd.block[0].dqcoeff;
-  return vp8_mbblock_error_mmx_impl(coeff_ptr, dcoef_ptr, dc);
+  return vp9_mbblock_error_mmx_impl(coeff_ptr, dcoef_ptr, dc);
 }
 
-int vp8_mbuverror_mmx_impl(short *s_ptr, short *d_ptr);
-int vp8_mbuverror_mmx(MACROBLOCK *mb) {
+int vp9_mbuverror_mmx_impl(short *s_ptr, short *d_ptr);
+int vp9_mbuverror_mmx(MACROBLOCK *mb) {
   short *s_ptr = &mb->coeff[256];
   short *d_ptr = &mb->e_mbd.dqcoeff[256];
-  return vp8_mbuverror_mmx_impl(s_ptr, d_ptr);
+  return vp9_mbuverror_mmx_impl(s_ptr, d_ptr);
 }
 
-void vp8_subtract_b_mmx_impl(unsigned char *z,  int src_stride,
+void vp9_subtract_b_mmx_impl(unsigned char *z,  int src_stride,
                              short *diff, unsigned char *predictor,
                              int pitch);
-void vp8_subtract_b_mmx(BLOCK *be, BLOCKD *bd, int pitch) {
+void vp9_subtract_b_mmx(BLOCK *be, BLOCKD *bd, int pitch) {
   unsigned char *z = *(be->base_src) + be->src;
   unsigned int  src_stride = be->src_stride;
   short *diff = &be->src_diff[0];
   unsigned char *predictor = &bd->predictor[0];
-  vp8_subtract_b_mmx_impl(z, src_stride, diff, predictor, pitch);
+  vp9_subtract_b_mmx_impl(z, src_stride, diff, predictor, pitch);
 }
 
 #endif
 
 #if HAVE_SSE2
-int vp8_mbblock_error_xmm_impl(short *coeff_ptr, short *dcoef_ptr, int dc);
-int vp8_mbblock_error_xmm(MACROBLOCK *mb, int dc) {
+int vp9_mbblock_error_xmm_impl(short *coeff_ptr, short *dcoef_ptr, int dc);
+int vp9_mbblock_error_xmm(MACROBLOCK *mb, int dc) {
   short *coeff_ptr =  mb->block[0].coeff;
   short *dcoef_ptr =  mb->e_mbd.block[0].dqcoeff;
-  return vp8_mbblock_error_xmm_impl(coeff_ptr, dcoef_ptr, dc);
+  return vp9_mbblock_error_xmm_impl(coeff_ptr, dcoef_ptr, dc);
 }
 
-int vp8_mbuverror_xmm_impl(short *s_ptr, short *d_ptr);
-int vp8_mbuverror_xmm(MACROBLOCK *mb) {
+int vp9_mbuverror_xmm_impl(short *s_ptr, short *d_ptr);
+int vp9_mbuverror_xmm(MACROBLOCK *mb) {
   short *s_ptr = &mb->coeff[256];
   short *d_ptr = &mb->e_mbd.dqcoeff[256];
-  return vp8_mbuverror_xmm_impl(s_ptr, d_ptr);
+  return vp9_mbuverror_xmm_impl(s_ptr, d_ptr);
 }
 
-void vp8_subtract_b_sse2_impl(unsigned char *z,  int src_stride,
+void vp9_subtract_b_sse2_impl(unsigned char *z,  int src_stride,
                               short *diff, unsigned char *predictor,
                               int pitch);
-void vp8_subtract_b_sse2(BLOCK *be, BLOCKD *bd, int pitch) {
+void vp9_subtract_b_sse2(BLOCK *be, BLOCKD *bd, int pitch) {
   unsigned char *z = *(be->base_src) + be->src;
   unsigned int  src_stride = be->src_stride;
   short *diff = &be->src_diff[0];
   unsigned char *predictor = &bd->predictor[0];
-  vp8_subtract_b_sse2_impl(z, src_stride, diff, predictor, pitch);
+  vp9_subtract_b_sse2_impl(z, src_stride, diff, predictor, pitch);
 }
 
 #endif
 
-void vp8_arch_x86_encoder_init(VP8_COMP *cpi) {
+void vp9_arch_x86_encoder_init(VP8_COMP *cpi) {
 #if CONFIG_RUNTIME_CPU_DETECT
   int flags = x86_simd_caps();
 
@@ -90,23 +90,23 @@ void vp8_arch_x86_encoder_init(VP8_COMP *cpi) {
   /* Override default functions with fastest ones for this CPU. */
 #if HAVE_SSE2
   if (flags & HAS_SSE2) {
-    cpi->rtcd.temporal.apply                 = vp8_temporal_filter_apply_sse2;
+    cpi->rtcd.temporal.apply                 = vp9_temporal_filter_apply_sse2;
 
   }
 #endif
 
 #if HAVE_SSE3
   if (flags & HAS_SSE3) {
-    cpi->rtcd.search.full_search             = vp8_full_search_sadx3;
-    cpi->rtcd.search.diamond_search          = vp8_diamond_search_sadx4;
-    cpi->rtcd.search.refining_search         = vp8_refining_search_sadx4;
+    cpi->rtcd.search.full_search             = vp9_full_search_sadx3;
+    cpi->rtcd.search.diamond_search          = vp9_diamond_search_sadx4;
+    cpi->rtcd.search.refining_search         = vp9_refining_search_sadx4;
   }
 #endif
 
 
 #if HAVE_SSE4_1
   if (flags & HAS_SSE4_1) {
-    cpi->rtcd.search.full_search             = vp8_full_search_sadx8;
+    cpi->rtcd.search.full_search             = vp9_full_search_sadx8;
   }
 #endif
 
diff --git a/vp8/vp8_cx_iface.c b/vp8/vp8_cx_iface.c
index e74eb78cf3a0aefd8383b97a1303613878b36948..1b8dba04f4216bdfc932230de24f950db7fa1664 100644
--- a/vp8/vp8_cx_iface.c
+++ b/vp8/vp8_cx_iface.c
@@ -357,14 +357,14 @@ static vpx_codec_err_t vp8e_set_config(vpx_codec_alg_priv_t       *ctx,
   if (!res) {
     ctx->cfg = *cfg;
     set_vp8e_config(&ctx->oxcf, ctx->cfg, ctx->vp8_cfg);
-    vp8_change_config(ctx->cpi, &ctx->oxcf);
+    vp9_change_config(ctx->cpi, &ctx->oxcf);
   }
 
   return res;
 }
 
 
-int vp8_reverse_trans(int);
+int vp9_reverse_trans(int);
 
 
 static vpx_codec_err_t get_param(vpx_codec_alg_priv_t *ctx,
@@ -378,8 +378,8 @@ static vpx_codec_err_t get_param(vpx_codec_alg_priv_t *ctx,
     return VPX_CODEC_INVALID_PARAM;
 
   switch (ctrl_id) {
-      MAP(VP8E_GET_LAST_QUANTIZER, vp8_get_quantizer(ctx->cpi));
-      MAP(VP8E_GET_LAST_QUANTIZER_64, vp8_reverse_trans(vp8_get_quantizer(ctx->cpi)));
+      MAP(VP8E_GET_LAST_QUANTIZER, vp9_get_quantizer(ctx->cpi));
+      MAP(VP8E_GET_LAST_QUANTIZER_64, vp9_reverse_trans(vp9_get_quantizer(ctx->cpi)));
   }
 
   return VPX_CODEC_OK;
@@ -418,7 +418,7 @@ static vpx_codec_err_t set_param(vpx_codec_alg_priv_t *ctx,
   if (!res) {
     ctx->vp8_cfg = xcfg;
     set_vp8e_config(&ctx->oxcf, ctx->cfg, ctx->vp8_cfg);
-    vp8_change_config(ctx->cpi, &ctx->oxcf);
+    vp9_change_config(ctx->cpi, &ctx->oxcf);
   }
 
   return res;
@@ -482,7 +482,7 @@ static vpx_codec_err_t vp8e_common_init(vpx_codec_ctx_t *ctx,
 
     priv->deprecated_mode = NO_MODE_SET;
 
-    vp8_initialize();
+    vp9_initialize();
 
     res = validate_config(priv, &priv->cfg, &priv->vp8_cfg);
 
@@ -490,7 +490,7 @@ static vpx_codec_err_t vp8e_common_init(vpx_codec_ctx_t *ctx,
       set_vp8e_config(&ctx->priv->alg_priv->oxcf,
                       ctx->priv->alg_priv->cfg,
                       ctx->priv->alg_priv->vp8_cfg);
-      optr = vp8_create_compressor(&ctx->priv->alg_priv->oxcf);
+      optr = vp9_create_compressor(&ctx->priv->alg_priv->oxcf);
 
       if (!optr)
         res = VPX_CODEC_MEM_ERROR;
@@ -518,7 +518,7 @@ static vpx_codec_err_t vp8e_exp_init(vpx_codec_ctx_t *ctx) {
 static vpx_codec_err_t vp8e_destroy(vpx_codec_alg_priv_t *ctx) {
 
   free(ctx->cx_data);
-  vp8_remove_compressor(&ctx->cpi);
+  vp9_remove_compressor(&ctx->cpi);
   free(ctx);
   return VPX_CODEC_OK;
 }
@@ -563,7 +563,7 @@ static void pick_quickcompress_mode(vpx_codec_alg_priv_t  *ctx,
 
   if (ctx->oxcf.Mode != new_qc) {
     ctx->oxcf.Mode = new_qc;
-    vp8_change_config(ctx->cpi, &ctx->oxcf);
+    vp9_change_config(ctx->cpi, &ctx->oxcf);
   }
 }
 
@@ -602,7 +602,7 @@ static vpx_codec_err_t vp8e_encode(vpx_codec_alg_priv_t  *ctx,
     if (flags & VP8_EFLAG_NO_REF_ARF)
       ref ^= VP8_ALT_FLAG;
 
-    vp8_use_as_reference(ctx->cpi, ref);
+    vp9_use_as_reference(ctx->cpi, ref);
   }
 
   if (flags & (VP8_EFLAG_NO_UPD_LAST | VP8_EFLAG_NO_UPD_GF
@@ -619,11 +619,11 @@ static vpx_codec_err_t vp8e_encode(vpx_codec_alg_priv_t  *ctx,
     if (flags & VP8_EFLAG_NO_UPD_ARF)
       upd ^= VP8_ALT_FLAG;
 
-    vp8_update_reference(ctx->cpi, upd);
+    vp9_update_reference(ctx->cpi, upd);
   }
 
   if (flags & VP8_EFLAG_NO_UPD_ENTROPY) {
-    vp8_update_entropy(ctx->cpi, 0);
+    vp9_update_entropy(ctx->cpi, 0);
   }
 
   /* Handle fixed keyframe intervals */
@@ -660,7 +660,7 @@ static vpx_codec_err_t vp8e_encode(vpx_codec_alg_priv_t  *ctx,
     if (img != NULL) {
       res = image2yuvconfig(img, &sd);
 
-      if (vp8_receive_raw_frame(ctx->cpi, ctx->next_frame_flag | lib_flags,
+      if (vp9_receive_raw_frame(ctx->cpi, ctx->next_frame_flag | lib_flags,
                                 &sd, dst_time_stamp, dst_end_time_stamp)) {
         VP8_COMP *cpi = (VP8_COMP *)ctx->cpi;
         res = update_error_state(ctx, &cpi->common.error);
@@ -675,7 +675,7 @@ static vpx_codec_err_t vp8e_encode(vpx_codec_alg_priv_t  *ctx,
     lib_flags = 0;
 
     while (cx_data_sz >= ctx->cx_data_sz / 2
-           && -1 != vp8_get_compressed_data(ctx->cpi, &lib_flags, &size, cx_data, &dst_time_stamp, &dst_end_time_stamp, !img)) {
+           && -1 != vp9_get_compressed_data(ctx->cpi, &lib_flags, &size, cx_data, &dst_time_stamp, &dst_end_time_stamp, !img)) {
       if (size) {
         vpx_codec_pts_t    round, delta;
         vpx_codec_cx_pkt_t pkt;
@@ -766,7 +766,7 @@ static vpx_codec_err_t vp8e_set_reference(vpx_codec_alg_priv_t *ctx,
     YV12_BUFFER_CONFIG sd;
 
     image2yuvconfig(&frame->img, &sd);
-    vp8_set_reference(ctx->cpi, frame->frame_type, &sd);
+    vp9_set_reference(ctx->cpi, frame->frame_type, &sd);
     return VPX_CODEC_OK;
   } else
     return VPX_CODEC_INVALID_PARAM;
@@ -784,7 +784,7 @@ static vpx_codec_err_t vp8e_get_reference(vpx_codec_alg_priv_t *ctx,
     YV12_BUFFER_CONFIG sd;
 
     image2yuvconfig(&frame->img, &sd);
-    vp8_get_reference(ctx->cpi, frame->frame_type, &sd);
+    vp9_get_reference(ctx->cpi, frame->frame_type, &sd);
     return VPX_CODEC_OK;
   } else
     return VPX_CODEC_INVALID_PARAM;
@@ -822,7 +822,7 @@ static vpx_image_t *vp8e_get_preview(vpx_codec_alg_priv_t *ctx) {
     flags.noise_level           = ctx->preview_ppcfg.noise_level;
   }
 
-  if (0 == vp8_get_preview_raw_frame(ctx->cpi, &sd, &flags)) {
+  if (0 == vp9_get_preview_raw_frame(ctx->cpi, &sd, &flags)) {
 
     /*
     vpx_img_wrap(&ctx->preview_img, VPX_IMG_FMT_YV12,
@@ -865,7 +865,7 @@ static vpx_codec_err_t vp8e_update_entropy(vpx_codec_alg_priv_t *ctx,
                                            int ctr_id,
                                            va_list args) {
   int update = va_arg(args, int);
-  vp8_update_entropy(ctx->cpi, update);
+  vp9_update_entropy(ctx->cpi, update);
   return VPX_CODEC_OK;
 
 }
@@ -874,7 +874,7 @@ static vpx_codec_err_t vp8e_update_reference(vpx_codec_alg_priv_t *ctx,
                                              int ctr_id,
                                              va_list args) {
   int update = va_arg(args, int);
-  vp8_update_reference(ctx->cpi, update);
+  vp9_update_reference(ctx->cpi, update);
   return VPX_CODEC_OK;
 }
 
@@ -882,7 +882,7 @@ static vpx_codec_err_t vp8e_use_reference(vpx_codec_alg_priv_t *ctx,
                                           int ctr_id,
                                           va_list args) {
   int reference_flag = va_arg(args, int);
-  vp8_use_as_reference(ctx->cpi, reference_flag);
+  vp9_use_as_reference(ctx->cpi, reference_flag);
   return VPX_CODEC_OK;
 }
 
@@ -894,7 +894,7 @@ static vpx_codec_err_t vp8e_set_roi_map(vpx_codec_alg_priv_t *ctx,
   if (data) {
     vpx_roi_map_t *roi = (vpx_roi_map_t *)data;
 
-    if (!vp8_set_roimap(ctx->cpi, roi->roi_map, roi->rows, roi->cols, roi->delta_q, roi->delta_lf, roi->static_threshold))
+    if (!vp9_set_roimap(ctx->cpi, roi->roi_map, roi->rows, roi->cols, roi->delta_q, roi->delta_lf, roi->static_threshold))
       return VPX_CODEC_OK;
     else
       return VPX_CODEC_INVALID_PARAM;
@@ -912,7 +912,7 @@ static vpx_codec_err_t vp8e_set_activemap(vpx_codec_alg_priv_t *ctx,
 
     vpx_active_map_t *map = (vpx_active_map_t *)data;
 
-    if (!vp8_set_active_map(ctx->cpi, map->active_map, map->rows, map->cols))
+    if (!vp9_set_active_map(ctx->cpi, map->active_map, map->rows, map->cols))
       return VPX_CODEC_OK;
     else
       return VPX_CODEC_INVALID_PARAM;
@@ -929,7 +929,7 @@ static vpx_codec_err_t vp8e_set_scalemode(vpx_codec_alg_priv_t *ctx,
   if (data) {
     int res;
     vpx_scaling_mode_t scalemode = *(vpx_scaling_mode_t *)data;
-    res = vp8_set_internal_size(ctx->cpi, scalemode.h_scaling_mode, scalemode.v_scaling_mode);
+    res = vp9_set_internal_size(ctx->cpi, scalemode.h_scaling_mode, scalemode.v_scaling_mode);
 
     if (!res) {
       /*force next frame a key frame to effect scaling mode */
diff --git a/vp8/vp8_dx_iface.c b/vp8/vp8_dx_iface.c
index 112c1c5d45d583fd27b0cf619f763e7a17bd3935..07926eafbe4c2282de6b5ca63acffac3bbae1f98 100644
--- a/vp8/vp8_dx_iface.c
+++ b/vp8/vp8_dx_iface.c
@@ -530,7 +530,7 @@ static vpx_codec_err_t image2yuvconfig(const vpx_image_t   *img,
 }
 
 
-static vpx_codec_err_t vp8_set_reference(vpx_codec_alg_priv_t *ctx,
+static vpx_codec_err_t vp9_set_reference(vpx_codec_alg_priv_t *ctx,
                                          int ctr_id,
                                          va_list args) {
 
@@ -548,7 +548,7 @@ static vpx_codec_err_t vp8_set_reference(vpx_codec_alg_priv_t *ctx,
 
 }
 
-static vpx_codec_err_t vp8_get_reference(vpx_codec_alg_priv_t *ctx,
+static vpx_codec_err_t vp9_get_reference(vpx_codec_alg_priv_t *ctx,
                                          int ctr_id,
                                          va_list args) {
 
@@ -639,8 +639,8 @@ static vpx_codec_err_t vp8_get_frame_corrupted(vpx_codec_alg_priv_t *ctx,
 }
 
 vpx_codec_ctrl_fn_map_t vp8_ctf_maps[] = {
-  {VP8_SET_REFERENCE,             vp8_set_reference},
-  {VP8_COPY_REFERENCE,            vp8_get_reference},
+  {VP8_SET_REFERENCE,             vp9_set_reference},
+  {VP8_COPY_REFERENCE,            vp9_get_reference},
   {VP8_SET_POSTPROC,              vp8_set_postproc},
   {VP8_SET_DBG_COLOR_REF_FRAME,   vp8_set_dbg_options},
   {VP8_SET_DBG_COLOR_MB_MODES,    vp8_set_dbg_options},