1. 13 Aug, 2015 1 commit
  2. 12 Aug, 2015 2 commits
  3. 21 Jul, 2015 1 commit
    • Yaowu Xu's avatar
      vpx_dsp/bitreader.h: vp9_->vpx_ · bf82514b
      Yaowu Xu authored
      Replace vp9_ in names to vpx_ as they are not codec specific.
      
      Change-Id: I2e583aa63dee769353ada4b42417aa15c4074ebb
      bf82514b
  4. 20 Jul, 2015 1 commit
  5. 17 Jul, 2015 1 commit
  6. 09 Jul, 2015 1 commit
  7. 07 Jul, 2015 1 commit
  8. 04 Jun, 2015 1 commit
  9. 02 Jun, 2015 1 commit
  10. 28 Apr, 2015 1 commit
  11. 19 Feb, 2015 1 commit
  12. 03 Feb, 2015 1 commit
    • Yunqing Wang's avatar
      vp9_dthread: pass frame counts to decoder functions · 85a9bc04
      Yunqing Wang authored
      The current multi-threaded tile decoder requires that the videoes
      are encoded with frame_parallel_decoding_mode = 1. This requirement
      is not necessary, and is better to be removed. This patch includes
      the first part of the work.
      
      Change-Id: Ic7695fb3cfe13f9022582c9f0edd2aa6e2e36d28
      85a9bc04
  13. 06 Mar, 2014 1 commit
  14. 05 Mar, 2014 1 commit
  15. 24 Jan, 2014 1 commit
  16. 20 Dec, 2013 1 commit
  17. 02 Dec, 2013 1 commit
    • Dmitry Kovalev's avatar
      Using local variable for token_cache. · 5ab920d2
      Dmitry Kovalev authored
      The difference with the old code is that originally the whole token_cache
      was initialized with zeros at the beginning of decode_coefs() function.
      Now we set several zero values explicitly with "token_cache[scan[c]] = 0".
      
      Change-Id: I88cc5031f01d13012d1a4491739c36cb44f9401e
      5ab920d2
  18. 14 Nov, 2013 1 commit
    • Deb Mukherjee's avatar
      Simplifies band-getting with a static array · cfcd5c4f
      Deb Mukherjee authored
      Simplifies the code by implementing band mapping with static arrays.
      A lot of the code complexity introduced in a previous patch
      disappears.
      
      Change-Id: Ia3fac36e594fb5ad2d55ae141c58bba4c55c2d28
      cfcd5c4f
  19. 12 Nov, 2013 2 commits
  20. 05 Nov, 2013 1 commit
    • Deb Mukherjee's avatar
      token_cache changes in decoder · 3a833ea3
      Deb Mukherjee authored
      Removes stack-alocation of token_cache in decode_coefs function
      
      Seems to achieve about 1% decode speed improvement as tested on
      25 480p videos.
      
      Change-Id: I8e7eb3361fa09d9654dfad0677a6d606701fdc6e
      3a833ea3
  21. 31 Oct, 2013 1 commit
    • Dmitry Kovalev's avatar
      Reducing the number of foreach_transformed_block() calls. · 47b6030d
      Dmitry Kovalev authored
      The change doesn't affect the bitstream. It changes the order or function
      calls and affects how we reconstruct intra- and inter-blocks. Speed up is
      about 1...1.5%.
      
      For intra-blocks:
        Before:
          for each transform block read tokens
          for each transform block do prediction
          for each transform block do inverse transform
        Now:
          for each transform block
            read tokens
            do prediction
            do inverse transform
      
      For inter-blocks:
        Before:
          for each transform block read tokens
          for each transform block do inverse transform
        Now:
          for each transform block
            read tokens
            do inverse transform
      
      Change-Id: I12a79bf1aa5a18c351b8010369bd3ff1deae1570
      47b6030d
  22. 21 Oct, 2013 1 commit
  23. 26 Aug, 2013 1 commit
  24. 25 Jul, 2013 1 commit
    • Dmitry Kovalev's avatar
      General cleanups. · 7131cb0e
      Dmitry Kovalev authored
      Removing unused constants, macros, and function declarations. Using
      ROUND_POWER_OF_TWO macro, vp9_zero, vp9_copy where possible. Moving
      #include from *.h to *.c. Merging for loops for motion vectors.
      
      Change-Id: Ic3bf841764a2bb177128bb3a6d7aa8f68229cd13
      7131cb0e
  25. 18 Jun, 2013 1 commit
  26. 07 May, 2013 1 commit
  27. 25 Apr, 2013 1 commit
    • John Koleszar's avatar
      Fix incorrect dequant used in detokenize · e40a7690
      John Koleszar authored
      The quantizer can vary per-plane, and the dequantization vector is
      available in the per-plane part of MACROBLOCKD. The previous code would
      incorrectly use the Y quantizer for the whole macroblock.
      
      Change-Id: I3ab418aef9168ea0ddcfa4b7c0be32ae48b536d7
      e40a7690
  28. 22 Apr, 2013 1 commit
  29. 18 Apr, 2013 1 commit
  30. 10 Apr, 2013 1 commit
  31. 04 Apr, 2013 1 commit
  32. 03 Apr, 2013 1 commit
    • John Koleszar's avatar
      Remove special case vp9_decode_coefs_4x4 · 1e5f25ec
      John Koleszar authored
      This code was only called in the BPRED case, but had no real special
      case associated with it. Made BPRED behave like all other modes. No
      bitstream change.
      
      Change-Id: I87ba11fe723928b6314d094979011228d5ba006f
      1e5f25ec
  33. 05 Mar, 2013 1 commit
    • Ronald S. Bultje's avatar
      Make superblocks independent of macroblock code and data. · 111ca421
      Ronald S. Bultje authored
      Split macroblock and superblock tokenization and detokenization
      functions and coefficient-related data structs so that the bitstream
      layout and related code of superblock coefficients looks less like it's
      a hack to fit macroblocks in superblocks.
      
      In addition, unify chroma transform size selection from luma transform
      size (i.e. always use the same size, as long as it fits the predictor);
      in practice, this means 32x32 and 64x64 superblocks using the 16x16 luma
      transform will now use the 16x16 (instead of the 8x8) chroma transform,
      and 64x64 superblocks using the 32x32 luma transform will now use the
      32x32 (instead of the 16x16) chroma transform.
      
      Lastly, add a trellis optimize function for 32x32 transform blocks.
      
      HD gains about 0.3%, STDHD about 0.15% and derf about 0.1%. There's
      a few negative points here and there that I might want to analyze
      a little closer.
      
      Change-Id: Ibad7c3ddfe1acfc52771dfc27c03e9783e054430
      111ca421
  34. 10 Jan, 2013 1 commit
  35. 08 Jan, 2013 1 commit
  36. 18 Dec, 2012 1 commit
  37. 07 Dec, 2012 1 commit
    • Ronald S. Bultje's avatar
      32x32 transform for superblocks. · c456b35f
      Ronald S. Bultje authored
      This adds Debargha's DCT/DWT hybrid and a regular 32x32 DCT, and adds
      code all over the place to wrap that in the bitstream/encoder/decoder/RD.
      
      Some implementation notes (these probably need careful review):
      - token range is extended by 1 bit, since the value range out of this
        transform is [-16384,16383].
      - the coefficients coming out of the FDCT are manually scaled back by
        1 bit, or else they won't fit in int16_t (they are 17 bits). Because
        of this, the RD error scoring does not right-shift the MSE score by
        two (unlike for 4x4/8x8/16x16).
      - to compensate for this loss in precision, the quantizer is halved
        also. This is currently a little hacky.
      - FDCT and IDCT is double-only right now. Needs a fixed-point impl.
      - There are no default probabilities for the 32x32 transform yet; I'm
        simply using the 16x16 luma ones. A future commit will add newly
        generated probabilities for all transforms.
      - No ADST version. I don't think we'll add one for this level; if an
        ADST is desired, transform-size selection can scale back to 16x16
        or lower, and use an ADST at that level.
      
      Additional notes specific to Debargha's DWT/DCT hybrid:
      - coefficient scale is different for the top/left 16x16 (DCT-over-DWT)
        block than for the rest (DWT pixel differences) of the block. Therefore,
        RD error scoring isn't easily scalable between coefficient and pixel
        domain. Thus, unfortunately, we need to compute the RD distortion in
        the pixel domain until we figure out how to scale these appropriately.
      
      Change-Id: I00386f20f35d7fabb19aba94c8162f8aee64ef2b
      c456b35f
  38. 30 Nov, 2012 1 commit