2010-05-18 17:58:33 +02:00
|
|
|
/*
|
2010-09-09 14:16:39 +02:00
|
|
|
* Copyright (c) 2010 The WebM project authors. All Rights Reserved.
|
2010-05-18 17:58:33 +02:00
|
|
|
*
|
2010-06-18 18:39:21 +02:00
|
|
|
* Use of this source code is governed by a BSD-style license
|
2010-06-04 22:19:40 +02:00
|
|
|
* that can be found in the LICENSE file in the root of the source
|
|
|
|
* tree. An additional intellectual property rights grant can be found
|
2010-06-18 18:39:21 +02:00
|
|
|
* in the file PATENTS. All contributing project authors may
|
2010-06-04 22:19:40 +02:00
|
|
|
* be found in the AUTHORS file in the root of the source tree.
|
2010-05-18 17:58:33 +02:00
|
|
|
*/
|
|
|
|
|
|
|
|
|
|
|
|
#ifndef tokenize_h
|
|
|
|
#define tokenize_h
|
|
|
|
|
2011-02-10 20:41:38 +01:00
|
|
|
#include "vp8/common/entropy.h"
|
2010-05-18 17:58:33 +02:00
|
|
|
#include "block.h"
|
|
|
|
|
2012-10-30 20:58:42 +01:00
|
|
|
void vp9_tokenize_initialize();
|
2010-05-18 17:58:33 +02:00
|
|
|
|
2012-07-14 00:21:29 +02:00
|
|
|
typedef struct {
|
|
|
|
short Token;
|
|
|
|
short Extra;
|
2010-09-16 16:00:04 +02:00
|
|
|
} TOKENVALUE;
|
|
|
|
|
2012-07-14 00:21:29 +02:00
|
|
|
typedef struct {
|
|
|
|
const vp8_prob *context_tree;
|
|
|
|
short Extra;
|
|
|
|
unsigned char Token;
|
|
|
|
unsigned char skip_eob_node;
|
2010-05-18 17:58:33 +02:00
|
|
|
} TOKENEXTRA;
|
|
|
|
|
|
|
|
int rd_cost_mby(MACROBLOCKD *);
|
|
|
|
|
2012-10-30 05:07:40 +01:00
|
|
|
extern int vp9_mby_is_skippable_4x4(MACROBLOCKD *xd, int has_y2_block);
|
|
|
|
extern int vp9_mbuv_is_skippable_4x4(MACROBLOCKD *xd);
|
|
|
|
extern int vp9_mby_is_skippable_8x8(MACROBLOCKD *xd, int has_y2_block);
|
|
|
|
extern int vp9_mbuv_is_skippable_8x8(MACROBLOCKD *xd);
|
|
|
|
extern int vp9_mby_is_skippable_16x16(MACROBLOCKD *xd);
|
2012-02-22 03:10:18 +01:00
|
|
|
|
2010-05-18 17:58:33 +02:00
|
|
|
#ifdef ENTROPY_STATS
|
|
|
|
void init_context_counters();
|
|
|
|
void print_context_counters();
|
|
|
|
|
2012-08-15 12:00:53 +02:00
|
|
|
extern INT64 context_counters[BLOCK_TYPES][COEF_BANDS]
|
|
|
|
[PREV_COEF_CONTEXTS][MAX_ENTROPY_TOKENS];
|
|
|
|
extern INT64 context_counters_8x8[BLOCK_TYPES_8X8][COEF_BANDS]
|
|
|
|
[PREV_COEF_CONTEXTS][MAX_ENTROPY_TOKENS];
|
|
|
|
extern INT64 context_counters_16x16[BLOCK_TYPES_16X16][COEF_BANDS]
|
|
|
|
[PREV_COEF_CONTEXTS][MAX_ENTROPY_TOKENS];
|
2012-08-03 02:03:14 +02:00
|
|
|
#endif
|
|
|
|
|
2012-10-31 01:12:12 +01:00
|
|
|
extern const int *vp9_dct_value_cost_ptr;
|
Add trellis quantization.
Replace the exponential search for optimal rounding during
quantization with a linear Viterbi trellis and enable it
by default when using --best.
Right now this operates on top of the output of the adaptive
zero-bin quantizer in vp8_regular_quantize_b() and gives a small
gain.
It can be tested as a replacement for that quantizer by
enabling the call to vp8_strict_quantize_b(), which uses
normal rounding and no zero bin offset.
Ultimately, the quantizer will have to become a function of lambda
in order to take advantage of activity masking, since there is
limited ability to change the quantization factor itself.
However, currently vp8_strict_quantize_b() plus the trellis
quantizer (which is lambda-dependent) loses to
vp8_regular_quantize_b() alone (which is not) on my test clip.
Patch Set 3:
Fix an issue related to the cost evaluation of successor
states when a coefficient is reduced to zero. With this
issue fixed, now the trellis search almost exactly matches
the exponential search.
Patch Set 2:
Overall, the goal of this patch set is to make "trellis"
search to produce encodings that match the exponential
search version. There are three main differences between
Patch Set 2 and 1:
a. Patch set 1 did not properly account for the scale of
2nd order error, so patch set 2 disable it all together
for 2nd blocks.
b. Patch set 1 was not consistent on when to enable the
the quantization optimization. Patch set 2 restore the
condition to be consistent.
c. Patch set 1 checks quantized level L-1, and L for any
input coefficient was quantized to L. Patch set 2 limits
the candidate coefficient to those that were rounded up
to L. It is worth noting here that a strategy to check
L and L+1 for coefficients that were truncated down to L
might work.
(a and b get trellis quant to basically match the exponential
search on all mid/low rate encodings on cif set, without
a, b, trellis quant can hurt the psnr by 0.2 to .3db at
200kbps for some cif clips)
(c gets trellis quant to match the exponential search
to match at Q0 encoding, without c, trellis quant can be
1.5 to 2db lower for encodings with fixed Q at 0 on most
derf cif clips)
Change-Id: Ib1a043b665d75fbf00cb0257b7c18e90eebab95e
2010-07-02 23:35:53 +02:00
|
|
|
/* TODO: The Token field should be broken out into a separate char array to
|
|
|
|
* improve cache locality, since it's needed for costing when the rest of the
|
|
|
|
* fields are not.
|
|
|
|
*/
|
2012-10-31 01:12:12 +01:00
|
|
|
extern const TOKENVALUE *vp9_dct_value_tokens_ptr;
|
2010-05-18 17:58:33 +02:00
|
|
|
|
|
|
|
#endif /* tokenize_h */
|