Merge "Framework changes in nzc to allow more flexibility" into experimental
This commit is contained in:
commit
e3955007df
@ -672,4 +672,10 @@ static TX_SIZE get_uv_tx_size(const MACROBLOCKD *xd) {
|
||||
}
|
||||
return tx_size_uv;
|
||||
}
|
||||
|
||||
#if CONFIG_CODE_NONZEROCOUNT
|
||||
static int get_nzc_used(TX_SIZE tx_size) {
|
||||
return (tx_size >= TX_16X16);
|
||||
}
|
||||
#endif
|
||||
#endif // VP9_COMMON_VP9_BLOCKD_H_
|
||||
|
@ -790,10 +790,10 @@ static const unsigned int default_nzc_counts_16x16[MAX_NZC_CONTEXTS]
|
||||
}
|
||||
}, {
|
||||
{
|
||||
{ 29408, 11758, 8023, 10123, 6705, 2468, 369, 17, 10, 5 },
|
||||
{ 19408, 31758, 16023, 10123, 6705, 2468, 369, 17, 10, 5 },
|
||||
{ 9612, 13874, 13329, 13022, 6500, 3250, 300, 12, 6, 3 },
|
||||
}, {
|
||||
{ 29408, 11758, 8023, 10123, 6705, 2468, 369, 17, 10, 5 },
|
||||
{ 22408, 11758, 8023, 10123, 6705, 2468, 369, 17, 10, 5 },
|
||||
{ 9612, 13874, 13329, 13022, 6500, 3250, 300, 12, 6, 3 },
|
||||
}
|
||||
}
|
||||
@ -821,7 +821,7 @@ static const unsigned int default_nzc_counts_32x32[MAX_NZC_CONTEXTS]
|
||||
}
|
||||
}, {
|
||||
{
|
||||
{ 29408, 11758, 8023, 10123, 6705, 2468, 369, 17, 10, 5, 2, 1 },
|
||||
{ 19408, 11758, 8023, 10123, 6705, 2468, 369, 17, 10, 5, 2, 1 },
|
||||
{ 9612, 13874, 13329, 13022, 6500, 3250, 300, 12, 6, 3, 2, 1 },
|
||||
}, {
|
||||
{ 29408, 11758, 8023, 10123, 6705, 2468, 369, 17, 10, 5, 2, 1 },
|
||||
|
@ -2646,8 +2646,8 @@ static unsigned int get_nzc_4x4_uv(VP9_COMMON *cm, MODE_INFO *m,
|
||||
if (m->mbmi.mb_skip_coeff || !mb_in_cur_tile(cm, mb_row - r, mb_col - c))
|
||||
return 0;
|
||||
else
|
||||
return get_nzc_4x4_uv_sb32(
|
||||
&m->mbmi, mb16_to_sb32_index(mb_row, mb_col, block));
|
||||
return get_nzc_4x4_uv_sb32(
|
||||
&m->mbmi, mb16_to_sb32_index(mb_row, mb_col, block));
|
||||
} else {
|
||||
return get_nzc_4x4_uv_mb16(mi, block);
|
||||
}
|
||||
@ -3291,6 +3291,7 @@ static void update_nzc(VP9_COMMON *cm,
|
||||
int ref,
|
||||
int type) {
|
||||
int e, c;
|
||||
if (!get_nzc_used(tx_size)) return;
|
||||
c = codenzc(nzc);
|
||||
if (tx_size == TX_32X32)
|
||||
cm->fc.nzc_counts_32x32[nzc_context][ref][type][c]++;
|
||||
@ -3604,10 +3605,10 @@ void vp9_adapt_coef_probs(VP9_COMMON *cm) {
|
||||
}
|
||||
|
||||
#if CONFIG_CODE_NONZEROCOUNT
|
||||
static void adapt_nzc_probs(VP9_COMMON *cm,
|
||||
int block_size,
|
||||
int count_sat,
|
||||
int update_factor) {
|
||||
static void adapt_nzc_probs_common(VP9_COMMON *cm,
|
||||
TX_SIZE tx_size,
|
||||
int count_sat,
|
||||
int update_factor) {
|
||||
int c, r, b, n;
|
||||
int count, factor;
|
||||
unsigned int nzc_branch_ct[NZC32X32_NODES][2];
|
||||
@ -3618,19 +3619,20 @@ static void adapt_nzc_probs(VP9_COMMON *cm,
|
||||
vp9_prob *pre_nzc_probs;
|
||||
unsigned int *nzc_counts;
|
||||
|
||||
if (block_size == 32) {
|
||||
if (!get_nzc_used(tx_size)) return;
|
||||
if (tx_size == TX_32X32) {
|
||||
tokens = NZC32X32_TOKENS;
|
||||
nzc_tree = vp9_nzc32x32_tree;
|
||||
dst_nzc_probs = cm->fc.nzc_probs_32x32[0][0][0];
|
||||
pre_nzc_probs = cm->fc.pre_nzc_probs_32x32[0][0][0];
|
||||
nzc_counts = cm->fc.nzc_counts_32x32[0][0][0];
|
||||
} else if (block_size == 16) {
|
||||
} else if (tx_size == TX_16X16) {
|
||||
tokens = NZC16X16_TOKENS;
|
||||
nzc_tree = vp9_nzc16x16_tree;
|
||||
dst_nzc_probs = cm->fc.nzc_probs_16x16[0][0][0];
|
||||
pre_nzc_probs = cm->fc.pre_nzc_probs_16x16[0][0][0];
|
||||
nzc_counts = cm->fc.nzc_counts_16x16[0][0][0];
|
||||
} else if (block_size == 8) {
|
||||
} else if (tx_size == TX_8X8) {
|
||||
tokens = NZC8X8_TOKENS;
|
||||
nzc_tree = vp9_nzc8x8_tree;
|
||||
dst_nzc_probs = cm->fc.nzc_probs_8x8[0][0][0];
|
||||
@ -3667,6 +3669,9 @@ static void adapt_nzc_probs(VP9_COMMON *cm,
|
||||
static void adapt_nzc_pcat(VP9_COMMON *cm, int count_sat, int update_factor) {
|
||||
int c, t;
|
||||
int count, factor;
|
||||
if (!(get_nzc_used(TX_4X4) || get_nzc_used(TX_8X8) ||
|
||||
get_nzc_used(TX_16X16) || get_nzc_used(TX_32X32)))
|
||||
return;
|
||||
for (c = 0; c < MAX_NZC_CONTEXTS; ++c) {
|
||||
for (t = 0; t < NZC_TOKENS_EXTRA; ++t) {
|
||||
int bits = vp9_extranzcbits[t + NZC_TOKENS_NOEXTRA];
|
||||
@ -3716,10 +3721,10 @@ void vp9_adapt_nzc_probs(VP9_COMMON *cm) {
|
||||
count_sat = COEF_COUNT_SAT;
|
||||
}
|
||||
|
||||
adapt_nzc_probs(cm, 4, count_sat, update_factor);
|
||||
adapt_nzc_probs(cm, 8, count_sat, update_factor);
|
||||
adapt_nzc_probs(cm, 16, count_sat, update_factor);
|
||||
adapt_nzc_probs(cm, 32, count_sat, update_factor);
|
||||
adapt_nzc_probs_common(cm, TX_4X4, count_sat, update_factor);
|
||||
adapt_nzc_probs_common(cm, TX_8X8, count_sat, update_factor);
|
||||
adapt_nzc_probs_common(cm, TX_16X16, count_sat, update_factor);
|
||||
adapt_nzc_probs_common(cm, TX_32X32, count_sat, update_factor);
|
||||
adapt_nzc_pcat(cm, count_sat, update_factor);
|
||||
}
|
||||
#endif // CONFIG_CODE_NONZEROCOUNT
|
||||
|
@ -1157,6 +1157,7 @@ static uint16_t read_nzc(VP9_COMMON *const cm,
|
||||
BOOL_DECODER* const bc) {
|
||||
int c, e;
|
||||
uint16_t nzc;
|
||||
if (!get_nzc_used(tx_size)) return 0;
|
||||
if (tx_size == TX_32X32) {
|
||||
c = treed_read(bc, vp9_nzc32x32_tree,
|
||||
cm->fc.nzc_probs_32x32[nzc_context][ref][type]);
|
||||
|
@ -151,12 +151,39 @@ static void mb_init_dequantizer(VP9D_COMP *pbi, MACROBLOCKD *mb) {
|
||||
}
|
||||
}
|
||||
|
||||
#if CONFIG_CODE_NONZEROCOUNT
|
||||
static void propagate_nzcs(VP9_COMMON *cm, MACROBLOCKD *xd) {
|
||||
MODE_INFO *m = xd->mode_info_context;
|
||||
BLOCK_SIZE_TYPE sb_type = m->mbmi.sb_type;
|
||||
const int mis = cm->mode_info_stride;
|
||||
int n;
|
||||
if (sb_type == BLOCK_SIZE_SB64X64) {
|
||||
for (n = 0; n < 16; ++n) {
|
||||
int i = n >> 2;
|
||||
int j = n & 3;
|
||||
if (i == 0 && j == 0) continue;
|
||||
vpx_memcpy((m + j + mis * i)->mbmi.nzcs, m->mbmi.nzcs,
|
||||
384 * sizeof(m->mbmi.nzcs[0]));
|
||||
}
|
||||
} else if (sb_type == BLOCK_SIZE_SB32X32) {
|
||||
for (n = 0; n < 4; ++n) {
|
||||
int i = n >> 1;
|
||||
int j = n & 1;
|
||||
if (i == 0 && j == 0) continue;
|
||||
vpx_memcpy((m + j + mis * i)->mbmi.nzcs, m->mbmi.nzcs,
|
||||
384 * sizeof(m->mbmi.nzcs[0]));
|
||||
}
|
||||
}
|
||||
}
|
||||
#endif
|
||||
|
||||
/* skip_recon_mb() is Modified: Instead of writing the result to predictor buffer and then copying it
|
||||
* to dst buffer, we can write the result directly to dst buffer. This eliminates unnecessary copy.
|
||||
*/
|
||||
static void skip_recon_mb(VP9D_COMP *pbi, MACROBLOCKD *xd,
|
||||
int mb_row, int mb_col) {
|
||||
BLOCK_SIZE_TYPE sb_type = xd->mode_info_context->mbmi.sb_type;
|
||||
MODE_INFO *m = xd->mode_info_context;
|
||||
BLOCK_SIZE_TYPE sb_type = m->mbmi.sb_type;
|
||||
|
||||
if (xd->mode_info_context->mbmi.ref_frame == INTRA_FRAME) {
|
||||
if (sb_type == BLOCK_SIZE_SB64X64) {
|
||||
@ -196,6 +223,10 @@ static void skip_recon_mb(VP9D_COMP *pbi, MACROBLOCKD *xd,
|
||||
mb_row, mb_col);
|
||||
}
|
||||
}
|
||||
#if CONFIG_CODE_NONZEROCOUNT
|
||||
vpx_memset(m->mbmi.nzcs, 0, 384 * sizeof(m->mbmi.nzcs[0]));
|
||||
propagate_nzcs(&pbi->common, xd);
|
||||
#endif
|
||||
}
|
||||
|
||||
static void decode_16x16(VP9D_COMP *pbi, MACROBLOCKD *xd,
|
||||
@ -468,16 +499,16 @@ static void decode_superblock64(VP9D_COMP *pbi, MACROBLOCKD *xd,
|
||||
MODE_INFO *mi = xd->mode_info_context;
|
||||
const int mis = pc->mode_info_stride;
|
||||
|
||||
assert(xd->mode_info_context->mbmi.sb_type == BLOCK_SIZE_SB64X64);
|
||||
assert(mi->mbmi.sb_type == BLOCK_SIZE_SB64X64);
|
||||
|
||||
if (pbi->common.frame_type != KEY_FRAME)
|
||||
vp9_setup_interp_filters(xd, xd->mode_info_context->mbmi.interp_filter, pc);
|
||||
vp9_setup_interp_filters(xd, mi->mbmi.interp_filter, pc);
|
||||
|
||||
// re-initialize macroblock dequantizer before detokenization
|
||||
if (xd->segmentation_enabled)
|
||||
mb_init_dequantizer(pbi, xd);
|
||||
|
||||
if (xd->mode_info_context->mbmi.mb_skip_coeff) {
|
||||
if (mi->mbmi.mb_skip_coeff) {
|
||||
vp9_reset_sb64_tokens_context(xd);
|
||||
|
||||
/* Special case: Force the loopfilter to skip when eobtotal and
|
||||
@ -632,6 +663,9 @@ static void decode_superblock64(VP9D_COMP *pbi, MACROBLOCKD *xd,
|
||||
default: assert(0);
|
||||
}
|
||||
}
|
||||
#if CONFIG_CODE_NONZEROCOUNT
|
||||
propagate_nzcs(&pbi->common, xd);
|
||||
#endif
|
||||
}
|
||||
|
||||
static void decode_superblock32(VP9D_COMP *pbi, MACROBLOCKD *xd,
|
||||
@ -639,18 +673,19 @@ static void decode_superblock32(VP9D_COMP *pbi, MACROBLOCKD *xd,
|
||||
BOOL_DECODER* const bc) {
|
||||
int n, eobtotal;
|
||||
VP9_COMMON *const pc = &pbi->common;
|
||||
MODE_INFO *mi = xd->mode_info_context;
|
||||
const int mis = pc->mode_info_stride;
|
||||
|
||||
assert(xd->mode_info_context->mbmi.sb_type == BLOCK_SIZE_SB32X32);
|
||||
assert(mi->mbmi.sb_type == BLOCK_SIZE_SB32X32);
|
||||
|
||||
if (pbi->common.frame_type != KEY_FRAME)
|
||||
vp9_setup_interp_filters(xd, xd->mode_info_context->mbmi.interp_filter, pc);
|
||||
vp9_setup_interp_filters(xd, mi->mbmi.interp_filter, pc);
|
||||
|
||||
// re-initialize macroblock dequantizer before detokenization
|
||||
if (xd->segmentation_enabled)
|
||||
mb_init_dequantizer(pbi, xd);
|
||||
|
||||
if (xd->mode_info_context->mbmi.mb_skip_coeff) {
|
||||
if (mi->mbmi.mb_skip_coeff) {
|
||||
vp9_reset_sb_tokens_context(xd);
|
||||
|
||||
/* Special case: Force the loopfilter to skip when eobtotal and
|
||||
@ -661,7 +696,7 @@ static void decode_superblock32(VP9D_COMP *pbi, MACROBLOCKD *xd,
|
||||
}
|
||||
|
||||
/* do prediction */
|
||||
if (xd->mode_info_context->mbmi.ref_frame == INTRA_FRAME) {
|
||||
if (mi->mbmi.ref_frame == INTRA_FRAME) {
|
||||
vp9_build_intra_predictors_sby_s(xd);
|
||||
vp9_build_intra_predictors_sbuv_s(xd);
|
||||
} else {
|
||||
@ -674,13 +709,13 @@ static void decode_superblock32(VP9D_COMP *pbi, MACROBLOCKD *xd,
|
||||
/* dequantization and idct */
|
||||
eobtotal = vp9_decode_sb_tokens(pbi, xd, bc);
|
||||
if (eobtotal == 0) { // skip loopfilter
|
||||
xd->mode_info_context->mbmi.mb_skip_coeff = 1;
|
||||
mi->mbmi.mb_skip_coeff = 1;
|
||||
if (mb_col + 1 < pc->mb_cols)
|
||||
xd->mode_info_context[1].mbmi.mb_skip_coeff = 1;
|
||||
mi[1].mbmi.mb_skip_coeff = 1;
|
||||
if (mb_row + 1 < pc->mb_rows) {
|
||||
xd->mode_info_context[mis].mbmi.mb_skip_coeff = 1;
|
||||
mi[mis].mbmi.mb_skip_coeff = 1;
|
||||
if (mb_col + 1 < pc->mb_cols)
|
||||
xd->mode_info_context[mis + 1].mbmi.mb_skip_coeff = 1;
|
||||
mi[mis + 1].mbmi.mb_skip_coeff = 1;
|
||||
}
|
||||
} else {
|
||||
switch (xd->mode_info_context->mbmi.txfm_size) {
|
||||
@ -793,6 +828,9 @@ static void decode_superblock32(VP9D_COMP *pbi, MACROBLOCKD *xd,
|
||||
default: assert(0);
|
||||
}
|
||||
}
|
||||
#if CONFIG_CODE_NONZEROCOUNT
|
||||
propagate_nzcs(&pbi->common, xd);
|
||||
#endif
|
||||
}
|
||||
|
||||
static void decode_macroblock(VP9D_COMP *pbi, MACROBLOCKD *xd,
|
||||
@ -1130,23 +1168,24 @@ static void init_frame(VP9D_COMP *pbi) {
|
||||
#if CONFIG_CODE_NONZEROCOUNT
|
||||
static void read_nzc_probs_common(VP9_COMMON *cm,
|
||||
BOOL_DECODER* const bc,
|
||||
int block_size) {
|
||||
TX_SIZE tx_size) {
|
||||
int c, r, b, t;
|
||||
int tokens, nodes;
|
||||
vp9_prob *nzc_probs;
|
||||
vp9_prob upd;
|
||||
|
||||
if (!get_nzc_used(tx_size)) return;
|
||||
if (!vp9_read_bit(bc)) return;
|
||||
|
||||
if (block_size == 32) {
|
||||
if (tx_size == TX_32X32) {
|
||||
tokens = NZC32X32_TOKENS;
|
||||
nzc_probs = cm->fc.nzc_probs_32x32[0][0][0];
|
||||
upd = NZC_UPDATE_PROB_32X32;
|
||||
} else if (block_size == 16) {
|
||||
} else if (tx_size == TX_16X16) {
|
||||
tokens = NZC16X16_TOKENS;
|
||||
nzc_probs = cm->fc.nzc_probs_16x16[0][0][0];
|
||||
upd = NZC_UPDATE_PROB_16X16;
|
||||
} else if (block_size == 8) {
|
||||
} else if (tx_size == TX_8X8) {
|
||||
tokens = NZC8X8_TOKENS;
|
||||
nzc_probs = cm->fc.nzc_probs_8x8[0][0][0];
|
||||
upd = NZC_UPDATE_PROB_8X8;
|
||||
@ -1175,6 +1214,9 @@ static void read_nzc_probs_common(VP9_COMMON *cm,
|
||||
static void read_nzc_pcat_probs(VP9_COMMON *cm, BOOL_DECODER* const bc) {
|
||||
int c, t, b;
|
||||
vp9_prob upd = NZC_UPDATE_PROB_PCAT;
|
||||
if (!(get_nzc_used(TX_4X4) || get_nzc_used(TX_8X8) ||
|
||||
get_nzc_used(TX_16X16) || get_nzc_used(TX_32X32)))
|
||||
return;
|
||||
if (!vp9_read_bit(bc)) {
|
||||
return;
|
||||
}
|
||||
@ -1193,13 +1235,13 @@ static void read_nzc_pcat_probs(VP9_COMMON *cm, BOOL_DECODER* const bc) {
|
||||
|
||||
static void read_nzc_probs(VP9_COMMON *cm,
|
||||
BOOL_DECODER* const bc) {
|
||||
read_nzc_probs_common(cm, bc, 4);
|
||||
read_nzc_probs_common(cm, bc, TX_4X4);
|
||||
if (cm->txfm_mode != ONLY_4X4)
|
||||
read_nzc_probs_common(cm, bc, 8);
|
||||
read_nzc_probs_common(cm, bc, TX_8X8);
|
||||
if (cm->txfm_mode > ALLOW_8X8)
|
||||
read_nzc_probs_common(cm, bc, 16);
|
||||
read_nzc_probs_common(cm, bc, TX_16X16);
|
||||
if (cm->txfm_mode > ALLOW_16X16)
|
||||
read_nzc_probs_common(cm, bc, 32);
|
||||
read_nzc_probs_common(cm, bc, TX_32X32);
|
||||
#ifdef NZC_PCAT_UPDATE
|
||||
read_nzc_pcat_probs(cm, bc);
|
||||
#endif
|
||||
@ -1208,7 +1250,7 @@ static void read_nzc_probs(VP9_COMMON *cm,
|
||||
|
||||
static void read_coef_probs_common(BOOL_DECODER* const bc,
|
||||
vp9_coeff_probs *coef_probs,
|
||||
int block_types) {
|
||||
TX_SIZE tx_size) {
|
||||
#if CONFIG_MODELCOEFPROB && MODEL_BASED_UPDATE
|
||||
const int entropy_nodes_update = UNCONSTRAINED_UPDATE_NODES;
|
||||
#else
|
||||
@ -1218,13 +1260,19 @@ static void read_coef_probs_common(BOOL_DECODER* const bc,
|
||||
int i, j, k, l, m;
|
||||
|
||||
if (vp9_read_bit(bc)) {
|
||||
for (i = 0; i < block_types; i++) {
|
||||
for (i = 0; i < BLOCK_TYPES; i++) {
|
||||
for (j = 0; j < REF_TYPES; j++) {
|
||||
for (k = 0; k < COEF_BANDS; k++) {
|
||||
for (l = 0; l < PREV_COEF_CONTEXTS; l++) {
|
||||
#if CONFIG_CODE_NONZEROCOUNT
|
||||
const int mstart = get_nzc_used(tx_size);
|
||||
#else
|
||||
const int mstart = 0;
|
||||
#endif
|
||||
if (l >= 3 && k == 0)
|
||||
continue;
|
||||
for (m = CONFIG_CODE_NONZEROCOUNT; m < entropy_nodes_update; m++) {
|
||||
|
||||
for (m = mstart; m < entropy_nodes_update; m++) {
|
||||
vp9_prob *const p = coef_probs[i][j][k][l] + m;
|
||||
|
||||
if (vp9_read(bc, vp9_coef_update_prob[m])) {
|
||||
@ -1245,16 +1293,16 @@ static void read_coef_probs_common(BOOL_DECODER* const bc,
|
||||
static void read_coef_probs(VP9D_COMP *pbi, BOOL_DECODER* const bc) {
|
||||
VP9_COMMON *const pc = &pbi->common;
|
||||
|
||||
read_coef_probs_common(bc, pc->fc.coef_probs_4x4, BLOCK_TYPES);
|
||||
read_coef_probs_common(bc, pc->fc.coef_probs_4x4, TX_4X4);
|
||||
|
||||
if (pbi->common.txfm_mode != ONLY_4X4)
|
||||
read_coef_probs_common(bc, pc->fc.coef_probs_8x8, BLOCK_TYPES);
|
||||
read_coef_probs_common(bc, pc->fc.coef_probs_8x8, TX_8X8);
|
||||
|
||||
if (pbi->common.txfm_mode > ALLOW_8X8)
|
||||
read_coef_probs_common(bc, pc->fc.coef_probs_16x16, BLOCK_TYPES);
|
||||
read_coef_probs_common(bc, pc->fc.coef_probs_16x16, TX_16X16);
|
||||
|
||||
if (pbi->common.txfm_mode > ALLOW_16X16)
|
||||
read_coef_probs_common(bc, pc->fc.coef_probs_32x32, BLOCK_TYPES);
|
||||
read_coef_probs_common(bc, pc->fc.coef_probs_32x32, TX_32X32);
|
||||
}
|
||||
|
||||
static void update_frame_size(VP9D_COMP *pbi) {
|
||||
|
@ -113,8 +113,10 @@ static int decode_coefs(VP9D_COMP *dx, const MACROBLOCKD *xd,
|
||||
vp9_coeff_count *coef_counts;
|
||||
const int ref = xd->mode_info_context->mbmi.ref_frame != INTRA_FRAME;
|
||||
#if CONFIG_CODE_NONZEROCOUNT
|
||||
const int nzc_used = get_nzc_used(txfm_size);
|
||||
uint16_t nzc = 0;
|
||||
uint16_t nzc_expected = xd->mode_info_context->mbmi.nzcs[block_idx];
|
||||
uint16_t nzc_expected =
|
||||
nzc_used ? xd->mode_info_context->mbmi.nzcs[block_idx] : 0;
|
||||
#endif
|
||||
const int *scan, *nb;
|
||||
uint8_t token_cache[1024];
|
||||
@ -245,24 +247,25 @@ static int decode_coefs(VP9D_COMP *dx, const MACROBLOCKD *xd,
|
||||
if (c >= seg_eob)
|
||||
break;
|
||||
#if CONFIG_CODE_NONZEROCOUNT
|
||||
if (nzc == nzc_expected)
|
||||
if (nzc_used && nzc == nzc_expected)
|
||||
break;
|
||||
#endif
|
||||
prob = coef_probs[type][ref][get_coef_band(scan, txfm_size, c)][pt];
|
||||
#if CONFIG_CODE_NONZEROCOUNT == 0
|
||||
fc->eob_branch_counts[txfm_size][type][ref]
|
||||
[get_coef_band(scan, txfm_size, c)][pt]++;
|
||||
if (!vp9_read(br, prob[EOB_CONTEXT_NODE]))
|
||||
break;
|
||||
#if CONFIG_CODE_NONZEROCOUNT
|
||||
if (!nzc_used)
|
||||
#endif
|
||||
if (!vp9_read(br, prob[EOB_CONTEXT_NODE]))
|
||||
break;
|
||||
SKIP_START:
|
||||
if (c >= seg_eob)
|
||||
break;
|
||||
#if CONFIG_CODE_NONZEROCOUNT
|
||||
if (nzc == nzc_expected)
|
||||
if (nzc_used && nzc == nzc_expected)
|
||||
break;
|
||||
// decode zero node only if there are zeros left
|
||||
if (seg_eob - nzc_expected - c + nzc > 0)
|
||||
if (!nzc_used || seg_eob - nzc_expected - c + nzc > 0)
|
||||
#endif
|
||||
if (!vp9_read(br, prob[ZERO_CONTEXT_NODE])) {
|
||||
INCREMENT_COUNT(ZERO_TOKEN);
|
||||
@ -330,10 +333,17 @@ SKIP_START:
|
||||
WRITE_COEF_CONTINUE(val, DCT_VAL_CATEGORY6);
|
||||
}
|
||||
|
||||
#if CONFIG_CODE_NONZEROCOUNT == 0
|
||||
if (c < seg_eob)
|
||||
coef_counts[type][ref][get_coef_band(scan, txfm_size, c)]
|
||||
[pt][DCT_EOB_TOKEN]++;
|
||||
#if CONFIG_CODE_NONZEROCOUNT
|
||||
if (!nzc_used)
|
||||
#endif
|
||||
if (c < seg_eob)
|
||||
coef_counts[type][ref][get_coef_band(scan, txfm_size, c)]
|
||||
[pt][DCT_EOB_TOKEN]++;
|
||||
#if CONFIG_CODE_NONZEROCOUNT
|
||||
if (!nzc_used)
|
||||
xd->mode_info_context->mbmi.nzcs[block_idx] = nzc;
|
||||
else
|
||||
assert(nzc == nzc_expected);
|
||||
#endif
|
||||
|
||||
A0[aidx] = L0[lidx] = c > 0;
|
||||
|
@ -1087,14 +1087,18 @@ static void write_mb_modes_kf(const VP9_COMP *cpi,
|
||||
}
|
||||
|
||||
#if CONFIG_CODE_NONZEROCOUNT
|
||||
static void write_nzc(VP9_COMMON *const cm,
|
||||
static void write_nzc(VP9_COMP *const cpi,
|
||||
uint16_t nzc,
|
||||
int nzc_context,
|
||||
TX_SIZE tx_size,
|
||||
int ref,
|
||||
int type,
|
||||
vp9_writer* const bc) {
|
||||
VP9_COMMON *const cm = &cpi->common;
|
||||
int c, e;
|
||||
// if (!cpi->dummy_packing && cm->current_video_frame == 27)
|
||||
// printf("nzc: %d, tx_size: %d\n", nzc, tx_size);
|
||||
if (!get_nzc_used(tx_size)) return;
|
||||
c = codenzc(nzc);
|
||||
if (tx_size == TX_32X32) {
|
||||
write_token(bc, vp9_nzc32x32_tree,
|
||||
@ -1152,44 +1156,44 @@ static void write_nzcs_sb64(VP9_COMP *cpi,
|
||||
case TX_32X32:
|
||||
for (j = 0; j < 256; j += 64) {
|
||||
nzc_context = vp9_get_nzc_context_y_sb64(cm, m, mb_row, mb_col, j);
|
||||
write_nzc(cm, m->mbmi.nzcs[j], nzc_context, TX_32X32, ref, 0, bc);
|
||||
write_nzc(cpi, m->mbmi.nzcs[j], nzc_context, TX_32X32, ref, 0, bc);
|
||||
}
|
||||
for (j = 256; j < 384; j += 64) {
|
||||
nzc_context = vp9_get_nzc_context_uv_sb64(cm, m, mb_row, mb_col, j);
|
||||
write_nzc(cm, m->mbmi.nzcs[j], nzc_context, TX_32X32, ref, 1, bc);
|
||||
write_nzc(cpi, m->mbmi.nzcs[j], nzc_context, TX_32X32, ref, 1, bc);
|
||||
}
|
||||
break;
|
||||
|
||||
case TX_16X16:
|
||||
for (j = 0; j < 256; j += 16) {
|
||||
nzc_context = vp9_get_nzc_context_y_sb64(cm, m, mb_row, mb_col, j);
|
||||
write_nzc(cm, m->mbmi.nzcs[j], nzc_context, TX_16X16, ref, 0, bc);
|
||||
write_nzc(cpi, m->mbmi.nzcs[j], nzc_context, TX_16X16, ref, 0, bc);
|
||||
}
|
||||
for (j = 256; j < 384; j += 16) {
|
||||
nzc_context = vp9_get_nzc_context_uv_sb64(cm, m, mb_row, mb_col, j);
|
||||
write_nzc(cm, m->mbmi.nzcs[j], nzc_context, TX_16X16, ref, 1, bc);
|
||||
write_nzc(cpi, m->mbmi.nzcs[j], nzc_context, TX_16X16, ref, 1, bc);
|
||||
}
|
||||
break;
|
||||
|
||||
case TX_8X8:
|
||||
for (j = 0; j < 256; j += 4) {
|
||||
nzc_context = vp9_get_nzc_context_y_sb64(cm, m, mb_row, mb_col, j);
|
||||
write_nzc(cm, m->mbmi.nzcs[j], nzc_context, TX_8X8, ref, 0, bc);
|
||||
write_nzc(cpi, m->mbmi.nzcs[j], nzc_context, TX_8X8, ref, 0, bc);
|
||||
}
|
||||
for (j = 256; j < 384; j += 4) {
|
||||
nzc_context = vp9_get_nzc_context_uv_sb64(cm, m, mb_row, mb_col, j);
|
||||
write_nzc(cm, m->mbmi.nzcs[j], nzc_context, TX_8X8, ref, 1, bc);
|
||||
write_nzc(cpi, m->mbmi.nzcs[j], nzc_context, TX_8X8, ref, 1, bc);
|
||||
}
|
||||
break;
|
||||
|
||||
case TX_4X4:
|
||||
for (j = 0; j < 256; ++j) {
|
||||
nzc_context = vp9_get_nzc_context_y_sb64(cm, m, mb_row, mb_col, j);
|
||||
write_nzc(cm, m->mbmi.nzcs[j], nzc_context, TX_4X4, ref, 0, bc);
|
||||
write_nzc(cpi, m->mbmi.nzcs[j], nzc_context, TX_4X4, ref, 0, bc);
|
||||
}
|
||||
for (j = 256; j < 384; ++j) {
|
||||
nzc_context = vp9_get_nzc_context_uv_sb64(cm, m, mb_row, mb_col, j);
|
||||
write_nzc(cm, m->mbmi.nzcs[j], nzc_context, TX_4X4, ref, 1, bc);
|
||||
write_nzc(cpi, m->mbmi.nzcs[j], nzc_context, TX_4X4, ref, 1, bc);
|
||||
}
|
||||
break;
|
||||
|
||||
@ -1219,44 +1223,44 @@ static void write_nzcs_sb32(VP9_COMP *cpi,
|
||||
case TX_32X32:
|
||||
for (j = 0; j < 64; j += 64) {
|
||||
nzc_context = vp9_get_nzc_context_y_sb32(cm, m, mb_row, mb_col, j);
|
||||
write_nzc(cm, m->mbmi.nzcs[j], nzc_context, TX_32X32, ref, 0, bc);
|
||||
write_nzc(cpi, m->mbmi.nzcs[j], nzc_context, TX_32X32, ref, 0, bc);
|
||||
}
|
||||
for (j = 64; j < 96; j += 16) {
|
||||
nzc_context = vp9_get_nzc_context_uv_sb32(cm, m, mb_row, mb_col, j);
|
||||
write_nzc(cm, m->mbmi.nzcs[j], nzc_context, TX_16X16, ref, 1, bc);
|
||||
write_nzc(cpi, m->mbmi.nzcs[j], nzc_context, TX_16X16, ref, 1, bc);
|
||||
}
|
||||
break;
|
||||
|
||||
case TX_16X16:
|
||||
for (j = 0; j < 64; j += 16) {
|
||||
nzc_context = vp9_get_nzc_context_y_sb32(cm, m, mb_row, mb_col, j);
|
||||
write_nzc(cm, m->mbmi.nzcs[j], nzc_context, TX_16X16, ref, 0, bc);
|
||||
write_nzc(cpi, m->mbmi.nzcs[j], nzc_context, TX_16X16, ref, 0, bc);
|
||||
}
|
||||
for (j = 64; j < 96; j += 16) {
|
||||
nzc_context = vp9_get_nzc_context_uv_sb32(cm, m, mb_row, mb_col, j);
|
||||
write_nzc(cm, m->mbmi.nzcs[j], nzc_context, TX_16X16, ref, 1, bc);
|
||||
write_nzc(cpi, m->mbmi.nzcs[j], nzc_context, TX_16X16, ref, 1, bc);
|
||||
}
|
||||
break;
|
||||
|
||||
case TX_8X8:
|
||||
for (j = 0; j < 64; j += 4) {
|
||||
nzc_context = vp9_get_nzc_context_y_sb32(cm, m, mb_row, mb_col, j);
|
||||
write_nzc(cm, m->mbmi.nzcs[j], nzc_context, TX_8X8, ref, 0, bc);
|
||||
write_nzc(cpi, m->mbmi.nzcs[j], nzc_context, TX_8X8, ref, 0, bc);
|
||||
}
|
||||
for (j = 64; j < 96; j += 4) {
|
||||
nzc_context = vp9_get_nzc_context_uv_sb32(cm, m, mb_row, mb_col, j);
|
||||
write_nzc(cm, m->mbmi.nzcs[j], nzc_context, TX_8X8, ref, 1, bc);
|
||||
write_nzc(cpi, m->mbmi.nzcs[j], nzc_context, TX_8X8, ref, 1, bc);
|
||||
}
|
||||
break;
|
||||
|
||||
case TX_4X4:
|
||||
for (j = 0; j < 64; ++j) {
|
||||
nzc_context = vp9_get_nzc_context_y_sb32(cm, m, mb_row, mb_col, j);
|
||||
write_nzc(cm, m->mbmi.nzcs[j], nzc_context, TX_4X4, ref, 0, bc);
|
||||
write_nzc(cpi, m->mbmi.nzcs[j], nzc_context, TX_4X4, ref, 0, bc);
|
||||
}
|
||||
for (j = 64; j < 96; ++j) {
|
||||
nzc_context = vp9_get_nzc_context_uv_sb32(cm, m, mb_row, mb_col, j);
|
||||
write_nzc(cm, m->mbmi.nzcs[j], nzc_context, TX_4X4, ref, 1, bc);
|
||||
write_nzc(cpi, m->mbmi.nzcs[j], nzc_context, TX_4X4, ref, 1, bc);
|
||||
}
|
||||
break;
|
||||
|
||||
@ -1286,28 +1290,28 @@ static void write_nzcs_mb16(VP9_COMP *cpi,
|
||||
case TX_16X16:
|
||||
for (j = 0; j < 16; j += 16) {
|
||||
nzc_context = vp9_get_nzc_context_y_mb16(cm, m, mb_row, mb_col, j);
|
||||
write_nzc(cm, m->mbmi.nzcs[j], nzc_context, TX_16X16, ref, 0, bc);
|
||||
write_nzc(cpi, m->mbmi.nzcs[j], nzc_context, TX_16X16, ref, 0, bc);
|
||||
}
|
||||
for (j = 16; j < 24; j += 4) {
|
||||
nzc_context = vp9_get_nzc_context_uv_mb16(cm, m, mb_row, mb_col, j);
|
||||
write_nzc(cm, m->mbmi.nzcs[j], nzc_context, TX_8X8, ref, 1, bc);
|
||||
write_nzc(cpi, m->mbmi.nzcs[j], nzc_context, TX_8X8, ref, 1, bc);
|
||||
}
|
||||
break;
|
||||
|
||||
case TX_8X8:
|
||||
for (j = 0; j < 16; j += 4) {
|
||||
nzc_context = vp9_get_nzc_context_y_mb16(cm, m, mb_row, mb_col, j);
|
||||
write_nzc(cm, m->mbmi.nzcs[j], nzc_context, TX_8X8, ref, 0, bc);
|
||||
write_nzc(cpi, m->mbmi.nzcs[j], nzc_context, TX_8X8, ref, 0, bc);
|
||||
}
|
||||
if (mi->mode == I8X8_PRED || mi->mode == SPLITMV) {
|
||||
for (j = 16; j < 24; ++j) {
|
||||
nzc_context = vp9_get_nzc_context_uv_mb16(cm, m, mb_row, mb_col, j);
|
||||
write_nzc(cm, m->mbmi.nzcs[j], nzc_context, TX_4X4, ref, 1, bc);
|
||||
write_nzc(cpi, m->mbmi.nzcs[j], nzc_context, TX_4X4, ref, 1, bc);
|
||||
}
|
||||
} else {
|
||||
for (j = 16; j < 24; j += 4) {
|
||||
nzc_context = vp9_get_nzc_context_uv_mb16(cm, m, mb_row, mb_col, j);
|
||||
write_nzc(cm, m->mbmi.nzcs[j], nzc_context, TX_8X8, ref, 1, bc);
|
||||
write_nzc(cpi, m->mbmi.nzcs[j], nzc_context, TX_8X8, ref, 1, bc);
|
||||
}
|
||||
}
|
||||
break;
|
||||
@ -1315,11 +1319,11 @@ static void write_nzcs_mb16(VP9_COMP *cpi,
|
||||
case TX_4X4:
|
||||
for (j = 0; j < 16; ++j) {
|
||||
nzc_context = vp9_get_nzc_context_y_mb16(cm, m, mb_row, mb_col, j);
|
||||
write_nzc(cm, m->mbmi.nzcs[j], nzc_context, TX_4X4, ref, 0, bc);
|
||||
write_nzc(cpi, m->mbmi.nzcs[j], nzc_context, TX_4X4, ref, 0, bc);
|
||||
}
|
||||
for (j = 16; j < 24; ++j) {
|
||||
nzc_context = vp9_get_nzc_context_uv_mb16(cm, m, mb_row, mb_col, j);
|
||||
write_nzc(cm, m->mbmi.nzcs[j], nzc_context, TX_4X4, ref, 1, bc);
|
||||
write_nzc(cpi, m->mbmi.nzcs[j], nzc_context, TX_4X4, ref, 1, bc);
|
||||
}
|
||||
break;
|
||||
|
||||
@ -1838,7 +1842,7 @@ static void build_coeff_contexts(VP9_COMP *cpi) {
|
||||
#if CONFIG_CODE_NONZEROCOUNT
|
||||
static void update_nzc_probs_common(VP9_COMP* cpi,
|
||||
vp9_writer* const bc,
|
||||
int block_size) {
|
||||
TX_SIZE tx_size) {
|
||||
VP9_COMMON *cm = &cpi->common;
|
||||
int c, r, b, t;
|
||||
int update[2] = {0, 0};
|
||||
@ -1851,7 +1855,8 @@ static void update_nzc_probs_common(VP9_COMP* cpi,
|
||||
unsigned int (*nzc_branch_ct)[2];
|
||||
vp9_prob upd;
|
||||
|
||||
if (block_size == 32) {
|
||||
if (!get_nzc_used(tx_size)) return;
|
||||
if (tx_size == TX_32X32) {
|
||||
tokens = NZC32X32_TOKENS;
|
||||
nzc_tree = vp9_nzc32x32_tree;
|
||||
old_nzc_probs = cm->fc.nzc_probs_32x32[0][0][0];
|
||||
@ -1859,7 +1864,7 @@ static void update_nzc_probs_common(VP9_COMP* cpi,
|
||||
nzc_counts = cm->fc.nzc_counts_32x32[0][0][0];
|
||||
nzc_branch_ct = cpi->frame_nzc_branch_ct_32x32[0][0][0];
|
||||
upd = NZC_UPDATE_PROB_32X32;
|
||||
} else if (block_size == 16) {
|
||||
} else if (tx_size == TX_16X16) {
|
||||
tokens = NZC16X16_TOKENS;
|
||||
nzc_tree = vp9_nzc16x16_tree;
|
||||
old_nzc_probs = cm->fc.nzc_probs_16x16[0][0][0];
|
||||
@ -1867,7 +1872,7 @@ static void update_nzc_probs_common(VP9_COMP* cpi,
|
||||
nzc_counts = cm->fc.nzc_counts_16x16[0][0][0];
|
||||
nzc_branch_ct = cpi->frame_nzc_branch_ct_16x16[0][0][0];
|
||||
upd = NZC_UPDATE_PROB_16X16;
|
||||
} else if (block_size == 8) {
|
||||
} else if (tx_size == TX_8X8) {
|
||||
tokens = NZC8X8_TOKENS;
|
||||
nzc_tree = vp9_nzc8x8_tree;
|
||||
old_nzc_probs = cm->fc.nzc_probs_8x8[0][0][0];
|
||||
@ -1974,6 +1979,9 @@ static void update_nzc_pcat_probs(VP9_COMP *cpi, vp9_writer* const bc) {
|
||||
int update[2] = {0, 0};
|
||||
int savings = 0;
|
||||
vp9_prob upd = NZC_UPDATE_PROB_PCAT;
|
||||
if (!(get_nzc_used(TX_4X4) || get_nzc_used(TX_8X8) ||
|
||||
get_nzc_used(TX_16X16) || get_nzc_used(TX_32X32)))
|
||||
return;
|
||||
for (c = 0; c < MAX_NZC_CONTEXTS; ++c) {
|
||||
for (t = 0; t < NZC_TOKENS_EXTRA; ++t) {
|
||||
int bits = vp9_extranzcbits[t + NZC_TOKENS_NOEXTRA];
|
||||
@ -2040,13 +2048,13 @@ static void update_nzc_pcat_probs(VP9_COMP *cpi, vp9_writer* const bc) {
|
||||
|
||||
static void update_nzc_probs(VP9_COMP* cpi,
|
||||
vp9_writer* const bc) {
|
||||
update_nzc_probs_common(cpi, bc, 4);
|
||||
update_nzc_probs_common(cpi, bc, TX_4X4);
|
||||
if (cpi->common.txfm_mode != ONLY_4X4)
|
||||
update_nzc_probs_common(cpi, bc, 8);
|
||||
update_nzc_probs_common(cpi, bc, TX_8X8);
|
||||
if (cpi->common.txfm_mode > ALLOW_8X8)
|
||||
update_nzc_probs_common(cpi, bc, 16);
|
||||
update_nzc_probs_common(cpi, bc, TX_16X16);
|
||||
if (cpi->common.txfm_mode > ALLOW_16X16)
|
||||
update_nzc_probs_common(cpi, bc, 32);
|
||||
update_nzc_probs_common(cpi, bc, TX_32X32);
|
||||
#ifdef NZC_PCAT_UPDATE
|
||||
update_nzc_pcat_probs(cpi, bc);
|
||||
#endif
|
||||
@ -2065,7 +2073,7 @@ static void update_coef_probs_common(vp9_writer* const bc,
|
||||
vp9_coeff_probs *new_frame_coef_probs,
|
||||
vp9_coeff_probs *old_frame_coef_probs,
|
||||
vp9_coeff_stats *frame_branch_ct,
|
||||
int block_types) {
|
||||
TX_SIZE tx_size) {
|
||||
int i, j, k, l, t;
|
||||
int update[2] = {0, 0};
|
||||
int savings;
|
||||
@ -2076,14 +2084,19 @@ static void update_coef_probs_common(vp9_writer* const bc,
|
||||
#endif
|
||||
// vp9_prob bestupd = find_coef_update_prob(cpi);
|
||||
|
||||
#if CONFIG_CODE_NONZEROCOUNT
|
||||
const int tstart = get_nzc_used(tx_size);
|
||||
#else
|
||||
const int tstart = 0;
|
||||
#endif
|
||||
/* dry run to see if there is any udpate at all needed */
|
||||
savings = 0;
|
||||
for (i = 0; i < block_types; ++i) {
|
||||
for (i = 0; i < BLOCK_TYPES; ++i) {
|
||||
for (j = 0; j < REF_TYPES; ++j) {
|
||||
for (k = 0; k < COEF_BANDS; ++k) {
|
||||
// int prev_coef_savings[ENTROPY_NODES] = {0};
|
||||
for (l = 0; l < PREV_COEF_CONTEXTS; ++l) {
|
||||
for (t = CONFIG_CODE_NONZEROCOUNT; t < entropy_nodes_update; ++t) {
|
||||
for (t = tstart; t < entropy_nodes_update; ++t) {
|
||||
vp9_prob newp = new_frame_coef_probs[i][j][k][l][t];
|
||||
const vp9_prob oldp = old_frame_coef_probs[i][j][k][l][t];
|
||||
const vp9_prob upd = vp9_coef_update_prob[t];
|
||||
@ -2131,13 +2144,13 @@ static void update_coef_probs_common(vp9_writer* const bc,
|
||||
return;
|
||||
}
|
||||
vp9_write_bit(bc, 1);
|
||||
for (i = 0; i < block_types; ++i) {
|
||||
for (i = 0; i < BLOCK_TYPES; ++i) {
|
||||
for (j = 0; j < REF_TYPES; ++j) {
|
||||
for (k = 0; k < COEF_BANDS; ++k) {
|
||||
// int prev_coef_savings[ENTROPY_NODES] = {0};
|
||||
for (l = 0; l < PREV_COEF_CONTEXTS; ++l) {
|
||||
// calc probs and branch cts for this frame only
|
||||
for (t = CONFIG_CODE_NONZEROCOUNT; t < entropy_nodes_update; ++t) {
|
||||
for (t = tstart; t < entropy_nodes_update; ++t) {
|
||||
vp9_prob newp = new_frame_coef_probs[i][j][k][l][t];
|
||||
vp9_prob *oldp = old_frame_coef_probs[i][j][k][l] + t;
|
||||
const vp9_prob upd = vp9_coef_update_prob[t];
|
||||
@ -2201,7 +2214,7 @@ static void update_coef_probs(VP9_COMP* const cpi, vp9_writer* const bc) {
|
||||
cpi->frame_coef_probs_4x4,
|
||||
cpi->common.fc.coef_probs_4x4,
|
||||
cpi->frame_branch_ct_4x4,
|
||||
BLOCK_TYPES);
|
||||
TX_4X4);
|
||||
|
||||
/* do not do this if not even allowed */
|
||||
if (cpi->common.txfm_mode != ONLY_4X4) {
|
||||
@ -2213,7 +2226,7 @@ static void update_coef_probs(VP9_COMP* const cpi, vp9_writer* const bc) {
|
||||
cpi->frame_coef_probs_8x8,
|
||||
cpi->common.fc.coef_probs_8x8,
|
||||
cpi->frame_branch_ct_8x8,
|
||||
BLOCK_TYPES);
|
||||
TX_8X8);
|
||||
}
|
||||
|
||||
if (cpi->common.txfm_mode > ALLOW_8X8) {
|
||||
@ -2225,7 +2238,7 @@ static void update_coef_probs(VP9_COMP* const cpi, vp9_writer* const bc) {
|
||||
cpi->frame_coef_probs_16x16,
|
||||
cpi->common.fc.coef_probs_16x16,
|
||||
cpi->frame_branch_ct_16x16,
|
||||
BLOCK_TYPES);
|
||||
TX_16X16);
|
||||
}
|
||||
|
||||
if (cpi->common.txfm_mode > ALLOW_16X16) {
|
||||
@ -2237,7 +2250,7 @@ static void update_coef_probs(VP9_COMP* const cpi, vp9_writer* const bc) {
|
||||
cpi->frame_coef_probs_32x32,
|
||||
cpi->common.fc.coef_probs_32x32,
|
||||
cpi->frame_branch_ct_32x32,
|
||||
BLOCK_TYPES);
|
||||
TX_32X32);
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -1960,9 +1960,10 @@ static void gather_nzcs_mb16(VP9_COMMON *const cm,
|
||||
|
||||
static void gather_nzcs_sb32(VP9_COMMON *const cm,
|
||||
MACROBLOCKD *xd) {
|
||||
int i, j;
|
||||
MODE_INFO *m = xd->mode_info_context;
|
||||
int mis = cm->mode_info_stride;
|
||||
int i, j;
|
||||
|
||||
vpx_memset(m->mbmi.nzcs, 0,
|
||||
384 * sizeof(xd->mode_info_context->mbmi.nzcs[0]));
|
||||
switch (xd->mode_info_context->mbmi.txfm_size) {
|
||||
@ -2004,9 +2005,10 @@ static void gather_nzcs_sb32(VP9_COMMON *const cm,
|
||||
|
||||
static void gather_nzcs_sb64(VP9_COMMON *const cm,
|
||||
MACROBLOCKD *xd) {
|
||||
int i, j;
|
||||
MODE_INFO *m = xd->mode_info_context;
|
||||
int mis = cm->mode_info_stride;
|
||||
int i, j;
|
||||
|
||||
vpx_memset(xd->mode_info_context->mbmi.nzcs, 0,
|
||||
384 * sizeof(xd->mode_info_context->mbmi.nzcs[0]));
|
||||
switch (xd->mode_info_context->mbmi.txfm_size) {
|
||||
|
@ -573,6 +573,7 @@ static void optimize_b(VP9_COMMON *const cm,
|
||||
// The current implementation uses a suboptimal approach to account for
|
||||
// the nzc rates somewhat, but in reality the optimization approach needs
|
||||
// to change substantially.
|
||||
const int nzc_used = get_nzc_used(tx_size);
|
||||
uint16_t nzc = xd->nzcs[ib];
|
||||
uint16_t nzc0, nzc1;
|
||||
uint16_t final_nzc = 0, final_nzc_exp;
|
||||
@ -649,7 +650,7 @@ static void optimize_b(VP9_COMMON *const cm,
|
||||
memset(best_index, 0, sizeof(best_index));
|
||||
/* Initialize the sentinel node of the trellis. */
|
||||
#if CONFIG_CODE_NONZEROCOUNT
|
||||
tokens[eob][0].rate = nzc_cost[nzc];
|
||||
tokens[eob][0].rate = nzc_used ? nzc_cost[nzc] : 0;
|
||||
#else
|
||||
tokens[eob][0].rate = 0;
|
||||
#endif
|
||||
@ -734,8 +735,10 @@ static void optimize_b(VP9_COMMON *const cm,
|
||||
#if CONFIG_CODE_NONZEROCOUNT
|
||||
// Account for rate drop because of the nzc change.
|
||||
// TODO(debargha): Find a better solution
|
||||
rate0 -= nzc_cost[nzc0] - nzc_cost[nzc0 - 1];
|
||||
rate1 -= nzc_cost[nzc1] - nzc_cost[nzc1 - 1];
|
||||
if (nzc_used) {
|
||||
rate0 -= nzc_cost[nzc0] - nzc_cost[nzc0 - 1];
|
||||
rate1 -= nzc_cost[nzc1] - nzc_cost[nzc1 - 1];
|
||||
}
|
||||
#endif
|
||||
} else {
|
||||
t0 = t1 = (vp9_dct_value_tokens_ptr + x)->Token;
|
||||
|
@ -149,10 +149,10 @@ const MODE_DEFINITION vp9_mode_order[MAX_MODES] = {
|
||||
|
||||
static void fill_token_costs(vp9_coeff_count *c,
|
||||
vp9_coeff_probs *p,
|
||||
int block_type_counts) {
|
||||
TX_SIZE tx_size) {
|
||||
int i, j, k, l;
|
||||
|
||||
for (i = 0; i < block_type_counts; i++)
|
||||
for (i = 0; i < BLOCK_TYPES; i++)
|
||||
for (j = 0; j < REF_TYPES; j++)
|
||||
for (k = 0; k < COEF_BANDS; k++)
|
||||
for (l = 0; l < PREV_COEF_CONTEXTS; l++) {
|
||||
@ -163,26 +163,26 @@ static void fill_token_costs(vp9_coeff_count *c,
|
||||
}
|
||||
|
||||
#if CONFIG_CODE_NONZEROCOUNT
|
||||
static void fill_nzc_costs(VP9_COMP *cpi, int block_size) {
|
||||
static void fill_nzc_costs(VP9_COMP *cpi, TX_SIZE tx_size) {
|
||||
int nzc_context, r, b, nzc, values;
|
||||
int cost[16];
|
||||
values = block_size * block_size + 1;
|
||||
values = (16 << (2 * tx_size)) + 1;
|
||||
|
||||
for (nzc_context = 0; nzc_context < MAX_NZC_CONTEXTS; ++nzc_context) {
|
||||
for (r = 0; r < REF_TYPES; ++r) {
|
||||
for (b = 0; b < BLOCK_TYPES; ++b) {
|
||||
unsigned int *nzc_costs;
|
||||
if (block_size == 4) {
|
||||
if (tx_size == TX_4X4) {
|
||||
vp9_cost_tokens(cost,
|
||||
cpi->common.fc.nzc_probs_4x4[nzc_context][r][b],
|
||||
vp9_nzc4x4_tree);
|
||||
nzc_costs = cpi->mb.nzc_costs_4x4[nzc_context][r][b];
|
||||
} else if (block_size == 8) {
|
||||
} else if (tx_size == TX_8X8) {
|
||||
vp9_cost_tokens(cost,
|
||||
cpi->common.fc.nzc_probs_8x8[nzc_context][r][b],
|
||||
vp9_nzc8x8_tree);
|
||||
nzc_costs = cpi->mb.nzc_costs_8x8[nzc_context][r][b];
|
||||
} else if (block_size == 16) {
|
||||
} else if (tx_size == TX_16X16) {
|
||||
vp9_cost_tokens(cost,
|
||||
cpi->common.fc.nzc_probs_16x16[nzc_context][r][b],
|
||||
vp9_nzc16x16_tree);
|
||||
@ -308,18 +308,18 @@ void vp9_initialize_rd_consts(VP9_COMP *cpi, int qindex) {
|
||||
}
|
||||
|
||||
fill_token_costs(cpi->mb.token_costs[TX_4X4],
|
||||
cpi->common.fc.coef_probs_4x4, BLOCK_TYPES);
|
||||
cpi->common.fc.coef_probs_4x4, TX_4X4);
|
||||
fill_token_costs(cpi->mb.token_costs[TX_8X8],
|
||||
cpi->common.fc.coef_probs_8x8, BLOCK_TYPES);
|
||||
cpi->common.fc.coef_probs_8x8, TX_8X8);
|
||||
fill_token_costs(cpi->mb.token_costs[TX_16X16],
|
||||
cpi->common.fc.coef_probs_16x16, BLOCK_TYPES);
|
||||
cpi->common.fc.coef_probs_16x16, TX_16X16);
|
||||
fill_token_costs(cpi->mb.token_costs[TX_32X32],
|
||||
cpi->common.fc.coef_probs_32x32, BLOCK_TYPES);
|
||||
cpi->common.fc.coef_probs_32x32, TX_32X32);
|
||||
#if CONFIG_CODE_NONZEROCOUNT
|
||||
fill_nzc_costs(cpi, 4);
|
||||
fill_nzc_costs(cpi, 8);
|
||||
fill_nzc_costs(cpi, 16);
|
||||
fill_nzc_costs(cpi, 32);
|
||||
fill_nzc_costs(cpi, TX_4X4);
|
||||
fill_nzc_costs(cpi, TX_8X8);
|
||||
fill_nzc_costs(cpi, TX_16X16);
|
||||
fill_nzc_costs(cpi, TX_32X32);
|
||||
#endif
|
||||
|
||||
/*rough estimate for costing*/
|
||||
@ -449,13 +449,13 @@ static INLINE int cost_coeffs(VP9_COMMON *const cm, MACROBLOCK *mb,
|
||||
sizeof(ENTROPY_CONTEXT_PLANES)/sizeof(ENTROPY_CONTEXT);
|
||||
|
||||
#if CONFIG_CODE_NONZEROCOUNT
|
||||
const int nzc_used = get_nzc_used(tx_size);
|
||||
int nzc_context = vp9_get_nzc_context(cm, xd, ib);
|
||||
unsigned int *nzc_cost;
|
||||
#else
|
||||
#endif
|
||||
const int segment_id = xd->mode_info_context->mbmi.segment_id;
|
||||
vp9_prob (*coef_probs)[REF_TYPES][COEF_BANDS][PREV_COEF_CONTEXTS]
|
||||
[ENTROPY_NODES];
|
||||
#endif
|
||||
int seg_eob, default_eob;
|
||||
uint8_t token_cache[1024];
|
||||
|
||||
@ -475,9 +475,8 @@ static INLINE int cost_coeffs(VP9_COMMON *const cm, MACROBLOCK *mb,
|
||||
l_ec = *l;
|
||||
#if CONFIG_CODE_NONZEROCOUNT
|
||||
nzc_cost = mb->nzc_costs_4x4[nzc_context][ref][type];
|
||||
#else
|
||||
coef_probs = cm->fc.coef_probs_4x4;
|
||||
#endif
|
||||
coef_probs = cm->fc.coef_probs_4x4;
|
||||
seg_eob = 16;
|
||||
if (tx_type == ADST_DCT) {
|
||||
scan = vp9_row_scan_4x4;
|
||||
@ -504,9 +503,8 @@ static INLINE int cost_coeffs(VP9_COMMON *const cm, MACROBLOCK *mb,
|
||||
}
|
||||
#if CONFIG_CODE_NONZEROCOUNT
|
||||
nzc_cost = mb->nzc_costs_8x8[nzc_context][ref][type];
|
||||
#else
|
||||
coef_probs = cm->fc.coef_probs_8x8;
|
||||
#endif
|
||||
coef_probs = cm->fc.coef_probs_8x8;
|
||||
seg_eob = 64;
|
||||
break;
|
||||
}
|
||||
@ -524,9 +522,8 @@ static INLINE int cost_coeffs(VP9_COMMON *const cm, MACROBLOCK *mb,
|
||||
}
|
||||
#if CONFIG_CODE_NONZEROCOUNT
|
||||
nzc_cost = mb->nzc_costs_16x16[nzc_context][ref][type];
|
||||
#else
|
||||
coef_probs = cm->fc.coef_probs_16x16;
|
||||
#endif
|
||||
coef_probs = cm->fc.coef_probs_16x16;
|
||||
seg_eob = 256;
|
||||
if (type == PLANE_TYPE_UV) {
|
||||
a_ec = (a[0] + a[1] + a1[0] + a1[1]) != 0;
|
||||
@ -541,9 +538,8 @@ static INLINE int cost_coeffs(VP9_COMMON *const cm, MACROBLOCK *mb,
|
||||
scan = vp9_default_zig_zag1d_32x32;
|
||||
#if CONFIG_CODE_NONZEROCOUNT
|
||||
nzc_cost = mb->nzc_costs_32x32[nzc_context][ref][type];
|
||||
#else
|
||||
coef_probs = cm->fc.coef_probs_32x32;
|
||||
#endif
|
||||
coef_probs = cm->fc.coef_probs_32x32;
|
||||
seg_eob = 1024;
|
||||
if (type == PLANE_TYPE_UV) {
|
||||
ENTROPY_CONTEXT *a2, *a3, *l2, *l3;
|
||||
@ -571,10 +567,11 @@ static INLINE int cost_coeffs(VP9_COMMON *const cm, MACROBLOCK *mb,
|
||||
nb = vp9_get_coef_neighbors_handle(scan, &pad);
|
||||
default_eob = seg_eob;
|
||||
|
||||
#if CONFIG_CODE_NONZEROCOUNT == 0
|
||||
if (vp9_segfeature_active(xd, segment_id, SEG_LVL_SKIP))
|
||||
seg_eob = 0;
|
||||
#if CONFIG_CODE_NONZEROCOUNT
|
||||
if (!nzc_used)
|
||||
#endif
|
||||
if (vp9_segfeature_active(xd, segment_id, SEG_LVL_SKIP))
|
||||
seg_eob = 0;
|
||||
|
||||
{
|
||||
#if CONFIG_CODE_NONZEROCOUNT
|
||||
@ -598,13 +595,14 @@ static INLINE int cost_coeffs(VP9_COMMON *const cm, MACROBLOCK *mb,
|
||||
pt = vp9_get_coef_context(scan, nb, pad, token_cache, c + 1, default_eob);
|
||||
}
|
||||
#if CONFIG_CODE_NONZEROCOUNT
|
||||
cost += nzc_cost[nzc];
|
||||
#else
|
||||
if (c < seg_eob)
|
||||
cost += mb->token_costs[tx_size][type][ref]
|
||||
[get_coef_band(scan, tx_size, c)]
|
||||
[pt][DCT_EOB_TOKEN];
|
||||
if (nzc_used)
|
||||
cost += nzc_cost[nzc];
|
||||
else
|
||||
#endif
|
||||
if (c < seg_eob)
|
||||
cost += mb->token_costs[tx_size][type][ref]
|
||||
[get_coef_band(scan, tx_size, c)]
|
||||
[pt][DCT_EOB_TOKEN];
|
||||
}
|
||||
|
||||
// is eob first coefficient;
|
||||
|
@ -134,7 +134,8 @@ static void tokenize_b(VP9_COMP *cpi,
|
||||
ENTROPY_CONTEXT *a, *l, *a1, *l1, *a2, *l2, *a3, *l3, a_ec, l_ec;
|
||||
uint8_t token_cache[1024];
|
||||
#if CONFIG_CODE_NONZEROCOUNT
|
||||
int zerosleft, nzc = 0;
|
||||
const int nzc_used = get_nzc_used(tx_size);
|
||||
int zerosleft = 0, nzc = 0;
|
||||
if (eob == 0)
|
||||
assert(xd->nzcs[ib] == 0);
|
||||
#endif
|
||||
@ -255,7 +256,8 @@ static void tokenize_b(VP9_COMP *cpi,
|
||||
int token;
|
||||
int v = 0;
|
||||
#if CONFIG_CODE_NONZEROCOUNT
|
||||
zerosleft = seg_eob - xd->nzcs[ib] - c + nzc;
|
||||
if (nzc_used)
|
||||
zerosleft = seg_eob - xd->nzcs[ib] - c + nzc;
|
||||
#endif
|
||||
if (c < eob) {
|
||||
const int rc = scan[c];
|
||||
@ -266,20 +268,22 @@ static void tokenize_b(VP9_COMP *cpi,
|
||||
token = vp9_dct_value_tokens_ptr[v].Token;
|
||||
} else {
|
||||
#if CONFIG_CODE_NONZEROCOUNT
|
||||
break;
|
||||
#else
|
||||
token = DCT_EOB_TOKEN;
|
||||
if (nzc_used)
|
||||
break;
|
||||
else
|
||||
#endif
|
||||
token = DCT_EOB_TOKEN;
|
||||
}
|
||||
|
||||
t->Token = token;
|
||||
t->context_tree = probs[type][ref][band][pt];
|
||||
#if CONFIG_CODE_NONZEROCOUNT
|
||||
// Skip zero node if there are no zeros left
|
||||
t->skip_eob_node = 1 + (zerosleft == 0);
|
||||
#else
|
||||
t->skip_eob_node = (c > 0) && (token_cache[c - 1] == 0);
|
||||
if (nzc_used)
|
||||
t->skip_eob_node = 1 + (zerosleft == 0);
|
||||
else
|
||||
#endif
|
||||
t->skip_eob_node = (c > 0) && (token_cache[c - 1] == 0);
|
||||
assert(vp9_coef_encodings[t->Token].Len - t->skip_eob_node > 0);
|
||||
if (!dry_run) {
|
||||
++counts[type][ref][band][pt][token];
|
||||
@ -975,14 +979,15 @@ static void stuff_b(VP9_COMP *cpi,
|
||||
int dry_run) {
|
||||
MB_MODE_INFO *mbmi = &xd->mode_info_context->mbmi;
|
||||
const BLOCK_SIZE_TYPE sb_type = mbmi->sb_type;
|
||||
#if CONFIG_CODE_NONZEROCOUNT == 0
|
||||
vp9_coeff_count *counts;
|
||||
vp9_coeff_probs *probs;
|
||||
int pt, band;
|
||||
TOKENEXTRA *t = *tp;
|
||||
const int ref = mbmi->ref_frame != INTRA_FRAME;
|
||||
#endif
|
||||
ENTROPY_CONTEXT *a, *l, *a1, *l1, *a2, *l2, *a3, *l3, a_ec, l_ec;
|
||||
#if CONFIG_CODE_NONZEROCOUNT
|
||||
const int nzc_used = get_nzc_used(tx_size);
|
||||
#endif
|
||||
|
||||
if (sb_type == BLOCK_SIZE_SB32X32) {
|
||||
a = (ENTROPY_CONTEXT *)xd->above_context +
|
||||
@ -1011,18 +1016,14 @@ static void stuff_b(VP9_COMP *cpi,
|
||||
case TX_4X4:
|
||||
a_ec = a[0];
|
||||
l_ec = l[0];
|
||||
#if CONFIG_CODE_NONZEROCOUNT == 0
|
||||
counts = cpi->coef_counts_4x4;
|
||||
probs = cpi->common.fc.coef_probs_4x4;
|
||||
#endif
|
||||
break;
|
||||
case TX_8X8:
|
||||
a_ec = (a[0] + a[1]) != 0;
|
||||
l_ec = (l[0] + l[1]) != 0;
|
||||
#if CONFIG_CODE_NONZEROCOUNT == 0
|
||||
counts = cpi->coef_counts_8x8;
|
||||
probs = cpi->common.fc.coef_probs_8x8;
|
||||
#endif
|
||||
break;
|
||||
case TX_16X16:
|
||||
if (type != PLANE_TYPE_UV) {
|
||||
@ -1032,10 +1033,8 @@ static void stuff_b(VP9_COMP *cpi,
|
||||
a_ec = (a[0] + a[1] + a1[0] + a1[1]) != 0;
|
||||
l_ec = (l[0] + l[1] + l1[0] + l1[1]) != 0;
|
||||
}
|
||||
#if CONFIG_CODE_NONZEROCOUNT == 0
|
||||
counts = cpi->coef_counts_16x16;
|
||||
probs = cpi->common.fc.coef_probs_16x16;
|
||||
#endif
|
||||
break;
|
||||
case TX_32X32:
|
||||
if (type != PLANE_TYPE_UV) {
|
||||
@ -1049,26 +1048,28 @@ static void stuff_b(VP9_COMP *cpi,
|
||||
l_ec = (l[0] + l[1] + l1[0] + l1[1] +
|
||||
l2[0] + l2[1] + l3[0] + l3[1]) != 0;
|
||||
}
|
||||
#if CONFIG_CODE_NONZEROCOUNT == 0
|
||||
counts = cpi->coef_counts_32x32;
|
||||
probs = cpi->common.fc.coef_probs_32x32;
|
||||
#endif
|
||||
break;
|
||||
}
|
||||
|
||||
#if CONFIG_CODE_NONZEROCOUNT == 0
|
||||
VP9_COMBINEENTROPYCONTEXTS(pt, a_ec, l_ec);
|
||||
band = 0;
|
||||
t->Token = DCT_EOB_TOKEN;
|
||||
t->context_tree = probs[type][ref][band][pt];
|
||||
t->skip_eob_node = 0;
|
||||
++t;
|
||||
*tp = t;
|
||||
if (!dry_run) {
|
||||
++counts[type][ref][band][pt][DCT_EOB_TOKEN];
|
||||
#if CONFIG_CODE_NONZEROCOUNT
|
||||
if (!nzc_used) {
|
||||
#endif
|
||||
VP9_COMBINEENTROPYCONTEXTS(pt, a_ec, l_ec);
|
||||
band = 0;
|
||||
t->Token = DCT_EOB_TOKEN;
|
||||
t->context_tree = probs[type][ref][band][pt];
|
||||
t->skip_eob_node = 0;
|
||||
++t;
|
||||
*tp = t;
|
||||
if (!dry_run) {
|
||||
++counts[type][ref][band][pt][DCT_EOB_TOKEN];
|
||||
}
|
||||
#if CONFIG_CODE_NONZEROCOUNT
|
||||
}
|
||||
#endif
|
||||
*a = *l = 0;
|
||||
*a = *l = 0;
|
||||
if (tx_size == TX_8X8) {
|
||||
a[1] = 0;
|
||||
l[1] = 0;
|
||||
|
Loading…
x
Reference in New Issue
Block a user