6189f2bcb1
This patch will use the dest buffer instead of the predictor buffer. This will allow us in future commits to remove the extra mem copy that occurs in the dequant functions when eob == 0. We should also be able to remove extra params that are passed into the dequant functions. Change-Id: I7241bc1ab797a430418b1f3a95b5476db7455f6a
254 lines
7.6 KiB
C
254 lines
7.6 KiB
C
/*
|
|
* Copyright (c) 2010 The WebM project authors. All Rights Reserved.
|
|
*
|
|
* Use of this source code is governed by a BSD-style license
|
|
* that can be found in the LICENSE file in the root of the source
|
|
* tree. An additional intellectual property rights grant can be found
|
|
* in the file PATENTS. All contributing project authors may
|
|
* be found in the AUTHORS file in the root of the source tree.
|
|
*/
|
|
|
|
#include "./vpx_config.h"
|
|
#include "vp9_rtcd.h"
|
|
#include "vp9/encoder/vp9_quantize.h"
|
|
#include "vp9/common/vp9_reconintra.h"
|
|
#include "vp9/encoder/vp9_encodemb.h"
|
|
#include "vp9/common/vp9_invtrans.h"
|
|
#include "vp9/encoder/vp9_encodeintra.h"
|
|
|
|
static void encode_intra4x4block(MACROBLOCK *x, int ib);
|
|
|
|
int vp9_encode_intra(VP9_COMP *cpi, MACROBLOCK *x, int use_16x16_pred) {
|
|
MB_MODE_INFO * mbmi = &x->e_mbd.mode_info_context->mbmi;
|
|
(void) cpi;
|
|
|
|
if (use_16x16_pred) {
|
|
mbmi->mode = DC_PRED;
|
|
mbmi->uv_mode = DC_PRED;
|
|
mbmi->ref_frame = INTRA_FRAME;
|
|
|
|
vp9_encode_intra16x16mby(&cpi->common, x);
|
|
} else {
|
|
int i;
|
|
|
|
for (i = 0; i < 16; i++) {
|
|
x->e_mbd.block[i].bmi.as_mode.first = B_DC_PRED;
|
|
encode_intra4x4block(x, i);
|
|
}
|
|
}
|
|
|
|
return vp9_get_mb_ss(x->src_diff);
|
|
}
|
|
|
|
static void encode_intra4x4block(MACROBLOCK *x, int ib) {
|
|
BLOCKD *b = &x->e_mbd.block[ib];
|
|
BLOCK *be = &x->block[ib];
|
|
MACROBLOCKD * const xd = &x->e_mbd;
|
|
TX_TYPE tx_type;
|
|
|
|
assert(ib < 16);
|
|
|
|
#if CONFIG_NEWBINTRAMODES
|
|
b->bmi.as_mode.context = vp9_find_bpred_context(&x->e_mbd, b);
|
|
#endif
|
|
|
|
vp9_intra4x4_predict(&x->e_mbd, b, b->bmi.as_mode.first, b->predictor, 16);
|
|
vp9_subtract_b(be, b, 16);
|
|
|
|
tx_type = get_tx_type_4x4(&x->e_mbd, ib);
|
|
if (tx_type != DCT_DCT) {
|
|
vp9_short_fht4x4(be->src_diff, be->coeff, 16, tx_type);
|
|
vp9_ht_quantize_b_4x4(x, ib, tx_type);
|
|
vp9_short_iht4x4(BLOCK_OFFSET(xd->plane[0].dqcoeff, ib, 16),
|
|
b->diff, 16, tx_type);
|
|
} else {
|
|
x->fwd_txm4x4(be->src_diff, be->coeff, 32);
|
|
x->quantize_b_4x4(x, ib, 16);
|
|
vp9_inverse_transform_b_4x4(&x->e_mbd, xd->plane[0].eobs[ib],
|
|
BLOCK_OFFSET(xd->plane[0].dqcoeff, ib, 16),
|
|
b->diff, 32);
|
|
}
|
|
|
|
vp9_recon_b(b->predictor, b->diff, *(b->base_dst) + b->dst, b->dst_stride);
|
|
}
|
|
|
|
void vp9_encode_intra4x4mby(MACROBLOCK *mb) {
|
|
int i;
|
|
|
|
for (i = 0; i < 16; i++)
|
|
encode_intra4x4block(mb, i);
|
|
}
|
|
|
|
void vp9_encode_intra16x16mby(VP9_COMMON *const cm, MACROBLOCK *x) {
|
|
MACROBLOCKD *xd = &x->e_mbd;
|
|
BLOCK *b = &x->block[0];
|
|
TX_SIZE tx_size = xd->mode_info_context->mbmi.txfm_size;
|
|
|
|
vp9_build_intra_predictors_mby(xd);
|
|
|
|
vp9_subtract_mby(x->src_diff, *(b->base_src), xd->predictor, b->src_stride);
|
|
|
|
switch (tx_size) {
|
|
case TX_16X16:
|
|
vp9_transform_mby_16x16(x);
|
|
vp9_quantize_mby_16x16(x);
|
|
if (x->optimize)
|
|
vp9_optimize_mby_16x16(cm, x);
|
|
vp9_inverse_transform_mby_16x16(xd);
|
|
break;
|
|
case TX_8X8:
|
|
vp9_transform_mby_8x8(x);
|
|
vp9_quantize_mby_8x8(x);
|
|
if (x->optimize)
|
|
vp9_optimize_mby_8x8(cm, x);
|
|
vp9_inverse_transform_mby_8x8(xd);
|
|
break;
|
|
default:
|
|
vp9_transform_mby_4x4(x);
|
|
vp9_quantize_mby_4x4(x);
|
|
if (x->optimize)
|
|
vp9_optimize_mby_4x4(cm, x);
|
|
vp9_inverse_transform_mby_4x4(xd);
|
|
break;
|
|
}
|
|
|
|
vp9_recon_mby(xd);
|
|
}
|
|
|
|
void vp9_encode_intra16x16mbuv(VP9_COMMON *const cm, MACROBLOCK *x) {
|
|
MACROBLOCKD *xd = &x->e_mbd;
|
|
TX_SIZE tx_size = xd->mode_info_context->mbmi.txfm_size;
|
|
|
|
vp9_build_intra_predictors_mbuv(xd);
|
|
|
|
vp9_subtract_mbuv(x->src_diff, x->src.u_buffer, x->src.v_buffer,
|
|
xd->predictor, x->src.uv_stride);
|
|
|
|
switch (tx_size) {
|
|
case TX_4X4:
|
|
vp9_transform_mbuv_4x4(x);
|
|
vp9_quantize_mbuv_4x4(x);
|
|
if (x->optimize)
|
|
vp9_optimize_mbuv_4x4(cm, x);
|
|
vp9_inverse_transform_mbuv_4x4(xd);
|
|
break;
|
|
default: // 16x16 or 8x8
|
|
vp9_transform_mbuv_8x8(x);
|
|
vp9_quantize_mbuv_8x8(x);
|
|
if (x->optimize)
|
|
vp9_optimize_mbuv_8x8(cm, x);
|
|
vp9_inverse_transform_mbuv_8x8(xd);
|
|
break;
|
|
}
|
|
|
|
vp9_recon_intra_mbuv(xd);
|
|
}
|
|
|
|
void vp9_encode_intra8x8(MACROBLOCK *x, int ib) {
|
|
MACROBLOCKD *xd = &x->e_mbd;
|
|
BLOCKD *b = &xd->block[ib];
|
|
BLOCK *be = &x->block[ib];
|
|
const int iblock[4] = {0, 1, 4, 5};
|
|
int i;
|
|
TX_TYPE tx_type;
|
|
|
|
vp9_intra8x8_predict(xd, b, b->bmi.as_mode.first, b->predictor, 16);
|
|
// generate residual blocks
|
|
vp9_subtract_4b_c(be, b, 16);
|
|
|
|
if (xd->mode_info_context->mbmi.txfm_size == TX_8X8) {
|
|
int idx = (ib & 0x02) ? (ib + 2) : ib;
|
|
int16_t * const dqcoeff = BLOCK_OFFSET(xd->plane[0].dqcoeff, idx, 16);
|
|
|
|
assert(idx < 16);
|
|
tx_type = get_tx_type_8x8(xd, ib);
|
|
if (tx_type != DCT_DCT) {
|
|
vp9_short_fht8x8(be->src_diff, (x->block + idx)->coeff, 16, tx_type);
|
|
x->quantize_b_8x8(x, idx, tx_type, 16);
|
|
vp9_short_iht8x8(dqcoeff, xd->block[ib].diff,
|
|
16, tx_type);
|
|
} else {
|
|
x->fwd_txm8x8(be->src_diff, (x->block + idx)->coeff, 32);
|
|
x->quantize_b_8x8(x, idx, DCT_DCT, 16);
|
|
vp9_short_idct8x8(dqcoeff, xd->block[ib].diff, 32);
|
|
}
|
|
} else {
|
|
for (i = 0; i < 4; i++) {
|
|
int idx = ib + iblock[i];
|
|
int16_t * const dqcoeff = BLOCK_OFFSET(xd->plane[0].dqcoeff, idx, 16);
|
|
|
|
assert(idx < 16);
|
|
b = &xd->block[ib + iblock[i]];
|
|
be = &x->block[ib + iblock[i]];
|
|
tx_type = get_tx_type_4x4(xd, ib + iblock[i]);
|
|
if (tx_type != DCT_DCT) {
|
|
vp9_short_fht4x4(be->src_diff, be->coeff, 16, tx_type);
|
|
vp9_ht_quantize_b_4x4(x, ib + iblock[i], tx_type);
|
|
vp9_short_iht4x4(dqcoeff, b->diff, 16, tx_type);
|
|
} else if (!(i & 1) &&
|
|
get_tx_type_4x4(xd, ib + iblock[i] + 1) == DCT_DCT) {
|
|
x->fwd_txm8x4(be->src_diff, be->coeff, 32);
|
|
x->quantize_b_4x4_pair(x, ib + iblock[i], ib + iblock[i] + 1, 16);
|
|
vp9_inverse_transform_b_4x4(xd, xd->plane[0].eobs[ib + iblock[i]],
|
|
dqcoeff, b->diff, 32);
|
|
vp9_inverse_transform_b_4x4(xd, xd->plane[0].eobs[ib + iblock[i] + 1],
|
|
dqcoeff + 16, (b + 1)->diff, 32);
|
|
i++;
|
|
} else {
|
|
x->fwd_txm4x4(be->src_diff, be->coeff, 32);
|
|
x->quantize_b_4x4(x, ib + iblock[i], 16);
|
|
vp9_inverse_transform_b_4x4(xd, xd->plane[0].eobs[ib + iblock[i]],
|
|
dqcoeff, b->diff, 32);
|
|
}
|
|
}
|
|
}
|
|
|
|
// reconstruct submacroblock
|
|
for (i = 0; i < 4; i++) {
|
|
b = &xd->block[ib + iblock[i]];
|
|
vp9_recon_b_c(b->predictor, b->diff, *(b->base_dst) + b->dst,
|
|
b->dst_stride);
|
|
}
|
|
}
|
|
|
|
void vp9_encode_intra8x8mby(MACROBLOCK *x) {
|
|
int i;
|
|
|
|
for (i = 0; i < 4; i++)
|
|
vp9_encode_intra8x8(x, vp9_i8x8_block[i]);
|
|
}
|
|
|
|
static void encode_intra_uv4x4(MACROBLOCK *x, int ib, int mode) {
|
|
MACROBLOCKD * const xd = &x->e_mbd;
|
|
BLOCKD *b = &x->e_mbd.block[ib];
|
|
BLOCK *be = &x->block[ib];
|
|
int16_t * const dqcoeff = MB_SUBBLOCK_FIELD(xd, dqcoeff, ib);
|
|
const int plane = ib < 20 ? 1 : 2;
|
|
const int block = ib < 20 ? ib - 16 : ib - 20;
|
|
|
|
assert(ib >= 16 && ib < 24);
|
|
vp9_intra_uv4x4_predict(&x->e_mbd, b, mode, b->predictor, 8);
|
|
|
|
vp9_subtract_b(be, b, 8);
|
|
|
|
x->fwd_txm4x4(be->src_diff, be->coeff, 16);
|
|
x->quantize_b_4x4(x, ib, 16);
|
|
vp9_inverse_transform_b_4x4(&x->e_mbd, xd->plane[plane].eobs[block],
|
|
dqcoeff, b->diff, 16);
|
|
|
|
vp9_recon_uv_b_c(b->predictor, b->diff, *(b->base_dst) + b->dst,
|
|
b->dst_stride);
|
|
}
|
|
|
|
void vp9_encode_intra8x8mbuv(MACROBLOCK *x) {
|
|
int i;
|
|
|
|
for (i = 0; i < 4; i++) {
|
|
BLOCKD *b = &x->e_mbd.block[vp9_i8x8_block[i]];
|
|
int mode = b->bmi.as_mode.first;
|
|
|
|
encode_intra_uv4x4(x, i + 16, mode); // u
|
|
encode_intra_uv4x4(x, i + 20, mode); // v
|
|
}
|
|
}
|