9ec0f658a1
Use the common sb functions instead. Change-Id: I4fa0a8ee3c6ada56271dd09bf895b97642f55858
260 lines
8.5 KiB
C
260 lines
8.5 KiB
C
/*
|
|
* Copyright (c) 2010 The WebM project authors. All Rights Reserved.
|
|
*
|
|
* Use of this source code is governed by a BSD-style license
|
|
* that can be found in the LICENSE file in the root of the source
|
|
* tree. An additional intellectual property rights grant can be found
|
|
* in the file PATENTS. All contributing project authors may
|
|
* be found in the AUTHORS file in the root of the source tree.
|
|
*/
|
|
|
|
#include "./vpx_config.h"
|
|
#include "vp9_rtcd.h"
|
|
#include "vp9/encoder/vp9_quantize.h"
|
|
#include "vp9/common/vp9_reconintra.h"
|
|
#include "vp9/encoder/vp9_encodemb.h"
|
|
#include "vp9/common/vp9_invtrans.h"
|
|
#include "vp9/encoder/vp9_encodeintra.h"
|
|
|
|
static void encode_intra4x4block(MACROBLOCK *x, int ib);
|
|
|
|
int vp9_encode_intra(VP9_COMP *cpi, MACROBLOCK *x, int use_16x16_pred) {
|
|
MB_MODE_INFO * mbmi = &x->e_mbd.mode_info_context->mbmi;
|
|
(void) cpi;
|
|
|
|
if (use_16x16_pred) {
|
|
mbmi->mode = DC_PRED;
|
|
mbmi->uv_mode = DC_PRED;
|
|
mbmi->ref_frame = INTRA_FRAME;
|
|
|
|
vp9_encode_intra16x16mby(&cpi->common, x);
|
|
} else {
|
|
int i;
|
|
|
|
for (i = 0; i < 16; i++) {
|
|
x->e_mbd.block[i].bmi.as_mode.first = B_DC_PRED;
|
|
encode_intra4x4block(x, i);
|
|
}
|
|
}
|
|
|
|
return vp9_get_mb_ss(x->src_diff);
|
|
}
|
|
|
|
static void encode_intra4x4block(MACROBLOCK *x, int ib) {
|
|
BLOCKD *b = &x->e_mbd.block[ib];
|
|
BLOCK *be = &x->block[ib];
|
|
MACROBLOCKD * const xd = &x->e_mbd;
|
|
TX_TYPE tx_type;
|
|
|
|
assert(ib < 16);
|
|
|
|
#if CONFIG_NEWBINTRAMODES
|
|
b->bmi.as_mode.context = vp9_find_bpred_context(&x->e_mbd, b);
|
|
#endif
|
|
|
|
vp9_intra4x4_predict(&x->e_mbd, b, b->bmi.as_mode.first,
|
|
*(b->base_dst) + b->dst, b->dst_stride);
|
|
vp9_subtract_b(be, b, 16);
|
|
|
|
tx_type = get_tx_type_4x4(&x->e_mbd, ib);
|
|
if (tx_type != DCT_DCT) {
|
|
vp9_short_fht4x4(be->src_diff, be->coeff, 16, tx_type);
|
|
vp9_ht_quantize_b_4x4(x, ib, tx_type);
|
|
vp9_short_iht4x4(BLOCK_OFFSET(xd->plane[0].dqcoeff, ib, 16),
|
|
b->diff, 16, tx_type);
|
|
} else {
|
|
x->fwd_txm4x4(be->src_diff, be->coeff, 32);
|
|
x->quantize_b_4x4(x, ib, 16);
|
|
vp9_inverse_transform_b_4x4(&x->e_mbd, xd->plane[0].eobs[ib],
|
|
BLOCK_OFFSET(xd->plane[0].dqcoeff, ib, 16),
|
|
b->diff, 32);
|
|
}
|
|
|
|
vp9_recon_b(*(b->base_dst) + b->dst, b->diff,
|
|
*(b->base_dst) + b->dst, b->dst_stride);
|
|
}
|
|
|
|
void vp9_encode_intra4x4mby(MACROBLOCK *mb) {
|
|
int i;
|
|
|
|
for (i = 0; i < 16; i++)
|
|
encode_intra4x4block(mb, i);
|
|
}
|
|
|
|
void vp9_encode_intra16x16mby(VP9_COMMON *const cm, MACROBLOCK *x) {
|
|
MACROBLOCKD *xd = &x->e_mbd;
|
|
TX_SIZE tx_size = xd->mode_info_context->mbmi.txfm_size;
|
|
|
|
vp9_build_intra_predictors_sby_s(xd, BLOCK_SIZE_MB16X16);
|
|
vp9_subtract_sby_s_c(x->src_diff,
|
|
x->src.y_buffer, x->src.y_stride,
|
|
xd->dst.y_buffer, xd->dst.y_stride,
|
|
BLOCK_SIZE_MB16X16);
|
|
|
|
switch (tx_size) {
|
|
case TX_16X16:
|
|
vp9_transform_sby_16x16(x, BLOCK_SIZE_MB16X16);
|
|
vp9_quantize_sby_16x16(x, BLOCK_SIZE_MB16X16);
|
|
if (x->optimize)
|
|
vp9_optimize_sby_16x16(cm, x, BLOCK_SIZE_MB16X16);
|
|
vp9_inverse_transform_sby_16x16(xd, BLOCK_SIZE_MB16X16);
|
|
break;
|
|
case TX_8X8:
|
|
vp9_transform_sby_8x8(x, BLOCK_SIZE_MB16X16);
|
|
vp9_quantize_sby_8x8(x, BLOCK_SIZE_MB16X16);
|
|
if (x->optimize)
|
|
vp9_optimize_sby_8x8(cm, x, BLOCK_SIZE_MB16X16);
|
|
vp9_inverse_transform_sby_8x8(xd, BLOCK_SIZE_MB16X16);
|
|
break;
|
|
default:
|
|
vp9_transform_sby_4x4(x, BLOCK_SIZE_MB16X16);
|
|
vp9_quantize_sby_4x4(x, BLOCK_SIZE_MB16X16);
|
|
if (x->optimize)
|
|
vp9_optimize_sby_4x4(cm, x, BLOCK_SIZE_MB16X16);
|
|
vp9_inverse_transform_sby_4x4(xd, BLOCK_SIZE_MB16X16);
|
|
break;
|
|
}
|
|
|
|
vp9_recon_sby(xd, BLOCK_SIZE_MB16X16);
|
|
}
|
|
|
|
void vp9_encode_intra16x16mbuv(VP9_COMMON *const cm, MACROBLOCK *x) {
|
|
MACROBLOCKD *xd = &x->e_mbd;
|
|
TX_SIZE tx_size = xd->mode_info_context->mbmi.txfm_size;
|
|
|
|
vp9_build_intra_predictors_sbuv_s(xd, BLOCK_SIZE_MB16X16);
|
|
vp9_subtract_sbuv_s_c(x->src_diff,
|
|
x->src.u_buffer, x->src.v_buffer, x->src.uv_stride,
|
|
xd->dst.u_buffer, xd->dst.v_buffer, xd->dst.uv_stride,
|
|
BLOCK_SIZE_MB16X16);
|
|
|
|
switch (tx_size) {
|
|
case TX_4X4:
|
|
vp9_transform_sbuv_4x4(x, BLOCK_SIZE_MB16X16);
|
|
vp9_quantize_sbuv_4x4(x, BLOCK_SIZE_MB16X16);
|
|
if (x->optimize)
|
|
vp9_optimize_sbuv_4x4(cm, x, BLOCK_SIZE_MB16X16);
|
|
vp9_inverse_transform_sbuv_4x4(xd, BLOCK_SIZE_MB16X16);
|
|
break;
|
|
default: // 16x16 or 8x8
|
|
vp9_transform_sbuv_8x8(x, BLOCK_SIZE_MB16X16);
|
|
vp9_quantize_sbuv_8x8(x, BLOCK_SIZE_MB16X16);
|
|
if (x->optimize)
|
|
vp9_optimize_sbuv_8x8(cm, x, BLOCK_SIZE_MB16X16);
|
|
vp9_inverse_transform_sbuv_8x8(xd, BLOCK_SIZE_MB16X16);
|
|
break;
|
|
}
|
|
|
|
vp9_recon_intra_mbuv(xd);
|
|
}
|
|
|
|
void vp9_encode_intra8x8(MACROBLOCK *x, int ib) {
|
|
MACROBLOCKD *xd = &x->e_mbd;
|
|
BLOCKD *b = &xd->block[ib];
|
|
BLOCK *be = &x->block[ib];
|
|
const int iblock[4] = {0, 1, 4, 5};
|
|
int i;
|
|
TX_TYPE tx_type;
|
|
|
|
vp9_intra8x8_predict(xd, b, b->bmi.as_mode.first,
|
|
*(b->base_dst) + b->dst, b->dst_stride);
|
|
// generate residual blocks
|
|
vp9_subtract_4b_c(be, b, 16);
|
|
|
|
if (xd->mode_info_context->mbmi.txfm_size == TX_8X8) {
|
|
int idx = (ib & 0x02) ? (ib + 2) : ib;
|
|
int16_t * const dqcoeff = BLOCK_OFFSET(xd->plane[0].dqcoeff, idx, 16);
|
|
|
|
assert(idx < 16);
|
|
tx_type = get_tx_type_8x8(xd, ib);
|
|
if (tx_type != DCT_DCT) {
|
|
vp9_short_fht8x8(be->src_diff, (x->block + idx)->coeff, 16, tx_type);
|
|
x->quantize_b_8x8(x, idx, tx_type, 16);
|
|
vp9_short_iht8x8(dqcoeff, xd->block[ib].diff,
|
|
16, tx_type);
|
|
} else {
|
|
x->fwd_txm8x8(be->src_diff, (x->block + idx)->coeff, 32);
|
|
x->quantize_b_8x8(x, idx, DCT_DCT, 16);
|
|
vp9_short_idct8x8(dqcoeff, xd->block[ib].diff, 32);
|
|
}
|
|
} else {
|
|
for (i = 0; i < 4; i++) {
|
|
int idx = ib + iblock[i];
|
|
int16_t * const dqcoeff = BLOCK_OFFSET(xd->plane[0].dqcoeff, idx, 16);
|
|
|
|
assert(idx < 16);
|
|
b = &xd->block[ib + iblock[i]];
|
|
be = &x->block[ib + iblock[i]];
|
|
tx_type = get_tx_type_4x4(xd, ib + iblock[i]);
|
|
if (tx_type != DCT_DCT) {
|
|
vp9_short_fht4x4(be->src_diff, be->coeff, 16, tx_type);
|
|
vp9_ht_quantize_b_4x4(x, ib + iblock[i], tx_type);
|
|
vp9_short_iht4x4(dqcoeff, b->diff, 16, tx_type);
|
|
} else if (!(i & 1) &&
|
|
get_tx_type_4x4(xd, ib + iblock[i] + 1) == DCT_DCT) {
|
|
x->fwd_txm8x4(be->src_diff, be->coeff, 32);
|
|
x->quantize_b_4x4_pair(x, ib + iblock[i], ib + iblock[i] + 1, 16);
|
|
vp9_inverse_transform_b_4x4(xd, xd->plane[0].eobs[ib + iblock[i]],
|
|
dqcoeff, b->diff, 32);
|
|
vp9_inverse_transform_b_4x4(xd, xd->plane[0].eobs[ib + iblock[i] + 1],
|
|
dqcoeff + 16, (b + 1)->diff, 32);
|
|
i++;
|
|
} else {
|
|
x->fwd_txm4x4(be->src_diff, be->coeff, 32);
|
|
x->quantize_b_4x4(x, ib + iblock[i], 16);
|
|
vp9_inverse_transform_b_4x4(xd, xd->plane[0].eobs[ib + iblock[i]],
|
|
dqcoeff, b->diff, 32);
|
|
}
|
|
}
|
|
}
|
|
|
|
// reconstruct submacroblock
|
|
for (i = 0; i < 4; i++) {
|
|
b = &xd->block[ib + iblock[i]];
|
|
vp9_recon_b_c(*(b->base_dst) + b->dst, b->diff, *(b->base_dst) + b->dst,
|
|
b->dst_stride);
|
|
}
|
|
}
|
|
|
|
void vp9_encode_intra8x8mby(MACROBLOCK *x) {
|
|
int i;
|
|
|
|
for (i = 0; i < 4; i++)
|
|
vp9_encode_intra8x8(x, vp9_i8x8_block[i]);
|
|
}
|
|
|
|
static void encode_intra_uv4x4(MACROBLOCK *x, int ib, int mode) {
|
|
MACROBLOCKD * const xd = &x->e_mbd;
|
|
BLOCKD *b = &x->e_mbd.block[ib];
|
|
BLOCK *be = &x->block[ib];
|
|
int16_t * const dqcoeff = MB_SUBBLOCK_FIELD(xd, dqcoeff, ib);
|
|
const int plane = ib < 20 ? 1 : 2;
|
|
const int block = ib < 20 ? ib - 16 : ib - 20;
|
|
|
|
assert(ib >= 16 && ib < 24);
|
|
vp9_intra_uv4x4_predict(&x->e_mbd, b, mode,
|
|
*(b->base_dst) + b->dst, b->dst_stride);
|
|
|
|
vp9_subtract_b(be, b, 8);
|
|
|
|
x->fwd_txm4x4(be->src_diff, be->coeff, 16);
|
|
x->quantize_b_4x4(x, ib, 16);
|
|
vp9_inverse_transform_b_4x4(&x->e_mbd, xd->plane[plane].eobs[block],
|
|
dqcoeff, b->diff, 16);
|
|
|
|
vp9_recon_uv_b_c(*(b->base_dst) + b->dst, b->diff, *(b->base_dst) + b->dst,
|
|
b->dst_stride);
|
|
}
|
|
|
|
void vp9_encode_intra8x8mbuv(MACROBLOCK *x) {
|
|
int i;
|
|
|
|
for (i = 0; i < 4; i++) {
|
|
BLOCKD *b = &x->e_mbd.block[vp9_i8x8_block[i]];
|
|
int mode = b->bmi.as_mode.first;
|
|
|
|
encode_intra_uv4x4(x, i + 16, mode); // u
|
|
encode_intra_uv4x4(x, i + 20, mode); // v
|
|
}
|
|
}
|