Refactoring on transform types

Prepare for adding more transform varieties (EXT_TX and TX_SKIP in nextgen).

Change-Id: I2dfe024f6be7a92078775917092ed62abc2e7d1e
This commit is contained in:
hui su 2015-08-13 16:27:19 -07:00
parent 41be09afee
commit d76e5b3652
6 changed files with 460 additions and 251 deletions

View File

@ -178,30 +178,73 @@ void vp10_idct32x32_add(const tran_low_t *input, uint8_t *dest, int stride,
vpx_idct32x32_1024_add(input, dest, stride); vpx_idct32x32_1024_add(input, dest, stride);
} }
// iht void vp10_inv_txfm_add_4x4(const tran_low_t *input, uint8_t *dest,
void vp10_iht4x4_add(TX_TYPE tx_type, const tran_low_t *input, uint8_t *dest, int stride, int eob, TX_TYPE tx_type,
int stride, int eob) { void (*itxm_add_4x4)(const tran_low_t *input,
if (tx_type == DCT_DCT) uint8_t *dest, int stride, int eob)) {
vp10_idct4x4_add(input, dest, stride, eob); switch (tx_type) {
else case DCT_DCT:
vp10_iht4x4_16_add(input, dest, stride, tx_type); itxm_add_4x4(input, dest, stride, eob);
} break;
case ADST_DCT:
void vp10_iht8x8_add(TX_TYPE tx_type, const tran_low_t *input, uint8_t *dest, case DCT_ADST:
int stride, int eob) { case ADST_ADST:
if (tx_type == DCT_DCT) { vp10_iht4x4_16_add(input, dest, stride, tx_type);
vp10_idct8x8_add(input, dest, stride, eob); break;
} else { default:
vp10_iht8x8_64_add(input, dest, stride, tx_type); assert(0);
break;
} }
} }
void vp10_iht16x16_add(TX_TYPE tx_type, const tran_low_t *input, uint8_t *dest, void vp10_inv_txfm_add_8x8(const tran_low_t *input, uint8_t *dest,
int stride, int eob) { int stride, int eob, TX_TYPE tx_type) {
if (tx_type == DCT_DCT) { switch (tx_type) {
vp10_idct16x16_add(input, dest, stride, eob); case DCT_DCT:
} else { vp10_idct8x8_add(input, dest, stride, eob);
vp10_iht16x16_256_add(input, dest, stride, tx_type); break;
case ADST_DCT:
case DCT_ADST:
case ADST_ADST:
vp10_iht8x8_64_add(input, dest, stride, tx_type);
break;
default:
assert(0);
break;
}
}
void vp10_inv_txfm_add_16x16(const tran_low_t *input, uint8_t *dest,
int stride, int eob, TX_TYPE tx_type) {
switch (tx_type) {
case DCT_DCT:
vp10_idct16x16_add(input, dest, stride, eob);
break;
case ADST_DCT:
case DCT_ADST:
case ADST_ADST:
vp10_iht16x16_256_add(input, dest, stride, tx_type);
break;
default:
assert(0);
break;
}
}
void vp10_inv_txfm_add_32x32(const tran_low_t *input, uint8_t *dest,
int stride, int eob, TX_TYPE tx_type) {
switch (tx_type) {
case DCT_DCT:
vp10_idct32x32_add(input, dest, stride, eob);
break;
case ADST_DCT:
case DCT_ADST:
case ADST_ADST:
assert(0);
break;
default:
assert(0);
break;
} }
} }
@ -373,30 +416,77 @@ void vp10_highbd_idct32x32_add(const tran_low_t *input, uint8_t *dest,
} }
} }
// iht void vp10_highbd_inv_txfm_add_4x4(const tran_low_t *input, uint8_t *dest,
void vp10_highbd_iht4x4_add(TX_TYPE tx_type, const tran_low_t *input, int stride, int eob, int bd, TX_TYPE tx_type,
uint8_t *dest, int stride, int eob, int bd) { void (*highbd_itxm_add_4x4)
if (tx_type == DCT_DCT) (const tran_low_t *input, uint8_t *dest,
vp10_highbd_idct4x4_add(input, dest, stride, eob, bd); int stride, int eob, int bd)) {
else switch (tx_type) {
vp10_highbd_iht4x4_16_add(input, dest, stride, tx_type, bd); case DCT_DCT:
} highbd_itxm_add_4x4(input, dest, stride, eob, bd);
break;
void vp10_highbd_iht8x8_add(TX_TYPE tx_type, const tran_low_t *input, case ADST_DCT:
uint8_t *dest, int stride, int eob, int bd) { case DCT_ADST:
if (tx_type == DCT_DCT) { case ADST_ADST:
vp10_highbd_idct8x8_add(input, dest, stride, eob, bd); vp10_highbd_iht4x4_16_add(input, dest, stride, tx_type, bd);
} else { break;
vp10_highbd_iht8x8_64_add(input, dest, stride, tx_type, bd); default:
assert(0);
break;
} }
} }
void vp10_highbd_iht16x16_add(TX_TYPE tx_type, const tran_low_t *input, void vp10_highbd_inv_txfm_add_8x8(const tran_low_t *input, uint8_t *dest,
uint8_t *dest, int stride, int eob, int bd) { int stride, int eob, int bd,
if (tx_type == DCT_DCT) { TX_TYPE tx_type) {
vp10_highbd_idct16x16_add(input, dest, stride, eob, bd); switch (tx_type) {
} else { case DCT_DCT:
vp10_highbd_iht16x16_256_add(input, dest, stride, tx_type, bd); vp10_highbd_idct8x8_add(input, dest, stride, eob, bd);
break;
case ADST_DCT:
case DCT_ADST:
case ADST_ADST:
vp10_highbd_iht8x8_64_add(input, dest, stride, tx_type, bd);
break;
default:
assert(0);
break;
}
}
void vp10_highbd_inv_txfm_add_16x16(const tran_low_t *input, uint8_t *dest,
int stride, int eob, int bd,
TX_TYPE tx_type) {
switch (tx_type) {
case DCT_DCT:
vp10_highbd_idct16x16_add(input, dest, stride, eob, bd);
break;
case ADST_DCT:
case DCT_ADST:
case ADST_ADST:
vp10_highbd_iht16x16_256_add(input, dest, stride, tx_type, bd);
break;
default:
assert(0);
break;
}
}
void vp10_highbd_inv_txfm_add_32x32(const tran_low_t *input, uint8_t *dest,
int stride, int eob, int bd,
TX_TYPE tx_type) {
switch (tx_type) {
case DCT_DCT:
vp10_highbd_idct32x32_add(input, dest, stride, eob, bd);
break;
case ADST_DCT:
case DCT_ADST:
case ADST_ADST:
assert(0);
break;
default:
assert(0);
break;
} }
} }
#endif // CONFIG_VP9_HIGHBITDEPTH #endif // CONFIG_VP9_HIGHBITDEPTH

View File

@ -42,19 +42,17 @@ void vp10_iwht4x4_add(const tran_low_t *input, uint8_t *dest, int stride,
int eob); int eob);
void vp10_idct4x4_add(const tran_low_t *input, uint8_t *dest, int stride, void vp10_idct4x4_add(const tran_low_t *input, uint8_t *dest, int stride,
int eob); int eob);
void vp10_idct8x8_add(const tran_low_t *input, uint8_t *dest, int stride,
int eob);
void vp10_idct16x16_add(const tran_low_t *input, uint8_t *dest, int stride,
int eob);
void vp10_idct32x32_add(const tran_low_t *input, uint8_t *dest, int stride,
int eob);
void vp10_iht4x4_add(TX_TYPE tx_type, const tran_low_t *input, uint8_t *dest, void vp10_inv_txfm_add_4x4(const tran_low_t *input, uint8_t *dest,
int stride, int eob); int stride, int eob, TX_TYPE tx_type,
void vp10_iht8x8_add(TX_TYPE tx_type, const tran_low_t *input, uint8_t *dest, void (*itxm_add_4x4)(const tran_low_t *input,
int stride, int eob); uint8_t *dest, int stride, int eob));
void vp10_iht16x16_add(TX_TYPE tx_type, const tran_low_t *input, uint8_t *dest, void vp10_inv_txfm_add_8x8(const tran_low_t *input, uint8_t *dest,
int stride, int eob); int stride, int eob, TX_TYPE tx_type);
void vp10_inv_txfm_add_16x16(const tran_low_t *input, uint8_t *dest,
int stride, int eob, TX_TYPE tx_type);
void vp10_inv_txfm_add_32x32(const tran_low_t *input, uint8_t *dest,
int stride, int eob, TX_TYPE tx_type);
#if CONFIG_VP9_HIGHBITDEPTH #if CONFIG_VP9_HIGHBITDEPTH
void vp10_highbd_iwht4x4_add(const tran_low_t *input, uint8_t *dest, int stride, void vp10_highbd_iwht4x4_add(const tran_low_t *input, uint8_t *dest, int stride,
@ -67,12 +65,19 @@ void vp10_highbd_idct16x16_add(const tran_low_t *input, uint8_t *dest,
int stride, int eob, int bd); int stride, int eob, int bd);
void vp10_highbd_idct32x32_add(const tran_low_t *input, uint8_t *dest, void vp10_highbd_idct32x32_add(const tran_low_t *input, uint8_t *dest,
int stride, int eob, int bd); int stride, int eob, int bd);
void vp10_highbd_iht4x4_add(TX_TYPE tx_type, const tran_low_t *input, void vp10_highbd_inv_txfm_add_4x4(const tran_low_t *input, uint8_t *dest,
uint8_t *dest, int stride, int eob, int bd); int stride, int eob, int bd, TX_TYPE tx_type,
void vp10_highbd_iht8x8_add(TX_TYPE tx_type, const tran_low_t *input, void (*highbd_itxm_add_4x4)
uint8_t *dest, int stride, int eob, int bd); (const tran_low_t *input, uint8_t *dest,
void vp10_highbd_iht16x16_add(TX_TYPE tx_type, const tran_low_t *input, int stride, int eob, int bd));
uint8_t *dest, int stride, int eob, int bd); void vp10_highbd_inv_txfm_add_8x8(const tran_low_t *input, uint8_t *dest,
int stride, int eob, int bd, TX_TYPE tx_type);
void vp10_highbd_inv_txfm_add_16x16(const tran_low_t *input, uint8_t *dest,
int stride, int eob, int bd,
TX_TYPE tx_type);
void vp10_highbd_inv_txfm_add_32x32(const tran_low_t *input, uint8_t *dest,
int stride, int eob, int bd,
TX_TYPE tx_type);
#endif // CONFIG_VP9_HIGHBITDEPTH #endif // CONFIG_VP9_HIGHBITDEPTH
#ifdef __cplusplus #ifdef __cplusplus
} // extern "C" } // extern "C"

View File

@ -186,76 +186,59 @@ static void read_mv_probs(nmv_context *ctx, int allow_hp, vpx_reader *r) {
static void inverse_transform_block_inter(MACROBLOCKD* xd, int plane, static void inverse_transform_block_inter(MACROBLOCKD* xd, int plane,
const TX_SIZE tx_size, const TX_SIZE tx_size,
uint8_t *dst, int stride, uint8_t *dst, int stride,
int eob) { int eob, int block) {
struct macroblockd_plane *const pd = &xd->plane[plane]; struct macroblockd_plane *const pd = &xd->plane[plane];
TX_TYPE tx_type = get_tx_type(pd->plane_type, xd, block);
if (eob > 0) { if (eob > 0) {
tran_low_t *const dqcoeff = pd->dqcoeff; tran_low_t *const dqcoeff = pd->dqcoeff;
#if CONFIG_VP9_HIGHBITDEPTH #if CONFIG_VP9_HIGHBITDEPTH
if (xd->cur_buf->flags & YV12_FLAG_HIGHBITDEPTH) { if (xd->cur_buf->flags & YV12_FLAG_HIGHBITDEPTH) {
if (xd->lossless) {
vp10_highbd_iwht4x4_add(dqcoeff, dst, stride, eob, xd->bd);
} else {
switch (tx_size) {
case TX_4X4:
vp10_highbd_idct4x4_add(dqcoeff, dst, stride, eob, xd->bd);
break;
case TX_8X8:
vp10_highbd_idct8x8_add(dqcoeff, dst, stride, eob, xd->bd);
break;
case TX_16X16:
vp10_highbd_idct16x16_add(dqcoeff, dst, stride, eob, xd->bd);
break;
case TX_32X32:
vp10_highbd_idct32x32_add(dqcoeff, dst, stride, eob, xd->bd);
break;
default:
assert(0 && "Invalid transform size");
}
}
} else {
if (xd->lossless) {
vp10_iwht4x4_add(dqcoeff, dst, stride, eob);
} else {
switch (tx_size) {
case TX_4X4:
vp10_idct4x4_add(dqcoeff, dst, stride, eob);
break;
case TX_8X8:
vp10_idct8x8_add(dqcoeff, dst, stride, eob);
break;
case TX_16X16:
vp10_idct16x16_add(dqcoeff, dst, stride, eob);
break;
case TX_32X32:
vp10_idct32x32_add(dqcoeff, dst, stride, eob);
break;
default:
assert(0 && "Invalid transform size");
return;
}
}
}
#else
if (xd->lossless) {
vp10_iwht4x4_add(dqcoeff, dst, stride, eob);
} else {
switch (tx_size) { switch (tx_size) {
case TX_4X4: case TX_4X4:
vp10_idct4x4_add(dqcoeff, dst, stride, eob); vp10_highbd_inv_txfm_add_4x4(dqcoeff, dst, stride, eob, xd->bd,
tx_type, xd->lossless ?
vp10_highbd_iwht4x4_add :
vp10_highbd_idct4x4_add);
break; break;
case TX_8X8: case TX_8X8:
vp10_idct8x8_add(dqcoeff, dst, stride, eob); vp10_highbd_inv_txfm_add_8x8(dqcoeff, dst, stride, eob, xd->bd,
tx_type);
break; break;
case TX_16X16: case TX_16X16:
vp10_idct16x16_add(dqcoeff, dst, stride, eob); vp10_highbd_inv_txfm_add_16x16(dqcoeff, dst, stride, eob, xd->bd,
tx_type);
break; break;
case TX_32X32: case TX_32X32:
vp10_idct32x32_add(dqcoeff, dst, stride, eob); vp10_highbd_inv_txfm_add_32x32(dqcoeff, dst, stride, eob, xd->bd,
tx_type);
break; break;
default: default:
assert(0 && "Invalid transform size"); assert(0 && "Invalid transform size");
return; return;
} }
} else {
#else // CONFIG_VP9_HIGHBITDEPTH
switch (tx_size) {
case TX_4X4:
vp10_inv_txfm_add_4x4(dqcoeff, dst, stride, eob, tx_type,
xd->lossless ? vp10_iwht4x4_add :
vp10_idct4x4_add);
break;
case TX_8X8:
vp10_inv_txfm_add_8x8(dqcoeff, dst, stride, eob, tx_type);
break;
case TX_16X16:
vp10_inv_txfm_add_16x16(dqcoeff, dst, stride, eob, tx_type);
break;
case TX_32X32:
vp10_inv_txfm_add_32x32(dqcoeff, dst, stride, eob, tx_type);
break;
default:
assert(0 && "Invalid transform size");
return;
}
#endif // CONFIG_VP9_HIGHBITDEPTH
#if CONFIG_VP9_HIGHBITDEPTH
} }
#endif // CONFIG_VP9_HIGHBITDEPTH #endif // CONFIG_VP9_HIGHBITDEPTH
@ -282,70 +265,52 @@ static void inverse_transform_block_intra(MACROBLOCKD* xd, int plane,
tran_low_t *const dqcoeff = pd->dqcoeff; tran_low_t *const dqcoeff = pd->dqcoeff;
#if CONFIG_VP9_HIGHBITDEPTH #if CONFIG_VP9_HIGHBITDEPTH
if (xd->cur_buf->flags & YV12_FLAG_HIGHBITDEPTH) { if (xd->cur_buf->flags & YV12_FLAG_HIGHBITDEPTH) {
if (xd->lossless) {
vp10_highbd_iwht4x4_add(dqcoeff, dst, stride, eob, xd->bd);
} else {
switch (tx_size) {
case TX_4X4:
vp10_highbd_iht4x4_add(tx_type, dqcoeff, dst, stride, eob, xd->bd);
break;
case TX_8X8:
vp10_highbd_iht8x8_add(tx_type, dqcoeff, dst, stride, eob, xd->bd);
break;
case TX_16X16:
vp10_highbd_iht16x16_add(tx_type, dqcoeff, dst, stride, eob, xd->bd);
break;
case TX_32X32:
vp10_highbd_idct32x32_add(dqcoeff, dst, stride, eob, xd->bd);
break;
default:
assert(0 && "Invalid transform size");
}
}
} else {
if (xd->lossless) {
vp10_iwht4x4_add(dqcoeff, dst, stride, eob);
} else {
switch (tx_size) {
case TX_4X4:
vp10_iht4x4_add(tx_type, dqcoeff, dst, stride, eob);
break;
case TX_8X8:
vp10_iht8x8_add(tx_type, dqcoeff, dst, stride, eob);
break;
case TX_16X16:
vp10_iht16x16_add(tx_type, dqcoeff, dst, stride, eob);
break;
case TX_32X32:
vp10_idct32x32_add(dqcoeff, dst, stride, eob);
break;
default:
assert(0 && "Invalid transform size");
return;
}
}
}
#else
if (xd->lossless) {
vp10_iwht4x4_add(dqcoeff, dst, stride, eob);
} else {
switch (tx_size) { switch (tx_size) {
case TX_4X4: case TX_4X4:
vp10_iht4x4_add(tx_type, dqcoeff, dst, stride, eob); vp10_highbd_inv_txfm_add_4x4(dqcoeff, dst, stride, eob, xd->bd,
tx_type, xd->lossless ?
vp10_highbd_iwht4x4_add :
vp10_highbd_idct4x4_add);
break; break;
case TX_8X8: case TX_8X8:
vp10_iht8x8_add(tx_type, dqcoeff, dst, stride, eob); vp10_highbd_inv_txfm_add_8x8(dqcoeff, dst, stride, eob, xd->bd,
tx_type);
break; break;
case TX_16X16: case TX_16X16:
vp10_iht16x16_add(tx_type, dqcoeff, dst, stride, eob); vp10_highbd_inv_txfm_add_16x16(dqcoeff, dst, stride, eob, xd->bd,
tx_type);
break; break;
case TX_32X32: case TX_32X32:
vp10_idct32x32_add(dqcoeff, dst, stride, eob); vp10_highbd_inv_txfm_add_32x32(dqcoeff, dst, stride, eob, xd->bd,
tx_type);
break; break;
default: default:
assert(0 && "Invalid transform size"); assert(0 && "Invalid transform size");
return; return;
} }
} else {
#else // CONFIG_VP9_HIGHBITDEPTH
switch (tx_size) {
case TX_4X4:
vp10_inv_txfm_add_4x4(dqcoeff, dst, stride, eob, tx_type,
xd->lossless ? vp10_iwht4x4_add :
vp10_idct4x4_add);
break;
case TX_8X8:
vp10_inv_txfm_add_8x8(dqcoeff, dst, stride, eob, tx_type);
break;
case TX_16X16:
vp10_inv_txfm_add_16x16(dqcoeff, dst, stride, eob, tx_type);
break;
case TX_32X32:
vp10_inv_txfm_add_32x32(dqcoeff, dst, stride, eob, tx_type);
break;
default:
assert(0 && "Invalid transform size");
return;
}
#endif // CONFIG_VP9_HIGHBITDEPTH
#if CONFIG_VP9_HIGHBITDEPTH
} }
#endif // CONFIG_VP9_HIGHBITDEPTH #endif // CONFIG_VP9_HIGHBITDEPTH
@ -406,7 +371,7 @@ static int reconstruct_inter_block(MACROBLOCKD *const xd, vpx_reader *r,
inverse_transform_block_inter(xd, plane, tx_size, inverse_transform_block_inter(xd, plane, tx_size,
&pd->dst.buf[4 * row * pd->dst.stride + 4 * col], &pd->dst.buf[4 * row * pd->dst.stride + 4 * col],
pd->dst.stride, eob); pd->dst.stride, eob, block_idx);
return eob; return eob;
} }

View File

@ -496,6 +496,146 @@ void vp10_xform_quant_dc(MACROBLOCK *x, int plane, int block,
} }
} }
void vp10_fwd_txfm_4x4(const int16_t *src_diff,
tran_low_t *coeff, int diff_stride, TX_TYPE tx_type,
void (*fwd_txm4x4)(const int16_t *input,
tran_low_t *output, int stride)) {
switch (tx_type) {
case DCT_DCT:
fwd_txm4x4(src_diff, coeff, diff_stride);
break;
case ADST_DCT:
case DCT_ADST:
case ADST_ADST:
vp10_fht4x4(src_diff, coeff, diff_stride, tx_type);
break;
default:
assert(0);
break;
}
}
static void fwd_txfm_8x8(const int16_t *src_diff, tran_low_t *coeff,
int diff_stride, TX_TYPE tx_type) {
switch (tx_type) {
case DCT_DCT:
case ADST_DCT:
case DCT_ADST:
case ADST_ADST:
vp10_fht8x8(src_diff, coeff, diff_stride, tx_type);
break;
default:
assert(0);
break;
}
}
static void fwd_txfm_16x16(const int16_t *src_diff, tran_low_t *coeff,
int diff_stride, TX_TYPE tx_type) {
switch (tx_type) {
case DCT_DCT:
case ADST_DCT:
case DCT_ADST:
case ADST_ADST:
vp10_fht16x16(src_diff, coeff, diff_stride, tx_type);
break;
default:
assert(0);
break;
}
}
static void fwd_txfm_32x32(int rd_transform, const int16_t *src_diff,
tran_low_t *coeff, int diff_stride,
TX_TYPE tx_type) {
switch (tx_type) {
case DCT_DCT:
fdct32x32(rd_transform, src_diff, coeff, diff_stride);
break;
case ADST_DCT:
case DCT_ADST:
case ADST_ADST:
assert(0);
break;
default:
assert(0);
break;
}
}
#if CONFIG_VP9_HIGHBITDEPTH
void vp10_highbd_fwd_txfm_4x4(const int16_t *src_diff, tran_low_t *coeff,
int diff_stride, TX_TYPE tx_type,
void (*highbd_fwd_txm4x4)(const int16_t *input,
tran_low_t *output, int stride)) {
switch (tx_type) {
case DCT_DCT:
highbd_fwd_txm4x4(src_diff, coeff, diff_stride);
break;
case ADST_DCT:
case DCT_ADST:
case ADST_ADST:
vp10_highbd_fht4x4(src_diff, coeff, diff_stride, tx_type);
break;
default:
assert(0);
break;
}
}
static void highbd_fwd_txfm_8x8(const int16_t *src_diff, tran_low_t *coeff,
int diff_stride, TX_TYPE tx_type) {
switch (tx_type) {
case DCT_DCT:
vpx_highbd_fdct8x8(src_diff, coeff, diff_stride);
break;
case ADST_DCT:
case DCT_ADST:
case ADST_ADST:
vp10_highbd_fht8x8(src_diff, coeff, diff_stride, tx_type);
break;
default:
assert(0);
break;
}
}
static void highbd_fwd_txfm_16x16(const int16_t *src_diff, tran_low_t *coeff,
int diff_stride, TX_TYPE tx_type) {
switch (tx_type) {
case DCT_DCT:
vpx_highbd_fdct16x16(src_diff, coeff, diff_stride);
break;
case ADST_DCT:
case DCT_ADST:
case ADST_ADST:
vp10_highbd_fht16x16(src_diff, coeff, diff_stride, tx_type);
break;
default:
assert(0);
break;
}
}
static void highbd_fwd_txfm_32x32(int rd_transform, const int16_t *src_diff,
tran_low_t *coeff, int diff_stride,
TX_TYPE tx_type) {
switch (tx_type) {
case DCT_DCT:
highbd_fdct32x32(rd_transform, src_diff, coeff, diff_stride);
break;
case ADST_DCT:
case DCT_ADST:
case ADST_ADST:
assert(0);
break;
default:
assert(0);
break;
}
}
#endif // CONFIG_VP9_HIGHBITDEPTH
void vp10_xform_quant(MACROBLOCK *x, int plane, int block, void vp10_xform_quant(MACROBLOCK *x, int plane, int block,
BLOCK_SIZE plane_bsize, TX_SIZE tx_size) { BLOCK_SIZE plane_bsize, TX_SIZE tx_size) {
MACROBLOCKD *const xd = &x->e_mbd; MACROBLOCKD *const xd = &x->e_mbd;
@ -518,28 +658,30 @@ void vp10_xform_quant(MACROBLOCK *x, int plane, int block,
if (xd->cur_buf->flags & YV12_FLAG_HIGHBITDEPTH) { if (xd->cur_buf->flags & YV12_FLAG_HIGHBITDEPTH) {
switch (tx_size) { switch (tx_size) {
case TX_32X32: case TX_32X32:
highbd_fdct32x32(x->use_lp32x32fdct, src_diff, coeff, diff_stride); highbd_fwd_txfm_32x32(x->use_lp32x32fdct, src_diff, coeff, diff_stride,
tx_type);
vpx_highbd_quantize_b_32x32(coeff, 1024, x->skip_block, p->zbin, vpx_highbd_quantize_b_32x32(coeff, 1024, x->skip_block, p->zbin,
p->round, p->quant, p->quant_shift, qcoeff, p->round, p->quant, p->quant_shift, qcoeff,
dqcoeff, pd->dequant, eob, dqcoeff, pd->dequant, eob,
scan_order->scan, scan_order->iscan); scan_order->scan, scan_order->iscan);
break; break;
case TX_16X16: case TX_16X16:
vpx_highbd_fdct16x16(src_diff, coeff, diff_stride); highbd_fwd_txfm_16x16(src_diff, coeff, diff_stride, tx_type);
vpx_highbd_quantize_b(coeff, 256, x->skip_block, p->zbin, p->round, vpx_highbd_quantize_b(coeff, 256, x->skip_block, p->zbin, p->round,
p->quant, p->quant_shift, qcoeff, dqcoeff, p->quant, p->quant_shift, qcoeff, dqcoeff,
pd->dequant, eob, pd->dequant, eob,
scan_order->scan, scan_order->iscan); scan_order->scan, scan_order->iscan);
break; break;
case TX_8X8: case TX_8X8:
vpx_highbd_fdct8x8(src_diff, coeff, diff_stride); highbd_fwd_txfm_8x8(src_diff, coeff, diff_stride, tx_type);
vpx_highbd_quantize_b(coeff, 64, x->skip_block, p->zbin, p->round, vpx_highbd_quantize_b(coeff, 64, x->skip_block, p->zbin, p->round,
p->quant, p->quant_shift, qcoeff, dqcoeff, p->quant, p->quant_shift, qcoeff, dqcoeff,
pd->dequant, eob, pd->dequant, eob,
scan_order->scan, scan_order->iscan); scan_order->scan, scan_order->iscan);
break; break;
case TX_4X4: case TX_4X4:
x->fwd_txm4x4(src_diff, coeff, diff_stride); vp10_highbd_fwd_txfm_4x4(src_diff, coeff, diff_stride, tx_type,
x->fwd_txm4x4);
vpx_highbd_quantize_b(coeff, 16, x->skip_block, p->zbin, p->round, vpx_highbd_quantize_b(coeff, 16, x->skip_block, p->zbin, p->round,
p->quant, p->quant_shift, qcoeff, dqcoeff, p->quant, p->quant_shift, qcoeff, dqcoeff,
pd->dequant, eob, pd->dequant, eob,
@ -554,28 +696,28 @@ void vp10_xform_quant(MACROBLOCK *x, int plane, int block,
switch (tx_size) { switch (tx_size) {
case TX_32X32: case TX_32X32:
fdct32x32(x->use_lp32x32fdct, src_diff, coeff, diff_stride); fwd_txfm_32x32(x->use_lp32x32fdct, src_diff, coeff, diff_stride, tx_type);
vpx_quantize_b_32x32(coeff, 1024, x->skip_block, p->zbin, p->round, vpx_quantize_b_32x32(coeff, 1024, x->skip_block, p->zbin, p->round,
p->quant, p->quant_shift, qcoeff, dqcoeff, p->quant, p->quant_shift, qcoeff, dqcoeff,
pd->dequant, eob, scan_order->scan, pd->dequant, eob, scan_order->scan,
scan_order->iscan); scan_order->iscan);
break; break;
case TX_16X16: case TX_16X16:
vpx_fdct16x16(src_diff, coeff, diff_stride); fwd_txfm_16x16(src_diff, coeff, diff_stride, tx_type);
vpx_quantize_b(coeff, 256, x->skip_block, p->zbin, p->round, vpx_quantize_b(coeff, 256, x->skip_block, p->zbin, p->round,
p->quant, p->quant_shift, qcoeff, dqcoeff, p->quant, p->quant_shift, qcoeff, dqcoeff,
pd->dequant, eob, pd->dequant, eob,
scan_order->scan, scan_order->iscan); scan_order->scan, scan_order->iscan);
break; break;
case TX_8X8: case TX_8X8:
vpx_fdct8x8(src_diff, coeff, diff_stride); fwd_txfm_8x8(src_diff, coeff, diff_stride, tx_type);
vpx_quantize_b(coeff, 64, x->skip_block, p->zbin, p->round, vpx_quantize_b(coeff, 64, x->skip_block, p->zbin, p->round,
p->quant, p->quant_shift, qcoeff, dqcoeff, p->quant, p->quant_shift, qcoeff, dqcoeff,
pd->dequant, eob, pd->dequant, eob,
scan_order->scan, scan_order->iscan); scan_order->scan, scan_order->iscan);
break; break;
case TX_4X4: case TX_4X4:
x->fwd_txm4x4(src_diff, coeff, diff_stride); vp10_fwd_txfm_4x4(src_diff, coeff, diff_stride, tx_type, x->fwd_txm4x4);
vpx_quantize_b(coeff, 16, x->skip_block, p->zbin, p->round, vpx_quantize_b(coeff, 16, x->skip_block, p->zbin, p->round,
p->quant, p->quant_shift, qcoeff, dqcoeff, p->quant, p->quant_shift, qcoeff, dqcoeff,
pd->dequant, eob, pd->dequant, eob,
@ -599,6 +741,7 @@ static void encode_block(int plane, int block, BLOCK_SIZE plane_bsize,
int i, j; int i, j;
uint8_t *dst; uint8_t *dst;
ENTROPY_CONTEXT *a, *l; ENTROPY_CONTEXT *a, *l;
TX_TYPE tx_type = get_tx_type(pd->plane_type, xd, block);
txfrm_block_to_raster_xy(plane_bsize, tx_size, block, &i, &j); txfrm_block_to_raster_xy(plane_bsize, tx_size, block, &i, &j);
dst = &pd->dst.buf[4 * j * pd->dst.stride + 4 * i]; dst = &pd->dst.buf[4 * j * pd->dst.stride + 4 * i];
a = &ctx->ta[plane][i]; a = &ctx->ta[plane][i];
@ -660,46 +803,53 @@ static void encode_block(int plane, int block, BLOCK_SIZE plane_bsize,
if (xd->cur_buf->flags & YV12_FLAG_HIGHBITDEPTH) { if (xd->cur_buf->flags & YV12_FLAG_HIGHBITDEPTH) {
switch (tx_size) { switch (tx_size) {
case TX_32X32: case TX_32X32:
vp10_highbd_idct32x32_add(dqcoeff, dst, pd->dst.stride, vp10_highbd_inv_txfm_add_32x32(dqcoeff, dst, pd->dst.stride,
p->eobs[block], xd->bd); p->eobs[block], xd->bd, tx_type);
break; break;
case TX_16X16: case TX_16X16:
vp10_highbd_idct16x16_add(dqcoeff, dst, pd->dst.stride, vp10_highbd_inv_txfm_add_16x16(dqcoeff, dst, pd->dst.stride,
p->eobs[block], xd->bd); p->eobs[block], xd->bd, tx_type);
break; break;
case TX_8X8: case TX_8X8:
vp10_highbd_idct8x8_add(dqcoeff, dst, pd->dst.stride, vp10_highbd_inv_txfm_add_8x8(dqcoeff, dst, pd->dst.stride,
p->eobs[block], xd->bd); p->eobs[block], xd->bd, tx_type);
break; break;
case TX_4X4: case TX_4X4:
// this is like vp10_short_idct4x4 but has a special case around eob<=1 // this is like vp10_short_idct4x4 but has a special case around eob<=1
// which is significant (not just an optimization) for the lossless // which is significant (not just an optimization) for the lossless
// case. // case.
x->highbd_itxm_add(dqcoeff, dst, pd->dst.stride, vp10_highbd_inv_txfm_add_4x4(dqcoeff, dst, pd->dst.stride,
p->eobs[block], xd->bd); p->eobs[block], xd->bd, tx_type,
x->highbd_itxm_add);
break; break;
default: default:
assert(0 && "Invalid transform size"); assert(0 && "Invalid transform size");
break;
} }
return; return;
} }
#endif // CONFIG_VP9_HIGHBITDEPTH #endif // CONFIG_VP9_HIGHBITDEPTH
switch (tx_size) { switch (tx_size) {
case TX_32X32: case TX_32X32:
vp10_idct32x32_add(dqcoeff, dst, pd->dst.stride, p->eobs[block]); vp10_inv_txfm_add_32x32(dqcoeff, dst, pd->dst.stride, p->eobs[block],
tx_type);
break; break;
case TX_16X16: case TX_16X16:
vp10_idct16x16_add(dqcoeff, dst, pd->dst.stride, p->eobs[block]); vp10_inv_txfm_add_16x16(dqcoeff, dst, pd->dst.stride, p->eobs[block],
tx_type);
break; break;
case TX_8X8: case TX_8X8:
vp10_idct8x8_add(dqcoeff, dst, pd->dst.stride, p->eobs[block]); vp10_inv_txfm_add_8x8(dqcoeff, dst, pd->dst.stride, p->eobs[block],
tx_type);
break; break;
case TX_4X4: case TX_4X4:
// this is like vp10_short_idct4x4 but has a special case around eob<=1 // this is like vp10_short_idct4x4 but has a special case around eob<=1
// which is significant (not just an optimization) for the lossless // which is significant (not just an optimization) for the lossless
// case. // case.
x->itxm_add(dqcoeff, dst, pd->dst.stride, p->eobs[block]); vp10_inv_txfm_add_4x4(dqcoeff, dst, pd->dst.stride, p->eobs[block],
tx_type, x->itxm_add);
break; break;
default: default:
assert(0 && "Invalid transform size"); assert(0 && "Invalid transform size");
@ -806,76 +956,63 @@ void vp10_encode_block_intra(int plane, int block, BLOCK_SIZE plane_bsize,
if (!x->skip_recode) { if (!x->skip_recode) {
vpx_highbd_subtract_block(32, 32, src_diff, diff_stride, vpx_highbd_subtract_block(32, 32, src_diff, diff_stride,
src, src_stride, dst, dst_stride, xd->bd); src, src_stride, dst, dst_stride, xd->bd);
highbd_fdct32x32(x->use_lp32x32fdct, src_diff, coeff, diff_stride); highbd_fwd_txfm_32x32(x->use_lp32x32fdct, src_diff, coeff,
diff_stride, tx_type);
vpx_highbd_quantize_b_32x32(coeff, 1024, x->skip_block, p->zbin, vpx_highbd_quantize_b_32x32(coeff, 1024, x->skip_block, p->zbin,
p->round, p->quant, p->quant_shift, p->round, p->quant, p->quant_shift,
qcoeff, dqcoeff, pd->dequant, eob, qcoeff, dqcoeff, pd->dequant, eob,
scan_order->scan, scan_order->iscan); scan_order->scan, scan_order->iscan);
} }
if (!x->skip_encode && *eob) { if (!x->skip_encode && *eob)
vp10_highbd_idct32x32_add(dqcoeff, dst, dst_stride, *eob, xd->bd); vp10_highbd_inv_txfm_add_32x32(dqcoeff, dst, dst_stride, *eob, xd->bd,
} tx_type);
break; break;
case TX_16X16: case TX_16X16:
if (!x->skip_recode) { if (!x->skip_recode) {
vpx_highbd_subtract_block(16, 16, src_diff, diff_stride, vpx_highbd_subtract_block(16, 16, src_diff, diff_stride,
src, src_stride, dst, dst_stride, xd->bd); src, src_stride, dst, dst_stride, xd->bd);
if (tx_type == DCT_DCT) highbd_fwd_txfm_16x16(src_diff, coeff, diff_stride, tx_type);
vpx_highbd_fdct16x16(src_diff, coeff, diff_stride);
else
vp10_highbd_fht16x16(src_diff, coeff, diff_stride, tx_type);
vpx_highbd_quantize_b(coeff, 256, x->skip_block, p->zbin, p->round, vpx_highbd_quantize_b(coeff, 256, x->skip_block, p->zbin, p->round,
p->quant, p->quant_shift, qcoeff, dqcoeff, p->quant, p->quant_shift, qcoeff, dqcoeff,
pd->dequant, eob, pd->dequant, eob,
scan_order->scan, scan_order->iscan); scan_order->scan, scan_order->iscan);
} }
if (!x->skip_encode && *eob) { if (!x->skip_encode && *eob)
vp10_highbd_iht16x16_add(tx_type, dqcoeff, dst, dst_stride, vp10_highbd_inv_txfm_add_16x16(dqcoeff, dst, dst_stride, *eob, xd->bd,
*eob, xd->bd); tx_type);
}
break; break;
case TX_8X8: case TX_8X8:
if (!x->skip_recode) { if (!x->skip_recode) {
vpx_highbd_subtract_block(8, 8, src_diff, diff_stride, vpx_highbd_subtract_block(8, 8, src_diff, diff_stride,
src, src_stride, dst, dst_stride, xd->bd); src, src_stride, dst, dst_stride, xd->bd);
if (tx_type == DCT_DCT) highbd_fwd_txfm_8x8(src_diff, coeff, diff_stride, tx_type);
vpx_highbd_fdct8x8(src_diff, coeff, diff_stride);
else
vp10_highbd_fht8x8(src_diff, coeff, diff_stride, tx_type);
vpx_highbd_quantize_b(coeff, 64, x->skip_block, p->zbin, p->round, vpx_highbd_quantize_b(coeff, 64, x->skip_block, p->zbin, p->round,
p->quant, p->quant_shift, qcoeff, dqcoeff, p->quant, p->quant_shift, qcoeff, dqcoeff,
pd->dequant, eob, pd->dequant, eob,
scan_order->scan, scan_order->iscan); scan_order->scan, scan_order->iscan);
} }
if (!x->skip_encode && *eob) { if (!x->skip_encode && *eob)
vp10_highbd_iht8x8_add(tx_type, dqcoeff, dst, dst_stride, *eob, vp10_highbd_inv_txfm_add_8x8(dqcoeff, dst, dst_stride, *eob, xd->bd,
xd->bd); tx_type);
}
break; break;
case TX_4X4: case TX_4X4:
if (!x->skip_recode) { if (!x->skip_recode) {
vpx_highbd_subtract_block(4, 4, src_diff, diff_stride, vpx_highbd_subtract_block(4, 4, src_diff, diff_stride,
src, src_stride, dst, dst_stride, xd->bd); src, src_stride, dst, dst_stride, xd->bd);
if (tx_type != DCT_DCT) vp10_highbd_fwd_txfm_4x4(src_diff, coeff, diff_stride, tx_type,
vp10_highbd_fht4x4(src_diff, coeff, diff_stride, tx_type); x->fwd_txm4x4);
else
x->fwd_txm4x4(src_diff, coeff, diff_stride);
vpx_highbd_quantize_b(coeff, 16, x->skip_block, p->zbin, p->round, vpx_highbd_quantize_b(coeff, 16, x->skip_block, p->zbin, p->round,
p->quant, p->quant_shift, qcoeff, dqcoeff, p->quant, p->quant_shift, qcoeff, dqcoeff,
pd->dequant, eob, pd->dequant, eob,
scan_order->scan, scan_order->iscan); scan_order->scan, scan_order->iscan);
} }
if (!x->skip_encode && *eob) { if (!x->skip_encode && *eob)
if (tx_type == DCT_DCT) { // this is like vp10_short_idct4x4 but has a special case around
// this is like vp10_short_idct4x4 but has a special case around // eob<=1 which is significant (not just an optimization) for the
// eob<=1 which is significant (not just an optimization) for the // lossless case.
// lossless case. vp10_highbd_inv_txfm_add_4x4(dqcoeff, dst, dst_stride, *eob, xd->bd,
x->highbd_itxm_add(dqcoeff, dst, dst_stride, *eob, xd->bd); tx_type, x->highbd_itxm_add);
} else {
vp10_highbd_iht4x4_16_add(dqcoeff, dst, dst_stride, tx_type, xd->bd);
}
}
break; break;
default: default:
assert(0); assert(0);
@ -892,49 +1029,47 @@ void vp10_encode_block_intra(int plane, int block, BLOCK_SIZE plane_bsize,
if (!x->skip_recode) { if (!x->skip_recode) {
vpx_subtract_block(32, 32, src_diff, diff_stride, vpx_subtract_block(32, 32, src_diff, diff_stride,
src, src_stride, dst, dst_stride); src, src_stride, dst, dst_stride);
fdct32x32(x->use_lp32x32fdct, src_diff, coeff, diff_stride); fwd_txfm_32x32(x->use_lp32x32fdct, src_diff, coeff, diff_stride,
tx_type);
vpx_quantize_b_32x32(coeff, 1024, x->skip_block, p->zbin, p->round, vpx_quantize_b_32x32(coeff, 1024, x->skip_block, p->zbin, p->round,
p->quant, p->quant_shift, qcoeff, dqcoeff, p->quant, p->quant_shift, qcoeff, dqcoeff,
pd->dequant, eob, scan_order->scan, pd->dequant, eob, scan_order->scan,
scan_order->iscan); scan_order->iscan);
} }
if (!x->skip_encode && *eob) if (!x->skip_encode && *eob)
vp10_idct32x32_add(dqcoeff, dst, dst_stride, *eob); vp10_inv_txfm_add_32x32(dqcoeff, dst, dst_stride, *eob, tx_type);
break; break;
case TX_16X16: case TX_16X16:
if (!x->skip_recode) { if (!x->skip_recode) {
vpx_subtract_block(16, 16, src_diff, diff_stride, vpx_subtract_block(16, 16, src_diff, diff_stride,
src, src_stride, dst, dst_stride); src, src_stride, dst, dst_stride);
vp10_fht16x16(src_diff, coeff, diff_stride, tx_type); fwd_txfm_16x16(src_diff, coeff, diff_stride, tx_type);
vpx_quantize_b(coeff, 256, x->skip_block, p->zbin, p->round, vpx_quantize_b(coeff, 256, x->skip_block, p->zbin, p->round,
p->quant, p->quant_shift, qcoeff, dqcoeff, p->quant, p->quant_shift, qcoeff, dqcoeff,
pd->dequant, eob, scan_order->scan, pd->dequant, eob, scan_order->scan,
scan_order->iscan); scan_order->iscan);
} }
if (!x->skip_encode && *eob) if (!x->skip_encode && *eob)
vp10_iht16x16_add(tx_type, dqcoeff, dst, dst_stride, *eob); vp10_inv_txfm_add_16x16(dqcoeff, dst, dst_stride, *eob, tx_type);
break; break;
case TX_8X8: case TX_8X8:
if (!x->skip_recode) { if (!x->skip_recode) {
vpx_subtract_block(8, 8, src_diff, diff_stride, vpx_subtract_block(8, 8, src_diff, diff_stride,
src, src_stride, dst, dst_stride); src, src_stride, dst, dst_stride);
vp10_fht8x8(src_diff, coeff, diff_stride, tx_type); fwd_txfm_8x8(src_diff, coeff, diff_stride, tx_type);
vpx_quantize_b(coeff, 64, x->skip_block, p->zbin, p->round, p->quant, vpx_quantize_b(coeff, 64, x->skip_block, p->zbin, p->round, p->quant,
p->quant_shift, qcoeff, dqcoeff, p->quant_shift, qcoeff, dqcoeff,
pd->dequant, eob, scan_order->scan, pd->dequant, eob, scan_order->scan,
scan_order->iscan); scan_order->iscan);
} }
if (!x->skip_encode && *eob) if (!x->skip_encode && *eob)
vp10_iht8x8_add(tx_type, dqcoeff, dst, dst_stride, *eob); vp10_inv_txfm_add_8x8(dqcoeff, dst, dst_stride, *eob, tx_type);
break; break;
case TX_4X4: case TX_4X4:
if (!x->skip_recode) { if (!x->skip_recode) {
vpx_subtract_block(4, 4, src_diff, diff_stride, vpx_subtract_block(4, 4, src_diff, diff_stride,
src, src_stride, dst, dst_stride); src, src_stride, dst, dst_stride);
if (tx_type != DCT_DCT) vp10_fwd_txfm_4x4(src_diff, coeff, diff_stride, tx_type, x->fwd_txm4x4);
vp10_fht4x4(src_diff, coeff, diff_stride, tx_type);
else
x->fwd_txm4x4(src_diff, coeff, diff_stride);
vpx_quantize_b(coeff, 16, x->skip_block, p->zbin, p->round, p->quant, vpx_quantize_b(coeff, 16, x->skip_block, p->zbin, p->round, p->quant,
p->quant_shift, qcoeff, dqcoeff, p->quant_shift, qcoeff, dqcoeff,
pd->dequant, eob, scan_order->scan, pd->dequant, eob, scan_order->scan,
@ -942,13 +1077,11 @@ void vp10_encode_block_intra(int plane, int block, BLOCK_SIZE plane_bsize,
} }
if (!x->skip_encode && *eob) { if (!x->skip_encode && *eob) {
if (tx_type == DCT_DCT) // this is like vp10_short_idct4x4 but has a special case around eob<=1
// this is like vp10_short_idct4x4 but has a special case around eob<=1 // which is significant (not just an optimization) for the lossless
// which is significant (not just an optimization) for the lossless // case.
// case. vp10_inv_txfm_add_4x4(dqcoeff, dst, dst_stride, *eob, tx_type,
x->itxm_add(dqcoeff, dst, dst_stride, *eob); x->itxm_add);
else
vp10_iht4x4_16_add(dqcoeff, dst, dst_stride, tx_type);
} }
break; break;
default: default:

View File

@ -39,6 +39,18 @@ void vp10_encode_block_intra(int plane, int block, BLOCK_SIZE plane_bsize,
void vp10_encode_intra_block_plane(MACROBLOCK *x, BLOCK_SIZE bsize, int plane); void vp10_encode_intra_block_plane(MACROBLOCK *x, BLOCK_SIZE bsize, int plane);
void vp10_fwd_txfm_4x4(const int16_t *src_diff,
tran_low_t *coeff, int diff_stride, TX_TYPE tx_type,
void (*fwd_txm4x4)(const int16_t *input,
tran_low_t *output, int stride));
#if CONFIG_VP9_HIGHBITDEPTH
void vp10_highbd_fwd_txfm_4x4(const int16_t *src_diff, tran_low_t *coeff,
int diff_stride, TX_TYPE tx_type,
void (*highbd_fwd_txm4x4)(const int16_t *input,
tran_low_t *output, int stride));
#endif // CONFIG_VP9_HIGHBITDEPTH
#ifdef __cplusplus #ifdef __cplusplus
} // extern "C" } // extern "C"
#endif #endif

View File

@ -799,24 +799,24 @@ static int64_t rd_pick_intra4x4block(VP10_COMP *cpi, MACROBLOCK *x,
if (xd->lossless) { if (xd->lossless) {
TX_TYPE tx_type = get_tx_type(PLANE_TYPE_Y, xd, block); TX_TYPE tx_type = get_tx_type(PLANE_TYPE_Y, xd, block);
const scan_order *so = get_scan(TX_4X4, tx_type); const scan_order *so = get_scan(TX_4X4, tx_type);
vp10_highbd_fwht4x4(src_diff, coeff, 8); vp10_highbd_fwd_txfm_4x4(src_diff, coeff, 8, DCT_DCT,
vp10_highbd_fwht4x4);
vp10_regular_quantize_b_4x4(x, 0, block, so->scan, so->iscan); vp10_regular_quantize_b_4x4(x, 0, block, so->scan, so->iscan);
ratey += cost_coeffs(x, 0, block, tempa + idx, templ + idy, TX_4X4, ratey += cost_coeffs(x, 0, block, tempa + idx, templ + idy, TX_4X4,
so->scan, so->neighbors, so->scan, so->neighbors,
cpi->sf.use_fast_coef_costing); cpi->sf.use_fast_coef_costing);
if (RDCOST(x->rdmult, x->rddiv, ratey, distortion) >= best_rd) if (RDCOST(x->rdmult, x->rddiv, ratey, distortion) >= best_rd)
goto next_highbd; goto next_highbd;
vp10_highbd_iwht4x4_add(BLOCK_OFFSET(pd->dqcoeff, block), vp10_highbd_inv_txfm_add_4x4(BLOCK_OFFSET(pd->dqcoeff, block),
dst, dst_stride, dst, dst_stride, p->eobs[block],
p->eobs[block], xd->bd); xd->bd, DCT_DCT,
vp10_highbd_iwht4x4_add);
} else { } else {
int64_t unused; int64_t unused;
TX_TYPE tx_type = get_tx_type(PLANE_TYPE_Y, xd, block); TX_TYPE tx_type = get_tx_type(PLANE_TYPE_Y, xd, block);
const scan_order *so = get_scan(TX_4X4, tx_type); const scan_order *so = get_scan(TX_4X4, tx_type);
if (tx_type == DCT_DCT) vp10_highbd_fwd_txfm_4x4(src_diff, coeff, 8, tx_type,
vpx_highbd_fdct4x4(src_diff, coeff, 8); vpx_highbd_fdct4x4);
else
vp10_highbd_fht4x4(src_diff, coeff, 8, tx_type);
vp10_regular_quantize_b_4x4(x, 0, block, so->scan, so->iscan); vp10_regular_quantize_b_4x4(x, 0, block, so->scan, so->iscan);
ratey += cost_coeffs(x, 0, block, tempa + idx, templ + idy, TX_4X4, ratey += cost_coeffs(x, 0, block, tempa + idx, templ + idy, TX_4X4,
so->scan, so->neighbors, so->scan, so->neighbors,
@ -826,8 +826,10 @@ static int64_t rd_pick_intra4x4block(VP10_COMP *cpi, MACROBLOCK *x,
16, &unused, xd->bd) >> 2; 16, &unused, xd->bd) >> 2;
if (RDCOST(x->rdmult, x->rddiv, ratey, distortion) >= best_rd) if (RDCOST(x->rdmult, x->rddiv, ratey, distortion) >= best_rd)
goto next_highbd; goto next_highbd;
vp10_highbd_iht4x4_add(tx_type, BLOCK_OFFSET(pd->dqcoeff, block), vp10_highbd_inv_txfm_add_4x4(BLOCK_OFFSET(pd->dqcoeff, block),
dst, dst_stride, p->eobs[block], xd->bd); dst, dst_stride, p->eobs[block],
xd->bd, tx_type,
vp10_highbd_idct4x4_add);
} }
} }
} }
@ -902,20 +904,21 @@ static int64_t rd_pick_intra4x4block(VP10_COMP *cpi, MACROBLOCK *x,
if (xd->lossless) { if (xd->lossless) {
TX_TYPE tx_type = get_tx_type(PLANE_TYPE_Y, xd, block); TX_TYPE tx_type = get_tx_type(PLANE_TYPE_Y, xd, block);
const scan_order *so = get_scan(TX_4X4, tx_type); const scan_order *so = get_scan(TX_4X4, tx_type);
vp10_fwht4x4(src_diff, coeff, 8); vp10_fwd_txfm_4x4(src_diff, coeff, 8, DCT_DCT, vp10_fwht4x4);
vp10_regular_quantize_b_4x4(x, 0, block, so->scan, so->iscan); vp10_regular_quantize_b_4x4(x, 0, block, so->scan, so->iscan);
ratey += cost_coeffs(x, 0, block, tempa + idx, templ + idy, TX_4X4, ratey += cost_coeffs(x, 0, block, tempa + idx, templ + idy, TX_4X4,
so->scan, so->neighbors, so->scan, so->neighbors,
cpi->sf.use_fast_coef_costing); cpi->sf.use_fast_coef_costing);
if (RDCOST(x->rdmult, x->rddiv, ratey, distortion) >= best_rd) if (RDCOST(x->rdmult, x->rddiv, ratey, distortion) >= best_rd)
goto next; goto next;
vp10_iwht4x4_add(BLOCK_OFFSET(pd->dqcoeff, block), dst, dst_stride, vp10_inv_txfm_add_4x4(BLOCK_OFFSET(pd->dqcoeff, block),
p->eobs[block]); dst, dst_stride, p->eobs[block], DCT_DCT,
vp10_iwht4x4_add);
} else { } else {
int64_t unused; int64_t unused;
TX_TYPE tx_type = get_tx_type(PLANE_TYPE_Y, xd, block); TX_TYPE tx_type = get_tx_type(PLANE_TYPE_Y, xd, block);
const scan_order *so = get_scan(TX_4X4, tx_type); const scan_order *so = get_scan(TX_4X4, tx_type);
vp10_fht4x4(src_diff, coeff, 8, tx_type); vp10_fwd_txfm_4x4(src_diff, coeff, 8, tx_type, vpx_fdct4x4);
vp10_regular_quantize_b_4x4(x, 0, block, so->scan, so->iscan); vp10_regular_quantize_b_4x4(x, 0, block, so->scan, so->iscan);
ratey += cost_coeffs(x, 0, block, tempa + idx, templ + idy, TX_4X4, ratey += cost_coeffs(x, 0, block, tempa + idx, templ + idy, TX_4X4,
so->scan, so->neighbors, so->scan, so->neighbors,
@ -924,8 +927,9 @@ static int64_t rd_pick_intra4x4block(VP10_COMP *cpi, MACROBLOCK *x,
16, &unused) >> 2; 16, &unused) >> 2;
if (RDCOST(x->rdmult, x->rddiv, ratey, distortion) >= best_rd) if (RDCOST(x->rdmult, x->rddiv, ratey, distortion) >= best_rd)
goto next; goto next;
vp10_iht4x4_add(tx_type, BLOCK_OFFSET(pd->dqcoeff, block), vp10_inv_txfm_add_4x4(BLOCK_OFFSET(pd->dqcoeff, block),
dst, dst_stride, p->eobs[block]); dst, dst_stride, p->eobs[block], tx_type,
vp10_idct4x4_add);
} }
} }
} }