/* * Copyright (c) 2016 The WebM project authors. All Rights Reserved. * * Use of this source code is governed by a BSD-style license * that can be found in the LICENSE file in the root of the source * tree. An additional intellectual property rights grant can be found * in the file PATENTS. All contributing project authors may * be found in the AUTHORS file in the root of the source tree. */ #include #include "vpx/vpx_integer.h" #include "vpx_ports/mem.h" #include "vpx_dsp/blend.h" #include "vpx_dsp/vpx_dsp_common.h" #include "./vpx_dsp_rtcd.h" // Blending with alpha mask. Mask values come from the range [0, 64], // as described for VPX_BLEND_A64 in vpx_dsp/blned.h. src0 or src1 can // be the same as dst, or dst can be different from both sources. void vpx_blend_a64_mask_c(uint8_t *dst, uint32_t dst_stride, const uint8_t *src0, uint32_t src0_stride, const uint8_t *src1, uint32_t src1_stride, const uint8_t *mask, uint32_t mask_stride, int h, int w, int subh, int subw) { int i, j; assert(IMPLIES(src0 == dst, src0_stride == dst_stride)); assert(IMPLIES(src1 == dst, src1_stride == dst_stride)); assert(h >= 1); assert(w >= 1); assert(IS_POWER_OF_TWO(h)); assert(IS_POWER_OF_TWO(w)); if (subw == 0 && subh == 0) { for (i = 0; i < h; ++i) { for (j = 0; j < w; ++j) { const int m = mask[i * mask_stride + j]; dst[i * dst_stride + j] = VPX_BLEND_A64(m, src0[i * src0_stride + j], src1[i * src1_stride + j]); } } } else if (subw == 1 && subh == 1) { for (i = 0; i < h; ++i) { for (j = 0; j < w; ++j) { const int m = ROUND_POWER_OF_TWO(mask[(2 * i) * mask_stride + (2 * j)] + mask[(2 * i + 1) * mask_stride + (2 * j)] + mask[(2 * i) * mask_stride + (2 * j + 1)] + mask[(2 * i + 1) * mask_stride + (2 * j + 1)], 2); dst[i * dst_stride + j] = VPX_BLEND_A64(m, src0[i * src0_stride + j], src1[i * src1_stride + j]); } } } else if (subw == 1 && subh == 0) { for (i = 0; i < h; ++i) { for (j = 0; j < w; ++j) { const int m = VPX_BLEND_AVG(mask[i * mask_stride + (2 * j)], mask[i * mask_stride + (2 * j + 1)]); dst[i * dst_stride + j] = VPX_BLEND_A64(m, src0[i * src0_stride + j], src1[i * src1_stride + j]); } } } else { for (i = 0; i < h; ++i) { for (j = 0; j < w; ++j) { const int m = VPX_BLEND_AVG(mask[(2 * i) * mask_stride + j], mask[(2 * i + 1) * mask_stride + j]); dst[i * dst_stride + j] = VPX_BLEND_A64(m, src0[i * src0_stride + j], src1[i * src1_stride + j]); } } } } #if CONFIG_VPX_HIGHBITDEPTH void vpx_highbd_blend_a64_mask_c(uint8_t *dst_8, uint32_t dst_stride, const uint8_t *src0_8, uint32_t src0_stride, const uint8_t *src1_8, uint32_t src1_stride, const uint8_t *mask, uint32_t mask_stride, int h, int w, int subh, int subw, int bd) { int i, j; uint16_t *dst = CONVERT_TO_SHORTPTR(dst_8); const uint16_t *src0 = CONVERT_TO_SHORTPTR(src0_8); const uint16_t *src1 = CONVERT_TO_SHORTPTR(src1_8); assert(IMPLIES(src0 == dst, src0_stride == dst_stride)); assert(IMPLIES(src1 == dst, src1_stride == dst_stride)); assert(h >= 1); assert(w >= 1); assert(IS_POWER_OF_TWO(h)); assert(IS_POWER_OF_TWO(w)); assert(bd == 8 || bd == 10 || bd == 12); if (subw == 0 && subh == 0) { for (i = 0; i < h; ++i) { for (j = 0; j < w; ++j) { const int m = mask[i * mask_stride + j]; dst[i * dst_stride + j] = VPX_BLEND_A64(m, src0[i * src0_stride + j], src1[i * src1_stride + j]); } } } else if (subw == 1 && subh == 1) { for (i = 0; i < h; ++i) { for (j = 0; j < w; ++j) { const int m = ROUND_POWER_OF_TWO(mask[(2 * i) * mask_stride + (2 * j)] + mask[(2 * i + 1) * mask_stride + (2 * j)] + mask[(2 * i) * mask_stride + (2 * j + 1)] + mask[(2 * i + 1) * mask_stride + (2 * j + 1)], 2); dst[i * dst_stride + j] = VPX_BLEND_A64(m, src0[i * src0_stride + j], src1[i * src1_stride + j]); } } } else if (subw == 1 && subh == 0) { for (i = 0; i < h; ++i) { for (j = 0; j < w; ++j) { const int m = VPX_BLEND_AVG(mask[i * mask_stride + (2 * j)], mask[i * mask_stride + (2 * j + 1)]); dst[i * dst_stride + j] = VPX_BLEND_A64(m, src0[i * src0_stride + j], src1[i * src1_stride + j]); } } } else { for (i = 0; i < h; ++i) { for (j = 0; j < w; ++j) { const int m = VPX_BLEND_AVG(mask[(2 * i) * mask_stride + j], mask[(2 * i + 1) * mask_stride + j]); dst[i * dst_stride + j] = VPX_BLEND_A64(m, src0[i * src0_stride + j], src1[i * src1_stride + j]); } } } } #endif // CONFIG_VPX_HIGHBITDEPTH