2c2e69b3a2
Signed-off-by: Paul B Mahol <onemda@gmail.com>
497 lines
21 KiB
C
497 lines
21 KiB
C
/*
|
|
* Copyright (c) 2013 Paul B Mahol
|
|
*
|
|
* This file is part of FFmpeg.
|
|
*
|
|
* FFmpeg is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with FFmpeg; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
#include "libavutil/imgutils.h"
|
|
#include "libavutil/eval.h"
|
|
#include "libavutil/opt.h"
|
|
#include "libavutil/pixfmt.h"
|
|
#include "avfilter.h"
|
|
#include "bufferqueue.h"
|
|
#include "formats.h"
|
|
#include "internal.h"
|
|
#include "video.h"
|
|
|
|
#define TOP 0
|
|
#define BOTTOM 1
|
|
|
|
enum BlendMode {
|
|
BLEND_UNSET = -1,
|
|
BLEND_NORMAL,
|
|
BLEND_ADDITION,
|
|
BLEND_AND,
|
|
BLEND_AVERAGE,
|
|
BLEND_BURN,
|
|
BLEND_DARKEN,
|
|
BLEND_DIFFERENCE,
|
|
BLEND_DIVIDE,
|
|
BLEND_DODGE,
|
|
BLEND_EXCLUSION,
|
|
BLEND_HARDLIGHT,
|
|
BLEND_LIGHTEN,
|
|
BLEND_MULTIPLY,
|
|
BLEND_NEGATION,
|
|
BLEND_OR,
|
|
BLEND_OVERLAY,
|
|
BLEND_PHOENIX,
|
|
BLEND_PINLIGHT,
|
|
BLEND_REFLECT,
|
|
BLEND_SCREEN,
|
|
BLEND_SOFTLIGHT,
|
|
BLEND_SUBTRACT,
|
|
BLEND_VIVIDLIGHT,
|
|
BLEND_XOR,
|
|
BLEND_NB
|
|
};
|
|
|
|
static const char *const var_names[] = { "X", "Y", "W", "H", "SW", "SH", "T", "N", "A", "B", "TOP", "BOTTOM", NULL };
|
|
enum { VAR_X, VAR_Y, VAR_W, VAR_H, VAR_SW, VAR_SH, VAR_T, VAR_N, VAR_A, VAR_B, VAR_TOP, VAR_BOTTOM, VAR_VARS_NB };
|
|
|
|
typedef struct FilterParams {
|
|
enum BlendMode mode;
|
|
double opacity;
|
|
AVExpr *e;
|
|
char *expr_str;
|
|
void (*blend)(const uint8_t *top, int top_linesize,
|
|
const uint8_t *bottom, int bottom_linesize,
|
|
uint8_t *dst, int dst_linesize,
|
|
int width, int start, int end,
|
|
struct FilterParams *param, double *values);
|
|
} FilterParams;
|
|
|
|
typedef struct ThreadData {
|
|
AVFrame *top, *bottom, *dst;
|
|
AVFilterLink *inlink;
|
|
int plane;
|
|
int w, h;
|
|
FilterParams *param;
|
|
} ThreadData;
|
|
|
|
typedef struct {
|
|
const AVClass *class;
|
|
struct FFBufQueue queue_top;
|
|
struct FFBufQueue queue_bottom;
|
|
int hsub, vsub; ///< chroma subsampling values
|
|
int nb_planes;
|
|
int frame_requested;
|
|
char *all_expr;
|
|
enum BlendMode all_mode;
|
|
double all_opacity;
|
|
|
|
FilterParams params[4];
|
|
} BlendContext;
|
|
|
|
#define OFFSET(x) offsetof(BlendContext, x)
|
|
#define FLAGS AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_VIDEO_PARAM
|
|
|
|
static const AVOption blend_options[] = {
|
|
{ "c0_mode", "set component #0 blend mode", OFFSET(params[0].mode), AV_OPT_TYPE_INT, {.i64=0}, 0, BLEND_NB-1, FLAGS, "mode"},
|
|
{ "c1_mode", "set component #1 blend mode", OFFSET(params[1].mode), AV_OPT_TYPE_INT, {.i64=0}, 0, BLEND_NB-1, FLAGS, "mode"},
|
|
{ "c2_mode", "set component #2 blend mode", OFFSET(params[2].mode), AV_OPT_TYPE_INT, {.i64=0}, 0, BLEND_NB-1, FLAGS, "mode"},
|
|
{ "c3_mode", "set component #3 blend mode", OFFSET(params[3].mode), AV_OPT_TYPE_INT, {.i64=0}, 0, BLEND_NB-1, FLAGS, "mode"},
|
|
{ "all_mode", "set blend mode for all components", OFFSET(all_mode), AV_OPT_TYPE_INT, {.i64=-1},-1, BLEND_NB-1, FLAGS, "mode"},
|
|
{ "addition", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_ADDITION}, 0, 0, FLAGS, "mode" },
|
|
{ "and", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_AND}, 0, 0, FLAGS, "mode" },
|
|
{ "average", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_AVERAGE}, 0, 0, FLAGS, "mode" },
|
|
{ "burn", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_BURN}, 0, 0, FLAGS, "mode" },
|
|
{ "darken", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_DARKEN}, 0, 0, FLAGS, "mode" },
|
|
{ "difference", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_DIFFERENCE}, 0, 0, FLAGS, "mode" },
|
|
{ "divide", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_DIVIDE}, 0, 0, FLAGS, "mode" },
|
|
{ "dodge", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_DODGE}, 0, 0, FLAGS, "mode" },
|
|
{ "exclusion", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_EXCLUSION}, 0, 0, FLAGS, "mode" },
|
|
{ "hardlight", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_HARDLIGHT}, 0, 0, FLAGS, "mode" },
|
|
{ "lighten", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_LIGHTEN}, 0, 0, FLAGS, "mode" },
|
|
{ "multiply", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_MULTIPLY}, 0, 0, FLAGS, "mode" },
|
|
{ "negation", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_NEGATION}, 0, 0, FLAGS, "mode" },
|
|
{ "normal", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_NORMAL}, 0, 0, FLAGS, "mode" },
|
|
{ "or", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_OR}, 0, 0, FLAGS, "mode" },
|
|
{ "overlay", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_OVERLAY}, 0, 0, FLAGS, "mode" },
|
|
{ "phoenix", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_PHOENIX}, 0, 0, FLAGS, "mode" },
|
|
{ "pinlight", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_PINLIGHT}, 0, 0, FLAGS, "mode" },
|
|
{ "reflect", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_REFLECT}, 0, 0, FLAGS, "mode" },
|
|
{ "screen", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_SCREEN}, 0, 0, FLAGS, "mode" },
|
|
{ "softlight", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_SOFTLIGHT}, 0, 0, FLAGS, "mode" },
|
|
{ "subtract", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_SUBTRACT}, 0, 0, FLAGS, "mode" },
|
|
{ "vividlight", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_VIVIDLIGHT}, 0, 0, FLAGS, "mode" },
|
|
{ "xor", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_XOR}, 0, 0, FLAGS, "mode" },
|
|
{ "c0_expr", "set color component #0 expression", OFFSET(params[0].expr_str), AV_OPT_TYPE_STRING, {.str=NULL}, CHAR_MIN, CHAR_MAX, FLAGS },
|
|
{ "c1_expr", "set color component #1 expression", OFFSET(params[1].expr_str), AV_OPT_TYPE_STRING, {.str=NULL}, CHAR_MIN, CHAR_MAX, FLAGS },
|
|
{ "c2_expr", "set color component #2 expression", OFFSET(params[2].expr_str), AV_OPT_TYPE_STRING, {.str=NULL}, CHAR_MIN, CHAR_MAX, FLAGS },
|
|
{ "c3_expr", "set color component #3 expression", OFFSET(params[3].expr_str), AV_OPT_TYPE_STRING, {.str=NULL}, CHAR_MIN, CHAR_MAX, FLAGS },
|
|
{ "all_expr", "set expression for all color components", OFFSET(all_expr), AV_OPT_TYPE_STRING, {.str=NULL}, CHAR_MIN, CHAR_MAX, FLAGS },
|
|
{ "c0_opacity", "set color component #0 opacity", OFFSET(params[0].opacity), AV_OPT_TYPE_DOUBLE, {.dbl=1}, 0, 1, FLAGS },
|
|
{ "c1_opacity", "set color component #1 opacity", OFFSET(params[1].opacity), AV_OPT_TYPE_DOUBLE, {.dbl=1}, 0, 1, FLAGS },
|
|
{ "c2_opacity", "set color component #2 opacity", OFFSET(params[2].opacity), AV_OPT_TYPE_DOUBLE, {.dbl=1}, 0, 1, FLAGS },
|
|
{ "c3_opacity", "set color component #3 opacity", OFFSET(params[3].opacity), AV_OPT_TYPE_DOUBLE, {.dbl=1}, 0, 1, FLAGS },
|
|
{ "all_opacity", "set opacity for all color components", OFFSET(all_opacity), AV_OPT_TYPE_DOUBLE, {.dbl=1}, 0, 1, FLAGS},
|
|
{ NULL },
|
|
};
|
|
|
|
AVFILTER_DEFINE_CLASS(blend);
|
|
|
|
static void blend_normal(const uint8_t *top, int top_linesize,
|
|
const uint8_t *bottom, int bottom_linesize,
|
|
uint8_t *dst, int dst_linesize,
|
|
int width, int start, int end,
|
|
FilterParams *param, double *values)
|
|
{
|
|
av_image_copy_plane(dst, dst_linesize, top, top_linesize, width, end - start);
|
|
}
|
|
|
|
#define DEFINE_BLEND(name, expr) \
|
|
static void blend_## name(const uint8_t *top, int top_linesize, \
|
|
const uint8_t *bottom, int bottom_linesize, \
|
|
uint8_t *dst, int dst_linesize, \
|
|
int width, int start, int end, \
|
|
FilterParams *param, double *values) \
|
|
{ \
|
|
double opacity = param->opacity; \
|
|
int i, j; \
|
|
\
|
|
for (i = start; i < end; i++) { \
|
|
for (j = 0; j < width; j++) { \
|
|
dst[j] = top[j] + ((expr) - top[j]) * opacity; \
|
|
} \
|
|
dst += dst_linesize; \
|
|
top += top_linesize; \
|
|
bottom += bottom_linesize; \
|
|
} \
|
|
}
|
|
|
|
#define A top[j]
|
|
#define B bottom[j]
|
|
|
|
#define MULTIPLY(x, a, b) (x * ((a * b) / 255))
|
|
#define SCREEN(x, a, b) (255 - x * ((255 - a) * (255 - b) / 255))
|
|
#define BURN(a, b) ((a == 0) ? a : FFMAX(0, 255 - ((255 - b) << 8) / a))
|
|
#define DODGE(a, b) ((a == 255) ? a : FFMIN(255, ((b << 8) / (255 - a))))
|
|
|
|
DEFINE_BLEND(addition, FFMIN(255, A + B))
|
|
DEFINE_BLEND(average, (A + B) / 2)
|
|
DEFINE_BLEND(subtract, FFMAX(0, A - B))
|
|
DEFINE_BLEND(multiply, MULTIPLY(1, A, B))
|
|
DEFINE_BLEND(negation, 255 - FFABS(255 - A - B))
|
|
DEFINE_BLEND(difference, FFABS(A - B))
|
|
DEFINE_BLEND(screen, SCREEN(1, A, B))
|
|
DEFINE_BLEND(overlay, (A < 128) ? MULTIPLY(2, A, B) : SCREEN(2, A, B))
|
|
DEFINE_BLEND(hardlight, (B < 128) ? MULTIPLY(2, B, A) : SCREEN(2, B, A))
|
|
DEFINE_BLEND(darken, FFMIN(A, B))
|
|
DEFINE_BLEND(lighten, FFMAX(A, B))
|
|
DEFINE_BLEND(divide, ((float)A / ((float)B) * 255))
|
|
DEFINE_BLEND(dodge, DODGE(A, B))
|
|
DEFINE_BLEND(burn, BURN(A, B))
|
|
DEFINE_BLEND(softlight, (A > 127) ? B + (255 - B) * (A - 127.5) / 127.5 * (0.5 - FFABS(B - 127.5) / 255): B - B * ((127.5 - A) / 127.5) * (0.5 - FFABS(B - 127.5)/255))
|
|
DEFINE_BLEND(exclusion, A + B - 2 * A * B / 255)
|
|
DEFINE_BLEND(pinlight, (B < 128) ? FFMIN(A, 2 * B) : FFMAX(A, 2 * (B - 128)))
|
|
DEFINE_BLEND(phoenix, FFMIN(A, B) - FFMAX(A, B) + 255)
|
|
DEFINE_BLEND(reflect, (B == 255) ? B : FFMIN(255, (A * A / (255 - B))))
|
|
DEFINE_BLEND(and, A & B)
|
|
DEFINE_BLEND(or, A | B)
|
|
DEFINE_BLEND(xor, A ^ B)
|
|
DEFINE_BLEND(vividlight, (B < 128) ? BURN(A, 2 * B) : DODGE(A, 2 * (B - 128)))
|
|
|
|
static void blend_expr(const uint8_t *top, int top_linesize,
|
|
const uint8_t *bottom, int bottom_linesize,
|
|
uint8_t *dst, int dst_linesize,
|
|
int width, int start, int end,
|
|
FilterParams *param, double *values)
|
|
{
|
|
AVExpr *e = param->e;
|
|
int y, x;
|
|
|
|
for (y = start; y < end; y++) {
|
|
values[VAR_Y] = y;
|
|
for (x = 0; x < width; x++) {
|
|
values[VAR_X] = x;
|
|
values[VAR_TOP] = values[VAR_A] = top[x];
|
|
values[VAR_BOTTOM] = values[VAR_B] = bottom[x];
|
|
dst[x] = av_expr_eval(e, values, NULL);
|
|
}
|
|
dst += dst_linesize;
|
|
top += top_linesize;
|
|
bottom += bottom_linesize;
|
|
}
|
|
}
|
|
|
|
static av_cold int init(AVFilterContext *ctx)
|
|
{
|
|
BlendContext *b = ctx->priv;
|
|
int ret, plane;
|
|
|
|
for (plane = 0; plane < FF_ARRAY_ELEMS(b->params); plane++) {
|
|
FilterParams *param = &b->params[plane];
|
|
|
|
if (b->all_mode >= 0)
|
|
param->mode = b->all_mode;
|
|
if (b->all_opacity < 1)
|
|
param->opacity = b->all_opacity;
|
|
|
|
switch (param->mode) {
|
|
case BLEND_ADDITION: param->blend = blend_addition; break;
|
|
case BLEND_AND: param->blend = blend_and; break;
|
|
case BLEND_AVERAGE: param->blend = blend_average; break;
|
|
case BLEND_BURN: param->blend = blend_burn; break;
|
|
case BLEND_DARKEN: param->blend = blend_darken; break;
|
|
case BLEND_DIFFERENCE: param->blend = blend_difference; break;
|
|
case BLEND_DIVIDE: param->blend = blend_divide; break;
|
|
case BLEND_DODGE: param->blend = blend_dodge; break;
|
|
case BLEND_EXCLUSION: param->blend = blend_exclusion; break;
|
|
case BLEND_HARDLIGHT: param->blend = blend_hardlight; break;
|
|
case BLEND_LIGHTEN: param->blend = blend_lighten; break;
|
|
case BLEND_MULTIPLY: param->blend = blend_multiply; break;
|
|
case BLEND_NEGATION: param->blend = blend_negation; break;
|
|
case BLEND_NORMAL: param->blend = blend_normal; break;
|
|
case BLEND_OR: param->blend = blend_or; break;
|
|
case BLEND_OVERLAY: param->blend = blend_overlay; break;
|
|
case BLEND_PHOENIX: param->blend = blend_phoenix; break;
|
|
case BLEND_PINLIGHT: param->blend = blend_pinlight; break;
|
|
case BLEND_REFLECT: param->blend = blend_reflect; break;
|
|
case BLEND_SCREEN: param->blend = blend_screen; break;
|
|
case BLEND_SOFTLIGHT: param->blend = blend_softlight; break;
|
|
case BLEND_SUBTRACT: param->blend = blend_subtract; break;
|
|
case BLEND_VIVIDLIGHT: param->blend = blend_vividlight; break;
|
|
case BLEND_XOR: param->blend = blend_xor; break;
|
|
}
|
|
|
|
if (b->all_expr && !param->expr_str) {
|
|
param->expr_str = av_strdup(b->all_expr);
|
|
if (!param->expr_str)
|
|
return AVERROR(ENOMEM);
|
|
}
|
|
if (param->expr_str) {
|
|
ret = av_expr_parse(¶m->e, param->expr_str, var_names,
|
|
NULL, NULL, NULL, NULL, 0, ctx);
|
|
if (ret < 0)
|
|
return ret;
|
|
param->blend = blend_expr;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int query_formats(AVFilterContext *ctx)
|
|
{
|
|
static const enum AVPixelFormat pix_fmts[] = {
|
|
AV_PIX_FMT_YUVA444P, AV_PIX_FMT_YUVA422P, AV_PIX_FMT_YUVA420P,
|
|
AV_PIX_FMT_YUVJ444P, AV_PIX_FMT_YUVJ440P, AV_PIX_FMT_YUVJ422P,AV_PIX_FMT_YUVJ420P, AV_PIX_FMT_YUVJ411P,
|
|
AV_PIX_FMT_YUV444P, AV_PIX_FMT_YUV440P, AV_PIX_FMT_YUV422P, AV_PIX_FMT_YUV420P, AV_PIX_FMT_YUV411P, AV_PIX_FMT_YUV410P,
|
|
AV_PIX_FMT_GBRP, AV_PIX_FMT_GBRAP, AV_PIX_FMT_GRAY8, AV_PIX_FMT_NONE
|
|
};
|
|
|
|
ff_set_common_formats(ctx, ff_make_format_list(pix_fmts));
|
|
return 0;
|
|
}
|
|
|
|
static int config_output(AVFilterLink *outlink)
|
|
{
|
|
AVFilterContext *ctx = outlink->src;
|
|
AVFilterLink *toplink = ctx->inputs[TOP];
|
|
AVFilterLink *bottomlink = ctx->inputs[BOTTOM];
|
|
BlendContext *b = ctx->priv;
|
|
const AVPixFmtDescriptor *pix_desc = av_pix_fmt_desc_get(toplink->format);
|
|
|
|
if (toplink->format != bottomlink->format) {
|
|
av_log(ctx, AV_LOG_ERROR, "inputs must be of same pixel format\n");
|
|
return AVERROR(EINVAL);
|
|
}
|
|
if (toplink->w != bottomlink->w ||
|
|
toplink->h != bottomlink->h ||
|
|
toplink->sample_aspect_ratio.num != bottomlink->sample_aspect_ratio.num ||
|
|
toplink->sample_aspect_ratio.den != bottomlink->sample_aspect_ratio.den) {
|
|
av_log(ctx, AV_LOG_ERROR, "First input link %s parameters "
|
|
"(size %dx%d, SAR %d:%d) do not match the corresponding "
|
|
"second input link %s parameters (%dx%d, SAR %d:%d)\n",
|
|
ctx->input_pads[TOP].name, toplink->w, toplink->h,
|
|
toplink->sample_aspect_ratio.num,
|
|
toplink->sample_aspect_ratio.den,
|
|
ctx->input_pads[BOTTOM].name, bottomlink->w, bottomlink->h,
|
|
bottomlink->sample_aspect_ratio.num,
|
|
bottomlink->sample_aspect_ratio.den);
|
|
return AVERROR(EINVAL);
|
|
}
|
|
|
|
outlink->w = toplink->w;
|
|
outlink->h = toplink->h;
|
|
outlink->time_base = toplink->time_base;
|
|
outlink->sample_aspect_ratio = toplink->sample_aspect_ratio;
|
|
outlink->frame_rate = toplink->frame_rate;
|
|
|
|
b->hsub = pix_desc->log2_chroma_w;
|
|
b->vsub = pix_desc->log2_chroma_h;
|
|
b->nb_planes = av_pix_fmt_count_planes(toplink->format);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static av_cold void uninit(AVFilterContext *ctx)
|
|
{
|
|
BlendContext *b = ctx->priv;
|
|
int i;
|
|
|
|
ff_bufqueue_discard_all(&b->queue_top);
|
|
ff_bufqueue_discard_all(&b->queue_bottom);
|
|
|
|
for (i = 0; i < FF_ARRAY_ELEMS(b->params); i++)
|
|
av_expr_free(b->params[i].e);
|
|
}
|
|
|
|
static int request_frame(AVFilterLink *outlink)
|
|
{
|
|
AVFilterContext *ctx = outlink->src;
|
|
BlendContext *b = ctx->priv;
|
|
int in, ret;
|
|
|
|
b->frame_requested = 1;
|
|
while (b->frame_requested) {
|
|
in = ff_bufqueue_peek(&b->queue_top, 0) ? BOTTOM : TOP;
|
|
ret = ff_request_frame(ctx->inputs[in]);
|
|
if (ret < 0)
|
|
return ret;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int filter_slice(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs)
|
|
{
|
|
ThreadData *td = arg;
|
|
int slice_start = (td->h * jobnr ) / nb_jobs;
|
|
int slice_end = (td->h * (jobnr+1)) / nb_jobs;
|
|
const uint8_t *top = td->top->data[td->plane];
|
|
const uint8_t *bottom = td->bottom->data[td->plane];
|
|
uint8_t *dst = td->dst->data[td->plane];
|
|
double values[VAR_VARS_NB];
|
|
|
|
values[VAR_N] = td->inlink->frame_count;
|
|
values[VAR_T] = td->dst->pts == AV_NOPTS_VALUE ? NAN : td->dst->pts * av_q2d(td->inlink->time_base);
|
|
values[VAR_W] = td->w;
|
|
values[VAR_H] = td->h;
|
|
values[VAR_SW] = td->w / (double)td->dst->width;
|
|
values[VAR_SH] = td->h / (double)td->dst->height;
|
|
|
|
td->param->blend(top + slice_start * td->top->linesize[td->plane],
|
|
td->top->linesize[td->plane],
|
|
bottom + slice_start * td->bottom->linesize[td->plane],
|
|
td->bottom->linesize[td->plane],
|
|
dst + slice_start * td->dst->linesize[td->plane],
|
|
td->dst->linesize[td->plane],
|
|
td->w, slice_start, slice_end, td->param, &values[0]);
|
|
return 0;
|
|
}
|
|
|
|
static void blend_frame(AVFilterContext *ctx,
|
|
AVFrame *top_buf,
|
|
AVFrame *bottom_buf,
|
|
AVFrame *dst_buf)
|
|
{
|
|
BlendContext *b = ctx->priv;
|
|
AVFilterLink *inlink = ctx->inputs[0];
|
|
int plane;
|
|
|
|
for (plane = 0; plane < b->nb_planes; plane++) {
|
|
int hsub = plane == 1 || plane == 2 ? b->hsub : 0;
|
|
int vsub = plane == 1 || plane == 2 ? b->vsub : 0;
|
|
int outw = FF_CEIL_RSHIFT(dst_buf->width, hsub);
|
|
int outh = FF_CEIL_RSHIFT(dst_buf->height, vsub);
|
|
FilterParams *param = &b->params[plane];
|
|
ThreadData td = { .top = top_buf, .bottom = bottom_buf, .dst = dst_buf,
|
|
.w = outw, .h = outh, .param = param, .plane = plane,
|
|
.inlink = inlink };
|
|
|
|
ctx->internal->execute(ctx, filter_slice, &td, NULL, FFMIN(outh, ctx->graph->nb_threads));
|
|
}
|
|
}
|
|
|
|
static int filter_frame(AVFilterLink *inlink, AVFrame *buf)
|
|
{
|
|
AVFilterContext *ctx = inlink->dst;
|
|
AVFilterLink *outlink = ctx->outputs[0];
|
|
BlendContext *b = ctx->priv;
|
|
|
|
int ret = 0;
|
|
int is_bottom = (inlink == ctx->inputs[BOTTOM]);
|
|
struct FFBufQueue *queue =
|
|
(is_bottom ? &b->queue_bottom : &b->queue_top);
|
|
ff_bufqueue_add(ctx, queue, buf);
|
|
|
|
while (1) {
|
|
AVFrame *top_buf, *bottom_buf, *out_buf;
|
|
|
|
if (!ff_bufqueue_peek(&b->queue_top, 0) ||
|
|
!ff_bufqueue_peek(&b->queue_bottom, 0)) break;
|
|
|
|
top_buf = ff_bufqueue_get(&b->queue_top);
|
|
bottom_buf = ff_bufqueue_get(&b->queue_bottom);
|
|
|
|
if (!ctx->is_disabled) {
|
|
out_buf = ff_get_video_buffer(outlink, outlink->w, outlink->h);
|
|
if (!out_buf)
|
|
return AVERROR(ENOMEM);
|
|
av_frame_copy_props(out_buf, top_buf);
|
|
blend_frame(ctx, top_buf, bottom_buf, out_buf);
|
|
} else {
|
|
out_buf = av_frame_clone(top_buf);
|
|
if (!out_buf)
|
|
return AVERROR(ENOMEM);
|
|
}
|
|
|
|
b->frame_requested = 0;
|
|
ret = ff_filter_frame(outlink, out_buf);
|
|
av_frame_free(&top_buf);
|
|
av_frame_free(&bottom_buf);
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
static const AVFilterPad blend_inputs[] = {
|
|
{
|
|
.name = "top",
|
|
.type = AVMEDIA_TYPE_VIDEO,
|
|
.filter_frame = filter_frame,
|
|
},{
|
|
.name = "bottom",
|
|
.type = AVMEDIA_TYPE_VIDEO,
|
|
.filter_frame = filter_frame,
|
|
},
|
|
{ NULL }
|
|
};
|
|
|
|
static const AVFilterPad blend_outputs[] = {
|
|
{
|
|
.name = "default",
|
|
.type = AVMEDIA_TYPE_VIDEO,
|
|
.config_props = config_output,
|
|
.request_frame = request_frame,
|
|
},
|
|
{ NULL }
|
|
};
|
|
|
|
AVFilter avfilter_vf_blend = {
|
|
.name = "blend",
|
|
.description = NULL_IF_CONFIG_SMALL("Blend two video frames into each other."),
|
|
.init = init,
|
|
.uninit = uninit,
|
|
.priv_size = sizeof(BlendContext),
|
|
.query_formats = query_formats,
|
|
.inputs = blend_inputs,
|
|
.outputs = blend_outputs,
|
|
.priv_class = &blend_class,
|
|
.flags = AVFILTER_FLAG_SUPPORT_TIMELINE_INTERNAL | AVFILTER_FLAG_SLICE_THREADS,
|
|
};
|