/* * Copyright (c) 2010 Mark Heath mjpeg0 @ silicontrip dot org * Copyright (c) 2014 Clément Bœsch * Copyright (c) 2014 Dave Rice @dericed * * This file is part of FFmpeg. * * FFmpeg is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public * License as published by the Free Software Foundation; either * version 2.1 of the License, or (at your option) any later version. * * FFmpeg is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public * License along with FFmpeg; if not, write to the Free Software * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA */ #include "libavutil/opt.h" #include "libavutil/pixdesc.h" #include "internal.h" enum FilterMode { FILTER_NONE = -1, FILTER_TOUT, FILTER_VREP, FILTER_BRNG, FILT_NUMB }; typedef struct { const AVClass *class; int chromah; // height of chroma plane int chromaw; // width of chroma plane int hsub; // horizontal subsampling int vsub; // vertical subsampling int fs; // pixel count per frame int cfs; // pixel count per frame of chroma planes int outfilter; // FilterMode int filters; AVFrame *frame_prev; uint8_t rgba_color[4]; int yuv_color[3]; int nb_jobs; int *jobs_rets; AVFrame *frame_sat; AVFrame *frame_hue; } SignalstatsContext; typedef struct ThreadData { const AVFrame *in; AVFrame *out; } ThreadData; typedef struct ThreadDataHueSatMetrics { const AVFrame *src; AVFrame *dst_sat, *dst_hue; } ThreadDataHueSatMetrics; #define OFFSET(x) offsetof(SignalstatsContext, x) #define FLAGS AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_VIDEO_PARAM static const AVOption signalstats_options[] = { {"stat", "set statistics filters", OFFSET(filters), AV_OPT_TYPE_FLAGS, {.i64=0}, 0, INT_MAX, FLAGS, "filters"}, {"tout", "analyze pixels for temporal outliers", 0, AV_OPT_TYPE_CONST, {.i64=1<<FILTER_TOUT}, 0, 0, FLAGS, "filters"}, {"vrep", "analyze video lines for vertical line repetition", 0, AV_OPT_TYPE_CONST, {.i64=1<<FILTER_VREP}, 0, 0, FLAGS, "filters"}, {"brng", "analyze for pixels outside of broadcast range", 0, AV_OPT_TYPE_CONST, {.i64=1<<FILTER_BRNG}, 0, 0, FLAGS, "filters"}, {"out", "set video filter", OFFSET(outfilter), AV_OPT_TYPE_INT, {.i64=FILTER_NONE}, -1, FILT_NUMB-1, FLAGS, "out"}, {"tout", "highlight pixels that depict temporal outliers", 0, AV_OPT_TYPE_CONST, {.i64=FILTER_TOUT}, 0, 0, FLAGS, "out"}, {"vrep", "highlight video lines that depict vertical line repetition", 0, AV_OPT_TYPE_CONST, {.i64=FILTER_VREP}, 0, 0, FLAGS, "out"}, {"brng", "highlight pixels that are outside of broadcast range", 0, AV_OPT_TYPE_CONST, {.i64=FILTER_BRNG}, 0, 0, FLAGS, "out"}, {"c", "set highlight color", OFFSET(rgba_color), AV_OPT_TYPE_COLOR, {.str="yellow"}, .flags=FLAGS}, {"color", "set highlight color", OFFSET(rgba_color), AV_OPT_TYPE_COLOR, {.str="yellow"}, .flags=FLAGS}, {NULL} }; AVFILTER_DEFINE_CLASS(signalstats); static av_cold int init(AVFilterContext *ctx) { uint8_t r, g, b; SignalstatsContext *s = ctx->priv; if (s->outfilter != FILTER_NONE) s->filters |= 1 << s->outfilter; r = s->rgba_color[0]; g = s->rgba_color[1]; b = s->rgba_color[2]; s->yuv_color[0] = (( 66*r + 129*g + 25*b + (1<<7)) >> 8) + 16; s->yuv_color[1] = ((-38*r + -74*g + 112*b + (1<<7)) >> 8) + 128; s->yuv_color[2] = ((112*r + -94*g + -18*b + (1<<7)) >> 8) + 128; return 0; } static av_cold void uninit(AVFilterContext *ctx) { SignalstatsContext *s = ctx->priv; av_frame_free(&s->frame_prev); av_frame_free(&s->frame_sat); av_frame_free(&s->frame_hue); av_freep(&s->jobs_rets); } static int query_formats(AVFilterContext *ctx) { // TODO: add more static const enum AVPixelFormat pix_fmts[] = { AV_PIX_FMT_YUV444P, AV_PIX_FMT_YUV422P, AV_PIX_FMT_YUV420P, AV_PIX_FMT_YUV411P, AV_PIX_FMT_YUV440P, AV_PIX_FMT_YUVJ422P, AV_PIX_FMT_YUVJ444P, AV_PIX_FMT_YUVJ420P, AV_PIX_FMT_YUVJ411P, AV_PIX_FMT_YUVJ440P, AV_PIX_FMT_NONE }; AVFilterFormats *fmts_list = ff_make_format_list(pix_fmts); if (!fmts_list) return AVERROR(ENOMEM); return ff_set_common_formats(ctx, fmts_list); } static AVFrame *alloc_frame(enum AVPixelFormat pixfmt, int w, int h) { AVFrame *frame = av_frame_alloc(); if (!frame) return NULL; frame->format = pixfmt; frame->width = w; frame->height = h; if (av_frame_get_buffer(frame, 32) < 0) { av_frame_free(&frame); return NULL; } return frame; } static int config_props(AVFilterLink *outlink) { AVFilterContext *ctx = outlink->src; SignalstatsContext *s = ctx->priv; AVFilterLink *inlink = outlink->src->inputs[0]; const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(outlink->format); s->hsub = desc->log2_chroma_w; s->vsub = desc->log2_chroma_h; outlink->w = inlink->w; outlink->h = inlink->h; s->chromaw = FF_CEIL_RSHIFT(inlink->w, s->hsub); s->chromah = FF_CEIL_RSHIFT(inlink->h, s->vsub); s->fs = inlink->w * inlink->h; s->cfs = s->chromaw * s->chromah; s->nb_jobs = FFMAX(1, FFMIN(inlink->h, ctx->graph->nb_threads)); s->jobs_rets = av_malloc_array(s->nb_jobs, sizeof(*s->jobs_rets)); if (!s->jobs_rets) return AVERROR(ENOMEM); s->frame_sat = alloc_frame(AV_PIX_FMT_GRAY8, inlink->w, inlink->h); s->frame_hue = alloc_frame(AV_PIX_FMT_GRAY16, inlink->w, inlink->h); if (!s->frame_sat || !s->frame_hue) return AVERROR(ENOMEM); return 0; } static void burn_frame(const SignalstatsContext *s, AVFrame *f, int x, int y) { const int chromax = x >> s->hsub; const int chromay = y >> s->vsub; f->data[0][y * f->linesize[0] + x] = s->yuv_color[0]; f->data[1][chromay * f->linesize[1] + chromax] = s->yuv_color[1]; f->data[2][chromay * f->linesize[2] + chromax] = s->yuv_color[2]; } static int filter_brng(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs) { ThreadData *td = arg; const SignalstatsContext *s = ctx->priv; const AVFrame *in = td->in; AVFrame *out = td->out; const int w = in->width; const int h = in->height; const int slice_start = (h * jobnr ) / nb_jobs; const int slice_end = (h * (jobnr+1)) / nb_jobs; int x, y, score = 0; for (y = slice_start; y < slice_end; y++) { const int yc = y >> s->vsub; const uint8_t *pluma = &in->data[0][y * in->linesize[0]]; const uint8_t *pchromau = &in->data[1][yc * in->linesize[1]]; const uint8_t *pchromav = &in->data[2][yc * in->linesize[2]]; for (x = 0; x < w; x++) { const int xc = x >> s->hsub; const int luma = pluma[x]; const int chromau = pchromau[xc]; const int chromav = pchromav[xc]; const int filt = luma < 16 || luma > 235 || chromau < 16 || chromau > 240 || chromav < 16 || chromav > 240; score += filt; if (out && filt) burn_frame(s, out, x, y); } } return score; } static int filter_tout_outlier(uint8_t x, uint8_t y, uint8_t z) { return ((abs(x - y) + abs (z - y)) / 2) - abs(z - x) > 4; // make 4 configurable? } static int filter_tout(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs) { ThreadData *td = arg; const SignalstatsContext *s = ctx->priv; const AVFrame *in = td->in; AVFrame *out = td->out; const int w = in->width; const int h = in->height; const int slice_start = (h * jobnr ) / nb_jobs; const int slice_end = (h * (jobnr+1)) / nb_jobs; const uint8_t *p = in->data[0]; int lw = in->linesize[0]; int x, y, score = 0, filt; for (y = slice_start; y < slice_end; y++) { if (y - 1 < 0 || y + 1 >= h) continue; // detect two pixels above and below (to eliminate interlace artefacts) // should check that video format is infact interlaced. #define FILTER(i, j) \ filter_tout_outlier(p[(y-j) * lw + x + i], \ p[ y * lw + x + i], \ p[(y+j) * lw + x + i]) #define FILTER3(j) (FILTER(-1, j) && FILTER(0, j) && FILTER(1, j)) if (y - 2 >= 0 && y + 2 < h) { for (x = 1; x < w - 1; x++) { filt = FILTER3(2) && FILTER3(1); score += filt; if (filt && out) burn_frame(s, out, x, y); } } else { for (x = 1; x < w - 1; x++) { filt = FILTER3(1); score += filt; if (filt && out) burn_frame(s, out, x, y); } } } return score; } #define VREP_START 4 static int filter_vrep(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs) { ThreadData *td = arg; const SignalstatsContext *s = ctx->priv; const AVFrame *in = td->in; AVFrame *out = td->out; const int w = in->width; const int h = in->height; const int slice_start = (h * jobnr ) / nb_jobs; const int slice_end = (h * (jobnr+1)) / nb_jobs; const uint8_t *p = in->data[0]; const int lw = in->linesize[0]; int x, y, score = 0; for (y = slice_start; y < slice_end; y++) { const int y2lw = (y - VREP_START) * lw; const int ylw = y * lw; int filt, totdiff = 0; if (y < VREP_START) continue; for (x = 0; x < w; x++) totdiff += abs(p[y2lw + x] - p[ylw + x]); filt = totdiff < w; score += filt; if (filt && out) for (x = 0; x < w; x++) burn_frame(s, out, x, y); } return score * w; } static const struct { const char *name; int (*process)(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs); } filters_def[] = { {"TOUT", filter_tout}, {"VREP", filter_vrep}, {"BRNG", filter_brng}, {NULL} }; #define DEPTH 256 static int compute_sat_hue_metrics(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs) { int i, j; ThreadDataHueSatMetrics *td = arg; const SignalstatsContext *s = ctx->priv; const AVFrame *src = td->src; AVFrame *dst_sat = td->dst_sat; AVFrame *dst_hue = td->dst_hue; const int slice_start = (s->chromah * jobnr ) / nb_jobs; const int slice_end = (s->chromah * (jobnr+1)) / nb_jobs; const int lsz_u = src->linesize[1]; const int lsz_v = src->linesize[2]; const uint8_t *p_u = src->data[1] + slice_start * lsz_u; const uint8_t *p_v = src->data[2] + slice_start * lsz_v; const int lsz_sat = dst_sat->linesize[0]; const int lsz_hue = dst_hue->linesize[0]; uint8_t *p_sat = dst_sat->data[0] + slice_start * lsz_sat; uint8_t *p_hue = dst_hue->data[0] + slice_start * lsz_hue; for (j = slice_start; j < slice_end; j++) { for (i = 0; i < s->chromaw; i++) { const int yuvu = p_u[i]; const int yuvv = p_v[i]; p_sat[i] = hypot(yuvu - 128, yuvv - 128); // int or round? ((int16_t*)p_hue)[i] = floor((180 / M_PI) * atan2f(yuvu-128, yuvv-128) + 180); } p_u += lsz_u; p_v += lsz_v; p_sat += lsz_sat; p_hue += lsz_hue; } return 0; } static int filter_frame(AVFilterLink *link, AVFrame *in) { AVFilterContext *ctx = link->dst; SignalstatsContext *s = ctx->priv; AVFilterLink *outlink = ctx->outputs[0]; AVFrame *out = in; int i, j; int w = 0, cw = 0, // in pw = 0, cpw = 0; // prev int fil; char metabuf[128]; unsigned int histy[DEPTH] = {0}, histu[DEPTH] = {0}, histv[DEPTH] = {0}, histhue[360] = {0}, histsat[DEPTH] = {0}; // limited to 8 bit data. int miny = -1, minu = -1, minv = -1; int maxy = -1, maxu = -1, maxv = -1; int lowy = -1, lowu = -1, lowv = -1; int highy = -1, highu = -1, highv = -1; int minsat = -1, maxsat = -1, lowsat = -1, highsat = -1; int lowp, highp, clowp, chighp; int accy, accu, accv; int accsat, acchue = 0; int medhue, maxhue; int toty = 0, totu = 0, totv = 0, totsat=0; int tothue = 0; int dify = 0, difu = 0, difv = 0; int filtot[FILT_NUMB] = {0}; AVFrame *prev; AVFrame *sat = s->frame_sat; AVFrame *hue = s->frame_hue; const uint8_t *p_sat = sat->data[0]; const uint8_t *p_hue = hue->data[0]; const int lsz_sat = sat->linesize[0]; const int lsz_hue = hue->linesize[0]; ThreadDataHueSatMetrics td_huesat = { .src = in, .dst_sat = sat, .dst_hue = hue, }; if (!s->frame_prev) s->frame_prev = av_frame_clone(in); prev = s->frame_prev; if (s->outfilter != FILTER_NONE) { out = av_frame_clone(in); av_frame_make_writable(out); } ctx->internal->execute(ctx, compute_sat_hue_metrics, &td_huesat, NULL, FFMIN(s->chromah, ctx->graph->nb_threads)); // Calculate luma histogram and difference with previous frame or field. for (j = 0; j < link->h; j++) { for (i = 0; i < link->w; i++) { const int yuv = in->data[0][w + i]; histy[yuv]++; dify += abs(yuv - prev->data[0][pw + i]); } w += in->linesize[0]; pw += prev->linesize[0]; } // Calculate chroma histogram and difference with previous frame or field. for (j = 0; j < s->chromah; j++) { for (i = 0; i < s->chromaw; i++) { const int yuvu = in->data[1][cw+i]; const int yuvv = in->data[2][cw+i]; histu[yuvu]++; difu += abs(yuvu - prev->data[1][cpw+i]); histv[yuvv]++; difv += abs(yuvv - prev->data[2][cpw+i]); histsat[p_sat[i]]++; histhue[((int16_t*)p_hue)[i]]++; } cw += in->linesize[1]; cpw += prev->linesize[1]; p_sat += lsz_sat; p_hue += lsz_hue; } for (fil = 0; fil < FILT_NUMB; fil ++) { if (s->filters & 1<<fil) { ThreadData td = { .in = in, .out = out != in && s->outfilter == fil ? out : NULL, }; memset(s->jobs_rets, 0, s->nb_jobs * sizeof(*s->jobs_rets)); ctx->internal->execute(ctx, filters_def[fil].process, &td, s->jobs_rets, s->nb_jobs); for (i = 0; i < s->nb_jobs; i++) filtot[fil] += s->jobs_rets[i]; } } // find low / high based on histogram percentile // these only need to be calculated once. lowp = lrint(s->fs * 10 / 100.); highp = lrint(s->fs * 90 / 100.); clowp = lrint(s->cfs * 10 / 100.); chighp = lrint(s->cfs * 90 / 100.); accy = accu = accv = accsat = 0; for (fil = 0; fil < DEPTH; fil++) { if (miny < 0 && histy[fil]) miny = fil; if (minu < 0 && histu[fil]) minu = fil; if (minv < 0 && histv[fil]) minv = fil; if (minsat < 0 && histsat[fil]) minsat = fil; if (histy[fil]) maxy = fil; if (histu[fil]) maxu = fil; if (histv[fil]) maxv = fil; if (histsat[fil]) maxsat = fil; toty += histy[fil] * fil; totu += histu[fil] * fil; totv += histv[fil] * fil; totsat += histsat[fil] * fil; accy += histy[fil]; accu += histu[fil]; accv += histv[fil]; accsat += histsat[fil]; if (lowy == -1 && accy >= lowp) lowy = fil; if (lowu == -1 && accu >= clowp) lowu = fil; if (lowv == -1 && accv >= clowp) lowv = fil; if (lowsat == -1 && accsat >= clowp) lowsat = fil; if (highy == -1 && accy >= highp) highy = fil; if (highu == -1 && accu >= chighp) highu = fil; if (highv == -1 && accv >= chighp) highv = fil; if (highsat == -1 && accsat >= chighp) highsat = fil; } maxhue = histhue[0]; medhue = -1; for (fil = 0; fil < 360; fil++) { tothue += histhue[fil] * fil; acchue += histhue[fil]; if (medhue == -1 && acchue > s->cfs / 2) medhue = fil; if (histhue[fil] > maxhue) { maxhue = histhue[fil]; } } av_frame_free(&s->frame_prev); s->frame_prev = av_frame_clone(in); #define SET_META(key, fmt, val) do { \ snprintf(metabuf, sizeof(metabuf), fmt, val); \ av_dict_set(&out->metadata, "lavfi.signalstats." key, metabuf, 0); \ } while (0) SET_META("YMIN", "%d", miny); SET_META("YLOW", "%d", lowy); SET_META("YAVG", "%g", 1.0 * toty / s->fs); SET_META("YHIGH", "%d", highy); SET_META("YMAX", "%d", maxy); SET_META("UMIN", "%d", minu); SET_META("ULOW", "%d", lowu); SET_META("UAVG", "%g", 1.0 * totu / s->cfs); SET_META("UHIGH", "%d", highu); SET_META("UMAX", "%d", maxu); SET_META("VMIN", "%d", minv); SET_META("VLOW", "%d", lowv); SET_META("VAVG", "%g", 1.0 * totv / s->cfs); SET_META("VHIGH", "%d", highv); SET_META("VMAX", "%d", maxv); SET_META("SATMIN", "%d", minsat); SET_META("SATLOW", "%d", lowsat); SET_META("SATAVG", "%g", 1.0 * totsat / s->cfs); SET_META("SATHIGH", "%d", highsat); SET_META("SATMAX", "%d", maxsat); SET_META("HUEMED", "%d", medhue); SET_META("HUEAVG", "%g", 1.0 * tothue / s->cfs); SET_META("YDIF", "%g", 1.0 * dify / s->fs); SET_META("UDIF", "%g", 1.0 * difu / s->cfs); SET_META("VDIF", "%g", 1.0 * difv / s->cfs); for (fil = 0; fil < FILT_NUMB; fil ++) { if (s->filters & 1<<fil) { char metaname[128]; snprintf(metabuf, sizeof(metabuf), "%g", 1.0 * filtot[fil] / s->fs); snprintf(metaname, sizeof(metaname), "lavfi.signalstats.%s", filters_def[fil].name); av_dict_set(&out->metadata, metaname, metabuf, 0); } } if (in != out) av_frame_free(&in); return ff_filter_frame(outlink, out); } static const AVFilterPad signalstats_inputs[] = { { .name = "default", .type = AVMEDIA_TYPE_VIDEO, .filter_frame = filter_frame, }, { NULL } }; static const AVFilterPad signalstats_outputs[] = { { .name = "default", .config_props = config_props, .type = AVMEDIA_TYPE_VIDEO, }, { NULL } }; AVFilter ff_vf_signalstats = { .name = "signalstats", .description = "Generate statistics from video analysis.", .init = init, .uninit = uninit, .query_formats = query_formats, .priv_size = sizeof(SignalstatsContext), .inputs = signalstats_inputs, .outputs = signalstats_outputs, .priv_class = &signalstats_class, .flags = AVFILTER_FLAG_SLICE_THREADS, };