47 int jobnr,
int nb_jobs);
55 for (i = 0; i < 9; ++
i) {
59 diff += sqrt((du * du + dv * dv) / (255.0 * 255.0 * 2));
64 if (ctx->
blend > 0.0001) {
65 return av_clipd((diff - ctx->
similarity) / ctx->
blend, 0.0, 1.0) * 255.0;
77 for (i = 0; i < 9; ++
i) {
81 diff += sqrt((du * du + dv * dv) / (max * max * 2));
86 if (ctx->
blend > 0.0001) {
95 if (x < 0 || x >= frame->
width || y < 0 || y >= frame->
height)
107 if (x < 0 || x >= frame->
width || y < 0 || y >= frame->
height)
121 const int slice_start = (frame->
height * jobnr) / nb_jobs;
132 for (y = slice_start; y <
slice_end; ++y) {
133 for (x = 0; x < frame->
width; ++x) {
134 for (yo = 0; yo < 3; ++yo) {
135 for (xo = 0; xo < 3; ++xo) {
151 const int slice_start = (frame->
height * jobnr) / nb_jobs;
159 for (
int i = 0;
i < 9;
i++) {
164 for (y = slice_start; y <
slice_end; ++y) {
165 for (x = 0; x < frame->
width; ++x) {
166 uint16_t *dst = (uint16_t *)(frame->
data[3] + frame->
linesize[3] * y);
168 for (yo = 0; yo < 3; ++yo) {
169 for (xo = 0; xo < 3; ++xo) {
185 const int slice_start = ((frame->
height >> ctx->
vsub_log2) * jobnr) / nb_jobs;
190 for (y = slice_start; y <
slice_end; ++y) {
200 diff = sqrt((du * du + dv * dv) / (255.0 * 255.0));
203 if (ctx->
blend > 0.0001) {
206 frame->
data[1][frame->
linesize[1] * y + x] = 128 + (u - 128) * f;
207 frame->
data[2][frame->
linesize[2] * y + x] = 128 + (v - 128) * f;
222 const int slice_start = ((frame->
height >> ctx->
vsub_log2) * jobnr) / nb_jobs;
229 for (y = slice_start; y <
slice_end; ++y) {
239 diff = sqrt((du * du + dv * dv) / (max * max));
242 if (ctx->
blend > 0.0001) {
269 #define FIXNUM(x) lrint((x) * (1 << 10)) 270 #define RGB_TO_U(rgb) (((- FIXNUM(0.16874) * rgb[0] - FIXNUM(0.33126) * rgb[1] + FIXNUM(0.50000) * rgb[2] + (1 << 9) - 1) >> 10) + 128) 271 #define RGB_TO_V(rgb) ((( FIXNUM(0.50000) * rgb[0] - FIXNUM(0.41869) * rgb[1] - FIXNUM(0.08131) * rgb[2] + (1 << 9) - 1) >> 10) + 128) 284 factor = 1 << (ctx->
depth - 8);
294 if (!strcmp(avctx->
filter->
name,
"chromakey")) {
357 char *res,
int res_len,
int flags)
388 #define OFFSET(x) offsetof(ChromakeyContext, x) 389 #define FLAGS AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_VIDEO_PARAM|AV_OPT_FLAG_RUNTIME_PARAM 403 .description =
NULL_IF_CONFIG_SMALL(
"Turns a certain color into transparency. Operates on YUV colors."),
405 .priv_class = &chromakey_class,
407 .
inputs = chromakey_inputs,
444 .
name =
"chromahold",
447 .priv_class = &chromahold_class,
449 .
inputs = chromahold_inputs,
#define AV_PIX_FMT_YUVA422P16
#define AV_PIX_FMT_YUVA422P9
const AVPixFmtDescriptor * av_pix_fmt_desc_get(enum AVPixelFormat pix_fmt)
This structure describes decoded (raw) audio or video data.
#define AV_PIX_FMT_YUVA420P10
#define AV_PIX_FMT_YUV444P14
#define AV_PIX_FMT_YUVA422P10
planar YUV 4:4:4, 24bpp, (1 Cr & Cb sample per 1x1 Y samples)
Main libavfilter public API header.
int(* do_slice)(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs)
AVFILTER_DEFINE_CLASS(chromakey)
#define AV_PIX_FMT_YUV420P12
uint8_t log2_chroma_w
Amount to shift the luma width right to find the chroma width.
static const AVOption chromahold_options[]
static av_always_inline void get_pixel16_uv(AVFrame *frame, int hsub_log2, int vsub_log2, int x, int y, uint16_t *u, uint16_t *v)
#define AVFILTER_FLAG_SUPPORT_TIMELINE_GENERIC
Some filters support a generic "enable" expression option that can be used to enable or disable a fil...
const char * name
Pad name.
int ff_filter_frame(AVFilterLink *link, AVFrame *frame)
Send a frame of data to the next filter.
planar YUV 4:2:0, 20bpp, (1 Cr & Cb sample per 2x2 Y & A samples)
AVComponentDescriptor comp[4]
Parameters that describe how pixels are packed.
AVFilter ff_vf_chromahold
static const AVFilterPad chromahold_inputs[]
static const AVFilterPad chromakey_outputs[]
#define u(width, name, range_min, range_max)
#define AV_PIX_FMT_YUVA420P9
#define AV_PIX_FMT_YUV444P16
static int process_command(AVFilterContext *ctx, const char *cmd, const char *args, char *res, int res_len, int flags)
#define AV_PIX_FMT_YUV422P12
#define AV_PIX_FMT_YUVA420P16
A filter pad used for either input or output.
static const AVOption chromakey_options[]
A link between two filters.
planar YUV 4:2:2 24bpp, (1 Cr & Cb sample per 2x1 Y & A samples)
#define i(width, name, range_min, range_max)
uint8_t log2_chroma_h
Amount to shift the luma height right to find the chroma height.
static uint8_t do_chromakey_pixel(ChromakeyContext *ctx, uint8_t u[9], uint8_t v[9])
#define NULL_IF_CONFIG_SMALL(x)
Return NULL if CONFIG_SMALL is true, otherwise the argument without modification. ...
int ff_filter_process_command(AVFilterContext *ctx, const char *cmd, const char *arg, char *res, int res_len, int flags)
Generic processing of user supplied commands that are set in the same way as the filter options...
void * priv
private data for use by the filter
#define AVFILTER_FLAG_SLICE_THREADS
The filter supports multithreading by splitting frames into multiple parts and processing them concur...
#define AV_PIX_FMT_YUVA444P16
#define AV_PIX_FMT_YUV444P10
planar YUV 4:2:2, 16bpp, (1 Cr & Cb sample per 2x1 Y samples)
static const AVFilterPad chromahold_outputs[]
#define AV_PIX_FMT_YUV422P9
int ff_filter_get_nb_threads(AVFilterContext *ctx)
Get number of threads for current filter instance.
#define AV_PIX_FMT_YUVA444P12
uint8_t chromakey_rgba[4]
static int do_chromahold16_slice(AVFilterContext *avctx, void *arg, int jobnr, int nb_jobs)
AVFilterContext * src
source filter
#define AV_PIX_FMT_YUVA444P10
static int filter_frame(AVFilterLink *link, AVFrame *frame)
static const AVFilterPad inputs[]
#define AV_PIX_FMT_YUV444P9
static const AVFilterPad outputs[]
int format
agreed upon media format
#define AV_PIX_FMT_YUV420P16
#define AV_PIX_FMT_YUV420P14
static int do_chromakey16_slice(AVFilterContext *avctx, void *arg, int jobnr, int nb_jobs)
int linesize[AV_NUM_DATA_POINTERS]
For video, size in bytes of each picture line.
planar YUV 4:4:4 32bpp, (1 Cr & Cb sample per 1x1 Y & A samples)
Descriptor that unambiguously describes how the bits of a pixel are stored in the up to 4 data planes...
static const int16_t alpha[]
#define AV_PIX_FMT_YUV420P10
Describe the class of an AVClass context structure.
static const int factor[16]
const char * name
Filter name.
#define AV_PIX_FMT_YUV420P9
AVFilterLink ** outputs
array of pointers to output links
#define AV_PIX_FMT_YUV422P14
static uint16_t do_chromakey_pixel16(ChromakeyContext *ctx, uint16_t u[9], uint16_t v[9])
#define flags(name, subs,...)
AVFilterInternal * internal
An opaque struct for libavfilter internal use.
#define AV_PIX_FMT_YUV422P10
#define AV_PIX_FMT_YUV444P12
uint8_t * data[AV_NUM_DATA_POINTERS]
pointer to the picture/channel planes.
static av_cold int query_formats(AVFilterContext *avctx)
static const AVFilterPad chromakey_inputs[]
static av_always_inline void get_pixel_uv(AVFrame *frame, int hsub_log2, int vsub_log2, int x, int y, uint8_t *u, uint8_t *v)
planar YUV 4:2:0, 12bpp, (1 Cr & Cb sample per 2x2 Y samples)
#define AV_PIX_FMT_YUVA444P9
avfilter_execute_func * execute
static av_cold int config_input(AVFilterLink *inlink)
static int do_chromakey_slice(AVFilterContext *avctx, void *arg, int jobnr, int nb_jobs)
static int slice_end(AVCodecContext *avctx, AVFrame *pict)
Handle slice ends.
static av_always_inline int diff(const uint32_t a, const uint32_t b)
AVFilterContext * dst
dest filter
static av_cold int config_output(AVFilterLink *outlink)
static int do_chromahold_slice(AVFilterContext *avctx, void *arg, int jobnr, int nb_jobs)
int depth
Number of bits in the component.
AVPixelFormat
Pixel format.
const AVFilter * filter
the AVFilter of which this is an instance
#define AV_PIX_FMT_YUV422P16
#define AV_PIX_FMT_YUVA422P12