mirror of
https://git.ffmpeg.org/ffmpeg.git
synced 2024-12-27 01:42:20 +00:00
20be5e0a0e
* commit '7950e519bb094897f957b9a9531cc60ba46cbc91': Disable deprecation warnings for cases where a replacement is available Conflicts: libavcodec/avpacket.c libavcodec/pthread.c libavcodec/utils.c libavdevice/v4l2.c libavfilter/avfiltergraph.c libavfilter/buffersrc.c Merged-by: Michael Niedermayer <michaelni@gmx.at>
552 lines
18 KiB
C
552 lines
18 KiB
C
/*
|
|
* Copyright (c) 2008 Vitor Sessak
|
|
*
|
|
* This file is part of FFmpeg.
|
|
*
|
|
* FFmpeg is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with FFmpeg; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
/**
|
|
* @file
|
|
* memory buffer source filter
|
|
*/
|
|
|
|
#include <float.h>
|
|
|
|
#include "libavutil/channel_layout.h"
|
|
#include "libavutil/common.h"
|
|
#include "libavutil/fifo.h"
|
|
#include "libavutil/frame.h"
|
|
#include "libavutil/imgutils.h"
|
|
#include "libavutil/internal.h"
|
|
#include "libavutil/opt.h"
|
|
#include "libavutil/samplefmt.h"
|
|
#include "audio.h"
|
|
#include "avfilter.h"
|
|
#include "buffersrc.h"
|
|
#include "formats.h"
|
|
#include "internal.h"
|
|
#include "video.h"
|
|
#include "avcodec.h"
|
|
|
|
typedef struct {
|
|
const AVClass *class;
|
|
AVFifoBuffer *fifo;
|
|
AVRational time_base; ///< time_base to set in the output link
|
|
AVRational frame_rate; ///< frame_rate to set in the output link
|
|
unsigned nb_failed_requests;
|
|
unsigned warning_limit;
|
|
|
|
/* video only */
|
|
int w, h;
|
|
enum AVPixelFormat pix_fmt;
|
|
AVRational pixel_aspect;
|
|
char *sws_param;
|
|
|
|
/* audio only */
|
|
int sample_rate;
|
|
enum AVSampleFormat sample_fmt;
|
|
char *sample_fmt_str;
|
|
int channels;
|
|
uint64_t channel_layout;
|
|
char *channel_layout_str;
|
|
|
|
int eof;
|
|
} BufferSourceContext;
|
|
|
|
#define CHECK_VIDEO_PARAM_CHANGE(s, c, width, height, format)\
|
|
if (c->w != width || c->h != height || c->pix_fmt != format) {\
|
|
av_log(s, AV_LOG_INFO, "Changing frame properties on the fly is not supported by all filters.\n");\
|
|
}
|
|
|
|
#define CHECK_AUDIO_PARAM_CHANGE(s, c, srate, ch_layout, ch_count, format)\
|
|
if (c->sample_fmt != format || c->sample_rate != srate ||\
|
|
c->channel_layout != ch_layout || c->channels != ch_count) {\
|
|
av_log(s, AV_LOG_ERROR, "Changing frame properties on the fly is not supported.\n");\
|
|
return AVERROR(EINVAL);\
|
|
}
|
|
|
|
int attribute_align_arg av_buffersrc_write_frame(AVFilterContext *ctx, const AVFrame *frame)
|
|
{
|
|
return av_buffersrc_add_frame_flags(ctx, (AVFrame *)frame,
|
|
AV_BUFFERSRC_FLAG_KEEP_REF);
|
|
}
|
|
|
|
int attribute_align_arg av_buffersrc_add_frame(AVFilterContext *ctx, AVFrame *frame)
|
|
{
|
|
return av_buffersrc_add_frame_flags(ctx, frame, 0);
|
|
}
|
|
|
|
static int av_buffersrc_add_frame_internal(AVFilterContext *ctx,
|
|
AVFrame *frame, int flags);
|
|
|
|
int attribute_align_arg av_buffersrc_add_frame_flags(AVFilterContext *ctx, AVFrame *frame, int flags)
|
|
{
|
|
AVFrame *copy = NULL;
|
|
int ret = 0;
|
|
|
|
if (frame && frame->channel_layout &&
|
|
av_get_channel_layout_nb_channels(frame->channel_layout) != av_frame_get_channels(frame)) {
|
|
av_log(0, AV_LOG_ERROR, "Layout indicates a different number of channels than actually present\n");
|
|
return AVERROR(EINVAL);
|
|
}
|
|
|
|
if (!(flags & AV_BUFFERSRC_FLAG_KEEP_REF) || !frame)
|
|
return av_buffersrc_add_frame_internal(ctx, frame, flags);
|
|
|
|
if (!(copy = av_frame_alloc()))
|
|
return AVERROR(ENOMEM);
|
|
ret = av_frame_ref(copy, frame);
|
|
if (ret >= 0)
|
|
ret = av_buffersrc_add_frame_internal(ctx, copy, flags);
|
|
|
|
av_frame_free(©);
|
|
return ret;
|
|
}
|
|
|
|
static int av_buffersrc_add_frame_internal(AVFilterContext *ctx,
|
|
AVFrame *frame, int flags)
|
|
{
|
|
BufferSourceContext *s = ctx->priv;
|
|
AVFrame *copy;
|
|
int ret;
|
|
|
|
s->nb_failed_requests = 0;
|
|
|
|
if (!frame) {
|
|
s->eof = 1;
|
|
return 0;
|
|
} else if (s->eof)
|
|
return AVERROR(EINVAL);
|
|
|
|
if (!(flags & AV_BUFFERSRC_FLAG_NO_CHECK_FORMAT)) {
|
|
|
|
switch (ctx->outputs[0]->type) {
|
|
case AVMEDIA_TYPE_VIDEO:
|
|
CHECK_VIDEO_PARAM_CHANGE(ctx, s, frame->width, frame->height,
|
|
frame->format);
|
|
break;
|
|
case AVMEDIA_TYPE_AUDIO:
|
|
/* For layouts unknown on input but known on link after negotiation. */
|
|
if (!frame->channel_layout)
|
|
frame->channel_layout = s->channel_layout;
|
|
CHECK_AUDIO_PARAM_CHANGE(ctx, s, frame->sample_rate, frame->channel_layout,
|
|
av_frame_get_channels(frame), frame->format);
|
|
break;
|
|
default:
|
|
return AVERROR(EINVAL);
|
|
}
|
|
|
|
}
|
|
|
|
if (!av_fifo_space(s->fifo) &&
|
|
(ret = av_fifo_realloc2(s->fifo, av_fifo_size(s->fifo) +
|
|
sizeof(copy))) < 0)
|
|
return ret;
|
|
|
|
if (!(copy = av_frame_alloc()))
|
|
return AVERROR(ENOMEM);
|
|
av_frame_move_ref(copy, frame);
|
|
|
|
if ((ret = av_fifo_generic_write(s->fifo, ©, sizeof(copy), NULL)) < 0) {
|
|
av_frame_move_ref(frame, copy);
|
|
av_frame_free(©);
|
|
return ret;
|
|
}
|
|
|
|
if ((flags & AV_BUFFERSRC_FLAG_PUSH))
|
|
if ((ret = ctx->output_pads[0].request_frame(ctx->outputs[0])) < 0)
|
|
return ret;
|
|
|
|
return 0;
|
|
}
|
|
|
|
#if FF_API_AVFILTERBUFFER
|
|
FF_DISABLE_DEPRECATION_WARNINGS
|
|
static void compat_free_buffer(void *opaque, uint8_t *data)
|
|
{
|
|
AVFilterBufferRef *buf = opaque;
|
|
AV_NOWARN_DEPRECATED(
|
|
avfilter_unref_buffer(buf);
|
|
)
|
|
}
|
|
|
|
static void compat_unref_buffer(void *opaque, uint8_t *data)
|
|
{
|
|
AVBufferRef *buf = opaque;
|
|
AV_NOWARN_DEPRECATED(
|
|
av_buffer_unref(&buf);
|
|
)
|
|
}
|
|
|
|
int av_buffersrc_add_ref(AVFilterContext *ctx, AVFilterBufferRef *buf,
|
|
int flags)
|
|
{
|
|
BufferSourceContext *s = ctx->priv;
|
|
AVFrame *frame = NULL;
|
|
AVBufferRef *dummy_buf = NULL;
|
|
int ret = 0, planes, i;
|
|
|
|
if (!buf) {
|
|
s->eof = 1;
|
|
return 0;
|
|
} else if (s->eof)
|
|
return AVERROR(EINVAL);
|
|
|
|
frame = av_frame_alloc();
|
|
if (!frame)
|
|
return AVERROR(ENOMEM);
|
|
|
|
dummy_buf = av_buffer_create(NULL, 0, compat_free_buffer, buf,
|
|
(buf->perms & AV_PERM_WRITE) ? 0 : AV_BUFFER_FLAG_READONLY);
|
|
if (!dummy_buf) {
|
|
ret = AVERROR(ENOMEM);
|
|
goto fail;
|
|
}
|
|
|
|
AV_NOWARN_DEPRECATED(
|
|
if ((ret = avfilter_copy_buf_props(frame, buf)) < 0)
|
|
goto fail;
|
|
)
|
|
|
|
#define WRAP_PLANE(ref_out, data, data_size) \
|
|
do { \
|
|
AVBufferRef *dummy_ref = av_buffer_ref(dummy_buf); \
|
|
if (!dummy_ref) { \
|
|
ret = AVERROR(ENOMEM); \
|
|
goto fail; \
|
|
} \
|
|
ref_out = av_buffer_create(data, data_size, compat_unref_buffer, \
|
|
dummy_ref, (buf->perms & AV_PERM_WRITE) ? 0 : AV_BUFFER_FLAG_READONLY); \
|
|
if (!ref_out) { \
|
|
av_frame_unref(frame); \
|
|
ret = AVERROR(ENOMEM); \
|
|
goto fail; \
|
|
} \
|
|
} while (0)
|
|
|
|
if (ctx->outputs[0]->type == AVMEDIA_TYPE_VIDEO) {
|
|
const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(frame->format);
|
|
|
|
planes = av_pix_fmt_count_planes(frame->format);
|
|
if (!desc || planes <= 0) {
|
|
ret = AVERROR(EINVAL);
|
|
goto fail;
|
|
}
|
|
|
|
for (i = 0; i < planes; i++) {
|
|
int v_shift = (i == 1 || i == 2) ? desc->log2_chroma_h : 0;
|
|
int plane_size = (frame->height >> v_shift) * frame->linesize[i];
|
|
|
|
WRAP_PLANE(frame->buf[i], frame->data[i], plane_size);
|
|
}
|
|
} else {
|
|
int planar = av_sample_fmt_is_planar(frame->format);
|
|
int channels = av_get_channel_layout_nb_channels(frame->channel_layout);
|
|
|
|
planes = planar ? channels : 1;
|
|
|
|
if (planes > FF_ARRAY_ELEMS(frame->buf)) {
|
|
frame->nb_extended_buf = planes - FF_ARRAY_ELEMS(frame->buf);
|
|
frame->extended_buf = av_mallocz(sizeof(*frame->extended_buf) *
|
|
frame->nb_extended_buf);
|
|
if (!frame->extended_buf) {
|
|
ret = AVERROR(ENOMEM);
|
|
goto fail;
|
|
}
|
|
}
|
|
|
|
for (i = 0; i < FFMIN(planes, FF_ARRAY_ELEMS(frame->buf)); i++)
|
|
WRAP_PLANE(frame->buf[i], frame->extended_data[i], frame->linesize[0]);
|
|
|
|
for (i = 0; i < planes - FF_ARRAY_ELEMS(frame->buf); i++)
|
|
WRAP_PLANE(frame->extended_buf[i],
|
|
frame->extended_data[i + FF_ARRAY_ELEMS(frame->buf)],
|
|
frame->linesize[0]);
|
|
}
|
|
|
|
ret = av_buffersrc_add_frame_flags(ctx, frame, flags);
|
|
|
|
fail:
|
|
av_buffer_unref(&dummy_buf);
|
|
av_frame_free(&frame);
|
|
|
|
return ret;
|
|
}
|
|
FF_ENABLE_DEPRECATION_WARNINGS
|
|
|
|
int av_buffersrc_buffer(AVFilterContext *ctx, AVFilterBufferRef *buf)
|
|
{
|
|
return av_buffersrc_add_ref(ctx, buf, 0);
|
|
}
|
|
#endif
|
|
|
|
static av_cold int init_video(AVFilterContext *ctx)
|
|
{
|
|
BufferSourceContext *c = ctx->priv;
|
|
|
|
if (c->pix_fmt == AV_PIX_FMT_NONE || !c->w || !c->h || av_q2d(c->time_base) <= 0) {
|
|
av_log(ctx, AV_LOG_ERROR, "Invalid parameters provided.\n");
|
|
return AVERROR(EINVAL);
|
|
}
|
|
|
|
if (!(c->fifo = av_fifo_alloc(sizeof(AVFrame*))))
|
|
return AVERROR(ENOMEM);
|
|
|
|
av_log(ctx, AV_LOG_VERBOSE, "w:%d h:%d pixfmt:%s tb:%d/%d fr:%d/%d sar:%d/%d sws_param:%s\n",
|
|
c->w, c->h, av_get_pix_fmt_name(c->pix_fmt),
|
|
c->time_base.num, c->time_base.den, c->frame_rate.num, c->frame_rate.den,
|
|
c->pixel_aspect.num, c->pixel_aspect.den, (char *)av_x_if_null(c->sws_param, ""));
|
|
c->warning_limit = 100;
|
|
return 0;
|
|
}
|
|
|
|
unsigned av_buffersrc_get_nb_failed_requests(AVFilterContext *buffer_src)
|
|
{
|
|
return ((BufferSourceContext *)buffer_src->priv)->nb_failed_requests;
|
|
}
|
|
|
|
#define OFFSET(x) offsetof(BufferSourceContext, x)
|
|
#define A AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_AUDIO_PARAM
|
|
#define V AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_VIDEO_PARAM
|
|
|
|
static const AVOption buffer_options[] = {
|
|
{ "width", NULL, OFFSET(w), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, V },
|
|
{ "video_size", NULL, OFFSET(w), AV_OPT_TYPE_IMAGE_SIZE, .flags = V },
|
|
{ "height", NULL, OFFSET(h), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, V },
|
|
{ "pix_fmt", NULL, OFFSET(pix_fmt), AV_OPT_TYPE_PIXEL_FMT, .flags = V },
|
|
#if FF_API_OLD_FILTER_OPTS
|
|
/* those 4 are for compatibility with the old option passing system where each filter
|
|
* did its own parsing */
|
|
{ "time_base_num", "deprecated, do not use", OFFSET(time_base.num), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, V },
|
|
{ "time_base_den", "deprecated, do not use", OFFSET(time_base.den), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, V },
|
|
{ "sar_num", "deprecated, do not use", OFFSET(pixel_aspect.num), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, V },
|
|
{ "sar_den", "deprecated, do not use", OFFSET(pixel_aspect.den), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, V },
|
|
#endif
|
|
{ "sar", "sample aspect ratio", OFFSET(pixel_aspect), AV_OPT_TYPE_RATIONAL, { .dbl = 1 }, 0, DBL_MAX, V },
|
|
{ "pixel_aspect", "sample aspect ratio", OFFSET(pixel_aspect), AV_OPT_TYPE_RATIONAL, { .dbl = 1 }, 0, DBL_MAX, V },
|
|
{ "time_base", NULL, OFFSET(time_base), AV_OPT_TYPE_RATIONAL, { .dbl = 0 }, 0, DBL_MAX, V },
|
|
{ "frame_rate", NULL, OFFSET(frame_rate), AV_OPT_TYPE_RATIONAL, { .dbl = 0 }, 0, DBL_MAX, V },
|
|
{ "sws_param", NULL, OFFSET(sws_param), AV_OPT_TYPE_STRING, .flags = V },
|
|
{ NULL },
|
|
};
|
|
|
|
AVFILTER_DEFINE_CLASS(buffer);
|
|
|
|
static const AVOption abuffer_options[] = {
|
|
{ "time_base", NULL, OFFSET(time_base), AV_OPT_TYPE_RATIONAL, { .dbl = 0 }, 0, INT_MAX, A },
|
|
{ "sample_rate", NULL, OFFSET(sample_rate), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, A },
|
|
{ "sample_fmt", NULL, OFFSET(sample_fmt_str), AV_OPT_TYPE_STRING, .flags = A },
|
|
{ "channel_layout", NULL, OFFSET(channel_layout_str), AV_OPT_TYPE_STRING, .flags = A },
|
|
{ "channels", NULL, OFFSET(channels), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, A },
|
|
{ NULL },
|
|
};
|
|
|
|
AVFILTER_DEFINE_CLASS(abuffer);
|
|
|
|
static av_cold int init_audio(AVFilterContext *ctx)
|
|
{
|
|
BufferSourceContext *s = ctx->priv;
|
|
int ret = 0;
|
|
|
|
s->sample_fmt = av_get_sample_fmt(s->sample_fmt_str);
|
|
if (s->sample_fmt == AV_SAMPLE_FMT_NONE) {
|
|
av_log(ctx, AV_LOG_ERROR, "Invalid sample format %s\n",
|
|
s->sample_fmt_str);
|
|
return AVERROR(EINVAL);
|
|
}
|
|
|
|
if (s->channel_layout_str) {
|
|
int n;
|
|
/* TODO reindent */
|
|
s->channel_layout = av_get_channel_layout(s->channel_layout_str);
|
|
if (!s->channel_layout) {
|
|
av_log(ctx, AV_LOG_ERROR, "Invalid channel layout %s.\n",
|
|
s->channel_layout_str);
|
|
return AVERROR(EINVAL);
|
|
}
|
|
n = av_get_channel_layout_nb_channels(s->channel_layout);
|
|
if (s->channels) {
|
|
if (n != s->channels) {
|
|
av_log(ctx, AV_LOG_ERROR,
|
|
"Mismatching channel count %d and layout '%s' "
|
|
"(%d channels)\n",
|
|
s->channels, s->channel_layout_str, n);
|
|
return AVERROR(EINVAL);
|
|
}
|
|
}
|
|
s->channels = n;
|
|
} else if (!s->channels) {
|
|
av_log(ctx, AV_LOG_ERROR, "Neither number of channels nor "
|
|
"channel layout specified\n");
|
|
return AVERROR(EINVAL);
|
|
}
|
|
|
|
if (!(s->fifo = av_fifo_alloc(sizeof(AVFrame*))))
|
|
return AVERROR(ENOMEM);
|
|
|
|
if (!s->time_base.num)
|
|
s->time_base = (AVRational){1, s->sample_rate};
|
|
|
|
av_log(ctx, AV_LOG_VERBOSE,
|
|
"tb:%d/%d samplefmt:%s samplerate:%d chlayout:%s\n",
|
|
s->time_base.num, s->time_base.den, s->sample_fmt_str,
|
|
s->sample_rate, s->channel_layout_str);
|
|
s->warning_limit = 100;
|
|
|
|
return ret;
|
|
}
|
|
|
|
static av_cold void uninit(AVFilterContext *ctx)
|
|
{
|
|
BufferSourceContext *s = ctx->priv;
|
|
while (s->fifo && av_fifo_size(s->fifo)) {
|
|
AVFrame *frame;
|
|
av_fifo_generic_read(s->fifo, &frame, sizeof(frame), NULL);
|
|
av_frame_free(&frame);
|
|
}
|
|
av_fifo_free(s->fifo);
|
|
s->fifo = NULL;
|
|
}
|
|
|
|
static int query_formats(AVFilterContext *ctx)
|
|
{
|
|
BufferSourceContext *c = ctx->priv;
|
|
AVFilterChannelLayouts *channel_layouts = NULL;
|
|
AVFilterFormats *formats = NULL;
|
|
AVFilterFormats *samplerates = NULL;
|
|
|
|
switch (ctx->outputs[0]->type) {
|
|
case AVMEDIA_TYPE_VIDEO:
|
|
ff_add_format(&formats, c->pix_fmt);
|
|
ff_set_common_formats(ctx, formats);
|
|
break;
|
|
case AVMEDIA_TYPE_AUDIO:
|
|
ff_add_format(&formats, c->sample_fmt);
|
|
ff_set_common_formats(ctx, formats);
|
|
|
|
ff_add_format(&samplerates, c->sample_rate);
|
|
ff_set_common_samplerates(ctx, samplerates);
|
|
|
|
ff_add_channel_layout(&channel_layouts,
|
|
c->channel_layout ? c->channel_layout :
|
|
FF_COUNT2LAYOUT(c->channels));
|
|
ff_set_common_channel_layouts(ctx, channel_layouts);
|
|
break;
|
|
default:
|
|
return AVERROR(EINVAL);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int config_props(AVFilterLink *link)
|
|
{
|
|
BufferSourceContext *c = link->src->priv;
|
|
|
|
switch (link->type) {
|
|
case AVMEDIA_TYPE_VIDEO:
|
|
link->w = c->w;
|
|
link->h = c->h;
|
|
link->sample_aspect_ratio = c->pixel_aspect;
|
|
break;
|
|
case AVMEDIA_TYPE_AUDIO:
|
|
if (!c->channel_layout)
|
|
c->channel_layout = link->channel_layout;
|
|
break;
|
|
default:
|
|
return AVERROR(EINVAL);
|
|
}
|
|
|
|
link->time_base = c->time_base;
|
|
link->frame_rate = c->frame_rate;
|
|
return 0;
|
|
}
|
|
|
|
static int request_frame(AVFilterLink *link)
|
|
{
|
|
BufferSourceContext *c = link->src->priv;
|
|
AVFrame *frame;
|
|
|
|
if (!av_fifo_size(c->fifo)) {
|
|
if (c->eof)
|
|
return AVERROR_EOF;
|
|
c->nb_failed_requests++;
|
|
return AVERROR(EAGAIN);
|
|
}
|
|
av_fifo_generic_read(c->fifo, &frame, sizeof(frame), NULL);
|
|
|
|
return ff_filter_frame(link, frame);
|
|
}
|
|
|
|
static int poll_frame(AVFilterLink *link)
|
|
{
|
|
BufferSourceContext *c = link->src->priv;
|
|
int size = av_fifo_size(c->fifo);
|
|
if (!size && c->eof)
|
|
return AVERROR_EOF;
|
|
return size/sizeof(AVFrame*);
|
|
}
|
|
|
|
static const AVFilterPad avfilter_vsrc_buffer_outputs[] = {
|
|
{
|
|
.name = "default",
|
|
.type = AVMEDIA_TYPE_VIDEO,
|
|
.request_frame = request_frame,
|
|
.poll_frame = poll_frame,
|
|
.config_props = config_props,
|
|
},
|
|
{ NULL }
|
|
};
|
|
|
|
AVFilter avfilter_vsrc_buffer = {
|
|
.name = "buffer",
|
|
.description = NULL_IF_CONFIG_SMALL("Buffer video frames, and make them accessible to the filterchain."),
|
|
.priv_size = sizeof(BufferSourceContext),
|
|
.query_formats = query_formats,
|
|
|
|
.init = init_video,
|
|
.uninit = uninit,
|
|
|
|
.inputs = NULL,
|
|
.outputs = avfilter_vsrc_buffer_outputs,
|
|
.priv_class = &buffer_class,
|
|
};
|
|
|
|
static const AVFilterPad avfilter_asrc_abuffer_outputs[] = {
|
|
{
|
|
.name = "default",
|
|
.type = AVMEDIA_TYPE_AUDIO,
|
|
.request_frame = request_frame,
|
|
.poll_frame = poll_frame,
|
|
.config_props = config_props,
|
|
},
|
|
{ NULL }
|
|
};
|
|
|
|
AVFilter avfilter_asrc_abuffer = {
|
|
.name = "abuffer",
|
|
.description = NULL_IF_CONFIG_SMALL("Buffer audio frames, and make them accessible to the filterchain."),
|
|
.priv_size = sizeof(BufferSourceContext),
|
|
.query_formats = query_formats,
|
|
|
|
.init = init_audio,
|
|
.uninit = uninit,
|
|
|
|
.inputs = NULL,
|
|
.outputs = avfilter_asrc_abuffer_outputs,
|
|
.priv_class = &abuffer_class,
|
|
};
|