mirror of https://git.ffmpeg.org/ffmpeg.git
956 lines
26 KiB
C
956 lines
26 KiB
C
/*
|
|
* This file is part of FFmpeg.
|
|
*
|
|
* FFmpeg is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with FFmpeg; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
#ifndef FFTOOLS_FFMPEG_H
|
|
#define FFTOOLS_FFMPEG_H
|
|
|
|
#include "config.h"
|
|
|
|
#include <stdatomic.h>
|
|
#include <stdint.h>
|
|
#include <stdio.h>
|
|
#include <signal.h>
|
|
|
|
#include "cmdutils.h"
|
|
#include "sync_queue.h"
|
|
|
|
#include "libavformat/avformat.h"
|
|
#include "libavformat/avio.h"
|
|
|
|
#include "libavcodec/avcodec.h"
|
|
#include "libavcodec/bsf.h"
|
|
|
|
#include "libavfilter/avfilter.h"
|
|
|
|
#include "libavutil/avutil.h"
|
|
#include "libavutil/dict.h"
|
|
#include "libavutil/eval.h"
|
|
#include "libavutil/fifo.h"
|
|
#include "libavutil/hwcontext.h"
|
|
#include "libavutil/pixfmt.h"
|
|
#include "libavutil/rational.h"
|
|
#include "libavutil/thread.h"
|
|
#include "libavutil/threadmessage.h"
|
|
|
|
#include "libswresample/swresample.h"
|
|
|
|
// deprecated features
|
|
#define FFMPEG_OPT_PSNR 1
|
|
#define FFMPEG_OPT_MAP_CHANNEL 1
|
|
#define FFMPEG_OPT_MAP_SYNC 1
|
|
#define FFMPEG_ROTATION_METADATA 1
|
|
#define FFMPEG_OPT_QPHIST 1
|
|
#define FFMPEG_OPT_ADRIFT_THRESHOLD 1
|
|
#define FFMPEG_OPT_ENC_TIME_BASE_NUM 1
|
|
#define FFMPEG_OPT_TOP 1
|
|
#define FFMPEG_OPT_FORCE_KF_SOURCE_NO_DROP 1
|
|
|
|
enum VideoSyncMethod {
|
|
VSYNC_AUTO = -1,
|
|
VSYNC_PASSTHROUGH,
|
|
VSYNC_CFR,
|
|
VSYNC_VFR,
|
|
VSYNC_VSCFR,
|
|
VSYNC_DROP,
|
|
};
|
|
|
|
enum EncTimeBase {
|
|
ENC_TIME_BASE_DEMUX = -1,
|
|
ENC_TIME_BASE_FILTER = -2,
|
|
};
|
|
|
|
enum HWAccelID {
|
|
HWACCEL_NONE = 0,
|
|
HWACCEL_AUTO,
|
|
HWACCEL_GENERIC,
|
|
};
|
|
|
|
typedef struct HWDevice {
|
|
const char *name;
|
|
enum AVHWDeviceType type;
|
|
AVBufferRef *device_ref;
|
|
} HWDevice;
|
|
|
|
/* select an input stream for an output stream */
|
|
typedef struct StreamMap {
|
|
int disabled; /* 1 is this mapping is disabled by a negative map */
|
|
int file_index;
|
|
int stream_index;
|
|
char *linklabel; /* name of an output link, for mapping lavfi outputs */
|
|
} StreamMap;
|
|
|
|
#if FFMPEG_OPT_MAP_CHANNEL
|
|
typedef struct {
|
|
int file_idx, stream_idx, channel_idx; // input
|
|
int ofile_idx, ostream_idx; // output
|
|
} AudioChannelMap;
|
|
#endif
|
|
|
|
typedef struct DemuxPktData {
|
|
// estimated dts in AV_TIME_BASE_Q,
|
|
// to be used when real dts is missing
|
|
int64_t dts_est;
|
|
} DemuxPktData;
|
|
|
|
typedef struct OptionsContext {
|
|
OptionGroup *g;
|
|
|
|
/* input/output options */
|
|
int64_t start_time;
|
|
int64_t start_time_eof;
|
|
int seek_timestamp;
|
|
const char *format;
|
|
|
|
SpecifierOpt *codec_names;
|
|
int nb_codec_names;
|
|
SpecifierOpt *audio_ch_layouts;
|
|
int nb_audio_ch_layouts;
|
|
SpecifierOpt *audio_channels;
|
|
int nb_audio_channels;
|
|
SpecifierOpt *audio_sample_rate;
|
|
int nb_audio_sample_rate;
|
|
SpecifierOpt *frame_rates;
|
|
int nb_frame_rates;
|
|
SpecifierOpt *max_frame_rates;
|
|
int nb_max_frame_rates;
|
|
SpecifierOpt *frame_sizes;
|
|
int nb_frame_sizes;
|
|
SpecifierOpt *frame_pix_fmts;
|
|
int nb_frame_pix_fmts;
|
|
|
|
/* input options */
|
|
int64_t input_ts_offset;
|
|
int loop;
|
|
int rate_emu;
|
|
float readrate;
|
|
double readrate_initial_burst;
|
|
int accurate_seek;
|
|
int thread_queue_size;
|
|
int input_sync_ref;
|
|
int find_stream_info;
|
|
|
|
SpecifierOpt *ts_scale;
|
|
int nb_ts_scale;
|
|
SpecifierOpt *dump_attachment;
|
|
int nb_dump_attachment;
|
|
SpecifierOpt *hwaccels;
|
|
int nb_hwaccels;
|
|
SpecifierOpt *hwaccel_devices;
|
|
int nb_hwaccel_devices;
|
|
SpecifierOpt *hwaccel_output_formats;
|
|
int nb_hwaccel_output_formats;
|
|
SpecifierOpt *autorotate;
|
|
int nb_autorotate;
|
|
|
|
/* output options */
|
|
StreamMap *stream_maps;
|
|
int nb_stream_maps;
|
|
#if FFMPEG_OPT_MAP_CHANNEL
|
|
AudioChannelMap *audio_channel_maps; /* one info entry per -map_channel */
|
|
int nb_audio_channel_maps; /* number of (valid) -map_channel settings */
|
|
#endif
|
|
const char **attachments;
|
|
int nb_attachments;
|
|
|
|
int chapters_input_file;
|
|
|
|
int64_t recording_time;
|
|
int64_t stop_time;
|
|
int64_t limit_filesize;
|
|
float mux_preload;
|
|
float mux_max_delay;
|
|
float shortest_buf_duration;
|
|
int shortest;
|
|
int bitexact;
|
|
|
|
int video_disable;
|
|
int audio_disable;
|
|
int subtitle_disable;
|
|
int data_disable;
|
|
|
|
// keys are stream indices
|
|
AVDictionary *streamid;
|
|
|
|
SpecifierOpt *metadata;
|
|
int nb_metadata;
|
|
SpecifierOpt *max_frames;
|
|
int nb_max_frames;
|
|
SpecifierOpt *bitstream_filters;
|
|
int nb_bitstream_filters;
|
|
SpecifierOpt *codec_tags;
|
|
int nb_codec_tags;
|
|
SpecifierOpt *sample_fmts;
|
|
int nb_sample_fmts;
|
|
SpecifierOpt *qscale;
|
|
int nb_qscale;
|
|
SpecifierOpt *forced_key_frames;
|
|
int nb_forced_key_frames;
|
|
SpecifierOpt *fps_mode;
|
|
int nb_fps_mode;
|
|
SpecifierOpt *force_fps;
|
|
int nb_force_fps;
|
|
SpecifierOpt *frame_aspect_ratios;
|
|
int nb_frame_aspect_ratios;
|
|
SpecifierOpt *display_rotations;
|
|
int nb_display_rotations;
|
|
SpecifierOpt *display_hflips;
|
|
int nb_display_hflips;
|
|
SpecifierOpt *display_vflips;
|
|
int nb_display_vflips;
|
|
SpecifierOpt *rc_overrides;
|
|
int nb_rc_overrides;
|
|
SpecifierOpt *intra_matrices;
|
|
int nb_intra_matrices;
|
|
SpecifierOpt *inter_matrices;
|
|
int nb_inter_matrices;
|
|
SpecifierOpt *chroma_intra_matrices;
|
|
int nb_chroma_intra_matrices;
|
|
#if FFMPEG_OPT_TOP
|
|
SpecifierOpt *top_field_first;
|
|
int nb_top_field_first;
|
|
#endif
|
|
SpecifierOpt *metadata_map;
|
|
int nb_metadata_map;
|
|
SpecifierOpt *presets;
|
|
int nb_presets;
|
|
SpecifierOpt *copy_initial_nonkeyframes;
|
|
int nb_copy_initial_nonkeyframes;
|
|
SpecifierOpt *copy_prior_start;
|
|
int nb_copy_prior_start;
|
|
SpecifierOpt *filters;
|
|
int nb_filters;
|
|
SpecifierOpt *filter_scripts;
|
|
int nb_filter_scripts;
|
|
SpecifierOpt *reinit_filters;
|
|
int nb_reinit_filters;
|
|
SpecifierOpt *fix_sub_duration;
|
|
int nb_fix_sub_duration;
|
|
SpecifierOpt *fix_sub_duration_heartbeat;
|
|
int nb_fix_sub_duration_heartbeat;
|
|
SpecifierOpt *canvas_sizes;
|
|
int nb_canvas_sizes;
|
|
SpecifierOpt *pass;
|
|
int nb_pass;
|
|
SpecifierOpt *passlogfiles;
|
|
int nb_passlogfiles;
|
|
SpecifierOpt *max_muxing_queue_size;
|
|
int nb_max_muxing_queue_size;
|
|
SpecifierOpt *muxing_queue_data_threshold;
|
|
int nb_muxing_queue_data_threshold;
|
|
SpecifierOpt *guess_layout_max;
|
|
int nb_guess_layout_max;
|
|
SpecifierOpt *apad;
|
|
int nb_apad;
|
|
SpecifierOpt *discard;
|
|
int nb_discard;
|
|
SpecifierOpt *disposition;
|
|
int nb_disposition;
|
|
SpecifierOpt *program;
|
|
int nb_program;
|
|
SpecifierOpt *time_bases;
|
|
int nb_time_bases;
|
|
SpecifierOpt *enc_time_bases;
|
|
int nb_enc_time_bases;
|
|
SpecifierOpt *autoscale;
|
|
int nb_autoscale;
|
|
SpecifierOpt *bits_per_raw_sample;
|
|
int nb_bits_per_raw_sample;
|
|
SpecifierOpt *enc_stats_pre;
|
|
int nb_enc_stats_pre;
|
|
SpecifierOpt *enc_stats_post;
|
|
int nb_enc_stats_post;
|
|
SpecifierOpt *mux_stats;
|
|
int nb_mux_stats;
|
|
SpecifierOpt *enc_stats_pre_fmt;
|
|
int nb_enc_stats_pre_fmt;
|
|
SpecifierOpt *enc_stats_post_fmt;
|
|
int nb_enc_stats_post_fmt;
|
|
SpecifierOpt *mux_stats_fmt;
|
|
int nb_mux_stats_fmt;
|
|
} OptionsContext;
|
|
|
|
typedef struct InputFilter {
|
|
struct FilterGraph *graph;
|
|
uint8_t *name;
|
|
} InputFilter;
|
|
|
|
typedef struct OutputFilter {
|
|
struct OutputStream *ost;
|
|
struct FilterGraph *graph;
|
|
uint8_t *name;
|
|
|
|
/* for filters that are not yet bound to an output stream,
|
|
* this stores the output linklabel, if any */
|
|
uint8_t *linklabel;
|
|
|
|
enum AVMediaType type;
|
|
|
|
/* pts of the last frame received from this filter, in AV_TIME_BASE_Q */
|
|
int64_t last_pts;
|
|
|
|
uint64_t nb_frames_dup;
|
|
uint64_t nb_frames_drop;
|
|
} OutputFilter;
|
|
|
|
typedef struct FilterGraph {
|
|
const AVClass *class;
|
|
int index;
|
|
|
|
AVFilterGraph *graph;
|
|
|
|
InputFilter **inputs;
|
|
int nb_inputs;
|
|
OutputFilter **outputs;
|
|
int nb_outputs;
|
|
} FilterGraph;
|
|
|
|
typedef struct Decoder Decoder;
|
|
|
|
typedef struct InputStream {
|
|
const AVClass *class;
|
|
|
|
int file_index;
|
|
int index;
|
|
|
|
AVStream *st;
|
|
int discard; /* true if stream data should be discarded */
|
|
int user_set_discard;
|
|
int decoding_needed; /* non zero if the packets must be decoded in 'raw_fifo', see DECODING_FOR_* */
|
|
#define DECODING_FOR_OST 1
|
|
#define DECODING_FOR_FILTER 2
|
|
|
|
/**
|
|
* Codec parameters - to be used by the decoding/streamcopy code.
|
|
* st->codecpar should not be accessed, because it may be modified
|
|
* concurrently by the demuxing thread.
|
|
*/
|
|
AVCodecParameters *par;
|
|
Decoder *decoder;
|
|
AVCodecContext *dec_ctx;
|
|
const AVCodec *dec;
|
|
const AVCodecDescriptor *codec_desc;
|
|
|
|
AVRational framerate_guessed;
|
|
|
|
int64_t nb_samples; /* number of samples in the last decoded audio frame before looping */
|
|
|
|
AVDictionary *decoder_opts;
|
|
AVRational framerate; /* framerate forced with -r */
|
|
#if FFMPEG_OPT_TOP
|
|
int top_field_first;
|
|
#endif
|
|
|
|
int autorotate;
|
|
|
|
int fix_sub_duration;
|
|
|
|
struct sub2video {
|
|
int w, h;
|
|
} sub2video;
|
|
|
|
/* decoded data from this stream goes into all those filters
|
|
* currently video and audio only */
|
|
InputFilter **filters;
|
|
int nb_filters;
|
|
|
|
/*
|
|
* Output targets that do not go through lavfi, i.e. subtitles or
|
|
* streamcopy. Those two cases are distinguished by the OutputStream
|
|
* having an encoder or not.
|
|
*/
|
|
struct OutputStream **outputs;
|
|
int nb_outputs;
|
|
|
|
int reinit_filters;
|
|
|
|
/* hwaccel options */
|
|
enum HWAccelID hwaccel_id;
|
|
enum AVHWDeviceType hwaccel_device_type;
|
|
char *hwaccel_device;
|
|
enum AVPixelFormat hwaccel_output_format;
|
|
|
|
/* stats */
|
|
// number of frames/samples retrieved from the decoder
|
|
uint64_t frames_decoded;
|
|
uint64_t samples_decoded;
|
|
uint64_t decode_errors;
|
|
} InputStream;
|
|
|
|
typedef struct LastFrameDuration {
|
|
int stream_idx;
|
|
int64_t duration;
|
|
} LastFrameDuration;
|
|
|
|
typedef struct InputFile {
|
|
const AVClass *class;
|
|
|
|
int index;
|
|
|
|
// input format has no timestamps
|
|
int format_nots;
|
|
|
|
AVFormatContext *ctx;
|
|
int eof_reached; /* true if eof reached */
|
|
int eagain; /* true if last read attempt returned EAGAIN */
|
|
int64_t input_ts_offset;
|
|
int input_sync_ref;
|
|
/**
|
|
* Effective format start time based on enabled streams.
|
|
*/
|
|
int64_t start_time_effective;
|
|
int64_t ts_offset;
|
|
int64_t start_time; /* user-specified start time in AV_TIME_BASE or AV_NOPTS_VALUE */
|
|
int64_t recording_time;
|
|
|
|
/* streams that ffmpeg is aware of;
|
|
* there may be extra streams in ctx that are not mapped to an InputStream
|
|
* if new streams appear dynamically during demuxing */
|
|
InputStream **streams;
|
|
int nb_streams;
|
|
|
|
float readrate;
|
|
int accurate_seek;
|
|
|
|
/* when looping the input file, this queue is used by decoders to report
|
|
* the last frame duration back to the demuxer thread */
|
|
AVThreadMessageQueue *audio_duration_queue;
|
|
int audio_duration_queue_size;
|
|
} InputFile;
|
|
|
|
enum forced_keyframes_const {
|
|
FKF_N,
|
|
FKF_N_FORCED,
|
|
FKF_PREV_FORCED_N,
|
|
FKF_PREV_FORCED_T,
|
|
FKF_T,
|
|
FKF_NB
|
|
};
|
|
|
|
#define ABORT_ON_FLAG_EMPTY_OUTPUT (1 << 0)
|
|
#define ABORT_ON_FLAG_EMPTY_OUTPUT_STREAM (1 << 1)
|
|
|
|
enum EncStatsType {
|
|
ENC_STATS_LITERAL = 0,
|
|
ENC_STATS_FILE_IDX,
|
|
ENC_STATS_STREAM_IDX,
|
|
ENC_STATS_FRAME_NUM,
|
|
ENC_STATS_FRAME_NUM_IN,
|
|
ENC_STATS_TIMEBASE,
|
|
ENC_STATS_TIMEBASE_IN,
|
|
ENC_STATS_PTS,
|
|
ENC_STATS_PTS_TIME,
|
|
ENC_STATS_PTS_IN,
|
|
ENC_STATS_PTS_TIME_IN,
|
|
ENC_STATS_DTS,
|
|
ENC_STATS_DTS_TIME,
|
|
ENC_STATS_SAMPLE_NUM,
|
|
ENC_STATS_NB_SAMPLES,
|
|
ENC_STATS_PKT_SIZE,
|
|
ENC_STATS_BITRATE,
|
|
ENC_STATS_AVG_BITRATE,
|
|
};
|
|
|
|
typedef struct EncStatsComponent {
|
|
enum EncStatsType type;
|
|
|
|
uint8_t *str;
|
|
size_t str_len;
|
|
} EncStatsComponent;
|
|
|
|
typedef struct EncStats {
|
|
EncStatsComponent *components;
|
|
int nb_components;
|
|
|
|
AVIOContext *io;
|
|
} EncStats;
|
|
|
|
extern const char *const forced_keyframes_const_names[];
|
|
|
|
typedef enum {
|
|
ENCODER_FINISHED = 1,
|
|
MUXER_FINISHED = 2,
|
|
} OSTFinished ;
|
|
|
|
enum {
|
|
KF_FORCE_SOURCE = 1,
|
|
#if FFMPEG_OPT_FORCE_KF_SOURCE_NO_DROP
|
|
KF_FORCE_SOURCE_NO_DROP = 2,
|
|
#endif
|
|
};
|
|
|
|
typedef struct KeyframeForceCtx {
|
|
int type;
|
|
|
|
int64_t ref_pts;
|
|
|
|
// timestamps of the forced keyframes, in AV_TIME_BASE_Q
|
|
int64_t *pts;
|
|
int nb_pts;
|
|
int index;
|
|
|
|
AVExpr *pexpr;
|
|
double expr_const_values[FKF_NB];
|
|
|
|
int dropped_keyframe;
|
|
} KeyframeForceCtx;
|
|
|
|
typedef struct Encoder Encoder;
|
|
|
|
typedef struct OutputStream {
|
|
const AVClass *class;
|
|
|
|
enum AVMediaType type;
|
|
|
|
int file_index; /* file index */
|
|
int index; /* stream index in the output file */
|
|
|
|
/**
|
|
* Codec parameters for packets submitted to the muxer (i.e. before
|
|
* bitstream filtering, if any).
|
|
*/
|
|
AVCodecParameters *par_in;
|
|
|
|
/* input stream that is the source for this output stream;
|
|
* may be NULL for streams with no well-defined source, e.g.
|
|
* attachments or outputs from complex filtergraphs */
|
|
InputStream *ist;
|
|
|
|
AVStream *st; /* stream in the output file */
|
|
/* dts of the last packet sent to the muxing queue, in AV_TIME_BASE_Q */
|
|
int64_t last_mux_dts;
|
|
|
|
AVRational enc_timebase;
|
|
|
|
Encoder *enc;
|
|
AVCodecContext *enc_ctx;
|
|
|
|
/* video only */
|
|
AVRational frame_rate;
|
|
AVRational max_frame_rate;
|
|
enum VideoSyncMethod vsync_method;
|
|
int is_cfr;
|
|
int force_fps;
|
|
#if FFMPEG_OPT_TOP
|
|
int top_field_first;
|
|
#endif
|
|
#if FFMPEG_ROTATION_METADATA
|
|
int rotate_overridden;
|
|
#endif
|
|
int autoscale;
|
|
int bitexact;
|
|
int bits_per_raw_sample;
|
|
#if FFMPEG_ROTATION_METADATA
|
|
double rotate_override_value;
|
|
#endif
|
|
|
|
AVRational frame_aspect_ratio;
|
|
|
|
KeyframeForceCtx kf;
|
|
|
|
/* audio only */
|
|
#if FFMPEG_OPT_MAP_CHANNEL
|
|
int *audio_channels_map; /* list of the channels id to pick from the source stream */
|
|
int audio_channels_mapped; /* number of channels in audio_channels_map */
|
|
#endif
|
|
|
|
char *logfile_prefix;
|
|
FILE *logfile;
|
|
|
|
OutputFilter *filter;
|
|
|
|
AVDictionary *encoder_opts;
|
|
AVDictionary *sws_dict;
|
|
AVDictionary *swr_opts;
|
|
char *apad;
|
|
OSTFinished finished; /* no more packets should be written for this stream */
|
|
int unavailable; /* true if the steram is unavailable (possibly temporarily) */
|
|
|
|
// init_output_stream() has been called for this stream
|
|
// The encoder and the bitstream filters have been initialized and the stream
|
|
// parameters are set in the AVStream.
|
|
int initialized;
|
|
|
|
const char *attachment_filename;
|
|
|
|
int keep_pix_fmt;
|
|
|
|
/* stats */
|
|
// number of packets send to the muxer
|
|
atomic_uint_least64_t packets_written;
|
|
// number of frames/samples sent to the encoder
|
|
uint64_t frames_encoded;
|
|
uint64_t samples_encoded;
|
|
|
|
/* packet quality factor */
|
|
int quality;
|
|
|
|
int sq_idx_encode;
|
|
int sq_idx_mux;
|
|
|
|
EncStats enc_stats_pre;
|
|
EncStats enc_stats_post;
|
|
|
|
/*
|
|
* bool on whether this stream should be utilized for splitting
|
|
* subtitles utilizing fix_sub_duration at random access points.
|
|
*/
|
|
unsigned int fix_sub_duration_heartbeat;
|
|
} OutputStream;
|
|
|
|
typedef struct OutputFile {
|
|
const AVClass *class;
|
|
|
|
int index;
|
|
|
|
const AVOutputFormat *format;
|
|
const char *url;
|
|
|
|
OutputStream **streams;
|
|
int nb_streams;
|
|
|
|
SyncQueue *sq_encode;
|
|
|
|
int64_t recording_time; ///< desired length of the resulting file in microseconds == AV_TIME_BASE units
|
|
int64_t start_time; ///< start time in microseconds == AV_TIME_BASE units
|
|
|
|
int shortest;
|
|
int bitexact;
|
|
} OutputFile;
|
|
|
|
// optionally attached as opaque_ref to decoded AVFrames
|
|
typedef struct FrameData {
|
|
// properties that come from the decoder
|
|
struct {
|
|
uint64_t frame_num;
|
|
|
|
int64_t pts;
|
|
AVRational tb;
|
|
} dec;
|
|
|
|
AVRational frame_rate_filter;
|
|
|
|
int bits_per_raw_sample;
|
|
} FrameData;
|
|
|
|
extern InputFile **input_files;
|
|
extern int nb_input_files;
|
|
|
|
extern OutputFile **output_files;
|
|
extern int nb_output_files;
|
|
|
|
extern FilterGraph **filtergraphs;
|
|
extern int nb_filtergraphs;
|
|
|
|
extern char *vstats_filename;
|
|
extern char *sdp_filename;
|
|
|
|
extern float dts_delta_threshold;
|
|
extern float dts_error_threshold;
|
|
|
|
extern enum VideoSyncMethod video_sync_method;
|
|
extern float frame_drop_threshold;
|
|
extern int do_benchmark;
|
|
extern int do_benchmark_all;
|
|
extern int do_hex_dump;
|
|
extern int do_pkt_dump;
|
|
extern int copy_ts;
|
|
extern int start_at_zero;
|
|
extern int copy_tb;
|
|
extern int debug_ts;
|
|
extern int exit_on_error;
|
|
extern int abort_on_flags;
|
|
extern int print_stats;
|
|
extern int64_t stats_period;
|
|
extern int stdin_interaction;
|
|
extern AVIOContext *progress_avio;
|
|
extern float max_error_rate;
|
|
|
|
extern char *filter_nbthreads;
|
|
extern int filter_complex_nbthreads;
|
|
extern int vstats_version;
|
|
extern int auto_conversion_filters;
|
|
|
|
extern const AVIOInterruptCB int_cb;
|
|
|
|
extern const OptionDef options[];
|
|
extern HWDevice *filter_hw_device;
|
|
|
|
extern unsigned nb_output_dumped;
|
|
|
|
extern int ignore_unknown_streams;
|
|
extern int copy_unknown_streams;
|
|
|
|
extern int recast_media;
|
|
|
|
extern FILE *vstats_file;
|
|
|
|
#if FFMPEG_OPT_PSNR
|
|
extern int do_psnr;
|
|
#endif
|
|
|
|
void term_init(void);
|
|
void term_exit(void);
|
|
|
|
void show_usage(void);
|
|
|
|
void remove_avoptions(AVDictionary **a, AVDictionary *b);
|
|
int check_avoptions(AVDictionary *m);
|
|
|
|
int assert_file_overwrite(const char *filename);
|
|
char *file_read(const char *filename);
|
|
AVDictionary *strip_specifiers(const AVDictionary *dict);
|
|
int find_codec(void *logctx, const char *name,
|
|
enum AVMediaType type, int encoder, const AVCodec **codec);
|
|
int parse_and_set_vsync(const char *arg, int *vsync_var, int file_idx, int st_idx, int is_global);
|
|
|
|
int check_filter_outputs(void);
|
|
int filtergraph_is_simple(const FilterGraph *fg);
|
|
int init_simple_filtergraph(InputStream *ist, OutputStream *ost,
|
|
char *graph_desc);
|
|
int init_complex_filtergraph(FilterGraph *fg);
|
|
|
|
int copy_av_subtitle(AVSubtitle *dst, const AVSubtitle *src);
|
|
int subtitle_wrap_frame(AVFrame *frame, AVSubtitle *subtitle, int copy);
|
|
|
|
/**
|
|
* Get our axiliary frame data attached to the frame, allocating it
|
|
* if needed.
|
|
*/
|
|
FrameData *frame_data(AVFrame *frame);
|
|
|
|
int ifilter_send_frame(InputFilter *ifilter, AVFrame *frame, int keep_reference);
|
|
int ifilter_send_eof(InputFilter *ifilter, int64_t pts, AVRational tb);
|
|
int ifilter_sub2video(InputFilter *ifilter, const AVFrame *frame);
|
|
void ifilter_sub2video_heartbeat(InputFilter *ifilter, int64_t pts, AVRational tb);
|
|
|
|
/**
|
|
* Set up fallback filtering parameters from a decoder context. They will only
|
|
* be used if no frames are ever sent on this input, otherwise the actual
|
|
* parameters are taken from the frame.
|
|
*/
|
|
int ifilter_parameters_from_dec(InputFilter *ifilter, const AVCodecContext *dec);
|
|
|
|
int ofilter_bind_ost(OutputFilter *ofilter, OutputStream *ost);
|
|
|
|
/**
|
|
* Create a new filtergraph in the global filtergraph list.
|
|
*
|
|
* @param graph_desc Graph description; an av_malloc()ed string, filtergraph
|
|
* takes ownership of it.
|
|
*/
|
|
int fg_create(FilterGraph **pfg, char *graph_desc);
|
|
|
|
void fg_free(FilterGraph **pfg);
|
|
|
|
/**
|
|
* Perform a step of transcoding for the specified filter graph.
|
|
*
|
|
* @param[in] graph filter graph to consider
|
|
* @param[out] best_ist input stream where a frame would allow to continue
|
|
* @return 0 for success, <0 for error
|
|
*/
|
|
int fg_transcode_step(FilterGraph *graph, InputStream **best_ist);
|
|
|
|
void fg_send_command(FilterGraph *fg, double time, const char *target,
|
|
const char *command, const char *arg, int all_filters);
|
|
|
|
/**
|
|
* Get and encode new output from specified filtergraph, without causing
|
|
* activity.
|
|
*
|
|
* @return 0 for success, <0 for severe errors
|
|
*/
|
|
int reap_filters(FilterGraph *fg, int flush);
|
|
|
|
int ffmpeg_parse_options(int argc, char **argv);
|
|
|
|
void enc_stats_write(OutputStream *ost, EncStats *es,
|
|
const AVFrame *frame, const AVPacket *pkt,
|
|
uint64_t frame_num);
|
|
|
|
HWDevice *hw_device_get_by_name(const char *name);
|
|
HWDevice *hw_device_get_by_type(enum AVHWDeviceType type);
|
|
int hw_device_init_from_string(const char *arg, HWDevice **dev);
|
|
int hw_device_init_from_type(enum AVHWDeviceType type,
|
|
const char *device,
|
|
HWDevice **dev_out);
|
|
void hw_device_free_all(void);
|
|
|
|
/**
|
|
* Get a hardware device to be used with this filtergraph.
|
|
* The returned reference is owned by the callee, the caller
|
|
* must ref it explicitly for long-term use.
|
|
*/
|
|
AVBufferRef *hw_device_for_filter(void);
|
|
|
|
int hwaccel_retrieve_data(AVCodecContext *avctx, AVFrame *input);
|
|
|
|
int dec_open(InputStream *ist);
|
|
void dec_free(Decoder **pdec);
|
|
|
|
/**
|
|
* Submit a packet for decoding
|
|
*
|
|
* When pkt==NULL and no_eof=0, there will be no more input. Flush decoders and
|
|
* mark all downstreams as finished.
|
|
*
|
|
* When pkt==NULL and no_eof=1, the stream was reset (e.g. after a seek). Flush
|
|
* decoders and await further input.
|
|
*/
|
|
int dec_packet(InputStream *ist, const AVPacket *pkt, int no_eof);
|
|
|
|
int enc_alloc(Encoder **penc, const AVCodec *codec);
|
|
void enc_free(Encoder **penc);
|
|
|
|
int enc_open(OutputStream *ost, const AVFrame *frame);
|
|
int enc_subtitle(OutputFile *of, OutputStream *ost, const AVSubtitle *sub);
|
|
int enc_frame(OutputStream *ost, AVFrame *frame);
|
|
int enc_flush(void);
|
|
|
|
/*
|
|
* Initialize muxing state for the given stream, should be called
|
|
* after the codec/streamcopy setup has been done.
|
|
*
|
|
* Open the muxer once all the streams have been initialized.
|
|
*/
|
|
int of_stream_init(OutputFile *of, OutputStream *ost);
|
|
int of_write_trailer(OutputFile *of);
|
|
int of_open(const OptionsContext *o, const char *filename);
|
|
void of_free(OutputFile **pof);
|
|
|
|
void of_enc_stats_close(void);
|
|
|
|
int of_output_packet(OutputFile *of, OutputStream *ost, AVPacket *pkt);
|
|
|
|
/**
|
|
* @param dts predicted packet dts in AV_TIME_BASE_Q
|
|
*/
|
|
int of_streamcopy(OutputStream *ost, const AVPacket *pkt, int64_t dts);
|
|
|
|
int64_t of_filesize(OutputFile *of);
|
|
|
|
int ifile_open(const OptionsContext *o, const char *filename);
|
|
void ifile_close(InputFile **f);
|
|
|
|
/**
|
|
* Get next input packet from the demuxer.
|
|
*
|
|
* @param pkt the packet is written here when this function returns 0
|
|
* @return
|
|
* - 0 when a packet has been read successfully
|
|
* - 1 when stream end was reached, but the stream is looped;
|
|
* caller should flush decoders and read from this demuxer again
|
|
* - a negative error code on failure
|
|
*/
|
|
int ifile_get_packet(InputFile *f, AVPacket **pkt);
|
|
|
|
int ist_output_add(InputStream *ist, OutputStream *ost);
|
|
int ist_filter_add(InputStream *ist, InputFilter *ifilter, int is_simple);
|
|
|
|
/**
|
|
* Find an unused input stream of given type.
|
|
*/
|
|
InputStream *ist_find_unused(enum AVMediaType type);
|
|
|
|
/* iterate over all input streams in all input files;
|
|
* pass NULL to start iteration */
|
|
InputStream *ist_iter(InputStream *prev);
|
|
|
|
/* iterate over all output streams in all output files;
|
|
* pass NULL to start iteration */
|
|
OutputStream *ost_iter(OutputStream *prev);
|
|
|
|
void close_output_stream(OutputStream *ost);
|
|
int trigger_fix_sub_duration_heartbeat(OutputStream *ost, const AVPacket *pkt);
|
|
int fix_sub_duration_heartbeat(InputStream *ist, int64_t signal_pts);
|
|
void update_benchmark(const char *fmt, ...);
|
|
|
|
/**
|
|
* Merge two return codes - return one of the error codes if at least one of
|
|
* them was negative, 0 otherwise.
|
|
* Currently just picks the first one, eventually we might want to do something
|
|
* more sophisticated, like sorting them by priority.
|
|
*/
|
|
static inline int err_merge(int err0, int err1)
|
|
{
|
|
return (err0 < 0) ? err0 : FFMIN(err1, 0);
|
|
}
|
|
|
|
#define SPECIFIER_OPT_FMT_str "%s"
|
|
#define SPECIFIER_OPT_FMT_i "%i"
|
|
#define SPECIFIER_OPT_FMT_i64 "%"PRId64
|
|
#define SPECIFIER_OPT_FMT_ui64 "%"PRIu64
|
|
#define SPECIFIER_OPT_FMT_f "%f"
|
|
#define SPECIFIER_OPT_FMT_dbl "%lf"
|
|
|
|
#define WARN_MULTIPLE_OPT_USAGE(name, type, so, st)\
|
|
{\
|
|
char namestr[128] = "";\
|
|
const char *spec = so->specifier && so->specifier[0] ? so->specifier : "";\
|
|
for (int _i = 0; opt_name_##name[_i]; _i++)\
|
|
av_strlcatf(namestr, sizeof(namestr), "-%s%s", opt_name_##name[_i], opt_name_##name[_i+1] ? (opt_name_##name[_i+2] ? ", " : " or ") : "");\
|
|
av_log(NULL, AV_LOG_WARNING, "Multiple %s options specified for stream %d, only the last option '-%s%s%s "SPECIFIER_OPT_FMT_##type"' will be used.\n",\
|
|
namestr, st->index, opt_name_##name[0], spec[0] ? ":" : "", spec, so->u.type);\
|
|
}
|
|
|
|
#define MATCH_PER_STREAM_OPT(name, type, outvar, fmtctx, st)\
|
|
{\
|
|
int _ret, _matches = 0;\
|
|
SpecifierOpt *so;\
|
|
for (int _i = 0; _i < o->nb_ ## name; _i++) {\
|
|
char *spec = o->name[_i].specifier;\
|
|
if ((_ret = check_stream_specifier(fmtctx, st, spec)) > 0) {\
|
|
outvar = o->name[_i].u.type;\
|
|
so = &o->name[_i];\
|
|
_matches++;\
|
|
} else if (_ret < 0)\
|
|
return _ret;\
|
|
}\
|
|
if (_matches > 1)\
|
|
WARN_MULTIPLE_OPT_USAGE(name, type, so, st);\
|
|
}
|
|
|
|
#define MATCH_PER_TYPE_OPT(name, type, outvar, fmtctx, mediatype)\
|
|
{\
|
|
int i;\
|
|
for (i = 0; i < o->nb_ ## name; i++) {\
|
|
char *spec = o->name[i].specifier;\
|
|
if (!strcmp(spec, mediatype))\
|
|
outvar = o->name[i].u.type;\
|
|
}\
|
|
}
|
|
|
|
extern const char * const opt_name_codec_names[];
|
|
extern const char * const opt_name_codec_tags[];
|
|
extern const char * const opt_name_frame_rates[];
|
|
#if FFMPEG_OPT_TOP
|
|
extern const char * const opt_name_top_field_first[];
|
|
#endif
|
|
|
|
static inline void pkt_move(void *dst, void *src)
|
|
{
|
|
av_packet_move_ref(dst, src);
|
|
}
|
|
|
|
static inline void frame_move(void *dst, void *src)
|
|
{
|
|
av_frame_move_ref(dst, src);
|
|
}
|
|
|
|
#endif /* FFTOOLS_FFMPEG_H */
|