2002-10-27 22:00:34 +00:00
|
|
|
|
\input texinfo @c -*- texinfo -*-
|
|
|
|
|
|
2011-04-23 13:19:17 +00:00
|
|
|
|
@settitle ffmpeg Documentation
|
2002-10-27 22:00:34 +00:00
|
|
|
|
@titlepage
|
2011-04-23 13:19:17 +00:00
|
|
|
|
@center @titlefont{ffmpeg Documentation}
|
2002-10-27 22:00:34 +00:00
|
|
|
|
@end titlepage
|
|
|
|
|
|
2011-01-29 12:24:13 +00:00
|
|
|
|
@top
|
|
|
|
|
|
|
|
|
|
@contents
|
|
|
|
|
|
2010-06-08 22:32:29 +00:00
|
|
|
|
@chapter Synopsis
|
|
|
|
|
|
2012-12-17 11:55:28 +00:00
|
|
|
|
ffmpeg [@var{global_options}] @{[@var{input_file_options}] -i @file{input_file}@} ... @{[@var{output_file_options}] @file{output_file}@} ...
|
2010-06-08 22:32:29 +00:00
|
|
|
|
|
2010-06-01 13:09:49 +00:00
|
|
|
|
@chapter Description
|
|
|
|
|
@c man begin DESCRIPTION
|
2002-10-27 22:00:34 +00:00
|
|
|
|
|
2011-04-23 14:40:01 +00:00
|
|
|
|
ffmpeg is a very fast video and audio converter that can also grab from
|
|
|
|
|
a live audio/video source. It can also convert between arbitrary sample
|
|
|
|
|
rates and resize video on the fly with a high quality polyphase filter.
|
2005-12-17 18:14:38 +00:00
|
|
|
|
|
2011-12-11 17:50:39 +00:00
|
|
|
|
ffmpeg reads from an arbitrary number of input "files" (which can be regular
|
|
|
|
|
files, pipes, network streams, grabbing devices, etc.), specified by the
|
|
|
|
|
@code{-i} option, and writes to an arbitrary number of output "files", which are
|
2012-01-25 22:01:34 +00:00
|
|
|
|
specified by a plain output filename. Anything found on the command line which
|
2011-12-11 17:50:39 +00:00
|
|
|
|
cannot be interpreted as an option is considered to be an output filename.
|
|
|
|
|
|
|
|
|
|
Each input or output file can in principle contain any number of streams of
|
|
|
|
|
different types (video/audio/subtitle/attachment/data). Allowed number and/or
|
|
|
|
|
types of streams can be limited by the container format. Selecting, which
|
|
|
|
|
streams from which inputs go into output, is done either automatically or with
|
|
|
|
|
the @code{-map} option (see the Stream selection chapter).
|
|
|
|
|
|
|
|
|
|
To refer to input files in options, you must use their indices (0-based). E.g.
|
|
|
|
|
the first input file is @code{0}, the second is @code{1} etc. Similarly, streams
|
|
|
|
|
within a file are referred to by their indices. E.g. @code{2:3} refers to the
|
|
|
|
|
fourth stream in the third input file. See also the Stream specifiers chapter.
|
|
|
|
|
|
2010-06-01 13:09:49 +00:00
|
|
|
|
As a general rule, options are applied to the next specified
|
|
|
|
|
file. Therefore, order is important, and you can have the same
|
|
|
|
|
option on the command line multiple times. Each occurrence is
|
|
|
|
|
then applied to the next input or output file.
|
2011-09-13 18:34:19 +00:00
|
|
|
|
Exceptions from this rule are the global options (e.g. verbosity level),
|
|
|
|
|
which should be specified first.
|
2010-06-01 13:09:49 +00:00
|
|
|
|
|
2011-12-11 17:50:39 +00:00
|
|
|
|
Do not mix input and output files -- first specify all input files, then all
|
|
|
|
|
output files. Also do not mix options which belong to different files. All
|
|
|
|
|
options apply ONLY to the next input or output file and are reset between files.
|
|
|
|
|
|
2011-01-16 13:04:56 +00:00
|
|
|
|
@itemize
|
|
|
|
|
@item
|
|
|
|
|
To set the video bitrate of the output file to 64kbit/s:
|
2010-06-01 13:09:49 +00:00
|
|
|
|
@example
|
2012-10-15 17:24:08 +00:00
|
|
|
|
ffmpeg -i input.avi -b:v 64k -bufsize 64k output.avi
|
2010-06-01 13:09:49 +00:00
|
|
|
|
@end example
|
|
|
|
|
|
2011-01-16 13:04:56 +00:00
|
|
|
|
@item
|
|
|
|
|
To force the frame rate of the output file to 24 fps:
|
2010-06-01 13:09:49 +00:00
|
|
|
|
@example
|
|
|
|
|
ffmpeg -i input.avi -r 24 output.avi
|
|
|
|
|
@end example
|
|
|
|
|
|
2011-01-16 13:04:56 +00:00
|
|
|
|
@item
|
|
|
|
|
To force the frame rate of the input file (valid for raw formats only)
|
2010-06-01 13:09:49 +00:00
|
|
|
|
to 1 fps and the frame rate of the output file to 24 fps:
|
|
|
|
|
@example
|
|
|
|
|
ffmpeg -r 1 -i input.m2v -r 24 output.avi
|
|
|
|
|
@end example
|
2011-01-16 13:04:56 +00:00
|
|
|
|
@end itemize
|
2010-06-01 13:09:49 +00:00
|
|
|
|
|
|
|
|
|
The format option may be needed for raw input files.
|
|
|
|
|
|
|
|
|
|
@c man end DESCRIPTION
|
|
|
|
|
|
2012-06-04 09:00:34 +00:00
|
|
|
|
@chapter Detailed description
|
|
|
|
|
@c man begin DETAILED DESCRIPTION
|
|
|
|
|
|
2012-06-04 20:54:15 +00:00
|
|
|
|
The transcoding process in @command{ffmpeg} for each output can be described by
|
2012-06-04 09:00:34 +00:00
|
|
|
|
the following diagram:
|
|
|
|
|
|
|
|
|
|
@example
|
|
|
|
|
_______ ______________ _________ ______________ ________
|
|
|
|
|
| | | | | | | | | |
|
|
|
|
|
| input | demuxer | encoded data | decoder | decoded | encoder | encoded data | muxer | output |
|
|
|
|
|
| file | ---------> | packets | ---------> | frames | ---------> | packets | -------> | file |
|
|
|
|
|
|_______| |______________| |_________| |______________| |________|
|
|
|
|
|
|
|
|
|
|
@end example
|
|
|
|
|
|
2012-06-04 20:54:15 +00:00
|
|
|
|
@command{ffmpeg} calls the libavformat library (containing demuxers) to read
|
2012-06-04 09:00:34 +00:00
|
|
|
|
input files and get packets containing encoded data from them. When there are
|
2012-06-04 20:54:15 +00:00
|
|
|
|
multiple input files, @command{ffmpeg} tries to keep them synchronized by
|
2012-06-04 09:00:34 +00:00
|
|
|
|
tracking lowest timestamp on any active input stream.
|
|
|
|
|
|
|
|
|
|
Encoded packets are then passed to the decoder (unless streamcopy is selected
|
|
|
|
|
for the stream, see further for a description). The decoder produces
|
|
|
|
|
uncompressed frames (raw video/PCM audio/...) which can be processed further by
|
|
|
|
|
filtering (see next section). After filtering the frames are passed to the
|
|
|
|
|
encoder, which encodes them and outputs encoded packets again. Finally those are
|
|
|
|
|
passed to the muxer, which writes the encoded packets to the output file.
|
|
|
|
|
|
|
|
|
|
@section Filtering
|
2012-06-04 20:54:15 +00:00
|
|
|
|
Before encoding, @command{ffmpeg} can process raw audio and video frames using
|
2012-06-04 09:00:34 +00:00
|
|
|
|
filters from the libavfilter library. Several chained filters form a filter
|
2012-06-04 20:54:15 +00:00
|
|
|
|
graph. @command{ffmpeg} distinguishes between two types of filtergraphs -
|
2012-06-04 09:00:34 +00:00
|
|
|
|
simple and complex.
|
|
|
|
|
|
|
|
|
|
@subsection Simple filtergraphs
|
|
|
|
|
Simple filtergraphs are those that have exactly one input and output, both of
|
|
|
|
|
the same type. In the above diagram they can be represented by simply inserting
|
|
|
|
|
an additional step between decoding and encoding:
|
|
|
|
|
|
|
|
|
|
@example
|
|
|
|
|
_________ __________ ______________
|
|
|
|
|
| | | | | |
|
|
|
|
|
| decoded | simple filtergraph | filtered | encoder | encoded data |
|
|
|
|
|
| frames | -------------------> | frames | ---------> | packets |
|
|
|
|
|
|_________| |__________| |______________|
|
|
|
|
|
|
|
|
|
|
@end example
|
|
|
|
|
|
|
|
|
|
Simple filtergraphs are configured with the per-stream @option{-filter} option
|
|
|
|
|
(with @option{-vf} and @option{-af} aliases for video and audio respectively).
|
|
|
|
|
A simple filtergraph for video can look for example like this:
|
|
|
|
|
|
|
|
|
|
@example
|
|
|
|
|
_______ _____________ _______ _____ ________
|
|
|
|
|
| | | | | | | | | |
|
|
|
|
|
| input | ---> | deinterlace | ---> | scale | ---> | fps | ---> | output |
|
|
|
|
|
|_______| |_____________| |_______| |_____| |________|
|
|
|
|
|
|
|
|
|
|
@end example
|
|
|
|
|
|
|
|
|
|
Note that some filters change frame properties but not frame contents. E.g. the
|
|
|
|
|
@code{fps} filter in the example above changes number of frames, but does not
|
|
|
|
|
touch the frame contents. Another example is the @code{setpts} filter, which
|
|
|
|
|
only sets timestamps and otherwise passes the frames unchanged.
|
|
|
|
|
|
|
|
|
|
@subsection Complex filtergraphs
|
|
|
|
|
Complex filtergraphs are those which cannot be described as simply a linear
|
|
|
|
|
processing chain applied to one stream. This is the case e.g. when the graph has
|
|
|
|
|
more than one input and/or output, or when output stream type is different from
|
|
|
|
|
input. They can be represented with the following diagram:
|
|
|
|
|
|
|
|
|
|
@example
|
|
|
|
|
_________
|
|
|
|
|
| |
|
|
|
|
|
| input 0 |\ __________
|
|
|
|
|
|_________| \ | |
|
|
|
|
|
\ _________ /| output 0 |
|
|
|
|
|
\ | | / |__________|
|
|
|
|
|
_________ \| complex | /
|
|
|
|
|
| | | |/
|
|
|
|
|
| input 1 |---->| filter |\
|
|
|
|
|
|_________| | | \ __________
|
|
|
|
|
/| graph | \ | |
|
|
|
|
|
/ | | \| output 1 |
|
|
|
|
|
_________ / |_________| |__________|
|
|
|
|
|
| | /
|
|
|
|
|
| input 2 |/
|
|
|
|
|
|_________|
|
|
|
|
|
|
|
|
|
|
@end example
|
|
|
|
|
|
|
|
|
|
Complex filtergraphs are configured with the @option{-filter_complex} option.
|
|
|
|
|
Note that this option is global, since a complex filtergraph by its nature
|
|
|
|
|
cannot be unambiguously associated with a single stream or file.
|
|
|
|
|
|
|
|
|
|
A trivial example of a complex filtergraph is the @code{overlay} filter, which
|
|
|
|
|
has two video inputs and one video output, containing one video overlaid on top
|
|
|
|
|
of the other. Its audio counterpart is the @code{amix} filter.
|
|
|
|
|
|
|
|
|
|
@section Stream copy
|
|
|
|
|
Stream copy is a mode selected by supplying the @code{copy} parameter to the
|
2012-06-04 20:54:15 +00:00
|
|
|
|
@option{-codec} option. It makes @command{ffmpeg} omit the decoding and encoding
|
2012-06-04 09:00:34 +00:00
|
|
|
|
step for the specified stream, so it does only demuxing and muxing. It is useful
|
|
|
|
|
for changing the container format or modifying container-level metadata. The
|
|
|
|
|
diagram above will in this case simplify to this:
|
|
|
|
|
|
|
|
|
|
@example
|
|
|
|
|
_______ ______________ ________
|
|
|
|
|
| | | | | |
|
|
|
|
|
| input | demuxer | encoded data | muxer | output |
|
|
|
|
|
| file | ---------> | packets | -------> | file |
|
|
|
|
|
|_______| |______________| |________|
|
|
|
|
|
|
|
|
|
|
@end example
|
|
|
|
|
|
|
|
|
|
Since there is no decoding or encoding, it is very fast and there is no quality
|
|
|
|
|
loss. However it might not work in some cases because of many factors. Applying
|
|
|
|
|
filters is obviously also impossible, since filters work on uncompressed data.
|
|
|
|
|
|
|
|
|
|
@c man end DETAILED DESCRIPTION
|
|
|
|
|
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@chapter Stream selection
|
|
|
|
|
@c man begin STREAM SELECTION
|
|
|
|
|
|
2011-11-17 17:08:10 +00:00
|
|
|
|
By default ffmpeg includes only one stream of each type (video, audio, subtitle)
|
|
|
|
|
present in the input files and adds them to each output file. It picks the
|
|
|
|
|
"best" of each based upon the following criteria; for video it is the stream
|
|
|
|
|
with the highest resolution, for audio the stream with the most channels, for
|
|
|
|
|
subtitle it's the first subtitle stream. In the case where several streams of
|
|
|
|
|
the same type rate equally, the lowest numbered stream is chosen.
|
2011-09-13 18:34:19 +00:00
|
|
|
|
|
|
|
|
|
You can disable some of those defaults by using @code{-vn/-an/-sn} options. For
|
|
|
|
|
full manual control, use the @code{-map} option, which disables the defaults just
|
|
|
|
|
described.
|
|
|
|
|
|
|
|
|
|
@c man end STREAM SELECTION
|
|
|
|
|
|
2010-06-08 22:32:29 +00:00
|
|
|
|
@chapter Options
|
2003-08-24 11:16:29 +00:00
|
|
|
|
@c man begin OPTIONS
|
2010-02-21 10:56:38 +00:00
|
|
|
|
|
2011-09-02 20:17:38 +00:00
|
|
|
|
@include avtools-common-opts.texi
|
2010-02-21 10:56:38 +00:00
|
|
|
|
|
2002-10-27 22:00:34 +00:00
|
|
|
|
@section Main options
|
|
|
|
|
|
2003-08-24 11:16:29 +00:00
|
|
|
|
@table @option
|
2009-12-12 16:18:11 +00:00
|
|
|
|
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@item -f @var{fmt} (@emph{input/output})
|
2011-11-29 16:54:49 +00:00
|
|
|
|
Force input or output file format. The format is normally auto detected for input
|
2011-09-28 00:57:34 +00:00
|
|
|
|
files and guessed from file extension for output files, so this option is not
|
|
|
|
|
needed in most cases.
|
2003-08-24 15:22:26 +00:00
|
|
|
|
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@item -i @var{filename} (@emph{input})
|
2008-06-06 07:51:36 +00:00
|
|
|
|
input file name
|
2002-10-27 22:00:34 +00:00
|
|
|
|
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@item -y (@emph{global})
|
2011-09-28 00:57:34 +00:00
|
|
|
|
Overwrite output files without asking.
|
2002-10-27 22:00:34 +00:00
|
|
|
|
|
2011-11-24 20:42:42 +00:00
|
|
|
|
@item -n (@emph{global})
|
|
|
|
|
Do not overwrite output files but exit if file exists.
|
|
|
|
|
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@item -c[:@var{stream_specifier}] @var{codec} (@emph{input/output,per-stream})
|
|
|
|
|
@itemx -codec[:@var{stream_specifier}] @var{codec} (@emph{input/output,per-stream})
|
|
|
|
|
Select an encoder (when used before an output file) or a decoder (when used
|
|
|
|
|
before an input file) for one or more streams. @var{codec} is the name of a
|
|
|
|
|
decoder/encoder or a special value @code{copy} (output only) to indicate that
|
2011-11-29 16:54:49 +00:00
|
|
|
|
the stream is not to be re-encoded.
|
2011-09-13 18:34:19 +00:00
|
|
|
|
|
|
|
|
|
For example
|
|
|
|
|
@example
|
|
|
|
|
ffmpeg -i INPUT -map 0 -c:v libx264 -c:a copy OUTPUT
|
|
|
|
|
@end example
|
|
|
|
|
encodes all video streams with libx264 and copies all audio streams.
|
|
|
|
|
|
|
|
|
|
For each stream, the last matching @code{c} option is applied, so
|
|
|
|
|
@example
|
|
|
|
|
ffmpeg -i INPUT -map 0 -c copy -c:v:1 libx264 -c:a:137 libvorbis OUTPUT
|
|
|
|
|
@end example
|
|
|
|
|
will copy all the streams except the second video, which will be encoded with
|
|
|
|
|
libx264, and the 138th audio, which will be encoded with libvorbis.
|
|
|
|
|
|
|
|
|
|
@item -t @var{duration} (@emph{output})
|
2011-09-28 00:57:34 +00:00
|
|
|
|
Stop writing the output after its duration reaches @var{duration}.
|
|
|
|
|
@var{duration} may be a number in seconds, or in @code{hh:mm:ss[.xxx]} form.
|
2002-10-27 22:00:34 +00:00
|
|
|
|
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@item -fs @var{limit_size} (@emph{output})
|
2012-01-18 12:13:41 +00:00
|
|
|
|
Set the file size limit, expressed in bytes.
|
2006-11-12 23:19:26 +00:00
|
|
|
|
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@item -ss @var{position} (@emph{input/output})
|
|
|
|
|
When used as an input option (before @code{-i}), seeks in this input file to
|
|
|
|
|
@var{position}. When used as an output option (before an output filename),
|
|
|
|
|
decodes but discards input until the timestamps reach @var{position}. This is
|
|
|
|
|
slower, but more accurate.
|
2004-03-08 19:13:23 +00:00
|
|
|
|
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@var{position} may be either in seconds or in @code{hh:mm:ss[.xxx]} form.
|
|
|
|
|
|
|
|
|
|
@item -itsoffset @var{offset} (@emph{input})
|
2006-11-12 23:19:26 +00:00
|
|
|
|
Set the input time offset in seconds.
|
|
|
|
|
@code{[-]hh:mm:ss[.xxx]} syntax is also supported.
|
|
|
|
|
The offset is added to the timestamps of the input files.
|
|
|
|
|
Specifying a positive offset means that the corresponding
|
2011-09-28 00:57:34 +00:00
|
|
|
|
streams are delayed by @var{offset} seconds.
|
2006-11-12 23:19:26 +00:00
|
|
|
|
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@item -timestamp @var{time} (@emph{output})
|
2010-06-22 22:54:36 +00:00
|
|
|
|
Set the recording timestamp in the container.
|
|
|
|
|
The syntax for @var{time} is:
|
|
|
|
|
@example
|
2012-04-23 12:56:01 +00:00
|
|
|
|
now|([(YYYY-MM-DD|YYYYMMDD)[T|t| ]]((HH:MM:SS[.m...])|(HHMMSS[.m...]))[Z|z])
|
2010-06-22 22:54:36 +00:00
|
|
|
|
@end example
|
|
|
|
|
If the value is "now" it takes the current time.
|
|
|
|
|
Time is local time unless 'Z' or 'z' is appended, in which case it is
|
|
|
|
|
interpreted as UTC.
|
|
|
|
|
If the year-month-day part is not specified it takes the current
|
|
|
|
|
year-month-day.
|
2006-11-12 23:19:26 +00:00
|
|
|
|
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@item -metadata[:metadata_specifier] @var{key}=@var{value} (@emph{output,per-metadata})
|
2009-06-28 20:25:08 +00:00
|
|
|
|
Set a metadata key/value pair.
|
2006-11-12 23:19:26 +00:00
|
|
|
|
|
2011-09-13 18:34:19 +00:00
|
|
|
|
An optional @var{metadata_specifier} may be given to set metadata
|
|
|
|
|
on streams or chapters. See @code{-map_metadata} documentation for
|
|
|
|
|
details.
|
|
|
|
|
|
|
|
|
|
This option overrides metadata set with @code{-map_metadata}. It is
|
|
|
|
|
also possible to delete metadata by using an empty value.
|
|
|
|
|
|
2009-07-01 20:11:53 +00:00
|
|
|
|
For example, for setting the title in the output file:
|
2009-06-27 13:26:35 +00:00
|
|
|
|
@example
|
2009-07-01 20:11:53 +00:00
|
|
|
|
ffmpeg -i in.avi -metadata title="my title" out.flv
|
2009-06-27 13:26:35 +00:00
|
|
|
|
@end example
|
2006-09-08 12:28:28 +00:00
|
|
|
|
|
2012-01-25 22:01:34 +00:00
|
|
|
|
To set the language of the first audio stream:
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@example
|
2012-01-25 22:01:34 +00:00
|
|
|
|
ffmpeg -i INPUT -metadata:s:a:1 language=eng OUTPUT
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@end example
|
|
|
|
|
|
|
|
|
|
@item -target @var{type} (@emph{output})
|
2011-09-28 00:57:34 +00:00
|
|
|
|
Specify target file type (@code{vcd}, @code{svcd}, @code{dvd}, @code{dv},
|
|
|
|
|
@code{dv50}). @var{type} may be prefixed with @code{pal-}, @code{ntsc-} or
|
|
|
|
|
@code{film-} to use the corresponding standard. All the format options
|
|
|
|
|
(bitrate, codecs, buffer sizes) are then set automatically. You can just type:
|
2003-12-10 19:17:15 +00:00
|
|
|
|
|
|
|
|
|
@example
|
|
|
|
|
ffmpeg -i myfile.avi -target vcd /tmp/vcd.mpg
|
|
|
|
|
@end example
|
|
|
|
|
|
2005-08-05 01:06:04 +00:00
|
|
|
|
Nevertheless you can specify additional options as long as you know
|
|
|
|
|
they do not conflict with the standard, as in:
|
2004-10-07 03:20:27 +00:00
|
|
|
|
|
|
|
|
|
@example
|
|
|
|
|
ffmpeg -i myfile.avi -target vcd -bf 2 /tmp/vcd.mpg
|
|
|
|
|
@end example
|
|
|
|
|
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@item -dframes @var{number} (@emph{output})
|
|
|
|
|
Set the number of data frames to record. This is an alias for @code{-frames:d}.
|
2006-11-12 23:19:26 +00:00
|
|
|
|
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@item -frames[:@var{stream_specifier}] @var{framecount} (@emph{output,per-stream})
|
|
|
|
|
Stop writing to the stream after @var{framecount} frames.
|
2006-11-12 23:19:26 +00:00
|
|
|
|
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@item -q[:@var{stream_specifier}] @var{q} (@emph{output,per-stream})
|
|
|
|
|
@itemx -qscale[:@var{stream_specifier}] @var{q} (@emph{output,per-stream})
|
|
|
|
|
Use fixed quality scale (VBR). The meaning of @var{q} is
|
|
|
|
|
codec-dependent.
|
2006-11-12 23:19:26 +00:00
|
|
|
|
|
2011-12-11 17:50:39 +00:00
|
|
|
|
@item -filter[:@var{stream_specifier}] @var{filter_graph} (@emph{output,per-stream})
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@var{filter_graph} is a description of the filter graph to apply to
|
|
|
|
|
the stream. Use @code{-filters} to show all the available filters
|
|
|
|
|
(including also sources and sinks).
|
2012-04-01 13:08:33 +00:00
|
|
|
|
|
|
|
|
|
See also the @option{-filter_complex} option if you want to create filter graphs
|
|
|
|
|
with multiple inputs and/or outputs.
|
2011-12-11 17:50:39 +00:00
|
|
|
|
@item -pre[:@var{stream_specifier}] @var{preset_name} (@emph{output,per-stream})
|
|
|
|
|
Specify the preset for matching stream(s).
|
|
|
|
|
|
|
|
|
|
@item -stats (@emph{global})
|
|
|
|
|
Print encoding progress/statistics. On by default.
|
|
|
|
|
|
2012-07-08 11:03:39 +00:00
|
|
|
|
@item -progress @var{url} (@emph{global})
|
|
|
|
|
Send program-friendly progress information to @var{url}.
|
|
|
|
|
|
|
|
|
|
Progress information is written approximately every second and at the end of
|
|
|
|
|
the encoding process. It is made of "@var{key}=@var{value}" lines. @var{key}
|
|
|
|
|
consists of only alphanumeric characters. The last key of a sequence of
|
|
|
|
|
progress information is always "progress".
|
|
|
|
|
|
2012-07-11 19:10:17 +00:00
|
|
|
|
@item -stdin
|
|
|
|
|
Enable interaction on standard input. On by default unless standard input is
|
2012-08-20 23:42:47 +00:00
|
|
|
|
used as an input. To explicitly disable interaction you need to specify
|
|
|
|
|
@code{-nostdin}.
|
2012-07-11 19:10:17 +00:00
|
|
|
|
|
2012-08-20 23:42:47 +00:00
|
|
|
|
Disabling interaction on standard input is useful, for example, if
|
|
|
|
|
ffmpeg is in the background process group. Roughly the same result can
|
|
|
|
|
be achieved with @code{ffmpeg ... < /dev/null} but it requires a
|
|
|
|
|
shell.
|
2012-07-11 19:10:17 +00:00
|
|
|
|
|
2012-01-20 16:06:26 +00:00
|
|
|
|
@item -debug_ts (@emph{global})
|
|
|
|
|
Print timestamp information. It is off by default. This option is
|
|
|
|
|
mostly useful for testing and debugging purposes, and the output
|
|
|
|
|
format may change from one version to another, so it should not be
|
|
|
|
|
employed by portable scripts.
|
|
|
|
|
|
|
|
|
|
See also the option @code{-fdebug ts}.
|
|
|
|
|
|
2011-12-11 17:50:39 +00:00
|
|
|
|
@item -attach @var{filename} (@emph{output})
|
|
|
|
|
Add an attachment to the output file. This is supported by a few formats
|
|
|
|
|
like Matroska for e.g. fonts used in rendering subtitles. Attachments
|
|
|
|
|
are implemented as a specific type of stream, so this option will add
|
|
|
|
|
a new stream to the file. It is then possible to use per-stream options
|
|
|
|
|
on this stream in the usual way. Attachment streams created with this
|
|
|
|
|
option will be created after all the other streams (i.e. those created
|
|
|
|
|
with @code{-map} or automatic mappings).
|
|
|
|
|
|
|
|
|
|
Note that for Matroska you also have to set the mimetype metadata tag:
|
|
|
|
|
@example
|
|
|
|
|
ffmpeg -i INPUT -attach DejaVuSans.ttf -metadata:s:2 mimetype=application/x-truetype-font out.mkv
|
|
|
|
|
@end example
|
|
|
|
|
(assuming that the attachment stream will be third in the output file).
|
|
|
|
|
|
|
|
|
|
@item -dump_attachment[:@var{stream_specifier}] @var{filename} (@emph{input,per-stream})
|
|
|
|
|
Extract the matching attachment stream into a file named @var{filename}. If
|
|
|
|
|
@var{filename} is empty, then the value of the @code{filename} metadata tag
|
|
|
|
|
will be used.
|
|
|
|
|
|
|
|
|
|
E.g. to extract the first attachment to a file named 'out.ttf':
|
|
|
|
|
@example
|
|
|
|
|
ffmpeg -dump_attachment:t:0 out.ttf INPUT
|
|
|
|
|
@end example
|
|
|
|
|
To extract all attachments to files determined by the @code{filename} tag:
|
|
|
|
|
@example
|
|
|
|
|
ffmpeg -dump_attachment:t "" INPUT
|
|
|
|
|
@end example
|
|
|
|
|
|
|
|
|
|
Technical note -- attachments are implemented as codec extradata, so this
|
|
|
|
|
option can actually be used to extract extradata from any stream, not just
|
|
|
|
|
attachments.
|
2004-06-24 18:39:24 +00:00
|
|
|
|
|
2002-10-27 22:00:34 +00:00
|
|
|
|
@end table
|
|
|
|
|
|
|
|
|
|
@section Video Options
|
|
|
|
|
|
2003-08-24 11:16:29 +00:00
|
|
|
|
@table @option
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@item -vframes @var{number} (@emph{output})
|
|
|
|
|
Set the number of video frames to record. This is an alias for @code{-frames:v}.
|
|
|
|
|
@item -r[:@var{stream_specifier}] @var{fps} (@emph{input/output,per-stream})
|
2012-05-26 14:52:58 +00:00
|
|
|
|
Set frame rate (Hz value, fraction or abbreviation).
|
|
|
|
|
|
|
|
|
|
As an input option, ignore any timestamps stored in the file and instead
|
|
|
|
|
generate timestamps assuming constant frame rate @var{fps}.
|
|
|
|
|
|
|
|
|
|
As an output option, duplicate or drop input frames to achieve constant output
|
2012-12-12 09:08:43 +00:00
|
|
|
|
frame rate @var{fps}.
|
2012-05-26 14:52:58 +00:00
|
|
|
|
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@item -s[:@var{stream_specifier}] @var{size} (@emph{input/output,per-stream})
|
2012-05-26 11:52:44 +00:00
|
|
|
|
Set frame size.
|
|
|
|
|
|
|
|
|
|
As an input option, this is a shortcut for the @option{video_size} private
|
|
|
|
|
option, recognized by some demuxers for which the frame size is either not
|
|
|
|
|
stored in the file or is configurable -- e.g. raw video or video grabbers.
|
|
|
|
|
|
|
|
|
|
As an output option, this inserts the @code{scale} video filter to the
|
|
|
|
|
@emph{end} of the corresponding filtergraph. Please use the @code{scale} filter
|
|
|
|
|
directly to insert it at the beginning or some other place.
|
|
|
|
|
|
2012-05-30 01:32:32 +00:00
|
|
|
|
The format is @samp{wxh} (default - same as source).
|
2003-08-24 15:22:26 +00:00
|
|
|
|
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@item -aspect[:@var{stream_specifier}] @var{aspect} (@emph{output,per-stream})
|
2011-03-26 15:17:18 +00:00
|
|
|
|
Set the video display aspect ratio specified by @var{aspect}.
|
|
|
|
|
|
|
|
|
|
@var{aspect} can be a floating point number string, or a string of the
|
|
|
|
|
form @var{num}:@var{den}, where @var{num} and @var{den} are the
|
|
|
|
|
numerator and denominator of the aspect ratio. For example "4:3",
|
|
|
|
|
"16:9", "1.3333", and "1.7777" are valid argument values.
|
|
|
|
|
|
2010-10-18 13:34:46 +00:00
|
|
|
|
@item -croptop @var{size}
|
|
|
|
|
@item -cropbottom @var{size}
|
|
|
|
|
@item -cropleft @var{size}
|
|
|
|
|
@item -cropright @var{size}
|
|
|
|
|
All the crop options have been removed. Use -vf
|
|
|
|
|
crop=width:height:x:y instead.
|
|
|
|
|
|
2008-04-08 16:39:30 +00:00
|
|
|
|
@item -padtop @var{size}
|
|
|
|
|
@item -padbottom @var{size}
|
|
|
|
|
@item -padleft @var{size}
|
|
|
|
|
@item -padright @var{size}
|
|
|
|
|
@item -padcolor @var{hex_color}
|
2010-07-04 20:59:00 +00:00
|
|
|
|
All the pad options have been removed. Use -vf
|
|
|
|
|
pad=width:height:x:y:color instead.
|
2011-09-13 18:34:19 +00:00
|
|
|
|
|
|
|
|
|
@item -vn (@emph{output})
|
2005-08-05 01:06:04 +00:00
|
|
|
|
Disable video recording.
|
2012-03-15 07:17:05 +00:00
|
|
|
|
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@item -vcodec @var{codec} (@emph{output})
|
|
|
|
|
Set the video codec. This is an alias for @code{-codec:v}.
|
|
|
|
|
|
2012-08-19 06:29:44 +00:00
|
|
|
|
@item -pass[:@var{stream_specifier}] @var{n} (@emph{output,per-stream})
|
2008-12-01 08:27:53 +00:00
|
|
|
|
Select the pass number (1 or 2). It is used to do two-pass
|
|
|
|
|
video encoding. The statistics of the video are recorded in the first
|
|
|
|
|
pass into a log file (see also the option -passlogfile),
|
|
|
|
|
and in the second pass that log file is used to generate the video
|
|
|
|
|
at the exact requested bitrate.
|
2008-08-13 19:49:37 +00:00
|
|
|
|
On pass 1, you may just deactivate audio and set output to null,
|
|
|
|
|
examples for Windows and Unix:
|
|
|
|
|
@example
|
2011-09-13 18:34:19 +00:00
|
|
|
|
ffmpeg -i foo.mov -c:v libxvid -pass 1 -an -f rawvideo -y NUL
|
|
|
|
|
ffmpeg -i foo.mov -c:v libxvid -pass 1 -an -f rawvideo -y /dev/null
|
2008-08-13 19:49:37 +00:00
|
|
|
|
@end example
|
2002-10-27 22:00:34 +00:00
|
|
|
|
|
2012-08-19 07:15:48 +00:00
|
|
|
|
@item -passlogfile[:@var{stream_specifier}] @var{prefix} (@emph{output,per-stream})
|
2008-12-26 13:40:02 +00:00
|
|
|
|
Set two-pass log file name prefix to @var{prefix}, the default file name
|
|
|
|
|
prefix is ``ffmpeg2pass''. The complete file name will be
|
|
|
|
|
@file{PREFIX-N.log}, where N is a number specific to the output
|
2011-09-13 18:34:19 +00:00
|
|
|
|
stream
|
2006-11-12 23:19:26 +00:00
|
|
|
|
|
2009-12-16 12:08:01 +00:00
|
|
|
|
@item -vlang @var{code}
|
|
|
|
|
Set the ISO 639 language code (3 letters) of the current video stream.
|
|
|
|
|
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@item -vf @var{filter_graph} (@emph{output})
|
2010-06-21 22:09:07 +00:00
|
|
|
|
@var{filter_graph} is a description of the filter graph to apply to
|
|
|
|
|
the input video.
|
|
|
|
|
Use the option "-filters" to show all the available filters (including
|
2011-09-13 18:34:19 +00:00
|
|
|
|
also sources and sinks). This is an alias for @code{-filter:v}.
|
2010-06-21 22:09:07 +00:00
|
|
|
|
|
2002-10-27 22:00:34 +00:00
|
|
|
|
@end table
|
|
|
|
|
|
2003-08-24 15:22:26 +00:00
|
|
|
|
@section Advanced Video Options
|
2002-10-27 22:00:34 +00:00
|
|
|
|
|
2003-08-24 11:16:29 +00:00
|
|
|
|
@table @option
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@item -pix_fmt[:@var{stream_specifier}] @var{format} (@emph{input/output,per-stream})
|
|
|
|
|
Set pixel format. Use @code{-pix_fmts} to show all the supported
|
2007-05-30 14:20:55 +00:00
|
|
|
|
pixel formats.
|
2012-04-27 08:55:12 +00:00
|
|
|
|
If the selected pixel format can not be selected, ffmpeg will print a
|
|
|
|
|
warning and select the best pixel format supported by the encoder.
|
|
|
|
|
If @var{pix_fmt} is prefixed by a @code{+}, ffmpeg will exit with an error
|
|
|
|
|
if the requested pixel format can not be selected, and automatic conversions
|
|
|
|
|
inside filter graphs are disabled.
|
|
|
|
|
If @var{pix_fmt} is a single @code{+}, ffmpeg selects the same pixel format
|
|
|
|
|
as the input (or graph output) and automatic conversions are disabled.
|
|
|
|
|
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@item -sws_flags @var{flags} (@emph{input/output})
|
2010-03-18 00:24:26 +00:00
|
|
|
|
Set SwScaler flags.
|
2008-04-08 16:39:30 +00:00
|
|
|
|
@item -vdt @var{n}
|
2006-11-12 23:19:26 +00:00
|
|
|
|
Discard threshold.
|
2010-11-03 19:43:55 +00:00
|
|
|
|
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@item -rc_override[:@var{stream_specifier}] @var{override} (@emph{output,per-stream})
|
2011-11-29 16:54:49 +00:00
|
|
|
|
Rate control override for specific intervals, formatted as "int,int,int"
|
2011-05-18 16:54:42 +00:00
|
|
|
|
list separated with slashes. Two first values are the beginning and
|
|
|
|
|
end frame numbers, last one is quantizer to use if positive, or quality
|
|
|
|
|
factor if negative.
|
2003-08-24 15:22:26 +00:00
|
|
|
|
|
|
|
|
|
@item -deinterlace
|
2005-08-05 01:06:04 +00:00
|
|
|
|
Deinterlace pictures.
|
2012-01-30 15:05:28 +00:00
|
|
|
|
This option is deprecated since the deinterlacing is very low quality.
|
|
|
|
|
Use the yadif filter with @code{-filter:v yadif}.
|
2006-09-28 19:06:57 +00:00
|
|
|
|
@item -ilme
|
2005-08-05 01:06:04 +00:00
|
|
|
|
Force interlacing support in encoder (MPEG-2 and MPEG-4 only).
|
|
|
|
|
Use this option if your input file is interlaced and you want
|
|
|
|
|
to keep the interlaced format for minimum losses.
|
|
|
|
|
The alternative is to deinterlace the input stream with
|
|
|
|
|
@option{-deinterlace}, but deinterlacing introduces losses.
|
2003-08-24 15:22:26 +00:00
|
|
|
|
@item -psnr
|
2005-08-05 01:06:04 +00:00
|
|
|
|
Calculate PSNR of compressed frames.
|
2003-08-24 15:22:26 +00:00
|
|
|
|
@item -vstats
|
2005-08-05 01:06:04 +00:00
|
|
|
|
Dump video coding statistics to @file{vstats_HHMMSS.log}.
|
2008-04-08 16:39:30 +00:00
|
|
|
|
@item -vstats_file @var{file}
|
2007-04-25 13:55:11 +00:00
|
|
|
|
Dump video coding statistics to @var{file}.
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@item -top[:@var{stream_specifier}] @var{n} (@emph{output,per-stream})
|
2006-11-12 23:19:26 +00:00
|
|
|
|
top=1/bottom=0/auto=-1 field first
|
2008-04-08 16:39:30 +00:00
|
|
|
|
@item -dc @var{precision}
|
2006-11-12 23:19:26 +00:00
|
|
|
|
Intra_dc_precision.
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@item -vtag @var{fourcc/tag} (@emph{output})
|
|
|
|
|
Force video tag/fourcc. This is an alias for @code{-tag:v}.
|
|
|
|
|
@item -qphist (@emph{global})
|
|
|
|
|
Show QP histogram
|
2008-04-08 16:39:30 +00:00
|
|
|
|
@item -vbsf @var{bitstream_filter}
|
2011-09-13 18:34:19 +00:00
|
|
|
|
Deprecated see -bsf
|
|
|
|
|
@item -force_key_frames[:@var{stream_specifier}] @var{time}[,@var{time}...] (@emph{output,per-stream})
|
2010-10-18 21:47:15 +00:00
|
|
|
|
Force key frames at the specified timestamps, more precisely at the first
|
|
|
|
|
frames after each specified time.
|
|
|
|
|
This option can be useful to ensure that a seek point is present at a
|
|
|
|
|
chapter mark or any other designated place in the output file.
|
|
|
|
|
The timestamps must be specified in ascending order.
|
2011-12-11 17:50:39 +00:00
|
|
|
|
|
|
|
|
|
@item -copyinkf[:@var{stream_specifier}] (@emph{output,per-stream})
|
|
|
|
|
When doing stream copy, copy also non-key frames found at the
|
|
|
|
|
beginning.
|
2003-08-24 15:22:26 +00:00
|
|
|
|
@end table
|
|
|
|
|
|
|
|
|
|
@section Audio Options
|
|
|
|
|
|
|
|
|
|
@table @option
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@item -aframes @var{number} (@emph{output})
|
|
|
|
|
Set the number of audio frames to record. This is an alias for @code{-frames:a}.
|
|
|
|
|
@item -ar[:@var{stream_specifier}] @var{freq} (@emph{input/output,per-stream})
|
2011-06-24 09:22:15 +00:00
|
|
|
|
Set the audio sampling frequency. For output streams it is set by
|
|
|
|
|
default to the frequency of the corresponding input stream. For input
|
|
|
|
|
streams this option only makes sense for audio grabbing devices and raw
|
|
|
|
|
demuxers and is mapped to the corresponding demuxer options.
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@item -aq @var{q} (@emph{output})
|
|
|
|
|
Set the audio quality (codec-specific, VBR). This is an alias for -q:a.
|
|
|
|
|
@item -ac[:@var{stream_specifier}] @var{channels} (@emph{input/output,per-stream})
|
2011-06-24 09:22:15 +00:00
|
|
|
|
Set the number of audio channels. For output streams it is set by
|
|
|
|
|
default to the number of input audio channels. For input streams
|
|
|
|
|
this option only makes sense for audio grabbing devices and raw demuxers
|
|
|
|
|
and is mapped to the corresponding demuxer options.
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@item -an (@emph{output})
|
2005-08-05 01:06:04 +00:00
|
|
|
|
Disable audio recording.
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@item -acodec @var{codec} (@emph{input/output})
|
|
|
|
|
Set the audio codec. This is an alias for @code{-codec:a}.
|
|
|
|
|
@item -sample_fmt[:@var{stream_specifier}] @var{sample_fmt} (@emph{output,per-stream})
|
2011-12-11 17:50:39 +00:00
|
|
|
|
Set the audio sample format. Use @code{-sample_fmts} to get a list
|
2011-09-13 18:34:19 +00:00
|
|
|
|
of supported sample formats.
|
2012-05-05 16:22:46 +00:00
|
|
|
|
@item -af @var{filter_graph} (@emph{output})
|
|
|
|
|
@var{filter_graph} is a description of the filter graph to apply to
|
|
|
|
|
the input audio.
|
|
|
|
|
Use the option "-filters" to show all the available filters (including
|
|
|
|
|
also sources and sinks). This is an alias for @code{-filter:a}.
|
2006-11-12 23:19:26 +00:00
|
|
|
|
@end table
|
|
|
|
|
|
|
|
|
|
@section Advanced Audio options:
|
|
|
|
|
|
|
|
|
|
@table @option
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@item -atag @var{fourcc/tag} (@emph{output})
|
|
|
|
|
Force audio tag/fourcc. This is an alias for @code{-tag:a}.
|
2008-04-08 16:39:30 +00:00
|
|
|
|
@item -absf @var{bitstream_filter}
|
2011-09-13 18:34:19 +00:00
|
|
|
|
Deprecated, see -bsf
|
2006-11-12 23:19:26 +00:00
|
|
|
|
@end table
|
|
|
|
|
|
|
|
|
|
@section Subtitle options:
|
|
|
|
|
|
|
|
|
|
@table @option
|
2008-04-08 16:39:30 +00:00
|
|
|
|
@item -slang @var{code}
|
2006-11-12 23:19:26 +00:00
|
|
|
|
Set the ISO 639 language code (3 letters) of the current subtitle stream.
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@item -scodec @var{codec} (@emph{input/output})
|
|
|
|
|
Set the subtitle codec. This is an alias for @code{-codec:s}.
|
|
|
|
|
@item -sn (@emph{output})
|
2009-10-01 13:48:23 +00:00
|
|
|
|
Disable subtitle recording.
|
2008-07-21 21:06:49 +00:00
|
|
|
|
@item -sbsf @var{bitstream_filter}
|
2011-09-13 18:34:19 +00:00
|
|
|
|
Deprecated, see -bsf
|
2003-08-24 15:22:26 +00:00
|
|
|
|
@end table
|
|
|
|
|
|
2012-08-09 09:44:17 +00:00
|
|
|
|
@section Advanced Subtitle options:
|
|
|
|
|
|
|
|
|
|
@table @option
|
|
|
|
|
|
|
|
|
|
@item -fix_sub_duration
|
|
|
|
|
Fix subtitles durations. For each subtitle, wait for the next packet in the
|
|
|
|
|
same stream and adjust the duration of the first to avoid overlap. This is
|
|
|
|
|
necessary with some subtitles codecs, especially DVB subtitles, because the
|
|
|
|
|
duration in the original packet is only a rough estimate and the end is
|
|
|
|
|
actually marked by an empty subtitle frame. Failing to use this option when
|
|
|
|
|
necessary can result in exaggerated durations or muxing failures due to
|
|
|
|
|
non-monotonic timestamps.
|
|
|
|
|
|
|
|
|
|
Note that this option will delay the output of all data until the next
|
|
|
|
|
subtitle packet is decoded: it may increase memory consumption and latency a
|
|
|
|
|
lot.
|
|
|
|
|
|
|
|
|
|
@end table
|
|
|
|
|
|
2003-08-24 15:22:26 +00:00
|
|
|
|
@section Advanced options
|
|
|
|
|
|
|
|
|
|
@table @option
|
2012-04-01 13:08:33 +00:00
|
|
|
|
@item -map [-]@var{input_file_id}[:@var{stream_specifier}][,@var{sync_file_id}[:@var{stream_specifier}]] | @var{[linklabel]} (@emph{output})
|
2011-03-20 07:01:41 +00:00
|
|
|
|
|
2011-09-13 18:34:19 +00:00
|
|
|
|
Designate one or more input streams as a source for the output file. Each input
|
2011-03-20 07:01:41 +00:00
|
|
|
|
stream is identified by the input file index @var{input_file_id} and
|
|
|
|
|
the input stream index @var{input_stream_id} within the input
|
2011-09-13 18:34:19 +00:00
|
|
|
|
file. Both indices start at 0. If specified,
|
|
|
|
|
@var{sync_file_id}:@var{stream_specifier} sets which input stream
|
2011-03-20 07:01:41 +00:00
|
|
|
|
is used as a presentation sync reference.
|
|
|
|
|
|
2011-09-13 18:34:19 +00:00
|
|
|
|
The first @code{-map} option on the command line specifies the
|
2011-03-20 07:01:41 +00:00
|
|
|
|
source for output stream 0, the second @code{-map} option specifies
|
|
|
|
|
the source for output stream 1, etc.
|
|
|
|
|
|
2011-09-13 18:34:19 +00:00
|
|
|
|
A @code{-} character before the stream identifier creates a "negative" mapping.
|
|
|
|
|
It disables matching streams from already created mappings.
|
|
|
|
|
|
2012-04-01 13:08:33 +00:00
|
|
|
|
An alternative @var{[linklabel]} form will map outputs from complex filter
|
|
|
|
|
graphs (see the @option{-filter_complex} option) to the output file.
|
|
|
|
|
@var{linklabel} must correspond to a defined output link label in the graph.
|
|
|
|
|
|
2011-09-13 18:34:19 +00:00
|
|
|
|
For example, to map ALL streams from the first input file to output
|
|
|
|
|
@example
|
|
|
|
|
ffmpeg -i INPUT -map 0 output
|
|
|
|
|
@end example
|
|
|
|
|
|
2011-03-20 07:01:41 +00:00
|
|
|
|
For example, if you have two audio streams in the first input file,
|
2011-09-13 18:34:19 +00:00
|
|
|
|
these streams are identified by "0:0" and "0:1". You can use
|
|
|
|
|
@code{-map} to select which streams to place in an output file. For
|
2011-03-20 07:01:41 +00:00
|
|
|
|
example:
|
|
|
|
|
@example
|
2011-09-13 18:34:19 +00:00
|
|
|
|
ffmpeg -i INPUT -map 0:1 out.wav
|
2011-03-20 07:01:41 +00:00
|
|
|
|
@end example
|
2011-09-13 18:34:19 +00:00
|
|
|
|
will map the input stream in @file{INPUT} identified by "0:1" to
|
2011-03-20 07:01:41 +00:00
|
|
|
|
the (single) output stream in @file{out.wav}.
|
|
|
|
|
|
|
|
|
|
For example, to select the stream with index 2 from input file
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@file{a.mov} (specified by the identifier "0:2"), and stream with
|
|
|
|
|
index 6 from input @file{b.mov} (specified by the identifier "1:6"),
|
2011-03-20 07:01:41 +00:00
|
|
|
|
and copy them to the output file @file{out.mov}:
|
|
|
|
|
@example
|
2011-09-13 18:34:19 +00:00
|
|
|
|
ffmpeg -i a.mov -i b.mov -c copy -map 0:2 -map 1:6 out.mov
|
2011-03-20 07:01:41 +00:00
|
|
|
|
@end example
|
|
|
|
|
|
2011-09-13 18:34:19 +00:00
|
|
|
|
To select all video and the third audio stream from an input file:
|
|
|
|
|
@example
|
|
|
|
|
ffmpeg -i INPUT -map 0:v -map 0:a:2 OUTPUT
|
|
|
|
|
@end example
|
2011-03-20 07:01:41 +00:00
|
|
|
|
|
2011-09-13 18:34:19 +00:00
|
|
|
|
To map all the streams except the second audio, use negative mappings
|
|
|
|
|
@example
|
|
|
|
|
ffmpeg -i INPUT -map 0 -map -0:a:1 OUTPUT
|
|
|
|
|
@end example
|
2011-02-02 20:45:16 +00:00
|
|
|
|
|
2011-09-13 18:34:19 +00:00
|
|
|
|
Note that using this option disables the default mappings for this output file.
|
|
|
|
|
|
2011-11-04 17:54:01 +00:00
|
|
|
|
@item -map_channel [@var{input_file_id}.@var{stream_specifier}.@var{channel_id}|-1][:@var{output_file_id}.@var{stream_specifier}]
|
|
|
|
|
Map an audio channel from a given input to an output. If
|
2012-03-03 18:19:40 +00:00
|
|
|
|
@var{output_file_id}.@var{stream_specifier} is not set, the audio channel will
|
2011-11-04 17:54:01 +00:00
|
|
|
|
be mapped on all the audio streams.
|
|
|
|
|
|
|
|
|
|
Using "-1" instead of
|
|
|
|
|
@var{input_file_id}.@var{stream_specifier}.@var{channel_id} will map a muted
|
|
|
|
|
channel.
|
|
|
|
|
|
|
|
|
|
For example, assuming @var{INPUT} is a stereo audio file, you can switch the
|
|
|
|
|
two audio channels with the following command:
|
|
|
|
|
@example
|
|
|
|
|
ffmpeg -i INPUT -map_channel 0.0.1 -map_channel 0.0.0 OUTPUT
|
|
|
|
|
@end example
|
|
|
|
|
|
|
|
|
|
If you want to mute the first channel and keep the second:
|
|
|
|
|
@example
|
|
|
|
|
ffmpeg -i INPUT -map_channel -1 -map_channel 0.0.1 OUTPUT
|
|
|
|
|
@end example
|
|
|
|
|
|
|
|
|
|
The order of the "-map_channel" option specifies the order of the channels in
|
|
|
|
|
the output stream. The output channel layout is guessed from the number of
|
|
|
|
|
channels mapped (mono if one "-map_channel", stereo if two, etc.). Using "-ac"
|
|
|
|
|
in combination of "-map_channel" makes the channel gain levels to be updated if
|
2012-03-03 18:19:40 +00:00
|
|
|
|
input and output channel layouts don't match (for instance two "-map_channel"
|
|
|
|
|
options and "-ac 6").
|
2011-11-04 17:54:01 +00:00
|
|
|
|
|
2012-03-03 18:19:40 +00:00
|
|
|
|
You can also extract each channel of an input to specific outputs; the following
|
|
|
|
|
command extracts two channels of the @var{INPUT} audio stream (file 0, stream 0)
|
|
|
|
|
to the respective @var{OUTPUT_CH0} and @var{OUTPUT_CH1} outputs:
|
2011-11-04 17:54:01 +00:00
|
|
|
|
@example
|
|
|
|
|
ffmpeg -i INPUT -map_channel 0.0.0 OUTPUT_CH0 -map_channel 0.0.1 OUTPUT_CH1
|
|
|
|
|
@end example
|
|
|
|
|
|
2012-03-03 18:19:40 +00:00
|
|
|
|
The following example splits the channels of a stereo input into two separate
|
|
|
|
|
streams, which are put into the same output file:
|
2011-11-22 20:51:20 +00:00
|
|
|
|
@example
|
|
|
|
|
ffmpeg -i stereo.wav -map 0:0 -map 0:0 -map_channel 0.0.0:0.0 -map_channel 0.0.1:0.1 -y out.ogg
|
|
|
|
|
@end example
|
|
|
|
|
|
2011-12-21 17:12:05 +00:00
|
|
|
|
Note that currently each output stream can only contain channels from a single
|
|
|
|
|
input stream; you can't for example use "-map_channel" to pick multiple input
|
|
|
|
|
audio channels contained in different streams (from the same or different files)
|
|
|
|
|
and merge them into a single output stream. It is therefore not currently
|
|
|
|
|
possible, for example, to turn two separate mono streams into a single stereo
|
2012-03-03 18:19:40 +00:00
|
|
|
|
stream. However splitting a stereo stream into two single channel mono streams
|
2011-12-21 17:12:05 +00:00
|
|
|
|
is possible.
|
2011-11-04 17:54:01 +00:00
|
|
|
|
|
2012-01-10 10:47:46 +00:00
|
|
|
|
If you need this feature, a possible workaround is to use the @emph{amerge}
|
|
|
|
|
filter. For example, if you need to merge a media (here @file{input.mkv}) with 2
|
|
|
|
|
mono audio streams into one single stereo channel audio stream (and keep the
|
2012-03-03 18:19:40 +00:00
|
|
|
|
video stream), you can use the following command:
|
2012-01-10 10:47:46 +00:00
|
|
|
|
@example
|
2012-07-03 10:42:51 +00:00
|
|
|
|
ffmpeg -i input.mkv -filter_complex "[0:1] [0:2] amerge" -c:a pcm_s16le -c:v copy output.mkv
|
2012-01-10 10:47:46 +00:00
|
|
|
|
@end example
|
|
|
|
|
|
2012-01-25 22:01:34 +00:00
|
|
|
|
@item -map_metadata[:@var{metadata_spec_out}] @var{infile}[:@var{metadata_spec_in}] (@emph{output,per-metadata})
|
2011-09-13 18:34:19 +00:00
|
|
|
|
Set metadata information of the next output file from @var{infile}. Note that
|
|
|
|
|
those are file indices (zero-based), not filenames.
|
2012-01-25 22:01:34 +00:00
|
|
|
|
Optional @var{metadata_spec_in/out} parameters specify, which metadata to copy.
|
|
|
|
|
A metadata specifier can have the following forms:
|
|
|
|
|
@table @option
|
|
|
|
|
@item @var{g}
|
|
|
|
|
global metadata, i.e. metadata that applies to the whole file
|
|
|
|
|
|
|
|
|
|
@item @var{s}[:@var{stream_spec}]
|
|
|
|
|
per-stream metadata. @var{stream_spec} is a stream specifier as described
|
|
|
|
|
in the @ref{Stream specifiers} chapter. In an input metadata specifier, the first
|
|
|
|
|
matching stream is copied from. In an output metadata specifier, all matching
|
|
|
|
|
streams are copied to.
|
|
|
|
|
|
|
|
|
|
@item @var{c}:@var{chapter_index}
|
|
|
|
|
per-chapter metadata. @var{chapter_index} is the zero-based chapter index.
|
|
|
|
|
|
|
|
|
|
@item @var{p}:@var{program_index}
|
|
|
|
|
per-program metadata. @var{program_index} is the zero-based program index.
|
|
|
|
|
@end table
|
|
|
|
|
If metadata specifier is omitted, it defaults to global.
|
2010-11-22 07:32:27 +00:00
|
|
|
|
|
2011-09-13 18:34:19 +00:00
|
|
|
|
By default, global metadata is copied from the first input file,
|
2010-11-22 07:32:27 +00:00
|
|
|
|
per-stream and per-chapter metadata is copied along with streams/chapters. These
|
|
|
|
|
default mappings are disabled by creating any mapping of the relevant type. A negative
|
|
|
|
|
file index can be used to create a dummy mapping that just disables automatic copying.
|
|
|
|
|
|
2010-11-02 06:01:28 +00:00
|
|
|
|
For example to copy metadata from the first stream of the input file to global metadata
|
|
|
|
|
of the output file:
|
|
|
|
|
@example
|
2011-09-13 18:34:19 +00:00
|
|
|
|
ffmpeg -i in.ogg -map_metadata 0:s:0 out.mp3
|
2010-11-02 06:01:28 +00:00
|
|
|
|
@end example
|
2012-01-25 22:01:34 +00:00
|
|
|
|
|
|
|
|
|
To do the reverse, i.e. copy global metadata to all audio streams:
|
|
|
|
|
@example
|
|
|
|
|
ffmpeg -i in.mkv -map_metadata:s:a 0:g out.mkv
|
|
|
|
|
@end example
|
|
|
|
|
Note that simple @code{0} would work as well in this example, since global
|
|
|
|
|
metadata is assumed by default.
|
|
|
|
|
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@item -map_chapters @var{input_file_index} (@emph{output})
|
|
|
|
|
Copy chapters from input file with index @var{input_file_index} to the next
|
|
|
|
|
output file. If no chapter mapping is specified, then chapters are copied from
|
|
|
|
|
the first input file with at least one chapter. Use a negative file index to
|
|
|
|
|
disable any chapter copying.
|
2012-11-02 00:21:05 +00:00
|
|
|
|
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@item -benchmark (@emph{global})
|
2010-02-22 22:21:58 +00:00
|
|
|
|
Show benchmarking information at the end of an encode.
|
|
|
|
|
Shows CPU time used and maximum memory consumption.
|
|
|
|
|
Maximum memory consumption is not supported on all systems,
|
|
|
|
|
it will usually display as 0 if not supported.
|
2012-04-11 17:26:09 +00:00
|
|
|
|
@item -benchmark_all (@emph{global})
|
|
|
|
|
Show benchmarking information during the encode.
|
|
|
|
|
Shows CPU time used in various steps (audio/video encode/decode).
|
2011-09-28 00:57:34 +00:00
|
|
|
|
@item -timelimit @var{duration} (@emph{global})
|
|
|
|
|
Exit after ffmpeg has been running for @var{duration} seconds.
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@item -dump (@emph{global})
|
2011-09-28 00:57:34 +00:00
|
|
|
|
Dump each input packet to stderr.
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@item -hex (@emph{global})
|
2006-11-12 23:19:26 +00:00
|
|
|
|
When dumping packets, also dump the payload.
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@item -re (@emph{input})
|
2005-08-05 01:06:04 +00:00
|
|
|
|
Read input at native frame rate. Mainly used to simulate a grab device.
|
2012-07-11 19:22:43 +00:00
|
|
|
|
By default @command{ffmpeg} attempts to read the input(s) as fast as possible.
|
|
|
|
|
This option will slow down the reading of the input(s) to the native frame rate
|
|
|
|
|
of the input(s). It is useful for real-time output (e.g. live streaming). If
|
|
|
|
|
your input(s) is coming from some other live streaming source (through HTTP or
|
|
|
|
|
UDP for example) the server might already be in real-time, thus the option will
|
|
|
|
|
likely not be required. On the other hand, this is meaningful if your input(s)
|
|
|
|
|
is a file you are trying to push in real-time.
|
2006-11-12 23:19:26 +00:00
|
|
|
|
@item -loop_input
|
2005-08-05 01:06:04 +00:00
|
|
|
|
Loop over the input stream. Currently it works only for image
|
|
|
|
|
streams. This option is used for automatic FFserver testing.
|
2011-08-03 09:45:02 +00:00
|
|
|
|
This option is deprecated, use -loop 1.
|
2008-04-08 16:39:30 +00:00
|
|
|
|
@item -loop_output @var{number_of_times}
|
2005-07-26 16:41:34 +00:00
|
|
|
|
Repeatedly loop output for formats that support looping such as animated GIF
|
2005-08-05 01:06:04 +00:00
|
|
|
|
(0 will loop the output infinitely).
|
2011-06-18 07:49:24 +00:00
|
|
|
|
This option is deprecated, use -loop.
|
2008-04-08 16:39:30 +00:00
|
|
|
|
@item -vsync @var{parameter}
|
2010-03-31 00:41:31 +00:00
|
|
|
|
Video sync method.
|
2012-01-28 18:56:13 +00:00
|
|
|
|
For compatibility reasons old values can be specified as numbers.
|
|
|
|
|
Newly added values will have to be specified as strings always.
|
2011-01-16 12:45:42 +00:00
|
|
|
|
|
|
|
|
|
@table @option
|
2012-01-17 12:32:17 +00:00
|
|
|
|
@item 0, passthrough
|
2011-01-16 12:45:42 +00:00
|
|
|
|
Each frame is passed with its timestamp from the demuxer to the muxer.
|
2012-01-17 12:32:17 +00:00
|
|
|
|
@item 1, cfr
|
2011-01-16 12:45:42 +00:00
|
|
|
|
Frames will be duplicated and dropped to achieve exactly the requested
|
|
|
|
|
constant framerate.
|
2012-01-17 12:32:17 +00:00
|
|
|
|
@item 2, vfr
|
2011-01-16 12:45:42 +00:00
|
|
|
|
Frames are passed through with their timestamp or dropped so as to
|
|
|
|
|
prevent 2 frames from having the same timestamp.
|
2012-01-28 18:56:13 +00:00
|
|
|
|
@item drop
|
|
|
|
|
As passthrough but destroys all timestamps, making the muxer generate
|
|
|
|
|
fresh timestamps based on frame-rate.
|
2012-01-17 12:32:17 +00:00
|
|
|
|
@item -1, auto
|
2011-01-16 12:45:42 +00:00
|
|
|
|
Chooses between 1 and 2 depending on muxer capabilities. This is the
|
|
|
|
|
default method.
|
|
|
|
|
@end table
|
|
|
|
|
|
|
|
|
|
With -map you can select from which stream the timestamps should be
|
|
|
|
|
taken. You can leave either video or audio unchanged and sync the
|
|
|
|
|
remaining stream(s) to the unchanged one.
|
|
|
|
|
|
2008-04-08 16:39:30 +00:00
|
|
|
|
@item -async @var{samples_per_second}
|
2006-08-19 21:15:02 +00:00
|
|
|
|
Audio sync method. "Stretches/squeezes" the audio stream to match the timestamps,
|
2006-08-19 20:00:22 +00:00
|
|
|
|
the parameter is the maximum samples per second by which the audio is changed.
|
|
|
|
|
-async 1 is a special case where only the start of the audio stream is corrected
|
|
|
|
|
without any later correction.
|
2012-12-13 17:31:10 +00:00
|
|
|
|
This option has been deprecated. Use the @code{aresample} audio filter instead.
|
2012-11-21 23:21:37 +00:00
|
|
|
|
|
2007-06-21 20:21:36 +00:00
|
|
|
|
@item -copyts
|
2012-11-21 23:21:37 +00:00
|
|
|
|
Do not process input timestamps, but keep their values without trying
|
|
|
|
|
to sanitize them. In particular, do not remove the initial start time
|
|
|
|
|
offset value.
|
|
|
|
|
|
|
|
|
|
Note that, depending on the @option{vsync} option or on specific muxer
|
|
|
|
|
processing, the output timestamps may mismatch with the input
|
|
|
|
|
timestamps even when this option is selected.
|
|
|
|
|
|
2012-03-24 15:11:46 +00:00
|
|
|
|
@item -copytb @var{mode}
|
|
|
|
|
Specify how to set the encoder timebase when stream copying. @var{mode} is an
|
|
|
|
|
integer numeric value, and can assume one of the following values:
|
|
|
|
|
|
|
|
|
|
@table @option
|
|
|
|
|
@item 1
|
|
|
|
|
Use the demuxer timebase.
|
|
|
|
|
|
|
|
|
|
The time base is copied to the output encoder from the corresponding input
|
|
|
|
|
demuxer. This is sometimes required to avoid non monotonically increasing
|
|
|
|
|
timestamps when copying video streams with variable frame rate.
|
|
|
|
|
|
|
|
|
|
@item 0
|
|
|
|
|
Use the decoder timebase.
|
|
|
|
|
|
|
|
|
|
The time base is copied to the output encoder from the corresponding input
|
|
|
|
|
decoder.
|
|
|
|
|
|
|
|
|
|
@item -1
|
|
|
|
|
Try to make the choice automatically, in order to generate a sane output.
|
|
|
|
|
@end table
|
|
|
|
|
|
|
|
|
|
Default value is -1.
|
|
|
|
|
|
2012-08-11 09:50:32 +00:00
|
|
|
|
@item -shortest (@emph{output})
|
2007-06-21 20:21:36 +00:00
|
|
|
|
Finish encoding when the shortest input stream ends.
|
|
|
|
|
@item -dts_delta_threshold
|
|
|
|
|
Timestamp discontinuity delta threshold.
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@item -muxdelay @var{seconds} (@emph{input})
|
2007-06-21 20:21:36 +00:00
|
|
|
|
Set the maximum demux-decode delay.
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@item -muxpreload @var{seconds} (@emph{input})
|
2007-06-21 20:21:36 +00:00
|
|
|
|
Set the initial demux-decode delay.
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@item -streamid @var{output-stream-index}:@var{new-value} (@emph{output})
|
2011-01-15 16:11:55 +00:00
|
|
|
|
Assign a new stream-id value to an output stream. This option should be
|
|
|
|
|
specified prior to the output filename to which it applies.
|
|
|
|
|
For the situation where multiple output files exist, a streamid
|
|
|
|
|
may be reassigned to a different value.
|
2010-06-30 22:35:57 +00:00
|
|
|
|
|
|
|
|
|
For example, to set the stream 0 PID to 33 and the stream 1 PID to 36 for
|
|
|
|
|
an output mpegts file:
|
|
|
|
|
@example
|
|
|
|
|
ffmpeg -i infile -streamid 0:33 -streamid 1:36 out.ts
|
|
|
|
|
@end example
|
2011-09-13 18:34:19 +00:00
|
|
|
|
|
|
|
|
|
@item -bsf[:@var{stream_specifier}] @var{bitstream_filters} (@emph{output,per-stream})
|
2012-12-18 01:46:42 +00:00
|
|
|
|
Set bitstream filters for matching streams. @var{bitstream_filters} is
|
2011-09-13 18:34:19 +00:00
|
|
|
|
a comma-separated list of bitstream filters. Use the @code{-bsfs} option
|
|
|
|
|
to get the list of bitstream filters.
|
|
|
|
|
@example
|
2012-02-25 03:00:43 +00:00
|
|
|
|
ffmpeg -i h264.mp4 -c:v copy -bsf:v h264_mp4toannexb -an out.h264
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@end example
|
|
|
|
|
@example
|
2012-02-25 03:00:43 +00:00
|
|
|
|
ffmpeg -i file.mov -an -vn -bsf:s mov2textsub -c:s copy -f rawvideo sub.txt
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@end example
|
|
|
|
|
|
2011-10-16 17:05:11 +00:00
|
|
|
|
@item -tag[:@var{stream_specifier}] @var{codec_tag} (@emph{per-stream})
|
2011-09-13 18:34:19 +00:00
|
|
|
|
Force a tag/fourcc for matching streams.
|
2012-01-03 09:47:42 +00:00
|
|
|
|
|
|
|
|
|
@item -timecode @var{hh}:@var{mm}:@var{ss}SEP@var{ff}
|
|
|
|
|
Specify Timecode for writing. @var{SEP} is ':' for non drop timecode and ';'
|
|
|
|
|
(or '.') for drop.
|
|
|
|
|
@example
|
|
|
|
|
ffmpeg -i input.mpg -timecode 01:02:03.04 -r 30000/1001 -s ntsc output.mpg
|
|
|
|
|
@end example
|
2012-04-01 13:08:33 +00:00
|
|
|
|
|
|
|
|
|
@item -filter_complex @var{filtergraph} (@emph{global})
|
|
|
|
|
Define a complex filter graph, i.e. one with arbitrary number of inputs and/or
|
|
|
|
|
outputs. For simple graphs -- those with one input and one output of the same
|
|
|
|
|
type -- see the @option{-filter} options. @var{filtergraph} is a description of
|
2012-12-16 12:05:08 +00:00
|
|
|
|
the filter graph, as described in the ``Filtergraph syntax'' section of the
|
|
|
|
|
ffmpeg-filters manual.
|
2012-04-01 13:08:33 +00:00
|
|
|
|
|
|
|
|
|
Input link labels must refer to input streams using the
|
|
|
|
|
@code{[file_index:stream_specifier]} syntax (i.e. the same as @option{-map}
|
|
|
|
|
uses). If @var{stream_specifier} matches multiple streams, the first one will be
|
|
|
|
|
used. An unlabeled input will be connected to the first unused input stream of
|
|
|
|
|
the matching type.
|
|
|
|
|
|
|
|
|
|
Output link labels are referred to with @option{-map}. Unlabeled outputs are
|
|
|
|
|
added to the first output file.
|
|
|
|
|
|
2012-08-05 06:37:43 +00:00
|
|
|
|
Note that with this option it is possible to use only lavfi sources without
|
|
|
|
|
normal input files.
|
|
|
|
|
|
2012-04-01 13:08:33 +00:00
|
|
|
|
For example, to overlay an image over video
|
|
|
|
|
@example
|
2012-04-17 02:01:17 +00:00
|
|
|
|
ffmpeg -i video.mkv -i image.png -filter_complex '[0:v][1:v]overlay[out]' -map
|
2012-04-01 13:08:33 +00:00
|
|
|
|
'[out]' out.mkv
|
|
|
|
|
@end example
|
|
|
|
|
Here @code{[0:v]} refers to the first video stream in the first input file,
|
|
|
|
|
which is linked to the first (main) input of the overlay filter. Similarly the
|
|
|
|
|
first video stream in the second input is linked to the second (overlay) input
|
|
|
|
|
of overlay.
|
|
|
|
|
|
|
|
|
|
Assuming there is only one video stream in each input file, we can omit input
|
|
|
|
|
labels, so the above is equivalent to
|
|
|
|
|
@example
|
2012-04-17 02:01:17 +00:00
|
|
|
|
ffmpeg -i video.mkv -i image.png -filter_complex 'overlay[out]' -map
|
2012-04-01 13:08:33 +00:00
|
|
|
|
'[out]' out.mkv
|
|
|
|
|
@end example
|
|
|
|
|
|
|
|
|
|
Furthermore we can omit the output label and the single output from the filter
|
|
|
|
|
graph will be added to the output file automatically, so we can simply write
|
|
|
|
|
@example
|
2012-04-17 02:01:17 +00:00
|
|
|
|
ffmpeg -i video.mkv -i image.png -filter_complex 'overlay' out.mkv
|
2012-04-01 13:08:33 +00:00
|
|
|
|
@end example
|
2012-08-05 06:37:43 +00:00
|
|
|
|
|
|
|
|
|
To generate 5 seconds of pure red video using lavfi @code{color} source:
|
|
|
|
|
@example
|
2012-08-09 17:09:39 +00:00
|
|
|
|
ffmpeg -filter_complex 'color=red' -t 5 out.mkv
|
2012-08-05 06:37:43 +00:00
|
|
|
|
@end example
|
2002-10-27 22:00:34 +00:00
|
|
|
|
@end table
|
2003-08-24 15:22:26 +00:00
|
|
|
|
|
2012-07-26 17:29:27 +00:00
|
|
|
|
As a special exception, you can use a bitmap subtitle stream as input: it
|
|
|
|
|
will be converted into a video with the same size as the largest video in
|
|
|
|
|
the file, or 720×576 if no video is present. Note that this is an
|
|
|
|
|
experimental and temporary solution. It will be removed once libavfilter has
|
|
|
|
|
proper support for subtitles.
|
|
|
|
|
|
|
|
|
|
For example, to hardcode subtitles on top of a DVB-T recording stored in
|
|
|
|
|
MPEG-TS format, delaying the subtitles by 1 second:
|
|
|
|
|
@example
|
|
|
|
|
ffmpeg -i input.ts -filter_complex \
|
|
|
|
|
'[#0x2ef] setpts=PTS+1/TB [sub] ; [#0x2d0] [sub] overlay' \
|
|
|
|
|
-sn -map '#0x2dc' output.mkv
|
|
|
|
|
@end example
|
|
|
|
|
(0x2d0, 0x2dc and 0x2ef are the MPEG-TS PIDs of respectively the video,
|
|
|
|
|
audio and subtitles streams; 0:0, 0:3 and 0:7 would have worked too)
|
|
|
|
|
|
2008-10-05 09:45:54 +00:00
|
|
|
|
@section Preset files
|
|
|
|
|
A preset file contains a sequence of @var{option}=@var{value} pairs,
|
|
|
|
|
one for each line, specifying a sequence of options which would be
|
|
|
|
|
awkward to specify on the command line. Lines starting with the hash
|
|
|
|
|
('#') character are ignored and are used to provide comments. Check
|
2011-10-20 00:23:36 +00:00
|
|
|
|
the @file{presets} directory in the FFmpeg source tree for examples.
|
2008-10-05 09:45:54 +00:00
|
|
|
|
|
2009-12-06 17:02:12 +00:00
|
|
|
|
Preset files are specified with the @code{vpre}, @code{apre},
|
|
|
|
|
@code{spre}, and @code{fpre} options. The @code{fpre} option takes the
|
|
|
|
|
filename of the preset instead of a preset name as input and can be
|
|
|
|
|
used for any kind of codec. For the @code{vpre}, @code{apre}, and
|
|
|
|
|
@code{spre} options, the options specified in a preset file are
|
2008-10-05 22:41:43 +00:00
|
|
|
|
applied to the currently selected codec of the same type as the preset
|
|
|
|
|
option.
|
2008-10-05 09:45:54 +00:00
|
|
|
|
|
2009-12-06 17:02:12 +00:00
|
|
|
|
The argument passed to the @code{vpre}, @code{apre}, and @code{spre}
|
|
|
|
|
preset options identifies the preset file to use according to the
|
|
|
|
|
following rules:
|
2008-10-05 09:45:54 +00:00
|
|
|
|
|
|
|
|
|
First ffmpeg searches for a file named @var{arg}.ffpreset in the
|
2010-04-20 17:37:06 +00:00
|
|
|
|
directories @file{$FFMPEG_DATADIR} (if set), and @file{$HOME/.ffmpeg}, and in
|
|
|
|
|
the datadir defined at configuration time (usually @file{PREFIX/share/ffmpeg})
|
2011-05-21 01:36:26 +00:00
|
|
|
|
or in a @file{ffpresets} folder along the executable on win32,
|
2012-06-19 18:18:30 +00:00
|
|
|
|
in that order. For example, if the argument is @code{libvpx-1080p}, it will
|
|
|
|
|
search for the file @file{libvpx-1080p.ffpreset}.
|
2008-10-05 09:45:54 +00:00
|
|
|
|
|
|
|
|
|
If no such file is found, then ffmpeg will search for a file named
|
|
|
|
|
@var{codec_name}-@var{arg}.ffpreset in the above-mentioned
|
|
|
|
|
directories, where @var{codec_name} is the name of the codec to which
|
|
|
|
|
the preset file options will be applied. For example, if you select
|
2012-06-19 18:18:30 +00:00
|
|
|
|
the video codec with @code{-vcodec libvpx} and use @code{-vpre 1080p},
|
|
|
|
|
then it will search for the file @file{libvpx-1080p.ffpreset}.
|
2011-09-13 18:34:19 +00:00
|
|
|
|
@c man end OPTIONS
|
2003-08-24 11:16:29 +00:00
|
|
|
|
|
2002-10-27 22:00:34 +00:00
|
|
|
|
@chapter Tips
|
2010-06-01 14:00:44 +00:00
|
|
|
|
@c man begin TIPS
|
2002-10-27 22:00:34 +00:00
|
|
|
|
|
|
|
|
|
@itemize
|
2010-06-01 14:00:44 +00:00
|
|
|
|
@item
|
|
|
|
|
For streaming at very low bitrate application, use a low frame rate
|
2005-08-05 01:06:04 +00:00
|
|
|
|
and a small GOP size. This is especially true for RealVideo where
|
2003-08-24 11:16:29 +00:00
|
|
|
|
the Linux player does not seem to be very fast, so it can miss
|
|
|
|
|
frames. An example is:
|
2002-10-27 22:00:34 +00:00
|
|
|
|
|
|
|
|
|
@example
|
2011-11-10 21:11:33 +00:00
|
|
|
|
ffmpeg -g 3 -r 3 -t 10 -b:v 50k -s qcif -f rv10 /tmp/b.rm
|
2002-10-27 22:00:34 +00:00
|
|
|
|
@end example
|
|
|
|
|
|
2010-06-01 14:00:44 +00:00
|
|
|
|
@item
|
|
|
|
|
The parameter 'q' which is displayed while encoding is the current
|
2005-08-05 01:06:04 +00:00
|
|
|
|
quantizer. The value 1 indicates that a very good quality could
|
|
|
|
|
be achieved. The value 31 indicates the worst quality. If q=31 appears
|
2003-08-24 11:16:29 +00:00
|
|
|
|
too often, it means that the encoder cannot compress enough to meet
|
2005-08-05 01:06:04 +00:00
|
|
|
|
your bitrate. You must either increase the bitrate, decrease the
|
2003-08-24 11:16:29 +00:00
|
|
|
|
frame rate or decrease the frame size.
|
2002-10-27 22:00:34 +00:00
|
|
|
|
|
2010-06-01 14:00:44 +00:00
|
|
|
|
@item
|
|
|
|
|
If your computer is not fast enough, you can speed up the
|
2003-08-24 11:16:29 +00:00
|
|
|
|
compression at the expense of the compression ratio. You can use
|
2012-05-07 18:57:11 +00:00
|
|
|
|
'-me zero' to speed up motion estimation, and '-g 0' to disable
|
2005-08-05 01:06:04 +00:00
|
|
|
|
motion estimation completely (you have only I-frames, which means it
|
2003-08-24 11:16:29 +00:00
|
|
|
|
is about as good as JPEG compression).
|
2002-10-27 22:00:34 +00:00
|
|
|
|
|
2010-06-01 14:00:44 +00:00
|
|
|
|
@item
|
|
|
|
|
To have very low audio bitrates, reduce the sampling frequency
|
2008-10-13 15:09:04 +00:00
|
|
|
|
(down to 22050 Hz for MPEG audio, 22050 or 11025 for AC-3).
|
2002-10-27 22:00:34 +00:00
|
|
|
|
|
2010-06-01 14:00:44 +00:00
|
|
|
|
@item
|
|
|
|
|
To have a constant quality (but a variable bitrate), use the option
|
2003-08-24 11:16:29 +00:00
|
|
|
|
'-qscale n' when 'n' is between 1 (excellent quality) and 31 (worst
|
|
|
|
|
quality).
|
2002-10-27 22:00:34 +00:00
|
|
|
|
|
|
|
|
|
@end itemize
|
2010-06-01 14:00:44 +00:00
|
|
|
|
@c man end TIPS
|
2002-10-27 22:00:34 +00:00
|
|
|
|
|
2010-06-03 20:11:36 +00:00
|
|
|
|
@chapter Examples
|
|
|
|
|
@c man begin EXAMPLES
|
|
|
|
|
|
2011-12-11 17:50:39 +00:00
|
|
|
|
@section Preset files
|
|
|
|
|
|
|
|
|
|
A preset file contains a sequence of @var{option=value} pairs, one for
|
|
|
|
|
each line, specifying a sequence of options which can be specified also on
|
|
|
|
|
the command line. Lines starting with the hash ('#') character are ignored and
|
|
|
|
|
are used to provide comments. Empty lines are also ignored. Check the
|
|
|
|
|
@file{presets} directory in the FFmpeg source tree for examples.
|
|
|
|
|
|
|
|
|
|
Preset files are specified with the @code{pre} option, this option takes a
|
|
|
|
|
preset name as input. FFmpeg searches for a file named @var{preset_name}.avpreset in
|
|
|
|
|
the directories @file{$AVCONV_DATADIR} (if set), and @file{$HOME/.ffmpeg}, and in
|
|
|
|
|
the data directory defined at configuration time (usually @file{$PREFIX/share/ffmpeg})
|
|
|
|
|
in that order. For example, if the argument is @code{libx264-max}, it will
|
|
|
|
|
search for the file @file{libx264-max.avpreset}.
|
|
|
|
|
|
2010-06-03 20:11:36 +00:00
|
|
|
|
@section Video and Audio grabbing
|
|
|
|
|
|
2011-04-23 14:40:01 +00:00
|
|
|
|
If you specify the input format and device then ffmpeg can grab video
|
|
|
|
|
and audio directly.
|
2010-06-03 20:11:36 +00:00
|
|
|
|
|
|
|
|
|
@example
|
|
|
|
|
ffmpeg -f oss -i /dev/dsp -f video4linux2 -i /dev/video0 /tmp/out.mpg
|
2011-10-30 01:07:38 +00:00
|
|
|
|
@end example
|
|
|
|
|
|
|
|
|
|
Or with an ALSA audio source (mono input, card id 1) instead of OSS:
|
|
|
|
|
@example
|
|
|
|
|
ffmpeg -f alsa -ac 1 -i hw:1 -f video4linux2 -i /dev/video0 /tmp/out.mpg
|
2010-06-03 20:11:36 +00:00
|
|
|
|
@end example
|
|
|
|
|
|
|
|
|
|
Note that you must activate the right video source and channel before
|
2011-07-08 13:33:17 +00:00
|
|
|
|
launching ffmpeg with any TV viewer such as
|
|
|
|
|
@uref{http://linux.bytesex.org/xawtv/, xawtv} by Gerd Knorr. You also
|
2010-06-03 20:11:36 +00:00
|
|
|
|
have to set the audio recording levels correctly with a
|
|
|
|
|
standard mixer.
|
|
|
|
|
|
|
|
|
|
@section X11 grabbing
|
|
|
|
|
|
2011-04-23 14:40:01 +00:00
|
|
|
|
Grab the X11 display with ffmpeg via
|
2010-06-03 20:11:36 +00:00
|
|
|
|
|
|
|
|
|
@example
|
2010-09-13 13:25:51 +00:00
|
|
|
|
ffmpeg -f x11grab -s cif -r 25 -i :0.0 /tmp/out.mpg
|
2010-06-03 20:11:36 +00:00
|
|
|
|
@end example
|
|
|
|
|
|
|
|
|
|
0.0 is display.screen number of your X11 server, same as
|
|
|
|
|
the DISPLAY environment variable.
|
|
|
|
|
|
|
|
|
|
@example
|
2010-09-13 13:25:51 +00:00
|
|
|
|
ffmpeg -f x11grab -s cif -r 25 -i :0.0+10,20 /tmp/out.mpg
|
2010-06-03 20:11:36 +00:00
|
|
|
|
@end example
|
|
|
|
|
|
2011-09-13 18:34:19 +00:00
|
|
|
|
0.0 is display.screen number of your X11 server, same as the DISPLAY environment
|
|
|
|
|
variable. 10 is the x-offset and 20 the y-offset for the grabbing.
|
2010-06-03 20:11:36 +00:00
|
|
|
|
|
|
|
|
|
@section Video and Audio file format conversion
|
|
|
|
|
|
2011-04-23 14:40:01 +00:00
|
|
|
|
Any supported file format and protocol can serve as input to ffmpeg:
|
2010-06-03 20:11:36 +00:00
|
|
|
|
|
|
|
|
|
Examples:
|
2011-01-16 13:04:56 +00:00
|
|
|
|
@itemize
|
|
|
|
|
@item
|
|
|
|
|
You can use YUV files as input:
|
2010-06-03 20:11:36 +00:00
|
|
|
|
|
|
|
|
|
@example
|
|
|
|
|
ffmpeg -i /tmp/test%d.Y /tmp/out.mpg
|
|
|
|
|
@end example
|
|
|
|
|
|
|
|
|
|
It will use the files:
|
|
|
|
|
@example
|
|
|
|
|
/tmp/test0.Y, /tmp/test0.U, /tmp/test0.V,
|
|
|
|
|
/tmp/test1.Y, /tmp/test1.U, /tmp/test1.V, etc...
|
|
|
|
|
@end example
|
|
|
|
|
|
|
|
|
|
The Y files use twice the resolution of the U and V files. They are
|
|
|
|
|
raw files, without header. They can be generated by all decent video
|
|
|
|
|
decoders. You must specify the size of the image with the @option{-s} option
|
2011-04-23 14:40:01 +00:00
|
|
|
|
if ffmpeg cannot guess it.
|
2010-06-03 20:11:36 +00:00
|
|
|
|
|
2011-01-16 13:04:56 +00:00
|
|
|
|
@item
|
|
|
|
|
You can input from a raw YUV420P file:
|
2010-06-03 20:11:36 +00:00
|
|
|
|
|
|
|
|
|
@example
|
|
|
|
|
ffmpeg -i /tmp/test.yuv /tmp/out.avi
|
|
|
|
|
@end example
|
|
|
|
|
|
|
|
|
|
test.yuv is a file containing raw YUV planar data. Each frame is composed
|
|
|
|
|
of the Y plane followed by the U and V planes at half vertical and
|
|
|
|
|
horizontal resolution.
|
|
|
|
|
|
2011-01-16 13:04:56 +00:00
|
|
|
|
@item
|
|
|
|
|
You can output to a raw YUV420P file:
|
2010-06-03 20:11:36 +00:00
|
|
|
|
|
|
|
|
|
@example
|
|
|
|
|
ffmpeg -i mydivx.avi hugefile.yuv
|
|
|
|
|
@end example
|
|
|
|
|
|
2011-01-16 13:04:56 +00:00
|
|
|
|
@item
|
|
|
|
|
You can set several input files and output files:
|
2010-06-03 20:11:36 +00:00
|
|
|
|
|
|
|
|
|
@example
|
|
|
|
|
ffmpeg -i /tmp/a.wav -s 640x480 -i /tmp/a.yuv /tmp/a.mpg
|
|
|
|
|
@end example
|
|
|
|
|
|
|
|
|
|
Converts the audio file a.wav and the raw YUV video file a.yuv
|
|
|
|
|
to MPEG file a.mpg.
|
|
|
|
|
|
2011-01-16 13:04:56 +00:00
|
|
|
|
@item
|
|
|
|
|
You can also do audio and video conversions at the same time:
|
2010-06-03 20:11:36 +00:00
|
|
|
|
|
|
|
|
|
@example
|
|
|
|
|
ffmpeg -i /tmp/a.wav -ar 22050 /tmp/a.mp2
|
|
|
|
|
@end example
|
|
|
|
|
|
|
|
|
|
Converts a.wav to MPEG audio at 22050 Hz sample rate.
|
|
|
|
|
|
2011-01-16 13:04:56 +00:00
|
|
|
|
@item
|
|
|
|
|
You can encode to several formats at the same time and define a
|
2010-06-03 20:11:36 +00:00
|
|
|
|
mapping from input stream to output streams:
|
|
|
|
|
|
|
|
|
|
@example
|
2011-11-10 21:11:33 +00:00
|
|
|
|
ffmpeg -i /tmp/a.wav -map 0:a -b:a 64k /tmp/a.mp2 -map 0:a -b:a 128k /tmp/b.mp2
|
2010-06-03 20:11:36 +00:00
|
|
|
|
@end example
|
|
|
|
|
|
|
|
|
|
Converts a.wav to a.mp2 at 64 kbits and to b.mp2 at 128 kbits. '-map
|
|
|
|
|
file:index' specifies which input stream is used for each output
|
|
|
|
|
stream, in the order of the definition of output streams.
|
|
|
|
|
|
2011-01-16 13:04:56 +00:00
|
|
|
|
@item
|
|
|
|
|
You can transcode decrypted VOBs:
|
2010-06-03 20:11:36 +00:00
|
|
|
|
|
|
|
|
|
@example
|
2011-09-13 18:34:19 +00:00
|
|
|
|
ffmpeg -i snatch_1.vob -f avi -c:v mpeg4 -b:v 800k -g 300 -bf 2 -c:a libmp3lame -b:a 128k snatch.avi
|
2010-06-03 20:11:36 +00:00
|
|
|
|
@end example
|
|
|
|
|
|
|
|
|
|
This is a typical DVD ripping example; the input is a VOB file, the
|
|
|
|
|
output an AVI file with MPEG-4 video and MP3 audio. Note that in this
|
|
|
|
|
command we use B-frames so the MPEG-4 stream is DivX5 compatible, and
|
|
|
|
|
GOP size is 300 which means one intra frame every 10 seconds for 29.97fps
|
|
|
|
|
input video. Furthermore, the audio stream is MP3-encoded so you need
|
|
|
|
|
to enable LAME support by passing @code{--enable-libmp3lame} to configure.
|
|
|
|
|
The mapping is particularly useful for DVD transcoding
|
|
|
|
|
to get the desired audio language.
|
|
|
|
|
|
|
|
|
|
NOTE: To see the supported input formats, use @code{ffmpeg -formats}.
|
|
|
|
|
|
2011-01-16 13:04:56 +00:00
|
|
|
|
@item
|
|
|
|
|
You can extract images from a video, or create a video from many images:
|
2010-06-03 20:11:36 +00:00
|
|
|
|
|
|
|
|
|
For extracting images from a video:
|
|
|
|
|
@example
|
|
|
|
|
ffmpeg -i foo.avi -r 1 -s WxH -f image2 foo-%03d.jpeg
|
|
|
|
|
@end example
|
|
|
|
|
|
|
|
|
|
This will extract one video frame per second from the video and will
|
|
|
|
|
output them in files named @file{foo-001.jpeg}, @file{foo-002.jpeg},
|
|
|
|
|
etc. Images will be rescaled to fit the new WxH values.
|
|
|
|
|
|
|
|
|
|
If you want to extract just a limited number of frames, you can use the
|
|
|
|
|
above command in combination with the -vframes or -t option, or in
|
|
|
|
|
combination with -ss to start extracting from a certain point in time.
|
|
|
|
|
|
|
|
|
|
For creating a video from many images:
|
|
|
|
|
@example
|
|
|
|
|
ffmpeg -f image2 -i foo-%03d.jpeg -r 12 -s WxH foo.avi
|
|
|
|
|
@end example
|
|
|
|
|
|
|
|
|
|
The syntax @code{foo-%03d.jpeg} specifies to use a decimal number
|
|
|
|
|
composed of three digits padded with zeroes to express the sequence
|
|
|
|
|
number. It is the same syntax supported by the C printf function, but
|
2012-04-05 22:21:43 +00:00
|
|
|
|
only formats accepting a normal integer are suitable.
|
|
|
|
|
|
2012-08-07 19:11:51 +00:00
|
|
|
|
When importing an image sequence, -i also supports expanding
|
|
|
|
|
shell-like wildcard patterns (globbing) internally, by selecting the
|
|
|
|
|
image2-specific @code{-pattern_type glob} option.
|
|
|
|
|
|
|
|
|
|
For example, for creating a video from filenames matching the glob pattern
|
|
|
|
|
@code{foo-*.jpeg}:
|
|
|
|
|
@example
|
|
|
|
|
ffmpeg -f image2 -pattern_type glob -i 'foo-*.jpeg' -r 12 -s WxH foo.avi
|
|
|
|
|
@end example
|
2010-06-03 20:11:36 +00:00
|
|
|
|
|
2011-01-16 13:04:56 +00:00
|
|
|
|
@item
|
|
|
|
|
You can put many streams of the same type in the output:
|
2010-06-03 20:11:36 +00:00
|
|
|
|
|
|
|
|
|
@example
|
2012-12-11 02:57:33 +00:00
|
|
|
|
ffmpeg -i test1.avi -i test2.avi -map 0:3 -map 0:2 -map 0:1 -map 0:0 -c copy test12.nut
|
2010-06-03 20:11:36 +00:00
|
|
|
|
@end example
|
|
|
|
|
|
2011-09-13 18:34:19 +00:00
|
|
|
|
The resulting output file @file{test12.avi} will contain first four streams from
|
|
|
|
|
the input file in reverse order.
|
2011-01-16 13:04:56 +00:00
|
|
|
|
|
2012-03-15 07:17:05 +00:00
|
|
|
|
@item
|
|
|
|
|
To force CBR video output:
|
|
|
|
|
@example
|
2012-03-20 23:15:18 +00:00
|
|
|
|
ffmpeg -i myfile.avi -b 4000k -minrate 4000k -maxrate 4000k -bufsize 1835k out.m2v
|
2012-03-15 07:17:05 +00:00
|
|
|
|
@end example
|
|
|
|
|
|
|
|
|
|
@item
|
|
|
|
|
The four options lmin, lmax, mblmin and mblmax use 'lambda' units,
|
|
|
|
|
but you may use the QP2LAMBDA constant to easily convert from 'q' units:
|
|
|
|
|
@example
|
2012-03-20 23:15:18 +00:00
|
|
|
|
ffmpeg -i src.ext -lmax 21*QP2LAMBDA dst.ext
|
2012-03-15 07:17:05 +00:00
|
|
|
|
@end example
|
|
|
|
|
|
2011-01-16 13:04:56 +00:00
|
|
|
|
@end itemize
|
2010-06-03 20:11:36 +00:00
|
|
|
|
@c man end EXAMPLES
|
|
|
|
|
|
2012-12-16 19:53:16 +00:00
|
|
|
|
@chapter See Also
|
|
|
|
|
|
|
|
|
|
@ifhtml
|
|
|
|
|
@url{ffplay.html,ffplay}, @url{ffprobe.html,ffprobe}, @url{ffserver.html,ffserver},
|
|
|
|
|
@url{ffmpeg-utils.html,ffmpeg-utils},
|
|
|
|
|
@url{ffmpeg-scaler.html,ffmpeg-scaler},
|
|
|
|
|
@url{ffmpeg-resampler.html,ffmpeg-resampler},
|
|
|
|
|
@url{ffmpeg-codecs.html,ffmpeg-codecs},
|
|
|
|
|
@url{ffmpeg-bitstream-filters,ffmpeg-bitstream-filters},
|
|
|
|
|
@url{ffmpeg-formats.html,ffmpeg-formats},
|
|
|
|
|
@url{ffmpeg-devices.html,ffmpeg-devices},
|
|
|
|
|
@url{ffmpeg-protocols.html,ffmpeg-protocols},
|
|
|
|
|
@url{ffmpeg-filters.html,ffmpeg-filters}
|
|
|
|
|
@end ifhtml
|
|
|
|
|
|
|
|
|
|
@ifnothtml
|
|
|
|
|
ffplay(1), ffprobe(1), ffserver(1),
|
|
|
|
|
ffmpeg-utils(1), ffmpeg-scaler(1), ffmpeg-resampler(1),
|
|
|
|
|
ffmpeg-codecs(1), ffmpeg-bitstream-filters(1), ffmpeg-formats(1),
|
|
|
|
|
ffmpeg-devices(1), ffmpeg-protocols(1), ffmpeg-filters(1)
|
|
|
|
|
@end ifnothtml
|
|
|
|
|
|
2012-12-16 17:04:37 +00:00
|
|
|
|
@include authors.texi
|
|
|
|
|
|
2010-06-03 19:07:13 +00:00
|
|
|
|
@ignore
|
|
|
|
|
|
|
|
|
|
@setfilename ffmpeg
|
2011-04-23 13:19:17 +00:00
|
|
|
|
@settitle ffmpeg video converter
|
2010-06-03 19:07:13 +00:00
|
|
|
|
|
|
|
|
|
@end ignore
|
|
|
|
|
|
2002-10-27 22:00:34 +00:00
|
|
|
|
@bye
|