mirror of https://git.ffmpeg.org/ffmpeg.git
512 lines
18 KiB
Plaintext
512 lines
18 KiB
Plaintext
@chapter Muxers
|
|
@c man begin MUXERS
|
|
|
|
Muxers are configured elements in FFmpeg which allow writing
|
|
multimedia streams to a particular type of file.
|
|
|
|
When you configure your FFmpeg build, all the supported muxers
|
|
are enabled by default. You can list all available muxers using the
|
|
configure option @code{--list-muxers}.
|
|
|
|
You can disable all the muxers with the configure option
|
|
@code{--disable-muxers} and selectively enable / disable single muxers
|
|
with the options @code{--enable-muxer=@var{MUXER}} /
|
|
@code{--disable-muxer=@var{MUXER}}.
|
|
|
|
The option @code{-formats} of the ff* tools will display the list of
|
|
enabled muxers.
|
|
|
|
A description of some of the currently available muxers follows.
|
|
|
|
@anchor{crc}
|
|
@section crc
|
|
|
|
CRC (Cyclic Redundancy Check) testing format.
|
|
|
|
This muxer computes and prints the Adler-32 CRC of all the input audio
|
|
and video frames. By default audio frames are converted to signed
|
|
16-bit raw audio and video frames to raw video before computing the
|
|
CRC.
|
|
|
|
The output of the muxer consists of a single line of the form:
|
|
CRC=0x@var{CRC}, where @var{CRC} is a hexadecimal number 0-padded to
|
|
8 digits containing the CRC for all the decoded input frames.
|
|
|
|
For example to compute the CRC of the input, and store it in the file
|
|
@file{out.crc}:
|
|
@example
|
|
ffmpeg -i INPUT -f crc out.crc
|
|
@end example
|
|
|
|
You can print the CRC to stdout with the command:
|
|
@example
|
|
ffmpeg -i INPUT -f crc -
|
|
@end example
|
|
|
|
You can select the output format of each frame with @command{ffmpeg} by
|
|
specifying the audio and video codec and format. For example to
|
|
compute the CRC of the input audio converted to PCM unsigned 8-bit
|
|
and the input video converted to MPEG-2 video, use the command:
|
|
@example
|
|
ffmpeg -i INPUT -c:a pcm_u8 -c:v mpeg2video -f crc -
|
|
@end example
|
|
|
|
See also the @ref{framecrc} muxer.
|
|
|
|
@anchor{framecrc}
|
|
@section framecrc
|
|
|
|
Per-packet CRC (Cyclic Redundancy Check) testing format.
|
|
|
|
This muxer computes and prints the Adler-32 CRC for each audio
|
|
and video packet. By default audio frames are converted to signed
|
|
16-bit raw audio and video frames to raw video before computing the
|
|
CRC.
|
|
|
|
The output of the muxer consists of a line for each audio and video
|
|
packet of the form:
|
|
@example
|
|
@var{stream_index}, @var{packet_dts}, @var{packet_pts}, @var{packet_duration}, @var{packet_size}, 0x@var{CRC}
|
|
@end example
|
|
|
|
@var{CRC} is a hexadecimal number 0-padded to 8 digits containing the
|
|
CRC of the packet.
|
|
|
|
For example to compute the CRC of the audio and video frames in
|
|
@file{INPUT}, converted to raw audio and video packets, and store it
|
|
in the file @file{out.crc}:
|
|
@example
|
|
ffmpeg -i INPUT -f framecrc out.crc
|
|
@end example
|
|
|
|
To print the information to stdout, use the command:
|
|
@example
|
|
ffmpeg -i INPUT -f framecrc -
|
|
@end example
|
|
|
|
With @command{ffmpeg}, you can select the output format to which the
|
|
audio and video frames are encoded before computing the CRC for each
|
|
packet by specifying the audio and video codec. For example, to
|
|
compute the CRC of each decoded input audio frame converted to PCM
|
|
unsigned 8-bit and of each decoded input video frame converted to
|
|
MPEG-2 video, use the command:
|
|
@example
|
|
ffmpeg -i INPUT -c:a pcm_u8 -c:v mpeg2video -f framecrc -
|
|
@end example
|
|
|
|
See also the @ref{crc} muxer.
|
|
|
|
@anchor{framemd5}
|
|
@section framemd5
|
|
|
|
Per-packet MD5 testing format.
|
|
|
|
This muxer computes and prints the MD5 hash for each audio
|
|
and video packet. By default audio frames are converted to signed
|
|
16-bit raw audio and video frames to raw video before computing the
|
|
hash.
|
|
|
|
The output of the muxer consists of a line for each audio and video
|
|
packet of the form:
|
|
@example
|
|
@var{stream_index}, @var{packet_dts}, @var{packet_pts}, @var{packet_duration}, @var{packet_size}, @var{MD5}
|
|
@end example
|
|
|
|
@var{MD5} is a hexadecimal number representing the computed MD5 hash
|
|
for the packet.
|
|
|
|
For example to compute the MD5 of the audio and video frames in
|
|
@file{INPUT}, converted to raw audio and video packets, and store it
|
|
in the file @file{out.md5}:
|
|
@example
|
|
ffmpeg -i INPUT -f framemd5 out.md5
|
|
@end example
|
|
|
|
To print the information to stdout, use the command:
|
|
@example
|
|
ffmpeg -i INPUT -f framemd5 -
|
|
@end example
|
|
|
|
See also the @ref{md5} muxer.
|
|
|
|
@anchor{image2}
|
|
@section image2
|
|
|
|
Image file muxer.
|
|
|
|
The image file muxer writes video frames to image files.
|
|
|
|
The output filenames are specified by a pattern, which can be used to
|
|
produce sequentially numbered series of files.
|
|
The pattern may contain the string "%d" or "%0@var{N}d", this string
|
|
specifies the position of the characters representing a numbering in
|
|
the filenames. If the form "%0@var{N}d" is used, the string
|
|
representing the number in each filename is 0-padded to @var{N}
|
|
digits. The literal character '%' can be specified in the pattern with
|
|
the string "%%".
|
|
|
|
If the pattern contains "%d" or "%0@var{N}d", the first filename of
|
|
the file list specified will contain the number 1, all the following
|
|
numbers will be sequential.
|
|
|
|
The pattern may contain a suffix which is used to automatically
|
|
determine the format of the image files to write.
|
|
|
|
For example the pattern "img-%03d.bmp" will specify a sequence of
|
|
filenames of the form @file{img-001.bmp}, @file{img-002.bmp}, ...,
|
|
@file{img-010.bmp}, etc.
|
|
The pattern "img%%-%d.jpg" will specify a sequence of filenames of the
|
|
form @file{img%-1.jpg}, @file{img%-2.jpg}, ..., @file{img%-10.jpg},
|
|
etc.
|
|
|
|
The following example shows how to use @command{ffmpeg} for creating a
|
|
sequence of files @file{img-001.jpeg}, @file{img-002.jpeg}, ...,
|
|
taking one image every second from the input video:
|
|
@example
|
|
ffmpeg -i in.avi -vsync 1 -r 1 -f image2 'img-%03d.jpeg'
|
|
@end example
|
|
|
|
Note that with @command{ffmpeg}, if the format is not specified with the
|
|
@code{-f} option and the output filename specifies an image file
|
|
format, the image2 muxer is automatically selected, so the previous
|
|
command can be written as:
|
|
@example
|
|
ffmpeg -i in.avi -vsync 1 -r 1 'img-%03d.jpeg'
|
|
@end example
|
|
|
|
Note also that the pattern must not necessarily contain "%d" or
|
|
"%0@var{N}d", for example to create a single image file
|
|
@file{img.jpeg} from the input video you can employ the command:
|
|
@example
|
|
ffmpeg -i in.avi -f image2 -frames:v 1 img.jpeg
|
|
@end example
|
|
|
|
The image muxer supports the .Y.U.V image file format. This format is
|
|
special in that that each image frame consists of three files, for
|
|
each of the YUV420P components. To read or write this image file format,
|
|
specify the name of the '.Y' file. The muxer will automatically open the
|
|
'.U' and '.V' files as required.
|
|
|
|
@anchor{md5}
|
|
@section md5
|
|
|
|
MD5 testing format.
|
|
|
|
This muxer computes and prints the MD5 hash of all the input audio
|
|
and video frames. By default audio frames are converted to signed
|
|
16-bit raw audio and video frames to raw video before computing the
|
|
hash.
|
|
|
|
The output of the muxer consists of a single line of the form:
|
|
MD5=@var{MD5}, where @var{MD5} is a hexadecimal number representing
|
|
the computed MD5 hash.
|
|
|
|
For example to compute the MD5 hash of the input converted to raw
|
|
audio and video, and store it in the file @file{out.md5}:
|
|
@example
|
|
ffmpeg -i INPUT -f md5 out.md5
|
|
@end example
|
|
|
|
You can print the MD5 to stdout with the command:
|
|
@example
|
|
ffmpeg -i INPUT -f md5 -
|
|
@end example
|
|
|
|
See also the @ref{framemd5} muxer.
|
|
|
|
@section MOV/MP4/ISMV
|
|
|
|
The mov/mp4/ismv muxer supports fragmentation. Normally, a MOV/MP4
|
|
file has all the metadata about all packets stored in one location
|
|
(written at the end of the file, it can be moved to the start for
|
|
better playback using the @command{qt-faststart} tool). A fragmented
|
|
file consists of a number of fragments, where packets and metadata
|
|
about these packets are stored together. Writing a fragmented
|
|
file has the advantage that the file is decodable even if the
|
|
writing is interrupted (while a normal MOV/MP4 is undecodable if
|
|
it is not properly finished), and it requires less memory when writing
|
|
very long files (since writing normal MOV/MP4 files stores info about
|
|
every single packet in memory until the file is closed). The downside
|
|
is that it is less compatible with other applications.
|
|
|
|
Fragmentation is enabled by setting one of the AVOptions that define
|
|
how to cut the file into fragments:
|
|
|
|
@table @option
|
|
@item -moov_size @var{bytes}
|
|
Reserves space for the moov atom at the beginning of the file instead of placing the
|
|
moov atom at the end. If the space reserved is insufficient, muxing will fail.
|
|
@item -movflags frag_keyframe
|
|
Start a new fragment at each video keyframe.
|
|
@item -frag_duration @var{duration}
|
|
Create fragments that are @var{duration} microseconds long.
|
|
@item -frag_size @var{size}
|
|
Create fragments that contain up to @var{size} bytes of payload data.
|
|
@item -movflags frag_custom
|
|
Allow the caller to manually choose when to cut fragments, by
|
|
calling @code{av_write_frame(ctx, NULL)} to write a fragment with
|
|
the packets written so far. (This is only useful with other
|
|
applications integrating libavformat, not from @command{ffmpeg}.)
|
|
@item -min_frag_duration @var{duration}
|
|
Don't create fragments that are shorter than @var{duration} microseconds long.
|
|
@end table
|
|
|
|
If more than one condition is specified, fragments are cut when
|
|
one of the specified conditions is fulfilled. The exception to this is
|
|
@code{-min_frag_duration}, which has to be fulfilled for any of the other
|
|
conditions to apply.
|
|
|
|
Additionally, the way the output file is written can be adjusted
|
|
through a few other options:
|
|
|
|
@table @option
|
|
@item -movflags empty_moov
|
|
Write an initial moov atom directly at the start of the file, without
|
|
describing any samples in it. Generally, an mdat/moov pair is written
|
|
at the start of the file, as a normal MOV/MP4 file, containing only
|
|
a short portion of the file. With this option set, there is no initial
|
|
mdat atom, and the moov atom only describes the tracks but has
|
|
a zero duration.
|
|
|
|
Files written with this option set do not work in QuickTime.
|
|
This option is implicitly set when writing ismv (Smooth Streaming) files.
|
|
@item -movflags separate_moof
|
|
Write a separate moof (movie fragment) atom for each track. Normally,
|
|
packets for all tracks are written in a moof atom (which is slightly
|
|
more efficient), but with this option set, the muxer writes one moof/mdat
|
|
pair for each track, making it easier to separate tracks.
|
|
|
|
This option is implicitly set when writing ismv (Smooth Streaming) files.
|
|
@end table
|
|
|
|
Smooth Streaming content can be pushed in real time to a publishing
|
|
point on IIS with this muxer. Example:
|
|
@example
|
|
ffmpeg -re @var{<normal input/transcoding options>} -movflags isml+frag_keyframe -f ismv http://server/publishingpoint.isml/Streams(Encoder1)
|
|
@end example
|
|
|
|
@section mpegts
|
|
|
|
MPEG transport stream muxer.
|
|
|
|
This muxer implements ISO 13818-1 and part of ETSI EN 300 468.
|
|
|
|
The muxer options are:
|
|
|
|
@table @option
|
|
@item -mpegts_original_network_id @var{number}
|
|
Set the original_network_id (default 0x0001). This is unique identifier
|
|
of a network in DVB. Its main use is in the unique identification of a
|
|
service through the path Original_Network_ID, Transport_Stream_ID.
|
|
@item -mpegts_transport_stream_id @var{number}
|
|
Set the transport_stream_id (default 0x0001). This identifies a
|
|
transponder in DVB.
|
|
@item -mpegts_service_id @var{number}
|
|
Set the service_id (default 0x0001) also known as program in DVB.
|
|
@item -mpegts_pmt_start_pid @var{number}
|
|
Set the first PID for PMT (default 0x1000, max 0x1f00).
|
|
@item -mpegts_start_pid @var{number}
|
|
Set the first PID for data packets (default 0x0100, max 0x0f00).
|
|
@end table
|
|
|
|
The recognized metadata settings in mpegts muxer are @code{service_provider}
|
|
and @code{service_name}. If they are not set the default for
|
|
@code{service_provider} is "FFmpeg" and the default for
|
|
@code{service_name} is "Service01".
|
|
|
|
@example
|
|
ffmpeg -i file.mpg -c copy \
|
|
-mpegts_original_network_id 0x1122 \
|
|
-mpegts_transport_stream_id 0x3344 \
|
|
-mpegts_service_id 0x5566 \
|
|
-mpegts_pmt_start_pid 0x1500 \
|
|
-mpegts_start_pid 0x150 \
|
|
-metadata service_provider="Some provider" \
|
|
-metadata service_name="Some Channel" \
|
|
-y out.ts
|
|
@end example
|
|
|
|
@section null
|
|
|
|
Null muxer.
|
|
|
|
This muxer does not generate any output file, it is mainly useful for
|
|
testing or benchmarking purposes.
|
|
|
|
For example to benchmark decoding with @command{ffmpeg} you can use the
|
|
command:
|
|
@example
|
|
ffmpeg -benchmark -i INPUT -f null out.null
|
|
@end example
|
|
|
|
Note that the above command does not read or write the @file{out.null}
|
|
file, but specifying the output file is required by the @command{ffmpeg}
|
|
syntax.
|
|
|
|
Alternatively you can write the command as:
|
|
@example
|
|
ffmpeg -benchmark -i INPUT -f null -
|
|
@end example
|
|
|
|
@section matroska
|
|
|
|
Matroska container muxer.
|
|
|
|
This muxer implements the matroska and webm container specs.
|
|
|
|
The recognized metadata settings in this muxer are:
|
|
|
|
@table @option
|
|
|
|
@item title=@var{title name}
|
|
Name provided to a single track
|
|
@end table
|
|
|
|
@table @option
|
|
|
|
@item language=@var{language name}
|
|
Specifies the language of the track in the Matroska languages form
|
|
@end table
|
|
|
|
@table @option
|
|
|
|
@item stereo_mode=@var{mode}
|
|
Stereo 3D video layout of two views in a single video track
|
|
@table @option
|
|
@item mono
|
|
video is not stereo
|
|
@item left_right
|
|
Both views are arranged side by side, Left-eye view is on the left
|
|
@item bottom_top
|
|
Both views are arranged in top-bottom orientation, Left-eye view is at bottom
|
|
@item top_bottom
|
|
Both views are arranged in top-bottom orientation, Left-eye view is on top
|
|
@item checkerboard_rl
|
|
Each view is arranged in a checkerboard interleaved pattern, Left-eye view being first
|
|
@item checkerboard_lr
|
|
Each view is arranged in a checkerboard interleaved pattern, Right-eye view being first
|
|
@item row_interleaved_rl
|
|
Each view is constituted by a row based interleaving, Right-eye view is first row
|
|
@item row_interleaved_lr
|
|
Each view is constituted by a row based interleaving, Left-eye view is first row
|
|
@item col_interleaved_rl
|
|
Both views are arranged in a column based interleaving manner, Right-eye view is first column
|
|
@item col_interleaved_lr
|
|
Both views are arranged in a column based interleaving manner, Left-eye view is first column
|
|
@item anaglyph_cyan_red
|
|
All frames are in anaglyph format viewable through red-cyan filters
|
|
@item right_left
|
|
Both views are arranged side by side, Right-eye view is on the left
|
|
@item anaglyph_green_magenta
|
|
All frames are in anaglyph format viewable through green-magenta filters
|
|
@item block_lr
|
|
Both eyes laced in one Block, Left-eye view is first
|
|
@item block_rl
|
|
Both eyes laced in one Block, Right-eye view is first
|
|
@end table
|
|
@end table
|
|
|
|
For example a 3D WebM clip can be created using the following command line:
|
|
@example
|
|
ffmpeg -i sample_left_right_clip.mpg -an -c:v libvpx -metadata stereo_mode=left_right -y stereo_clip.webm
|
|
@end example
|
|
|
|
@section segment, stream_segment, ssegment
|
|
|
|
Basic stream segmenter.
|
|
|
|
The segmenter muxer outputs streams to a number of separate files of nearly
|
|
fixed duration. Output filename pattern can be set in a fashion similar to
|
|
@ref{image2}.
|
|
|
|
@code{stream_segment} is a variant of the muxer used to write to
|
|
streaming output formats, i.e. which do not require global headers,
|
|
and is recommended for outputting e.g. to MPEG transport stream segments.
|
|
@code{ssegment} is a shorter alias for @code{stream_segment}.
|
|
|
|
Every segment starts with a video keyframe, if a video stream is present.
|
|
Note that if you want accurate splitting for a video file, you need to
|
|
make the input key frames correspond to the exact splitting times
|
|
expected by the segmenter, or the segment muxer will start the new
|
|
segment with the key frame found next after the specified start
|
|
time.
|
|
|
|
The segment muxer works best with a single constant frame rate video.
|
|
|
|
Optionally it can generate a flat list of the created segments, one segment
|
|
per line, by setting the option @var{segment_list}.
|
|
|
|
The segment muxer supports the following options:
|
|
|
|
@table @option
|
|
@item segment_format @var{format}
|
|
Override the inner container format, by default it is guessed by the filename
|
|
extension.
|
|
@item segment_time @var{t}
|
|
Set segment duration to @var{t} seconds. Default value is 2.
|
|
@item segment_list @var{name}
|
|
Generate also a listfile named @var{name}. If not specified no
|
|
listfile is generated.
|
|
@item segment_list_size @var{size}
|
|
Overwrite the listfile once it reaches @var{size} entries. If 0
|
|
the listfile is never overwritten. Default value is 5.
|
|
@item segment_wrap @var{limit}
|
|
Wrap around segment index once it reaches @var{limit}.
|
|
@end table
|
|
|
|
Some examples follow.
|
|
|
|
@itemize
|
|
@item
|
|
To remux the content of file @file{in.mkv} to a list of segments
|
|
@file{out-000.nut}, @file{out-001.nut}, etc., and write the list of
|
|
generated segments to @file{out.list}:
|
|
@example
|
|
ffmpeg -i in.mkv -codec copy -map 0 -f segment -segment_list out.list out%03d.nut
|
|
@end example
|
|
|
|
@item
|
|
To convert the @file{in.mkv} to TS segments using the @code{libx264}
|
|
and @code{libfaac} encoders:
|
|
@example
|
|
ffmpeg -i in.mkv -map 0 -codec:v libx264 -codec:a libfaac -f ssegment -segment_list out.list out%03d.ts
|
|
@end example
|
|
@end itemize
|
|
|
|
@section mp3
|
|
|
|
The MP3 muxer writes a raw MP3 stream with an ID3v2 header at the beginning and
|
|
optionally an ID3v1 tag at the end. ID3v2.3 and ID3v2.4 are supported, the
|
|
@code{id3v2_version} option controls which one is used. The legacy ID3v1 tag is
|
|
not written by default, but may be enabled with the @code{write_id3v1} option.
|
|
|
|
For seekable output the muxer also writes a Xing frame at the beginning, which
|
|
contains the number of frames in the file. It is useful for computing duration
|
|
of VBR files.
|
|
|
|
The muxer supports writing ID3v2 attached pictures (APIC frames). The pictures
|
|
are supplied to the muxer in form of a video stream with a single packet. There
|
|
can be any number of those streams, each will correspond to a single APIC frame.
|
|
The stream metadata tags @var{title} and @var{comment} map to APIC
|
|
@var{description} and @var{picture type} respectively. See
|
|
@url{http://id3.org/id3v2.4.0-frames} for allowed picture types.
|
|
|
|
Note that the APIC frames must be written at the beginning, so the muxer will
|
|
buffer the audio frames until it gets all the pictures. It is therefore advised
|
|
to provide the pictures as soon as possible to avoid excessive buffering.
|
|
|
|
Examples:
|
|
|
|
Write an mp3 with an ID3v2.3 header and an ID3v1 footer:
|
|
@example
|
|
ffmpeg -i INPUT -id3v2_version 3 -write_id3v1 1 out.mp3
|
|
@end example
|
|
|
|
Attach a picture to an mp3:
|
|
@example
|
|
ffmpeg -i input.mp3 -i cover.png -c copy -metadata:s:v title="Album cover"
|
|
-metadata:s:v comment="Cover (Front)" out.mp3
|
|
@end example
|
|
|
|
@c man end MUXERS
|