/* * FLV demuxer * Copyright (c) 2003 The FFmpeg Project * * This demuxer will generate a 1 byte extradata for VP6F content. * It is composed of: * - upper 4 bits: difference between encoded width and visible width * - lower 4 bits: difference between encoded height and visible height * * This file is part of FFmpeg. * * FFmpeg is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public * License as published by the Free Software Foundation; either * version 2.1 of the License, or (at your option) any later version. * * FFmpeg is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public * License along with FFmpeg; if not, write to the Free Software * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA */ #include "libavutil/avassert.h" #include "libavutil/avstring.h" #include "libavutil/channel_layout.h" #include "libavutil/dict.h" #include "libavutil/dict_internal.h" #include "libavutil/mem.h" #include "libavutil/opt.h" #include "libavutil/internal.h" #include "libavutil/intfloat.h" #include "libavutil/intreadwrite.h" #include "libavutil/mastering_display_metadata.h" #include "avformat.h" #include "demux.h" #include "internal.h" #include "flv.h" #define VALIDATE_INDEX_TS_THRESH 2500 #define RESYNC_BUFFER_SIZE (1<<20) #define MAX_DEPTH 16 ///< arbitrary limit to prevent unbounded recursion typedef struct FLVMasteringMeta { double r_x; double r_y; double g_x; double g_y; double b_x; double b_y; double white_x; double white_y; double max_luminance; double min_luminance; } FLVMasteringMeta; typedef struct FLVMetaVideoColor { uint64_t matrix_coefficients; uint64_t transfer_characteristics; uint64_t primaries; uint64_t max_cll; uint64_t max_fall; FLVMasteringMeta mastering_meta; } FLVMetaVideoColor; typedef struct FLVContext { const AVClass *class; ///< Class for private options. int trust_metadata; ///< configure streams according onMetaData int trust_datasize; ///< trust data size of FLVTag int dump_full_metadata; ///< Dump full metadata of the onMetadata int wrong_dts; ///< wrong dts due to negative cts uint8_t *new_extradata[FLV_STREAM_TYPE_NB]; int new_extradata_size[FLV_STREAM_TYPE_NB]; int last_sample_rate; int last_channels; struct { int64_t dts; int64_t pos; } validate_index[2]; int validate_next; int validate_count; int searched_for_end; uint8_t resync_buffer[2*RESYNC_BUFFER_SIZE]; int broken_sizes; int64_t sum_flv_tag_size; int last_keyframe_stream_index; int keyframe_count; int64_t video_bit_rate; int64_t audio_bit_rate; int64_t *keyframe_times; int64_t *keyframe_filepositions; AVRational framerate; int64_t last_ts; int64_t time_offset; int64_t time_pos; FLVMetaVideoColor *metaVideoColor; int meta_color_info_flag; } FLVContext; /* AMF date type */ typedef struct amf_date { double milliseconds; int16_t timezone; } amf_date; static int probe(const AVProbeData *p, int live) { const uint8_t *d = p->buf; unsigned offset = AV_RB32(d + 5); if (d[0] == 'F' && d[1] == 'L' && d[2] == 'V' && d[3] < 5 && d[5] == 0 && offset + 100 < p->buf_size && offset > 8) { int is_live = !memcmp(d + offset + 40, "NGINX RTMP", 10); if (live == is_live) return AVPROBE_SCORE_MAX; } return 0; } static int flv_probe(const AVProbeData *p) { return probe(p, 0); } static int live_flv_probe(const AVProbeData *p) { return probe(p, 1); } static int kux_probe(const AVProbeData *p) { const uint8_t *d = p->buf; if (d[0] == 'K' && d[1] == 'D' && d[2] == 'K' && d[3] == 0 && d[4] == 0) { return AVPROBE_SCORE_EXTENSION + 1; } return 0; } static void add_keyframes_index(AVFormatContext *s) { FLVContext *flv = s->priv_data; AVStream *stream = NULL; unsigned int i = 0; if (flv->last_keyframe_stream_index < 0) { av_log(s, AV_LOG_DEBUG, "keyframe stream hasn't been created\n"); return; } av_assert0(flv->last_keyframe_stream_index <= s->nb_streams); stream = s->streams[flv->last_keyframe_stream_index]; if (ffstream(stream)->nb_index_entries == 0) { for (i = 0; i < flv->keyframe_count; i++) { av_log(s, AV_LOG_TRACE, "keyframe filepositions = %"PRId64" times = %"PRId64"\n", flv->keyframe_filepositions[i], flv->keyframe_times[i]); av_add_index_entry(stream, flv->keyframe_filepositions[i], flv->keyframe_times[i], 0, 0, AVINDEX_KEYFRAME); } } else av_log(s, AV_LOG_WARNING, "Skipping duplicate index\n"); if (stream->codecpar->codec_type == AVMEDIA_TYPE_VIDEO) { av_freep(&flv->keyframe_times); av_freep(&flv->keyframe_filepositions); flv->keyframe_count = 0; } } static AVStream *create_stream(AVFormatContext *s, int codec_type) { FFFormatContext *const si = ffformatcontext(s); FLVContext *flv = s->priv_data; AVStream *st = avformat_new_stream(s, NULL); if (!st) return NULL; st->codecpar->codec_type = codec_type; if (s->nb_streams>=3 ||( s->nb_streams==2 && s->streams[0]->codecpar->codec_type != AVMEDIA_TYPE_SUBTITLE && s->streams[1]->codecpar->codec_type != AVMEDIA_TYPE_SUBTITLE && s->streams[0]->codecpar->codec_type != AVMEDIA_TYPE_DATA && s->streams[1]->codecpar->codec_type != AVMEDIA_TYPE_DATA)) s->ctx_flags &= ~AVFMTCTX_NOHEADER; if (codec_type == AVMEDIA_TYPE_AUDIO) { st->codecpar->bit_rate = flv->audio_bit_rate; si->missing_streams &= ~FLV_HEADER_FLAG_HASAUDIO; } if (codec_type == AVMEDIA_TYPE_VIDEO) { st->codecpar->bit_rate = flv->video_bit_rate; si->missing_streams &= ~FLV_HEADER_FLAG_HASVIDEO; st->avg_frame_rate = flv->framerate; } avpriv_set_pts_info(st, 32, 1, 1000); /* 32 bit pts in ms */ flv->last_keyframe_stream_index = s->nb_streams - 1; add_keyframes_index(s); return st; } static int flv_same_audio_codec(AVCodecParameters *apar, int flags) { int bits_per_coded_sample = (flags & FLV_AUDIO_SAMPLESIZE_MASK) ? 16 : 8; int flv_codecid = flags & FLV_AUDIO_CODECID_MASK; int codec_id; if (!apar->codec_id && !apar->codec_tag) return 1; if (apar->bits_per_coded_sample != bits_per_coded_sample) return 0; switch (flv_codecid) { // no distinction between S16 and S8 PCM codec flags case FLV_CODECID_PCM: codec_id = bits_per_coded_sample == 8 ? AV_CODEC_ID_PCM_U8 #if HAVE_BIGENDIAN : AV_CODEC_ID_PCM_S16BE; #else : AV_CODEC_ID_PCM_S16LE; #endif return codec_id == apar->codec_id; case FLV_CODECID_PCM_LE: codec_id = bits_per_coded_sample == 8 ? AV_CODEC_ID_PCM_U8 : AV_CODEC_ID_PCM_S16LE; return codec_id == apar->codec_id; case FLV_CODECID_AAC: return apar->codec_id == AV_CODEC_ID_AAC; case FLV_CODECID_ADPCM: return apar->codec_id == AV_CODEC_ID_ADPCM_SWF; case FLV_CODECID_SPEEX: return apar->codec_id == AV_CODEC_ID_SPEEX; case FLV_CODECID_MP3: return apar->codec_id == AV_CODEC_ID_MP3; case FLV_CODECID_NELLYMOSER_8KHZ_MONO: case FLV_CODECID_NELLYMOSER_16KHZ_MONO: case FLV_CODECID_NELLYMOSER: return apar->codec_id == AV_CODEC_ID_NELLYMOSER; case FLV_CODECID_PCM_MULAW: return apar->sample_rate == 8000 && apar->codec_id == AV_CODEC_ID_PCM_MULAW; case FLV_CODECID_PCM_ALAW: return apar->sample_rate == 8000 && apar->codec_id == AV_CODEC_ID_PCM_ALAW; default: return apar->codec_tag == (flv_codecid >> FLV_AUDIO_CODECID_OFFSET); } } static void flv_set_audio_codec(AVFormatContext *s, AVStream *astream, AVCodecParameters *apar, int flv_codecid) { switch (flv_codecid) { // no distinction between S16 and S8 PCM codec flags case FLV_CODECID_PCM: apar->codec_id = apar->bits_per_coded_sample == 8 ? AV_CODEC_ID_PCM_U8 #if HAVE_BIGENDIAN : AV_CODEC_ID_PCM_S16BE; #else : AV_CODEC_ID_PCM_S16LE; #endif break; case FLV_CODECID_PCM_LE: apar->codec_id = apar->bits_per_coded_sample == 8 ? AV_CODEC_ID_PCM_U8 : AV_CODEC_ID_PCM_S16LE; break; case FLV_CODECID_AAC: apar->codec_id = AV_CODEC_ID_AAC; break; case FLV_CODECID_ADPCM: apar->codec_id = AV_CODEC_ID_ADPCM_SWF; break; case FLV_CODECID_SPEEX: apar->codec_id = AV_CODEC_ID_SPEEX; apar->sample_rate = 16000; break; case FLV_CODECID_MP3: apar->codec_id = AV_CODEC_ID_MP3; ffstream(astream)->need_parsing = AVSTREAM_PARSE_FULL; break; case FLV_CODECID_NELLYMOSER_8KHZ_MONO: // in case metadata does not otherwise declare samplerate apar->sample_rate = 8000; apar->codec_id = AV_CODEC_ID_NELLYMOSER; break; case FLV_CODECID_NELLYMOSER_16KHZ_MONO: apar->sample_rate = 16000; apar->codec_id = AV_CODEC_ID_NELLYMOSER; break; case FLV_CODECID_NELLYMOSER: apar->codec_id = AV_CODEC_ID_NELLYMOSER; break; case FLV_CODECID_PCM_MULAW: apar->sample_rate = 8000; apar->codec_id = AV_CODEC_ID_PCM_MULAW; break; case FLV_CODECID_PCM_ALAW: apar->sample_rate = 8000; apar->codec_id = AV_CODEC_ID_PCM_ALAW; break; default: avpriv_request_sample(s, "Audio codec (%x)", flv_codecid >> FLV_AUDIO_CODECID_OFFSET); apar->codec_tag = flv_codecid >> FLV_AUDIO_CODECID_OFFSET; } } static int flv_same_video_codec(AVCodecParameters *vpar, uint32_t flv_codecid) { if (!vpar->codec_id && !vpar->codec_tag) return 1; switch (flv_codecid) { case MKBETAG('h', 'v', 'c', '1'): return vpar->codec_id == AV_CODEC_ID_HEVC; case MKBETAG('a', 'v', '0', '1'): return vpar->codec_id == AV_CODEC_ID_AV1; case MKBETAG('v', 'p', '0', '9'): return vpar->codec_id == AV_CODEC_ID_VP9; case FLV_CODECID_H263: return vpar->codec_id == AV_CODEC_ID_FLV1; case FLV_CODECID_SCREEN: return vpar->codec_id == AV_CODEC_ID_FLASHSV; case FLV_CODECID_SCREEN2: return vpar->codec_id == AV_CODEC_ID_FLASHSV2; case FLV_CODECID_VP6: return vpar->codec_id == AV_CODEC_ID_VP6F; case FLV_CODECID_VP6A: return vpar->codec_id == AV_CODEC_ID_VP6A; case FLV_CODECID_H264: return vpar->codec_id == AV_CODEC_ID_H264; default: return vpar->codec_tag == flv_codecid; } } static int flv_set_video_codec(AVFormatContext *s, AVStream *vstream, uint32_t flv_codecid, int read) { FFStream *const vstreami = ffstream(vstream); int ret = 0; AVCodecParameters *par = vstream->codecpar; enum AVCodecID old_codec_id = vstream->codecpar->codec_id; switch (flv_codecid) { case MKBETAG('h', 'v', 'c', '1'): par->codec_id = AV_CODEC_ID_HEVC; vstreami->need_parsing = AVSTREAM_PARSE_HEADERS; break; case MKBETAG('a', 'v', '0', '1'): par->codec_id = AV_CODEC_ID_AV1; vstreami->need_parsing = AVSTREAM_PARSE_HEADERS; break; case MKBETAG('v', 'p', '0', '9'): par->codec_id = AV_CODEC_ID_VP9; vstreami->need_parsing = AVSTREAM_PARSE_HEADERS; break; case FLV_CODECID_H263: par->codec_id = AV_CODEC_ID_FLV1; break; case FLV_CODECID_REALH263: par->codec_id = AV_CODEC_ID_H263; break; // Really mean it this time case FLV_CODECID_SCREEN: par->codec_id = AV_CODEC_ID_FLASHSV; break; case FLV_CODECID_SCREEN2: par->codec_id = AV_CODEC_ID_FLASHSV2; break; case FLV_CODECID_VP6: par->codec_id = AV_CODEC_ID_VP6F; case FLV_CODECID_VP6A: if (flv_codecid == FLV_CODECID_VP6A) par->codec_id = AV_CODEC_ID_VP6A; if (read) { if (par->extradata_size != 1) { ff_alloc_extradata(par, 1); } if (par->extradata) par->extradata[0] = avio_r8(s->pb); else avio_skip(s->pb, 1); } ret = 1; // 1 byte body size adjustment for flv_read_packet() break; case FLV_CODECID_H264: par->codec_id = AV_CODEC_ID_H264; vstreami->need_parsing = AVSTREAM_PARSE_HEADERS; break; case FLV_CODECID_MPEG4: par->codec_id = AV_CODEC_ID_MPEG4; break; default: avpriv_request_sample(s, "Video codec (%x)", flv_codecid); par->codec_tag = flv_codecid; } if (!vstreami->need_context_update && par->codec_id != old_codec_id) { avpriv_request_sample(s, "Changing the codec id midstream"); return AVERROR_PATCHWELCOME; } return ret; } static int amf_get_string(AVIOContext *ioc, char *buffer, int buffsize) { int ret; int length = avio_rb16(ioc); if (length >= buffsize) { avio_skip(ioc, length); return -1; } ret = avio_read(ioc, buffer, length); if (ret < 0) return ret; if (ret < length) return AVERROR_INVALIDDATA; buffer[length] = '\0'; return length; } static int parse_keyframes_index(AVFormatContext *s, AVIOContext *ioc, int64_t max_pos) { FLVContext *flv = s->priv_data; unsigned int timeslen = 0, fileposlen = 0, i; char str_val[256]; int64_t *times = NULL; int64_t *filepositions = NULL; int ret = AVERROR(ENOSYS); int64_t initial_pos = avio_tell(ioc); if (flv->keyframe_count > 0) { av_log(s, AV_LOG_DEBUG, "keyframes have been parsed\n"); return 0; } av_assert0(!flv->keyframe_times); av_assert0(!flv->keyframe_filepositions); if (s->flags & AVFMT_FLAG_IGNIDX) return 0; while (avio_tell(ioc) < max_pos - 2 && amf_get_string(ioc, str_val, sizeof(str_val)) > 0) { int64_t **current_array; unsigned int arraylen; int factor; // Expect array object in context if (avio_r8(ioc) != AMF_DATA_TYPE_ARRAY) break; arraylen = avio_rb32(ioc); if (arraylen>>28) break; if (!strcmp(KEYFRAMES_TIMESTAMP_TAG , str_val) && !times) { current_array = × timeslen = arraylen; factor = 1000; } else if (!strcmp(KEYFRAMES_BYTEOFFSET_TAG, str_val) && !filepositions) { current_array = &filepositions; fileposlen = arraylen; factor = 1; } else // unexpected metatag inside keyframes, will not use such // metadata for indexing break; if (!(*current_array = av_mallocz(sizeof(**current_array) * arraylen))) { ret = AVERROR(ENOMEM); goto finish; } for (i = 0; i < arraylen && avio_tell(ioc) < max_pos - 1; i++) { double d; if (avio_r8(ioc) != AMF_DATA_TYPE_NUMBER) goto invalid; d = av_int2double(avio_rb64(ioc)) * factor; if (isnan(d) || d < INT64_MIN || d > INT64_MAX) goto invalid; if (avio_feof(ioc)) goto invalid; current_array[0][i] = d; } if (times && filepositions) { // All done, exiting at a position allowing amf_parse_object // to finish parsing the object ret = 0; break; } } if (timeslen == fileposlen && fileposlen>1 && max_pos <= filepositions[0]) { for (i = 0; i < FFMIN(2,fileposlen); i++) { flv->validate_index[i].pos = filepositions[i]; flv->validate_index[i].dts = times[i]; flv->validate_count = i + 1; } flv->keyframe_times = times; flv->keyframe_filepositions = filepositions; flv->keyframe_count = timeslen; times = NULL; filepositions = NULL; } else { invalid: av_log(s, AV_LOG_WARNING, "Invalid keyframes object, skipping.\n"); } finish: av_freep(×); av_freep(&filepositions); avio_seek(ioc, initial_pos, SEEK_SET); return ret; } static int amf_parse_object(AVFormatContext *s, AVStream *astream, AVStream *vstream, const char *key, int64_t max_pos, int depth) { AVCodecParameters *apar, *vpar; FLVContext *flv = s->priv_data; AVIOContext *ioc; AMFDataType amf_type; FLVMetaVideoColor *meta_video_color = flv->metaVideoColor; char str_val[1024]; double num_val; amf_date date; if (depth > MAX_DEPTH) return AVERROR_PATCHWELCOME; num_val = 0; ioc = s->pb; if (avio_feof(ioc)) return AVERROR_EOF; amf_type = avio_r8(ioc); switch (amf_type) { case AMF_DATA_TYPE_NUMBER: num_val = av_int2double(avio_rb64(ioc)); break; case AMF_DATA_TYPE_BOOL: num_val = avio_r8(ioc); break; case AMF_DATA_TYPE_STRING: if (amf_get_string(ioc, str_val, sizeof(str_val)) < 0) { av_log(s, AV_LOG_ERROR, "AMF_DATA_TYPE_STRING parsing failed\n"); return -1; } break; case AMF_DATA_TYPE_OBJECT: if (key && (ioc->seekable & AVIO_SEEKABLE_NORMAL) && !strcmp(KEYFRAMES_TAG, key) && depth == 1) if (parse_keyframes_index(s, ioc, max_pos) < 0) av_log(s, AV_LOG_ERROR, "Keyframe index parsing failed\n"); else add_keyframes_index(s); while (avio_tell(ioc) < max_pos - 2 && amf_get_string(ioc, str_val, sizeof(str_val)) > 0) if (amf_parse_object(s, astream, vstream, str_val, max_pos, depth + 1) < 0) return -1; // if we couldn't skip, bomb out. if (avio_r8(ioc) != AMF_END_OF_OBJECT) { av_log(s, AV_LOG_ERROR, "Missing AMF_END_OF_OBJECT in AMF_DATA_TYPE_OBJECT\n"); return -1; } break; case AMF_DATA_TYPE_NULL: case AMF_DATA_TYPE_UNDEFINED: case AMF_DATA_TYPE_UNSUPPORTED: break; // these take up no additional space case AMF_DATA_TYPE_MIXEDARRAY: { unsigned v; avio_skip(ioc, 4); // skip 32-bit max array index while (avio_tell(ioc) < max_pos - 2 && amf_get_string(ioc, str_val, sizeof(str_val)) > 0) // this is the only case in which we would want a nested // parse to not skip over the object if (amf_parse_object(s, astream, vstream, str_val, max_pos, depth + 1) < 0) return -1; v = avio_r8(ioc); if (v != AMF_END_OF_OBJECT) { av_log(s, AV_LOG_ERROR, "Missing AMF_END_OF_OBJECT in AMF_DATA_TYPE_MIXEDARRAY, found %d\n", v); return -1; } break; } case AMF_DATA_TYPE_ARRAY: { unsigned int arraylen, i; arraylen = avio_rb32(ioc); for (i = 0; i < arraylen && avio_tell(ioc) < max_pos - 1; i++) if (amf_parse_object(s, NULL, NULL, NULL, max_pos, depth + 1) < 0) return -1; // if we couldn't skip, bomb out. } break; case AMF_DATA_TYPE_DATE: // timestamp (double) and UTC offset (int16) date.milliseconds = av_int2double(avio_rb64(ioc)); date.timezone = avio_rb16(ioc); break; default: // unsupported type, we couldn't skip av_log(s, AV_LOG_ERROR, "unsupported amf type %d\n", amf_type); return -1; } if (key) { apar = astream ? astream->codecpar : NULL; vpar = vstream ? vstream->codecpar : NULL; // stream info doesn't live any deeper than the first object if (depth == 1) { if (amf_type == AMF_DATA_TYPE_NUMBER || amf_type == AMF_DATA_TYPE_BOOL) { if (!strcmp(key, "duration")) s->duration = num_val * AV_TIME_BASE; else if (!strcmp(key, "videodatarate") && 0 <= (int)(num_val * 1024.0)) flv->video_bit_rate = num_val * 1024.0; else if (!strcmp(key, "audiodatarate") && 0 <= (int)(num_val * 1024.0)) flv->audio_bit_rate = num_val * 1024.0; else if (!strcmp(key, "framerate")) { flv->framerate = av_d2q(num_val, 1000); if (vstream) vstream->avg_frame_rate = flv->framerate; } else if (flv->trust_metadata) { if (!strcmp(key, "videocodecid") && vpar) { int ret = flv_set_video_codec(s, vstream, num_val, 0); if (ret < 0) return ret; } else if (!strcmp(key, "audiocodecid") && apar) { int id = ((int)num_val) << FLV_AUDIO_CODECID_OFFSET; flv_set_audio_codec(s, astream, apar, id); } else if (!strcmp(key, "audiosamplerate") && apar) { apar->sample_rate = num_val; } else if (!strcmp(key, "audiosamplesize") && apar) { apar->bits_per_coded_sample = num_val; } else if (!strcmp(key, "stereo") && apar) { av_channel_layout_default(&apar->ch_layout, num_val + 1); } else if (!strcmp(key, "width") && vpar) { vpar->width = num_val; } else if (!strcmp(key, "height") && vpar) { vpar->height = num_val; } else if (!strcmp(key, "datastream")) { AVStream *st = create_stream(s, AVMEDIA_TYPE_SUBTITLE); if (!st) return AVERROR(ENOMEM); st->codecpar->codec_id = AV_CODEC_ID_TEXT; } } } if (amf_type == AMF_DATA_TYPE_STRING) { if (!strcmp(key, "encoder")) { int version = -1; if (1 == sscanf(str_val, "Open Broadcaster Software v0.%d", &version)) { if (version > 0 && version <= 655) flv->broken_sizes = 1; } } else if (!strcmp(key, "metadatacreator")) { if ( !strcmp (str_val, "MEGA") || !strncmp(str_val, "FlixEngine", 10)) flv->broken_sizes = 1; } } } if (meta_video_color) { if (amf_type == AMF_DATA_TYPE_NUMBER || amf_type == AMF_DATA_TYPE_BOOL) { if (!strcmp(key, "colorPrimaries")) { meta_video_color->primaries = num_val; } else if (!strcmp(key, "transferCharacteristics")) { meta_video_color->transfer_characteristics = num_val; } else if (!strcmp(key, "matrixCoefficients")) { meta_video_color->matrix_coefficients = num_val; } else if (!strcmp(key, "maxFall")) { meta_video_color->max_fall = num_val; } else if (!strcmp(key, "maxCLL")) { meta_video_color->max_cll = num_val; } else if (!strcmp(key, "redX")) { meta_video_color->mastering_meta.r_x = num_val; } else if (!strcmp(key, "redY")) { meta_video_color->mastering_meta.r_y = num_val; } else if (!strcmp(key, "greenX")) { meta_video_color->mastering_meta.g_x = num_val; } else if (!strcmp(key, "greenY")) { meta_video_color->mastering_meta.g_y = num_val; } else if (!strcmp(key, "blueX")) { meta_video_color->mastering_meta.b_x = num_val; } else if (!strcmp(key, "blueY")) { meta_video_color->mastering_meta.b_y = num_val; } else if (!strcmp(key, "whitePointX")) { meta_video_color->mastering_meta.white_x = num_val; } else if (!strcmp(key, "whitePointY")) { meta_video_color->mastering_meta.white_y = num_val; } else if (!strcmp(key, "maxLuminance")) { meta_video_color->mastering_meta.max_luminance = num_val; } else if (!strcmp(key, "minLuminance")) { meta_video_color->mastering_meta.min_luminance = num_val; } } } if (amf_type == AMF_DATA_TYPE_OBJECT && s->nb_streams == 1 && ((!apar && !strcmp(key, "audiocodecid")) || (!vpar && !strcmp(key, "videocodecid")))) s->ctx_flags &= ~AVFMTCTX_NOHEADER; //If there is either audio/video missing, codecid will be an empty object if ((!strcmp(key, "duration") || !strcmp(key, "filesize") || !strcmp(key, "width") || !strcmp(key, "height") || !strcmp(key, "videodatarate") || !strcmp(key, "framerate") || !strcmp(key, "videocodecid") || !strcmp(key, "audiodatarate") || !strcmp(key, "audiosamplerate") || !strcmp(key, "audiosamplesize") || !strcmp(key, "stereo") || !strcmp(key, "audiocodecid") || !strcmp(key, "datastream")) && !flv->dump_full_metadata) return 0; s->event_flags |= AVFMT_EVENT_FLAG_METADATA_UPDATED; if (amf_type == AMF_DATA_TYPE_BOOL) { av_strlcpy(str_val, num_val > 0 ? "true" : "false", sizeof(str_val)); av_dict_set(&s->metadata, key, str_val, 0); } else if (amf_type == AMF_DATA_TYPE_NUMBER) { snprintf(str_val, sizeof(str_val), "%.f", num_val); av_dict_set(&s->metadata, key, str_val, 0); } else if (amf_type == AMF_DATA_TYPE_STRING) { av_dict_set(&s->metadata, key, str_val, 0); } else if ( amf_type == AMF_DATA_TYPE_DATE && isfinite(date.milliseconds) && date.milliseconds > INT64_MIN/1000 && date.milliseconds < INT64_MAX/1000 ) { // timezone is ignored, since there is no easy way to offset the UTC // timestamp into the specified timezone avpriv_dict_set_timestamp(&s->metadata, key, 1000 * (int64_t)date.milliseconds); } } return 0; } #define TYPE_ONTEXTDATA 1 #define TYPE_ONCAPTION 2 #define TYPE_ONCAPTIONINFO 3 #define TYPE_UNKNOWN 9 static int flv_read_metabody(AVFormatContext *s, int64_t next_pos) { FLVContext *flv = s->priv_data; AMFDataType type; AVStream *stream, *astream, *vstream; AVStream av_unused *dstream; AVIOContext *ioc; int i; char buffer[32]; astream = NULL; vstream = NULL; dstream = NULL; ioc = s->pb; // first object needs to be "onMetaData" string type = avio_r8(ioc); if (type != AMF_DATA_TYPE_STRING || amf_get_string(ioc, buffer, sizeof(buffer)) < 0) return TYPE_UNKNOWN; if (!strcmp(buffer, "onTextData")) return TYPE_ONTEXTDATA; if (!strcmp(buffer, "onCaption")) return TYPE_ONCAPTION; if (!strcmp(buffer, "onCaptionInfo")) return TYPE_ONCAPTIONINFO; if (strcmp(buffer, "onMetaData") && strcmp(buffer, "onCuePoint") && strcmp(buffer, "|RtmpSampleAccess")) { av_log(s, AV_LOG_DEBUG, "Unknown type %s\n", buffer); return TYPE_UNKNOWN; } // find the streams now so that amf_parse_object doesn't need to do // the lookup every time it is called. for (i = 0; i < s->nb_streams; i++) { stream = s->streams[i]; if (stream->codecpar->codec_type == AVMEDIA_TYPE_VIDEO) { vstream = stream; flv->last_keyframe_stream_index = i; } else if (stream->codecpar->codec_type == AVMEDIA_TYPE_AUDIO) { astream = stream; if (flv->last_keyframe_stream_index == -1) flv->last_keyframe_stream_index = i; } else if (stream->codecpar->codec_type == AVMEDIA_TYPE_SUBTITLE) dstream = stream; } // parse the second object (we want a mixed array) if (amf_parse_object(s, astream, vstream, buffer, next_pos, 0) < 0) return -1; return 0; } static int flv_read_header(AVFormatContext *s) { FFFormatContext *const si = ffformatcontext(s); int flags; FLVContext *flv = s->priv_data; int offset; int pre_tag_size = 0; /* Actual FLV data at 0xe40000 in KUX file */ if(!strcmp(s->iformat->name, "kux")) avio_skip(s->pb, 0xe40000); avio_skip(s->pb, 4); flags = avio_r8(s->pb); si->missing_streams = flags & (FLV_HEADER_FLAG_HASVIDEO | FLV_HEADER_FLAG_HASAUDIO); s->ctx_flags |= AVFMTCTX_NOHEADER; offset = avio_rb32(s->pb); avio_seek(s->pb, offset, SEEK_SET); /* Annex E. The FLV File Format * E.3 TheFLVFileBody * Field Type Comment * PreviousTagSize0 UI32 Always 0 * */ pre_tag_size = avio_rb32(s->pb); if (pre_tag_size) { av_log(s, AV_LOG_WARNING, "Read FLV header error, input file is not a standard flv format, first PreviousTagSize0 always is 0\n"); } s->start_time = 0; flv->sum_flv_tag_size = 0; flv->last_keyframe_stream_index = -1; return 0; } static int flv_read_close(AVFormatContext *s) { int i; FLVContext *flv = s->priv_data; for (i=0; inew_extradata[i]); av_freep(&flv->keyframe_times); av_freep(&flv->keyframe_filepositions); av_freep(&flv->metaVideoColor); return 0; } static int flv_get_extradata(AVFormatContext *s, AVStream *st, int size) { int ret; if (!size) return 0; if ((ret = ff_get_extradata(s, st->codecpar, s->pb, size)) < 0) return ret; ffstream(st)->need_context_update = 1; return 0; } static int flv_queue_extradata(FLVContext *flv, AVIOContext *pb, int stream, int size) { if (!size) return 0; av_free(flv->new_extradata[stream]); flv->new_extradata[stream] = av_mallocz(size + AV_INPUT_BUFFER_PADDING_SIZE); if (!flv->new_extradata[stream]) return AVERROR(ENOMEM); flv->new_extradata_size[stream] = size; avio_read(pb, flv->new_extradata[stream], size); return 0; } static void clear_index_entries(AVFormatContext *s, int64_t pos) { av_log(s, AV_LOG_WARNING, "Found invalid index entries, clearing the index.\n"); for (unsigned i = 0; i < s->nb_streams; i++) { FFStream *const sti = ffstream(s->streams[i]); int out = 0; /* Remove all index entries that point to >= pos */ for (int j = 0; j < sti->nb_index_entries; j++) if (sti->index_entries[j].pos < pos) sti->index_entries[out++] = sti->index_entries[j]; sti->nb_index_entries = out; } } static int amf_skip_tag(AVIOContext *pb, AMFDataType type, int depth) { int nb = -1, ret, parse_name = 1; if (depth > MAX_DEPTH) return AVERROR_PATCHWELCOME; if (avio_feof(pb)) return AVERROR_EOF; switch (type) { case AMF_DATA_TYPE_NUMBER: avio_skip(pb, 8); break; case AMF_DATA_TYPE_BOOL: avio_skip(pb, 1); break; case AMF_DATA_TYPE_STRING: avio_skip(pb, avio_rb16(pb)); break; case AMF_DATA_TYPE_ARRAY: parse_name = 0; case AMF_DATA_TYPE_MIXEDARRAY: nb = avio_rb32(pb); if (nb < 0) return AVERROR_INVALIDDATA; case AMF_DATA_TYPE_OBJECT: while(!pb->eof_reached && (nb-- > 0 || type != AMF_DATA_TYPE_ARRAY)) { if (parse_name) { int size = avio_rb16(pb); if (!size) { avio_skip(pb, 1); break; } avio_skip(pb, size); } if ((ret = amf_skip_tag(pb, avio_r8(pb), depth + 1)) < 0) return ret; } break; case AMF_DATA_TYPE_NULL: case AMF_DATA_TYPE_OBJECT_END: break; default: return AVERROR_INVALIDDATA; } return 0; } static int flv_data_packet(AVFormatContext *s, AVPacket *pkt, int64_t dts, int64_t next) { AVIOContext *pb = s->pb; AVStream *st = NULL; char buf[20]; int ret = AVERROR_INVALIDDATA; int i, length = -1; int array = 0; switch (avio_r8(pb)) { case AMF_DATA_TYPE_ARRAY: array = 1; case AMF_DATA_TYPE_MIXEDARRAY: avio_seek(pb, 4, SEEK_CUR); case AMF_DATA_TYPE_OBJECT: break; default: goto skip; } while (array || (ret = amf_get_string(pb, buf, sizeof(buf))) > 0) { AMFDataType type = avio_r8(pb); if (type == AMF_DATA_TYPE_STRING && (array || !strcmp(buf, "text"))) { length = avio_rb16(pb); ret = av_get_packet(pb, pkt, length); if (ret < 0) goto skip; else break; } else { if ((ret = amf_skip_tag(pb, type, 0)) < 0) goto skip; } } if (length < 0) { ret = AVERROR_INVALIDDATA; goto skip; } for (i = 0; i < s->nb_streams; i++) { st = s->streams[i]; if (st->codecpar->codec_type == AVMEDIA_TYPE_SUBTITLE) break; } if (i == s->nb_streams) { st = create_stream(s, AVMEDIA_TYPE_SUBTITLE); if (!st) return AVERROR(ENOMEM); st->codecpar->codec_id = AV_CODEC_ID_TEXT; } pkt->dts = dts; pkt->pts = dts; pkt->size = ret; pkt->stream_index = st->index; pkt->flags |= AV_PKT_FLAG_KEY; skip: avio_seek(s->pb, next + 4, SEEK_SET); return ret; } static int resync(AVFormatContext *s) { FLVContext *flv = s->priv_data; int64_t i; int64_t pos = avio_tell(s->pb); for (i=0; !avio_feof(s->pb); i++) { int j = i & (RESYNC_BUFFER_SIZE-1); int j1 = j + RESYNC_BUFFER_SIZE; flv->resync_buffer[j ] = flv->resync_buffer[j1] = avio_r8(s->pb); if (i >= 8 && pos) { uint8_t *d = flv->resync_buffer + j1 - 8; if (d[0] == 'F' && d[1] == 'L' && d[2] == 'V' && d[3] < 5 && d[5] == 0) { av_log(s, AV_LOG_WARNING, "Concatenated FLV detected, might fail to demux, decode and seek %"PRId64"\n", flv->last_ts); flv->time_offset = flv->last_ts + 1; flv->time_pos = avio_tell(s->pb); } } if (i > 22) { unsigned lsize2 = AV_RB32(flv->resync_buffer + j1 - 4); if (lsize2 >= 11 && lsize2 + 8LL < FFMIN(i, RESYNC_BUFFER_SIZE)) { unsigned size2 = AV_RB24(flv->resync_buffer + j1 - lsize2 + 1 - 4); unsigned lsize1 = AV_RB32(flv->resync_buffer + j1 - lsize2 - 8); if (lsize1 >= 11 && lsize1 + 8LL + lsize2 < FFMIN(i, RESYNC_BUFFER_SIZE)) { unsigned size1 = AV_RB24(flv->resync_buffer + j1 - lsize1 + 1 - lsize2 - 8); if (size1 == lsize1 - 11 && size2 == lsize2 - 11) { avio_seek(s->pb, pos + i - lsize1 - lsize2 - 8, SEEK_SET); return 1; } } } } } return AVERROR_EOF; } static int flv_parse_video_color_info(AVFormatContext *s, AVStream *st, int64_t next_pos) { FLVContext *flv = s->priv_data; AMFDataType type; AVIOContext *ioc; char buffer[32]; ioc = s->pb; // first object needs to be "colorInfo" string type = avio_r8(ioc); if (type != AMF_DATA_TYPE_STRING || amf_get_string(ioc, buffer, sizeof(buffer)) < 0) return TYPE_UNKNOWN; if (strcmp(buffer, "colorInfo")) { av_log(s, AV_LOG_DEBUG, "Unknown type %s\n", buffer); return TYPE_UNKNOWN; } av_free(flv->metaVideoColor); if (!(flv->metaVideoColor = av_mallocz(sizeof(FLVMetaVideoColor)))) { return AVERROR(ENOMEM); } flv->meta_color_info_flag = 1; amf_parse_object(s, NULL, NULL, buffer, next_pos, 0); // parse metadata return 0; } static int flv_update_video_color_info(AVFormatContext *s, AVStream *st) { FLVContext *flv = s->priv_data; const FLVMetaVideoColor* meta_video_color = flv->metaVideoColor; const FLVMasteringMeta *mastering_meta = &meta_video_color->mastering_meta; int has_mastering_primaries, has_mastering_luminance; // Mastering primaries are CIE 1931 coords, and must be > 0. has_mastering_primaries = mastering_meta->r_x > 0 && mastering_meta->r_y > 0 && mastering_meta->g_x > 0 && mastering_meta->g_y > 0 && mastering_meta->b_x > 0 && mastering_meta->b_y > 0 && mastering_meta->white_x > 0 && mastering_meta->white_y > 0; has_mastering_luminance = mastering_meta->max_luminance > 0 && mastering_meta->min_luminance > 0; if (meta_video_color->matrix_coefficients != AVCOL_SPC_RESERVED) st->codecpar->color_space = meta_video_color->matrix_coefficients; if (meta_video_color->primaries != AVCOL_PRI_RESERVED && meta_video_color->primaries != AVCOL_PRI_RESERVED0) st->codecpar->color_primaries = meta_video_color->primaries; if (meta_video_color->transfer_characteristics != AVCOL_TRC_RESERVED && meta_video_color->transfer_characteristics != AVCOL_TRC_RESERVED0) st->codecpar->color_trc = meta_video_color->transfer_characteristics; if (meta_video_color->max_cll && meta_video_color->max_fall) { size_t size = 0; AVContentLightMetadata *metadata = av_content_light_metadata_alloc(&size); if (!metadata) return AVERROR(ENOMEM); if (!av_packet_side_data_add(&st->codecpar->coded_side_data, &st->codecpar->nb_coded_side_data, AV_PKT_DATA_CONTENT_LIGHT_LEVEL, metadata, size, 0)) { av_freep(&metadata); return AVERROR(ENOMEM); } metadata->MaxCLL = meta_video_color->max_cll; metadata->MaxFALL = meta_video_color->max_fall; } if (has_mastering_primaries || has_mastering_luminance) { AVMasteringDisplayMetadata *metadata; AVPacketSideData *sd = av_packet_side_data_new(&st->codecpar->coded_side_data, &st->codecpar->nb_coded_side_data, AV_PKT_DATA_MASTERING_DISPLAY_METADATA, sizeof(AVMasteringDisplayMetadata), 0); if (!sd) return AVERROR(ENOMEM); metadata = (AVMasteringDisplayMetadata*)sd->data; memset(metadata, 0, sizeof(AVMasteringDisplayMetadata)); // hdrCll if (has_mastering_luminance) { metadata->max_luminance = av_d2q(mastering_meta->max_luminance, INT_MAX); metadata->min_luminance = av_d2q(mastering_meta->min_luminance, INT_MAX); metadata->has_luminance = 1; } // hdrMdcv if (has_mastering_primaries) { metadata->display_primaries[0][0] = av_d2q(mastering_meta->r_x, INT_MAX); metadata->display_primaries[0][1] = av_d2q(mastering_meta->r_y, INT_MAX); metadata->display_primaries[1][0] = av_d2q(mastering_meta->g_x, INT_MAX); metadata->display_primaries[1][1] = av_d2q(mastering_meta->g_y, INT_MAX); metadata->display_primaries[2][0] = av_d2q(mastering_meta->b_x, INT_MAX); metadata->display_primaries[2][1] = av_d2q(mastering_meta->b_y, INT_MAX); metadata->white_point[0] = av_d2q(mastering_meta->white_x, INT_MAX); metadata->white_point[1] = av_d2q(mastering_meta->white_y, INT_MAX); metadata->has_primaries = 1; } } return 0; } static int flv_read_packet(AVFormatContext *s, AVPacket *pkt) { FLVContext *flv = s->priv_data; int ret, i, size, flags; enum FlvTagType type; int stream_type=-1; int64_t next, pos, meta_pos; int64_t dts, pts = AV_NOPTS_VALUE; int av_uninit(channels); int av_uninit(sample_rate); AVStream *st = NULL; int last = -1; int orig_size; int enhanced_flv = 0; uint32_t video_codec_id = 0; retry: /* pkt size is repeated at end. skip it */ pos = avio_tell(s->pb); type = (avio_r8(s->pb) & 0x1F); orig_size = size = avio_rb24(s->pb); flv->sum_flv_tag_size += size + 11LL; dts = avio_rb24(s->pb); dts |= (unsigned)avio_r8(s->pb) << 24; av_log(s, AV_LOG_TRACE, "type:%d, size:%d, last:%d, dts:%"PRId64" pos:%"PRId64"\n", type, size, last, dts, avio_tell(s->pb)); if (avio_feof(s->pb)) return AVERROR_EOF; avio_skip(s->pb, 3); /* stream id, always 0 */ flags = 0; if (flv->validate_next < flv->validate_count) { int64_t validate_pos = flv->validate_index[flv->validate_next].pos; if (pos == validate_pos) { if (FFABS(dts - flv->validate_index[flv->validate_next].dts) <= VALIDATE_INDEX_TS_THRESH) { flv->validate_next++; } else { clear_index_entries(s, validate_pos); flv->validate_count = 0; } } else if (pos > validate_pos) { clear_index_entries(s, validate_pos); flv->validate_count = 0; } } if (size == 0) { ret = FFERROR_REDO; goto leave; } next = size + avio_tell(s->pb); if (type == FLV_TAG_TYPE_AUDIO) { stream_type = FLV_STREAM_TYPE_AUDIO; flags = avio_r8(s->pb); size--; } else if (type == FLV_TAG_TYPE_VIDEO) { stream_type = FLV_STREAM_TYPE_VIDEO; flags = avio_r8(s->pb); video_codec_id = flags & FLV_VIDEO_CODECID_MASK; /* * Reference Enhancing FLV 2023-03-v1.0.0-B.8 * https://github.com/veovera/enhanced-rtmp/blob/main/enhanced-rtmp-v1.pdf * */ enhanced_flv = (flags >> 7) & 1; size--; if (enhanced_flv) { video_codec_id = avio_rb32(s->pb); size -= 4; } if (enhanced_flv && stream_type == FLV_STREAM_TYPE_VIDEO && (flags & FLV_VIDEO_FRAMETYPE_MASK) == FLV_FRAME_VIDEO_INFO_CMD) { int pkt_type = flags & 0x0F; if (pkt_type == PacketTypeMetadata) { int ret = flv_parse_video_color_info(s, st, next); av_log(s, AV_LOG_DEBUG, "enhanced flv parse metadata ret %d and skip\n", ret); } goto skip; } else if ((flags & FLV_VIDEO_FRAMETYPE_MASK) == FLV_FRAME_VIDEO_INFO_CMD) { goto skip; } } else if (type == FLV_TAG_TYPE_META) { stream_type=FLV_STREAM_TYPE_SUBTITLE; if (size > 13 + 1 + 4) { // Header-type metadata stuff int type; meta_pos = avio_tell(s->pb); type = flv_read_metabody(s, next); if (type == 0 && dts == 0 || type < 0) { if (type < 0 && flv->validate_count && flv->validate_index[0].pos > next && flv->validate_index[0].pos - 4 < next) { av_log(s, AV_LOG_WARNING, "Adjusting next position due to index mismatch\n"); next = flv->validate_index[0].pos - 4; } goto skip; } else if (type == TYPE_ONTEXTDATA) { avpriv_request_sample(s, "OnTextData packet"); return flv_data_packet(s, pkt, dts, next); } else if (type == TYPE_ONCAPTION) { return flv_data_packet(s, pkt, dts, next); } else if (type == TYPE_UNKNOWN) { stream_type = FLV_STREAM_TYPE_DATA; } avio_seek(s->pb, meta_pos, SEEK_SET); } } else { av_log(s, AV_LOG_DEBUG, "Skipping flv packet: type %d, size %d, flags %d.\n", type, size, flags); skip: if (avio_seek(s->pb, next, SEEK_SET) != next) { // This can happen if flv_read_metabody above read past // next, on a non-seekable input, and the preceding data has // been flushed out from the IO buffer. av_log(s, AV_LOG_ERROR, "Unable to seek to the next packet\n"); return AVERROR_INVALIDDATA; } ret = FFERROR_REDO; goto leave; } /* skip empty data packets */ if (!size) { ret = FFERROR_REDO; goto leave; } /* now find stream */ for (i = 0; i < s->nb_streams; i++) { st = s->streams[i]; if (stream_type == FLV_STREAM_TYPE_AUDIO) { if (st->codecpar->codec_type == AVMEDIA_TYPE_AUDIO && (s->audio_codec_id || flv_same_audio_codec(st->codecpar, flags))) break; } else if (stream_type == FLV_STREAM_TYPE_VIDEO) { if (st->codecpar->codec_type == AVMEDIA_TYPE_VIDEO && (s->video_codec_id || flv_same_video_codec(st->codecpar, video_codec_id))) break; } else if (stream_type == FLV_STREAM_TYPE_SUBTITLE) { if (st->codecpar->codec_type == AVMEDIA_TYPE_SUBTITLE) break; } else if (stream_type == FLV_STREAM_TYPE_DATA) { if (st->codecpar->codec_type == AVMEDIA_TYPE_DATA) break; } } if (i == s->nb_streams) { static const enum AVMediaType stream_types[] = {AVMEDIA_TYPE_VIDEO, AVMEDIA_TYPE_AUDIO, AVMEDIA_TYPE_SUBTITLE, AVMEDIA_TYPE_DATA}; st = create_stream(s, stream_types[stream_type]); if (!st) return AVERROR(ENOMEM); } av_log(s, AV_LOG_TRACE, "%d %X %d \n", stream_type, flags, st->discard); if (flv->time_pos <= pos) { dts += flv->time_offset; } if ((s->pb->seekable & AVIO_SEEKABLE_NORMAL) && ((flags & FLV_VIDEO_FRAMETYPE_MASK) == FLV_FRAME_KEY || stream_type == FLV_STREAM_TYPE_AUDIO)) av_add_index_entry(st, pos, dts, size, 0, AVINDEX_KEYFRAME); if ((st->discard >= AVDISCARD_NONKEY && !((flags & FLV_VIDEO_FRAMETYPE_MASK) == FLV_FRAME_KEY || stream_type == FLV_STREAM_TYPE_AUDIO)) || (st->discard >= AVDISCARD_BIDIR && ((flags & FLV_VIDEO_FRAMETYPE_MASK) == FLV_FRAME_DISP_INTER && stream_type == FLV_STREAM_TYPE_VIDEO)) || st->discard >= AVDISCARD_ALL) { avio_seek(s->pb, next, SEEK_SET); ret = FFERROR_REDO; goto leave; } // if not streamed and no duration from metadata then seek to end to find // the duration from the timestamps if ((s->pb->seekable & AVIO_SEEKABLE_NORMAL) && (!s->duration || s->duration == AV_NOPTS_VALUE) && !flv->searched_for_end) { int size; const int64_t pos = avio_tell(s->pb); // Read the last 4 bytes of the file, this should be the size of the // previous FLV tag. Use the timestamp of its payload as duration. int64_t fsize = avio_size(s->pb); retry_duration: avio_seek(s->pb, fsize - 4, SEEK_SET); size = avio_rb32(s->pb); if (size > 0 && size < fsize) { // Seek to the start of the last FLV tag at position (fsize - 4 - size) // but skip the byte indicating the type. avio_seek(s->pb, fsize - 3 - size, SEEK_SET); if (size == avio_rb24(s->pb) + 11) { uint32_t ts = avio_rb24(s->pb); ts |= (unsigned)avio_r8(s->pb) << 24; if (ts) s->duration = ts * (int64_t)AV_TIME_BASE / 1000; else if (fsize >= 8 && fsize - 8 >= size) { fsize -= size+4; goto retry_duration; } } } avio_seek(s->pb, pos, SEEK_SET); flv->searched_for_end = 1; } if (stream_type == FLV_STREAM_TYPE_AUDIO) { int bits_per_coded_sample; channels = (flags & FLV_AUDIO_CHANNEL_MASK) == FLV_STEREO ? 2 : 1; sample_rate = 44100 << ((flags & FLV_AUDIO_SAMPLERATE_MASK) >> FLV_AUDIO_SAMPLERATE_OFFSET) >> 3; bits_per_coded_sample = (flags & FLV_AUDIO_SAMPLESIZE_MASK) ? 16 : 8; if (!av_channel_layout_check(&st->codecpar->ch_layout) || !st->codecpar->sample_rate || !st->codecpar->bits_per_coded_sample) { av_channel_layout_default(&st->codecpar->ch_layout, channels); st->codecpar->sample_rate = sample_rate; st->codecpar->bits_per_coded_sample = bits_per_coded_sample; } if (!st->codecpar->codec_id) { flv_set_audio_codec(s, st, st->codecpar, flags & FLV_AUDIO_CODECID_MASK); flv->last_sample_rate = sample_rate = st->codecpar->sample_rate; flv->last_channels = channels = st->codecpar->ch_layout.nb_channels; } else { AVCodecParameters *par = avcodec_parameters_alloc(); if (!par) { ret = AVERROR(ENOMEM); goto leave; } par->sample_rate = sample_rate; par->bits_per_coded_sample = bits_per_coded_sample; flv_set_audio_codec(s, st, par, flags & FLV_AUDIO_CODECID_MASK); sample_rate = par->sample_rate; avcodec_parameters_free(&par); } } else if (stream_type == FLV_STREAM_TYPE_VIDEO) { int ret = flv_set_video_codec(s, st, video_codec_id, 1); if (ret < 0) return ret; size -= ret; } else if (stream_type == FLV_STREAM_TYPE_SUBTITLE) { st->codecpar->codec_id = AV_CODEC_ID_TEXT; } else if (stream_type == FLV_STREAM_TYPE_DATA) { st->codecpar->codec_id = AV_CODEC_ID_NONE; // Opaque AMF data } if (st->codecpar->codec_id == AV_CODEC_ID_AAC || st->codecpar->codec_id == AV_CODEC_ID_H264 || st->codecpar->codec_id == AV_CODEC_ID_MPEG4 || st->codecpar->codec_id == AV_CODEC_ID_HEVC || st->codecpar->codec_id == AV_CODEC_ID_AV1 || st->codecpar->codec_id == AV_CODEC_ID_VP9) { int type = 0; if (enhanced_flv && stream_type == FLV_STREAM_TYPE_VIDEO) { type = flags & 0x0F; } else { type = avio_r8(s->pb); size--; } if (size < 0) { ret = AVERROR_INVALIDDATA; goto leave; } if (enhanced_flv && stream_type == FLV_STREAM_TYPE_VIDEO && flv->meta_color_info_flag) { flv_update_video_color_info(s, st); // update av packet side data flv->meta_color_info_flag = 0; } if (st->codecpar->codec_id == AV_CODEC_ID_H264 || st->codecpar->codec_id == AV_CODEC_ID_MPEG4 || (st->codecpar->codec_id == AV_CODEC_ID_HEVC && type == PacketTypeCodedFrames)) { // sign extension int32_t cts = (avio_rb24(s->pb) + 0xff800000) ^ 0xff800000; pts = av_sat_add64(dts, cts); if (cts < 0) { // dts might be wrong if (!flv->wrong_dts) av_log(s, AV_LOG_WARNING, "Negative cts, previous timestamps might be wrong.\n"); flv->wrong_dts = 1; } else if (FFABS(dts - pts) > 1000*60*15) { av_log(s, AV_LOG_WARNING, "invalid timestamps %"PRId64" %"PRId64"\n", dts, pts); dts = pts = AV_NOPTS_VALUE; } size -= 3; } if (type == 0 && (!st->codecpar->extradata || st->codecpar->codec_id == AV_CODEC_ID_AAC || st->codecpar->codec_id == AV_CODEC_ID_H264 || st->codecpar->codec_id == AV_CODEC_ID_HEVC || st->codecpar->codec_id == AV_CODEC_ID_AV1 || st->codecpar->codec_id == AV_CODEC_ID_VP9)) { AVDictionaryEntry *t; if (st->codecpar->extradata) { if ((ret = flv_queue_extradata(flv, s->pb, stream_type, size)) < 0) return ret; ret = FFERROR_REDO; goto leave; } if ((ret = flv_get_extradata(s, st, size)) < 0) return ret; /* Workaround for buggy Omnia A/XE encoder */ t = av_dict_get(s->metadata, "Encoder", NULL, 0); if (st->codecpar->codec_id == AV_CODEC_ID_AAC && t && !strcmp(t->value, "Omnia A/XE")) st->codecpar->extradata_size = 2; ret = FFERROR_REDO; goto leave; } } /* skip empty data packets */ if (!size) { ret = FFERROR_REDO; goto leave; } ret = av_get_packet(s->pb, pkt, size); if (ret < 0) return ret; pkt->dts = dts; pkt->pts = pts == AV_NOPTS_VALUE ? dts : pts; pkt->stream_index = st->index; pkt->pos = pos; if (flv->new_extradata[stream_type]) { int ret = av_packet_add_side_data(pkt, AV_PKT_DATA_NEW_EXTRADATA, flv->new_extradata[stream_type], flv->new_extradata_size[stream_type]); if (ret >= 0) { flv->new_extradata[stream_type] = NULL; flv->new_extradata_size[stream_type] = 0; } } if (stream_type == FLV_STREAM_TYPE_AUDIO && (sample_rate != flv->last_sample_rate || channels != flv->last_channels)) { flv->last_sample_rate = sample_rate; flv->last_channels = channels; ff_add_param_change(pkt, channels, 0, sample_rate, 0, 0); } if (stream_type == FLV_STREAM_TYPE_AUDIO || (flags & FLV_VIDEO_FRAMETYPE_MASK) == FLV_FRAME_KEY || stream_type == FLV_STREAM_TYPE_SUBTITLE || stream_type == FLV_STREAM_TYPE_DATA) pkt->flags |= AV_PKT_FLAG_KEY; leave: last = avio_rb32(s->pb); if (!flv->trust_datasize) { if (last != orig_size + 11 && last != orig_size + 10 && !avio_feof(s->pb) && (last != orig_size || !last) && last != flv->sum_flv_tag_size && !flv->broken_sizes) { av_log(s, AV_LOG_ERROR, "Packet mismatch %d %d %"PRId64"\n", last, orig_size + 11, flv->sum_flv_tag_size); avio_seek(s->pb, pos + 1, SEEK_SET); ret = resync(s); av_packet_unref(pkt); if (ret >= 0) { goto retry; } } } if (ret >= 0) flv->last_ts = pkt->dts; return ret; } static int flv_read_seek(AVFormatContext *s, int stream_index, int64_t ts, int flags) { FLVContext *flv = s->priv_data; flv->validate_count = 0; return avio_seek_time(s->pb, stream_index, ts, flags); } #define OFFSET(x) offsetof(FLVContext, x) #define VD AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_DECODING_PARAM static const AVOption options[] = { { "flv_metadata", "Allocate streams according to the onMetaData array", OFFSET(trust_metadata), AV_OPT_TYPE_BOOL, { .i64 = 0 }, 0, 1, VD }, { "flv_full_metadata", "Dump full metadata of the onMetadata", OFFSET(dump_full_metadata), AV_OPT_TYPE_BOOL, { .i64 = 0 }, 0, 1, VD }, { "flv_ignore_prevtag", "Ignore the Size of previous tag", OFFSET(trust_datasize), AV_OPT_TYPE_BOOL, { .i64 = 0 }, 0, 1, VD }, { NULL } }; static const AVClass flv_kux_class = { .class_name = "(live) flv/kux demuxer", .item_name = av_default_item_name, .option = options, .version = LIBAVUTIL_VERSION_INT, }; const FFInputFormat ff_flv_demuxer = { .p.name = "flv", .p.long_name = NULL_IF_CONFIG_SMALL("FLV (Flash Video)"), .p.extensions = "flv", .p.priv_class = &flv_kux_class, .priv_data_size = sizeof(FLVContext), .read_probe = flv_probe, .read_header = flv_read_header, .read_packet = flv_read_packet, .read_seek = flv_read_seek, .read_close = flv_read_close, }; const FFInputFormat ff_live_flv_demuxer = { .p.name = "live_flv", .p.long_name = NULL_IF_CONFIG_SMALL("live RTMP FLV (Flash Video)"), .p.extensions = "flv", .p.priv_class = &flv_kux_class, .p.flags = AVFMT_TS_DISCONT, .priv_data_size = sizeof(FLVContext), .read_probe = live_flv_probe, .read_header = flv_read_header, .read_packet = flv_read_packet, .read_seek = flv_read_seek, .read_close = flv_read_close, }; const FFInputFormat ff_kux_demuxer = { .p.name = "kux", .p.long_name = NULL_IF_CONFIG_SMALL("KUX (YouKu)"), .p.extensions = "kux", .p.priv_class = &flv_kux_class, .priv_data_size = sizeof(FLVContext), .read_probe = kux_probe, .read_header = flv_read_header, .read_packet = flv_read_packet, .read_seek = flv_read_seek, .read_close = flv_read_close, };