ffmpeg/libavformat/seek.c

579 lines
20 KiB
C

/*
* Utility functions for seeking for use within FFmpeg format handlers.
*
* Copyright (c) 2009 Ivan Schreter
*
* This file is part of FFmpeg.
*
* FFmpeg is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2.1 of the License, or (at your option) any later version.
*
* FFmpeg is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with FFmpeg; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
#include "seek.h"
#include "libavutil/mem.h"
// NOTE: implementation should be moved here in another patch, to keep patches
// separated.
extern void av_read_frame_flush(AVFormatContext *s);
/**
* Helper structure to store parser state of AVStream.
*/
typedef struct AVStreamState {
// Saved members of AVStream
AVCodecParserContext *parser;
AVPacket cur_pkt;
int64_t last_IP_pts;
int64_t cur_dts;
int64_t reference_dts;
const uint8_t *cur_ptr;
int cur_len;
int probe_packets;
} AVStreamState;
/**
* Helper structure to store parser state of AVFormat.
*/
struct AVParserState {
int64_t fpos; ///< File position at the time of call.
// Saved members of AVFormatContext
AVStream *cur_st; ///< Current stream.
AVPacketList *packet_buffer; ///< Packet buffer of original state.
AVPacketList *raw_packet_buffer; ///< Raw packet buffer of original state.
int raw_packet_buffer_remaining_size; ///< Remaining size available for raw_packet_buffer.
// Saved info for streams.
int nb_streams; ///< Number of streams with stored state.
AVStreamState *stream_states; ///< States of individual streams (array).
};
/**
* Helper structure describing keyframe search state of one stream.
*/
typedef struct {
int64_t pos_lo; ///< Position of the frame with low timestamp in file or INT64_MAX if not found (yet).
int64_t ts_lo; ///< Frame presentation timestamp or same as pos_lo for byte seeking.
int64_t pos_hi; ///< Position of the frame with high timestamp in file or INT64_MAX if not found (yet).
int64_t ts_hi; ///< Frame presentation timestamp or same as pos_hi for byte seeking.
int64_t last_pos; ///< Last known position of a frame, for multi-frame packets.
int64_t term_ts; ///< Termination timestamp (which TS we already read).
AVRational term_ts_tb; ///< Timebase for term_ts.
int64_t first_ts; ///< First packet timestamp in this iteration (to fill term_ts later).
AVRational first_ts_tb;///< Timebase for first_ts.
int terminated; ///< Termination flag for current iteration.
} AVSyncPoint;
/**
* Compare two timestamps exactly, taking into account their respective time bases.
*
* @param ts_a timestamp A.
* @param tb_a time base for timestamp A.
* @param ts_b timestamp B.
* @param tb_b time base for timestamp A.
* @return -1. 0 or 1 if timestamp A is less than, equal or greater than timestamp B.
*/
static int compare_ts(int64_t ts_a, AVRational tb_a, int64_t ts_b, AVRational tb_b)
{
int64_t a, b, res;
if (ts_a == INT64_MIN)
return ts_a < ts_b ? -1 : 0;
if (ts_a == INT64_MAX)
return ts_a > ts_b ? 1 : 0;
if (ts_b == INT64_MIN)
return ts_a > ts_b ? 1 : 0;
if (ts_b == INT64_MAX)
return ts_a < ts_b ? -1 : 0;
a = ts_a * tb_a.num * tb_b.den;
b = ts_b * tb_b.num * tb_a.den;
res = a - b;
if (res == 0)
return 0;
else
return (res >> 63) | 1;
}
/**
* Compute a distance between timestamps.
*
* Distances are only comparable, if same time bases are used for computing
* distances.
*
* @param ts_hi high timestamp.
* @param tb_hi high timestamp time base.
* @param ts_lo low timestamp.
* @param tb_lo low timestamp time base.
* @return representation of distance between high and low timestamps.
*/
static int64_t ts_distance(int64_t ts_hi, AVRational tb_hi, int64_t ts_lo, AVRational tb_lo)
{
int64_t hi, lo;
hi = ts_hi * tb_hi.num * tb_lo.den;
lo = ts_lo * tb_lo.num * tb_hi.den;
return hi - lo;
}
/**
* Partial search for keyframes in multiple streams.
*
* This routine searches for the next lower and next higher timestamp to
* given target timestamp in each stream, starting at current file position
* and ending at position, where all streams have already been examined
* (or when all higher key frames found in first iteration).
*
* This routine is called iteratively with exponential backoff to find lower
* timestamp.
*
* @param s format context.
* @param timestamp target timestamp (or position, if AVSEEK_FLAG_BYTE).
* @param timebase time base for timestamps.
* @param flags seeking flags.
* @param sync array with information per stream.
* @param keyframes_to_find count of keyframes to find in total.
* @param found_lo pointer to count of already found low timestamp keyframes.
* @param found_hi pointer to count of already found high timestamp keyframes.
* @param first_iter flag for first iteration.
*/
static void search_hi_lo_keyframes(AVFormatContext *s,
int64_t timestamp,
AVRational timebase,
int flags,
AVSyncPoint *sync,
int keyframes_to_find,
int *found_lo,
int *found_hi,
int first_iter)
{
AVPacket pkt;
AVSyncPoint *sp;
AVStream *st;
int idx;
int flg;
int terminated_count = 0;
int64_t pos;
int64_t pts, dts; // PTS/DTS from stream
int64_t ts; // PTS in stream-local time base or position for byte seeking
AVRational ts_tb; // Time base of the stream or 1:1 for byte seeking
for (;;) {
if (av_read_frame(s, &pkt) < 0) {
// EOF or error, make sure high flags are set
for (idx = 0; idx < s->nb_streams; ++idx) {
if (s->streams[idx]->discard < AVDISCARD_ALL) {
sp = &sync[idx];
if (sp->pos_hi == INT64_MAX) {
// No high frame exists for this stream
(*found_hi)++;
sp->ts_hi = INT64_MAX;
sp->pos_hi = INT64_MAX - 1;
}
}
}
break;
}
idx = pkt.stream_index;
st = s->streams[idx];
if (st->discard >= AVDISCARD_ALL) {
// This stream is not active, skip packet.
continue;
}
sp = &sync[idx];
flg = pkt.flags;
pos = pkt.pos;
pts = pkt.pts;
dts = pkt.dts;
if (pts == AV_NOPTS_VALUE) {
// Some formats don't provide PTS, only DTS.
pts = dts;
}
av_free_packet(&pkt);
// Multi-frame packets only return position for the very first frame.
// Other frames are read with position == -1. Therefore, we note down
// last known position of a frame and use it if a frame without
// position arrives. In this way, it's possible to seek to proper
// position. Additionally, for parsers not providing position at all,
// an approximation will be used (starting position of this iteration).
if (pos < 0) {
pos = sp->last_pos;
} else {
sp->last_pos = pos;
}
// Evaluate key frames with known TS (or any frames, if AVSEEK_FLAG_ANY set).
if (pts != AV_NOPTS_VALUE && ((flg & PKT_FLAG_KEY) || (flags & AVSEEK_FLAG_ANY))) {
if (flags & AVSEEK_FLAG_BYTE) {
// For byte seeking, use position as timestamp.
ts = pos;
ts_tb.num = 1;
ts_tb.den = 1;
} else {
// Get stream time_base.
ts = pts;
ts_tb = st->time_base;
}
if (sp->first_ts == AV_NOPTS_VALUE) {
// Note down termination timestamp for the next iteration - when
// we encounter a packet with the same timestamp, we will ignore
// any further packets for this stream in next iteration (as they
// are already evaluated).
sp->first_ts = ts;
sp->first_ts_tb = ts_tb;
}
if (sp->term_ts != AV_NOPTS_VALUE && compare_ts(ts, ts_tb, sp->term_ts, sp->term_ts_tb) > 0) {
// We are past the end position from last iteration, ignore packet.
if (!sp->terminated) {
sp->terminated = 1;
++terminated_count;
if (sp->pos_hi == INT64_MAX) {
// No high frame exists for this stream
(*found_hi)++;
sp->ts_hi = INT64_MAX;
sp->pos_hi = INT64_MAX - 1;
}
if (terminated_count == keyframes_to_find)
break; // all terminated, iteration done
}
continue;
}
if (compare_ts(ts, ts_tb, timestamp, timebase) <= 0) {
// Keyframe found before target timestamp.
if (sp->pos_lo == INT64_MAX) {
// Found first keyframe lower than target timestamp.
(*found_lo)++;
sp->ts_lo = ts;
sp->pos_lo = pos;
} else if (sp->ts_lo < ts) {
// Found a better match (closer to target timestamp).
sp->ts_lo = ts;
sp->pos_lo = pos;
}
}
if (compare_ts(ts, ts_tb, timestamp, timebase) >= 0) {
// Keyframe found after target timestamp.
if (sp->pos_hi == INT64_MAX) {
// Found first keyframe higher than target timestamp.
(*found_hi)++;
sp->ts_hi = ts;
sp->pos_hi = pos;
if (*found_hi >= keyframes_to_find && first_iter) {
// We found high frame for all. They may get updated
// to TS closer to target TS in later iterations (which
// will stop at start position of previous iteration).
break;
}
} else if (sp->ts_hi > ts) {
// Found a better match (actually, shouldn't happen).
sp->ts_hi = ts;
sp->pos_hi = pos;
}
}
}
}
// Clean up the parser.
av_read_frame_flush(s);
}
int64_t ff_gen_syncpoint_search(AVFormatContext *s,
int stream_index,
int64_t pos,
int64_t ts_min,
int64_t ts,
int64_t ts_max,
int flags)
{
AVSyncPoint *sync, *sp;
AVStream *st;
int i;
int keyframes_to_find = 0;
int64_t curpos;
int64_t step;
int found_lo = 0, found_hi = 0;
int64_t min_distance, distance;
int64_t min_pos = 0;
int first_iter = 1;
AVRational time_base;
if (flags & AVSEEK_FLAG_BYTE) {
/* For byte seeking, we have exact 1:1 "timestamps" - positions */
time_base.num = 1;
time_base.den = 1;
} else {
if (stream_index >= 0) {
/* We have a reference stream, which time base we use */
st = s->streams[stream_index];
time_base = st->time_base;
} else {
/* No reference stream, use AV_TIME_BASE as reference time base */
time_base.num = 1;
time_base.den = AV_TIME_BASE;
}
}
// Initialize syncpoint structures for each stream.
sync = (AVSyncPoint*) av_malloc(s->nb_streams * sizeof(AVSyncPoint));
if (!sync) {
// cannot allocate helper structure
return -1;
}
for (i = 0; i < s->nb_streams; ++i) {
st = s->streams[i];
sp = &sync[i];
sp->pos_lo = INT64_MAX;
sp->ts_lo = INT64_MAX;
sp->pos_hi = INT64_MAX;
sp->ts_hi = INT64_MAX;
sp->terminated = 0;
sp->first_ts = AV_NOPTS_VALUE;
sp->term_ts = ts_max;
sp->term_ts_tb = time_base;
sp->last_pos = pos;
st->cur_dts = AV_NOPTS_VALUE;
if (st->discard < AVDISCARD_ALL)
++keyframes_to_find;
}
if (keyframes_to_find == 0) {
// No stream active, error.
av_free(sync);
return -1;
}
// Find keyframes in all active streams with timestamp/position just before
// and just after requested timestamp/position.
step = 1024;
curpos = pos;
for (;;) {
url_fseek(s->pb, curpos, SEEK_SET);
search_hi_lo_keyframes(s,
ts, time_base,
flags,
sync,
keyframes_to_find,
&found_lo, &found_hi,
first_iter);
if (found_lo == keyframes_to_find && found_hi == keyframes_to_find)
break; // have all keyframes we wanted
if (curpos == 0)
break; // cannot go back anymore
curpos = pos - step;
if (curpos < 0)
curpos = 0;
step *= 2;
// switch termination positions
for (i = 0; i < s->nb_streams; ++i) {
st = s->streams[i];
st->cur_dts = AV_NOPTS_VALUE;
sp = &sync[i];
if (sp->first_ts != AV_NOPTS_VALUE) {
sp->term_ts = sp->first_ts;
sp->term_ts_tb = sp->first_ts_tb;
sp->first_ts = AV_NOPTS_VALUE;
}
sp->terminated = 0;
sp->last_pos = curpos;
}
first_iter = 0;
}
// Find actual position to start decoding so that decoder synchronizes
// closest to ts and between ts_min and ts_max.
pos = INT64_MAX;
for (i = 0; i < s->nb_streams; ++i) {
st = s->streams[i];
if (st->discard < AVDISCARD_ALL) {
sp = &sync[i];
min_distance = INT64_MAX;
// Find timestamp closest to requested timestamp within min/max limits.
if (sp->pos_lo != INT64_MAX
&& compare_ts(ts_min, time_base, sp->ts_lo, st->time_base) <= 0
&& compare_ts(sp->ts_lo, st->time_base, ts_max, time_base) <= 0) {
// low timestamp is in range
min_distance = ts_distance(ts, time_base, sp->ts_lo, st->time_base);
min_pos = sp->pos_lo;
}
if (sp->pos_hi != INT64_MAX
&& compare_ts(ts_min, time_base, sp->ts_hi, st->time_base) <= 0
&& compare_ts(sp->ts_hi, st->time_base, ts_max, time_base) <= 0) {
// high timestamp is in range, check distance
distance = ts_distance(sp->ts_hi, st->time_base, ts, time_base);
if (distance < min_distance) {
min_distance = distance;
min_pos = sp->pos_hi;
}
}
if (min_distance == INT64_MAX) {
// no timestamp is in range, cannot seek
av_free(sync);
return -1;
}
if (min_pos < pos)
pos = min_pos;
}
}
url_fseek(s->pb, pos, SEEK_SET);
av_free(sync);
return pos;
}
AVParserState *ff_store_parser_state(AVFormatContext *s)
{
int i;
AVStream *st;
AVStreamState *ss;
AVParserState *state = (AVParserState*) av_malloc(sizeof(AVParserState));
if (!state)
return NULL;
state->stream_states = (AVStreamState*) av_malloc(sizeof(AVStreamState) * s->nb_streams);
if (!state->stream_states) {
av_free(state);
return NULL;
}
state->fpos = url_ftell(s->pb);
// copy context structures
state->cur_st = s->cur_st;
state->packet_buffer = s->packet_buffer;
state->raw_packet_buffer = s->raw_packet_buffer;
state->raw_packet_buffer_remaining_size = s->raw_packet_buffer_remaining_size;
s->cur_st = NULL;
s->packet_buffer = NULL;
s->raw_packet_buffer = NULL;
s->raw_packet_buffer_remaining_size = RAW_PACKET_BUFFER_SIZE;
// copy stream structures
state->nb_streams = s->nb_streams;
for (i = 0; i < s->nb_streams; i++) {
st = s->streams[i];
ss = &state->stream_states[i];
ss->parser = st->parser;
ss->last_IP_pts = st->last_IP_pts;
ss->cur_dts = st->cur_dts;
ss->reference_dts = st->reference_dts;
ss->cur_ptr = st->cur_ptr;
ss->cur_len = st->cur_len;
ss->probe_packets = st->probe_packets;
ss->cur_pkt = st->cur_pkt;
st->parser = NULL;
st->last_IP_pts = AV_NOPTS_VALUE;
st->cur_dts = AV_NOPTS_VALUE;
st->reference_dts = AV_NOPTS_VALUE;
st->cur_ptr = NULL;
st->cur_len = 0;
st->probe_packets = MAX_PROBE_PACKETS;
av_init_packet(&st->cur_pkt);
}
return state;
}
void ff_restore_parser_state(AVFormatContext *s, AVParserState *state)
{
int i;
AVStream *st;
AVStreamState *ss;
av_read_frame_flush(s);
if (!state)
return;
url_fseek(s->pb, state->fpos, SEEK_SET);
// copy context structures
s->cur_st = state->cur_st;
s->packet_buffer = state->packet_buffer;
s->raw_packet_buffer = state->raw_packet_buffer;
s->raw_packet_buffer_remaining_size = state->raw_packet_buffer_remaining_size;
// copy stream structures
for (i = 0; i < state->nb_streams; i++) {
st = s->streams[i];
ss = &state->stream_states[i];
st->parser = ss->parser;
st->last_IP_pts = ss->last_IP_pts;
st->cur_dts = ss->cur_dts;
st->reference_dts = ss->reference_dts;
st->cur_ptr = ss->cur_ptr;
st->cur_len = ss->cur_len;
st->probe_packets = ss->probe_packets;
st->cur_pkt = ss->cur_pkt;
}
av_free(state->stream_states);
av_free(state);
}
static void free_packet_list(AVPacketList *pktl)
{
AVPacketList *cur;
while (pktl) {
cur = pktl;
pktl = cur->next;
av_free_packet(&cur->pkt);
av_free(cur);
}
}
void ff_free_parser_state(AVFormatContext *s, AVParserState *state)
{
int i;
AVStreamState *ss;
if (!state)
return;
for (i = 0; i < state->nb_streams; i++) {
ss = &state->stream_states[i];
if (ss->parser)
av_parser_close(ss->parser);
av_free_packet(&ss->cur_pkt);
}
free_packet_list(state->packet_buffer);
free_packet_list(state->raw_packet_buffer);
av_free(state->stream_states);
av_free(state);
}