mpv/libmpcodecs/vd_ffmpeg.c

840 lines
29 KiB
C

/*
* This file is part of MPlayer.
*
* MPlayer is free software; you can redistribute it and/or modify
* it under the terms of the GNU General Public License as published by
* the Free Software Foundation; either version 2 of the License, or
* (at your option) any later version.
*
* MPlayer is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*
* You should have received a copy of the GNU General Public License along
* with MPlayer; if not, write to the Free Software Foundation, Inc.,
* 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
*/
#include <stdio.h>
#include <stdlib.h>
#include <assert.h>
#include <time.h>
#include <stdbool.h>
#include <libavutil/common.h>
#include <libavutil/opt.h>
#include <libavutil/intreadwrite.h>
#include "talloc.h"
#include "config.h"
#include "mp_msg.h"
#include "options.h"
#include "av_opts.h"
#include "mpbswap.h"
#include "fmt-conversion.h"
#include "vd.h"
#include "img_format.h"
#include "libmpdemux/stheader.h"
#include "libmpdemux/demux_packet.h"
#include "codec-cfg.h"
#include "osdep/numcores.h"
static const vd_info_t info = {
"FFmpeg's libavcodec codec family",
"ffmpeg",
"A'rpi",
"A'rpi, Michael, Alex",
"native codecs"
};
#include "libavcodec/avcodec.h"
#if AVPALETTE_SIZE > 1024
#error palette too large, adapt libmpcodecs/vf.c:vf_get_image
#endif
typedef struct {
AVCodecContext *avctx;
AVFrame *pic;
enum PixelFormat pix_fmt;
int do_slices;
int do_dr1;
int vo_initialized;
int best_csp;
int qp_stat[32];
double qp_sum;
double inv_qp_sum;
int ip_count;
int b_count;
AVRational last_sample_aspect_ratio;
enum AVDiscard skip_frame;
} vd_ffmpeg_ctx;
#include "m_option.h"
static int get_buffer(AVCodecContext *avctx, AVFrame *pic);
static void release_buffer(AVCodecContext *avctx, AVFrame *pic);
static void draw_slice(struct AVCodecContext *s, const AVFrame *src,
int offset[4], int y, int type, int height);
static enum PixelFormat get_format(struct AVCodecContext *avctx,
const enum PixelFormat *pix_fmt);
static void uninit(struct sh_video *sh);
const m_option_t lavc_decode_opts_conf[] = {
OPT_INTRANGE("bug", lavc_param.workaround_bugs, 0, -1, 999999),
OPT_FLAG_ON("gray", lavc_param.gray, 0),
OPT_INTRANGE("idct", lavc_param.idct_algo, 0, 0, 99),
OPT_INTRANGE("ec", lavc_param.error_concealment, 0, 0, 99),
OPT_FLAG_ON("vstats", lavc_param.vstats, 0),
OPT_INTRANGE("debug", lavc_param.debug, 0, 0, 9999999),
OPT_INTRANGE("vismv", lavc_param.vismv, 0, 0, 9999999),
OPT_INTRANGE("st", lavc_param.skip_top, 0, 0, 999),
OPT_INTRANGE("sb", lavc_param.skip_bottom, 0, 0, 999),
OPT_FLAG_CONSTANTS("fast", lavc_param.fast, 0, 0, CODEC_FLAG2_FAST),
OPT_STRING("lowres", lavc_param.lowres_str, 0),
OPT_STRING("skiploopfilter", lavc_param.skip_loop_filter_str, 0),
OPT_STRING("skipidct", lavc_param.skip_idct_str, 0),
OPT_STRING("skipframe", lavc_param.skip_frame_str, 0),
OPT_INTRANGE("threads", lavc_param.threads, 0, 0, 16),
OPT_FLAG_CONSTANTS("bitexact", lavc_param.bitexact, 0, 0, CODEC_FLAG_BITEXACT),
OPT_STRING("o", lavc_param.avopt, 0),
{NULL, NULL, 0, 0, 0, 0, NULL}
};
static enum AVDiscard str2AVDiscard(char *str)
{
if (!str) return AVDISCARD_DEFAULT;
if (strcasecmp(str, "none" ) == 0) return AVDISCARD_NONE;
if (strcasecmp(str, "default") == 0) return AVDISCARD_DEFAULT;
if (strcasecmp(str, "nonref" ) == 0) return AVDISCARD_NONREF;
if (strcasecmp(str, "bidir" ) == 0) return AVDISCARD_BIDIR;
if (strcasecmp(str, "nonkey" ) == 0) return AVDISCARD_NONKEY;
if (strcasecmp(str, "all" ) == 0) return AVDISCARD_ALL;
mp_msg(MSGT_DECVIDEO, MSGL_ERR, "Unknown discard value %s\n", str);
return AVDISCARD_DEFAULT;
}
static int init(sh_video_t *sh)
{
struct lavc_param *lavc_param = &sh->opts->lavc_param;
AVCodecContext *avctx;
vd_ffmpeg_ctx *ctx;
AVCodec *lavc_codec;
int do_vis_debug = lavc_param->vismv ||
(lavc_param->debug & (FF_DEBUG_VIS_MB_TYPE | FF_DEBUG_VIS_QP));
ctx = sh->context = talloc_zero(NULL, vd_ffmpeg_ctx);
lavc_codec = avcodec_find_decoder_by_name(sh->codec->dll);
if (!lavc_codec) {
mp_tmsg(MSGT_DECVIDEO, MSGL_ERR,
"Cannot find codec '%s' in libavcodec...\n", sh->codec->dll);
uninit(sh);
return 0;
}
if (sh->opts->vd_use_slices
&& (lavc_codec->capabilities & CODEC_CAP_DRAW_HORIZ_BAND)
&& !do_vis_debug)
ctx->do_slices = 1;
if (lavc_codec->capabilities & CODEC_CAP_DR1 && !do_vis_debug
&& lavc_codec->id != CODEC_ID_H264
&& lavc_codec->id != CODEC_ID_INTERPLAY_VIDEO
&& lavc_codec->id != CODEC_ID_ROQ && lavc_codec->id != CODEC_ID_VP8
&& lavc_codec->id != CODEC_ID_LAGARITH)
ctx->do_dr1 = 1;
ctx->ip_count = ctx->b_count = 0;
ctx->pic = avcodec_alloc_frame();
ctx->avctx = avcodec_alloc_context3(lavc_codec);
avctx = ctx->avctx;
avctx->opaque = sh;
avctx->codec_type = AVMEDIA_TYPE_VIDEO;
avctx->codec_id = lavc_codec->id;
if (lavc_codec->capabilities & CODEC_CAP_HWACCEL // XvMC
|| lavc_codec->capabilities & CODEC_CAP_HWACCEL_VDPAU) {
ctx->do_dr1 = true;
ctx->do_slices = true;
lavc_param->threads = 1;
avctx->get_format = get_format;
avctx->get_buffer = get_buffer;
avctx->release_buffer = release_buffer;
avctx->reget_buffer = get_buffer;
avctx->draw_horiz_band = draw_slice;
if (lavc_codec->capabilities & CODEC_CAP_HWACCEL)
mp_msg(MSGT_DECVIDEO, MSGL_V, "[VD_FFMPEG] XVMC-accelerated "
"MPEG-2.\n");
if (lavc_codec->capabilities & CODEC_CAP_HWACCEL_VDPAU)
mp_msg(MSGT_DECVIDEO, MSGL_V, "[VD_FFMPEG] VDPAU hardware "
"decoding.\n");
avctx->slice_flags = SLICE_FLAG_CODED_ORDER | SLICE_FLAG_ALLOW_FIELD;
}
if (lavc_param->threads == 0) {
int threads = default_thread_count();
if (threads < 1) {
mp_msg(MSGT_DECVIDEO, MSGL_WARN, "[VD_FFMPEG] Could not determine "
"thread count to use, defaulting to 1.\n");
threads = 1;
}
threads = FFMIN(threads, 16);
lavc_param->threads = threads;
}
/* Our get_buffer and draw_horiz_band callbacks are not safe to call
* from other threads. */
if (lavc_param->threads > 1) {
ctx->do_dr1 = false;
ctx->do_slices = false;
mp_tmsg(MSGT_DECVIDEO, MSGL_INFO, "Asking decoder to use "
"%d threads if supported.\n", lavc_param->threads);
}
if (ctx->do_dr1) {
avctx->flags |= CODEC_FLAG_EMU_EDGE;
avctx->get_buffer = get_buffer;
avctx->release_buffer = release_buffer;
avctx->reget_buffer = get_buffer;
}
avctx->flags |= lavc_param->bitexact;
avctx->coded_width = sh->disp_w;
avctx->coded_height = sh->disp_h;
avctx->workaround_bugs = lavc_param->workaround_bugs;
if (lavc_param->gray)
avctx->flags |= CODEC_FLAG_GRAY;
avctx->flags2 |= lavc_param->fast;
avctx->codec_tag = sh->format;
avctx->stream_codec_tag = sh->video.fccHandler;
avctx->idct_algo = lavc_param->idct_algo;
avctx->error_concealment = lavc_param->error_concealment;
avctx->debug = lavc_param->debug;
if (lavc_param->debug)
av_log_set_level(AV_LOG_DEBUG);
avctx->debug_mv = lavc_param->vismv;
avctx->skip_top = lavc_param->skip_top;
avctx->skip_bottom = lavc_param->skip_bottom;
if (lavc_param->lowres_str != NULL) {
int lowres, lowres_w;
sscanf(lavc_param->lowres_str, "%d,%d", &lowres, &lowres_w);
if (lowres < 1 || lowres > 16 ||
lowres_w > 0 && avctx->width < lowres_w)
lowres = 0;
avctx->lowres = lowres;
}
avctx->skip_loop_filter = str2AVDiscard(lavc_param->skip_loop_filter_str);
avctx->skip_idct = str2AVDiscard(lavc_param->skip_idct_str);
avctx->skip_frame = str2AVDiscard(lavc_param->skip_frame_str);
if (lavc_param->avopt) {
if (parse_avopts(avctx, lavc_param->avopt) < 0) {
mp_msg(MSGT_DECVIDEO, MSGL_ERR,
"Your options /%s/ look like gibberish to me pal\n",
lavc_param->avopt);
uninit(sh);
return 0;
}
}
// Do this after the above avopt handling in case it changes values
ctx->skip_frame = avctx->skip_frame;
mp_dbg(MSGT_DECVIDEO, MSGL_DBG2,
"libavcodec.size: %d x %d\n", avctx->width, avctx->height);
switch (sh->format) {
case mmioFOURCC('S','V','Q','3'):
/* SVQ3 extradata can show up as sh->ImageDesc if demux_mov is used, or
* in the phony AVI header if demux_lavf is used. The first case is
* handled here; the second case falls through to the next section. */
if (sh->ImageDesc) {
avctx->extradata_size = (*(int *)sh->ImageDesc) - sizeof(int);
avctx->extradata = av_mallocz(avctx->extradata_size +
FF_INPUT_BUFFER_PADDING_SIZE);
memcpy(avctx->extradata, ((int *)sh->ImageDesc) + 1,
avctx->extradata_size);
break;
}
/* fallthrough */
case mmioFOURCC('A','V','R','n'):
case mmioFOURCC('M','J','P','G'):
/* AVRn stores huffman table in AVI header */
/* Pegasus MJPEG stores it also in AVI header, but it uses the common
* MJPG fourcc :( */
if (!sh->bih || sh->bih->biSize <= sizeof(*sh->bih))
break;
av_opt_set_int(avctx, "extern_huff", 1, AV_OPT_SEARCH_CHILDREN);
avctx->extradata_size = sh->bih->biSize - sizeof(*sh->bih);
avctx->extradata = av_mallocz(avctx->extradata_size +
FF_INPUT_BUFFER_PADDING_SIZE);
memcpy(avctx->extradata, sh->bih + 1, avctx->extradata_size);
break;
case mmioFOURCC('R','V','1','0'):
case mmioFOURCC('R','V','1','3'):
case mmioFOURCC('R','V','2','0'):
case mmioFOURCC('R','V','3','0'):
case mmioFOURCC('R','V','4','0'):
if (sh->bih->biSize < sizeof(*sh->bih) + 8) {
// only 1 packet per frame & sub_id from fourcc
avctx->extradata_size = 8;
avctx->extradata = av_mallocz(avctx->extradata_size +
FF_INPUT_BUFFER_PADDING_SIZE);
((uint32_t *)avctx->extradata)[0] = 0;
((uint32_t *)avctx->extradata)[1] =
sh->format == mmioFOURCC('R','V','1','3') ?
0x10003001 : 0x10000000;
} else {
// has extra slice header (demux_rm or rm->avi streamcopy)
avctx->extradata_size = sh->bih->biSize - sizeof(*sh->bih);
avctx->extradata = av_mallocz(avctx->extradata_size +
FF_INPUT_BUFFER_PADDING_SIZE);
memcpy(avctx->extradata, sh->bih + 1, avctx->extradata_size);
}
break;
default:
if (!sh->bih || sh->bih->biSize <= sizeof(*sh->bih))
break;
avctx->extradata_size = sh->bih->biSize - sizeof(*sh->bih);
avctx->extradata = av_mallocz(avctx->extradata_size +
FF_INPUT_BUFFER_PADDING_SIZE);
memcpy(avctx->extradata, sh->bih + 1, avctx->extradata_size);
break;
}
if (sh->bih)
avctx->bits_per_coded_sample = sh->bih->biBitCount;
avctx->thread_count = lavc_param->threads;
/* open it */
if (avcodec_open2(avctx, lavc_codec, NULL) < 0) {
mp_tmsg(MSGT_DECVIDEO, MSGL_ERR, "Could not open codec.\n");
uninit(sh);
return 0;
}
mp_msg(MSGT_DECVIDEO, MSGL_V, "INFO: libavcodec init OK!\n");
return 1; //mpcodecs_config_vo(sh, sh->disp_w, sh->disp_h, IMGFMT_YV12);
}
static void uninit(sh_video_t *sh)
{
vd_ffmpeg_ctx *ctx = sh->context;
AVCodecContext *avctx = ctx->avctx;
if (sh->opts->lavc_param.vstats && avctx->coded_frame) {
for (int i = 1; i < 32; i++)
mp_msg(MSGT_DECVIDEO, MSGL_INFO,
"QP: %d, count: %d\n", i, ctx->qp_stat[i]);
mp_tmsg(MSGT_DECVIDEO, MSGL_INFO, "[VD_FFMPEG] Arithmetic mean of QP: "
"%2.4f, Harmonic mean of QP: %2.4f\n",
ctx->qp_sum / avctx->coded_frame->coded_picture_number,
1.0 / (ctx->inv_qp_sum / avctx->coded_frame->coded_picture_number));
}
if (avctx) {
if (avctx->codec && avcodec_close(avctx) < 0)
mp_tmsg(MSGT_DECVIDEO, MSGL_ERR, "Could not close codec.\n");
av_freep(&avctx->extradata);
av_freep(&avctx->slice_offset);
}
av_freep(&avctx);
av_freep(&ctx->pic);
talloc_free(ctx);
}
static void draw_slice(struct AVCodecContext *s,
const AVFrame *src, int offset[4],
int y, int type, int height)
{
sh_video_t *sh = s->opaque;
uint8_t *source[MP_MAX_PLANES] = {
src->data[0] + offset[0], src->data[1] + offset[1],
src->data[2] + offset[2]
};
int strides[MP_MAX_PLANES] = {
src->linesize[0], src->linesize[1], src->linesize[2]
};
if (height < 0) {
int i;
height = -height;
y -= height;
for (i = 0; i < MP_MAX_PLANES; i++) {
strides[i] = -strides[i];
source[i] -= strides[i];
}
}
if (y < sh->disp_h) {
height = FFMIN(height, sh->disp_h - y);
mpcodecs_draw_slice(sh, source, strides, sh->disp_w, height, 0, y);
}
}
static int init_vo(sh_video_t *sh, enum PixelFormat pix_fmt)
{
vd_ffmpeg_ctx *ctx = sh->context;
AVCodecContext *avctx = ctx->avctx;
float aspect = av_q2d(avctx->sample_aspect_ratio) *
avctx->width / avctx->height;
int width, height;
width = avctx->width;
height = avctx->height;
// HACK!
// if sh->ImageDesc is non-NULL, it means we decode QuickTime(tm) video.
// use dimensions from BIH to avoid black borders at the right and bottom.
if (sh->bih && sh->ImageDesc) {
width = sh->bih->biWidth >> avctx->lowres;
height = sh->bih->biHeight >> avctx->lowres;
}
/* Reconfiguring filter/VO chain may invalidate direct rendering buffers
* we have allocated for libavcodec (including the VDPAU HW decoding
* case). Is it guaranteed that the code below only triggers in a situation
* with no busy direct rendering buffers for reference frames?
*/
if (av_cmp_q(avctx->sample_aspect_ratio, ctx->last_sample_aspect_ratio) ||
width != sh->disp_w || height != sh->disp_h ||
pix_fmt != ctx->pix_fmt || !ctx->vo_initialized) {
ctx->vo_initialized = 0;
mp_msg(MSGT_DECVIDEO, MSGL_V, "[ffmpeg] aspect_ratio: %f\n", aspect);
// Do not overwrite s->aspect on the first call, so that a container
// aspect if available is preferred.
// But set it even if the sample aspect did not change, since a
// resolution change can cause an aspect change even if the
// _sample_ aspect is unchanged.
if (sh->aspect == 0 || ctx->last_sample_aspect_ratio.den)
sh->aspect = aspect;
ctx->last_sample_aspect_ratio = avctx->sample_aspect_ratio;
sh->disp_w = width;
sh->disp_h = height;
ctx->pix_fmt = pix_fmt;
ctx->best_csp = pixfmt2imgfmt(pix_fmt);
const unsigned int *supported_fmts;
if (ctx->best_csp == IMGFMT_YV12)
supported_fmts = (const unsigned int[]){
IMGFMT_YV12, IMGFMT_I420, IMGFMT_IYUV, 0xffffffff
};
else if (ctx->best_csp == IMGFMT_422P)
supported_fmts = (const unsigned int[]){
IMGFMT_422P, IMGFMT_YV12, IMGFMT_I420, IMGFMT_IYUV, 0xffffffff
};
else
supported_fmts = (const unsigned int[]){ctx->best_csp, 0xffffffff};
if (!mpcodecs_config_vo2(sh, sh->disp_w, sh->disp_h, supported_fmts,
ctx->best_csp))
return -1;
ctx->vo_initialized = 1;
}
return 0;
}
static int get_buffer(AVCodecContext *avctx, AVFrame *pic)
{
sh_video_t *sh = avctx->opaque;
vd_ffmpeg_ctx *ctx = sh->context;
mp_image_t *mpi = NULL;
int flags = MP_IMGFLAG_ACCEPT_ALIGNED_STRIDE |
MP_IMGFLAG_PREFER_ALIGNED_STRIDE;
int type = MP_IMGTYPE_IPB;
int width = avctx->width;
int height = avctx->height;
// special case to handle reget_buffer without buffer hints
if (pic->opaque && pic->data[0] && !pic->buffer_hints)
return 0;
avcodec_align_dimensions(avctx, &width, &height);
if (pic->buffer_hints) {
mp_msg(MSGT_DECVIDEO, MSGL_DBG2, "Buffer hints: %u\n",
pic->buffer_hints);
type = MP_IMGTYPE_TEMP;
if (pic->buffer_hints & FF_BUFFER_HINTS_READABLE)
flags |= MP_IMGFLAG_READABLE;
if (pic->buffer_hints & FF_BUFFER_HINTS_PRESERVE) {
type = MP_IMGTYPE_STATIC;
flags |= MP_IMGFLAG_PRESERVE;
}
if (pic->buffer_hints & FF_BUFFER_HINTS_REUSABLE) {
type = MP_IMGTYPE_STATIC;
flags |= MP_IMGFLAG_PRESERVE;
}
flags |= ctx->do_slices ? MP_IMGFLAG_DRAW_CALLBACK : 0;
mp_msg(MSGT_DECVIDEO, MSGL_DBG2,
type == MP_IMGTYPE_STATIC ? "using STATIC\n" : "using TEMP\n");
} else {
if (!pic->reference) {
ctx->b_count++;
flags |= ctx->do_slices ? MP_IMGFLAG_DRAW_CALLBACK : 0;
} else {
ctx->ip_count++;
flags |= MP_IMGFLAG_PRESERVE | MP_IMGFLAG_READABLE
| (ctx->do_slices ? MP_IMGFLAG_DRAW_CALLBACK : 0);
}
}
if (init_vo(sh, avctx->pix_fmt) < 0) {
avctx->release_buffer = avcodec_default_release_buffer;
avctx->get_buffer = avcodec_default_get_buffer;
avctx->reget_buffer = avcodec_default_reget_buffer;
if (pic->data[0])
release_buffer(avctx, pic);
return avctx->get_buffer(avctx, pic);
}
if (IMGFMT_IS_HWACCEL(ctx->best_csp))
type = MP_IMGTYPE_NUMBERED | (0xffff << 16);
else if (!pic->buffer_hints) {
if (ctx->b_count > 1 || ctx->ip_count > 2) {
mp_tmsg(MSGT_DECVIDEO, MSGL_WARN, "[VD_FFMPEG] DRI failure.\n");
ctx->do_dr1 = 0; //FIXME
avctx->get_buffer = avcodec_default_get_buffer;
avctx->reget_buffer = avcodec_default_reget_buffer;
if (pic->data[0])
release_buffer(avctx, pic);
return avctx->get_buffer(avctx, pic);
}
if (avctx->has_b_frames || ctx->b_count)
type = MP_IMGTYPE_IPB;
else
type = MP_IMGTYPE_IP;
mp_msg(MSGT_DECVIDEO, MSGL_DBG2,
type == MP_IMGTYPE_IPB ? "using IPB\n" : "using IP\n");
}
if (ctx->best_csp == IMGFMT_RGB8 || ctx->best_csp == IMGFMT_BGR8)
flags |= MP_IMGFLAG_RGB_PALETTE;
mpi = mpcodecs_get_image(sh, type, flags, width, height);
if (!mpi)
return -1;
// ok, let's see what did we get:
if (mpi->flags & MP_IMGFLAG_DRAW_CALLBACK &&
!(mpi->flags & MP_IMGFLAG_DIRECT)) {
// nice, filter/vo likes draw_callback :)
avctx->draw_horiz_band = draw_slice;
} else
avctx->draw_horiz_band = NULL;
if (IMGFMT_IS_HWACCEL(mpi->imgfmt))
avctx->draw_horiz_band = draw_slice;
pic->data[0] = mpi->planes[0];
pic->data[1] = mpi->planes[1];
pic->data[2] = mpi->planes[2];
pic->data[3] = mpi->planes[3];
/* Note: some (many) codecs in libavcodec require
* linesize[1] == linesize[2] and no changes between frames.
* Lavc will check that and die with an error message if it's not true.
*/
pic->linesize[0] = mpi->stride[0];
pic->linesize[1] = mpi->stride[1];
pic->linesize[2] = mpi->stride[2];
pic->linesize[3] = mpi->stride[3];
pic->opaque = mpi;
pic->type = FF_BUFFER_TYPE_USER;
/* The libavcodec reordered_opaque functionality is implemented by
* a similar copy in avcodec_default_get_buffer() and without a
* workaround like this it'd stop working when a custom buffer
* callback is used.
*/
pic->reordered_opaque = avctx->reordered_opaque;
return 0;
}
static void release_buffer(struct AVCodecContext *avctx, AVFrame *pic)
{
mp_image_t *mpi = pic->opaque;
sh_video_t *sh = avctx->opaque;
vd_ffmpeg_ctx *ctx = sh->context;
if (ctx->ip_count <= 2 && ctx->b_count <= 1) {
if (mpi->flags & MP_IMGFLAG_PRESERVE)
ctx->ip_count--;
else
ctx->b_count--;
}
if (mpi) {
// Palette support: free palette buffer allocated in get_buffer
if (mpi->bpp == 8)
av_freep(&mpi->planes[1]);
// release mpi (in case MPI_IMGTYPE_NUMBERED is used, e.g. for VDPAU)
mpi->usage_count--;
}
if (pic->type != FF_BUFFER_TYPE_USER) {
avcodec_default_release_buffer(avctx, pic);
return;
}
for (int i = 0; i < 4; i++)
pic->data[i] = NULL;
}
static av_unused void swap_palette(void *pal)
{
int i;
uint32_t *p = pal;
for (i = 0; i < AVPALETTE_COUNT; i++)
p[i] = le2me_32(p[i]);
}
static struct mp_image *decode(struct sh_video *sh, struct demux_packet *packet,
void *data, int len, int flags,
double *reordered_pts)
{
int got_picture = 0;
int ret;
vd_ffmpeg_ctx *ctx = sh->context;
AVFrame *pic = ctx->pic;
AVCodecContext *avctx = ctx->avctx;
struct lavc_param *lavc_param = &sh->opts->lavc_param;
mp_image_t *mpi = NULL;
int dr1 = ctx->do_dr1;
AVPacket pkt;
if (!dr1)
avctx->draw_horiz_band = NULL;
if (flags & 2)
avctx->skip_frame = AVDISCARD_ALL;
else if (flags & 1)
avctx->skip_frame = AVDISCARD_NONREF;
else
avctx->skip_frame = ctx->skip_frame;
av_init_packet(&pkt);
pkt.data = data;
pkt.size = len;
// HACK: make PNGs decode normally instead of as CorePNG delta frames
pkt.flags = AV_PKT_FLAG_KEY;
if (packet && packet->avpacket) {
pkt.side_data = packet->avpacket->side_data;
pkt.side_data_elems = packet->avpacket->side_data_elems;
}
// The avcodec opaque field stupidly supports only int64_t type
union pts { int64_t i; double d; };
avctx->reordered_opaque = (union pts){.d = *reordered_pts}.i;
ret = avcodec_decode_video2(avctx, pic, &got_picture, &pkt);
*reordered_pts = (union pts){.i = pic->reordered_opaque}.d;
dr1 = ctx->do_dr1;
if (ret < 0)
mp_msg(MSGT_DECVIDEO, MSGL_WARN, "Error while decoding frame!\n");
//-- vstats generation
while (lavc_param->vstats) { // always one time loop
static FILE *fvstats = NULL;
char filename[20];
static long long int all_len = 0;
static int frame_number = 0;
static double all_frametime = 0.0;
AVFrame *pic = avctx->coded_frame;
double quality = 0.0;
if (!pic)
break;
if (!fvstats) {
time_t today2;
struct tm *today;
today2 = time(NULL);
today = localtime(&today2);
sprintf(filename, "vstats_%02d%02d%02d.log", today->tm_hour,
today->tm_min, today->tm_sec);
fvstats = fopen(filename, "w");
if (!fvstats) {
perror("fopen");
lavc_param->vstats = 0; // disable block
break;
/*exit(1);*/
}
}
// average MB quantizer
{
int x, y;
int w = ((avctx->width << avctx->lowres) + 15) >> 4;
int h = ((avctx->height << avctx->lowres) + 15) >> 4;
int8_t *q = pic->qscale_table;
for (y = 0; y < h; y++) {
for (x = 0; x < w; x++)
quality += (double)*(q + x);
q += pic->qstride;
}
quality /= w * h;
}
all_len += len;
all_frametime += sh->frametime;
fprintf(fvstats, "frame= %5d q= %2.2f f_size= %6d s_size= %8.0fkB ",
++frame_number, quality, len, (double)all_len / 1024);
fprintf(fvstats, "time= %0.3f br= %7.1fkbits/s avg_br= %7.1fkbits/s ",
all_frametime, (double)(len * 8) / sh->frametime / 1000.0,
(double)(all_len * 8) / all_frametime / 1000.0);
switch (pic->pict_type) {
case AV_PICTURE_TYPE_I:
fprintf(fvstats, "type= I\n");
break;
case AV_PICTURE_TYPE_P:
fprintf(fvstats, "type= P\n");
break;
case AV_PICTURE_TYPE_S:
fprintf(fvstats, "type= S\n");
break;
case AV_PICTURE_TYPE_B:
fprintf(fvstats, "type= B\n");
break;
default:
fprintf(fvstats, "type= ? (%d)\n", pic->pict_type);
break;
}
ctx->qp_stat[(int)(quality + 0.5)]++;
ctx->qp_sum += quality;
ctx->inv_qp_sum += 1.0 / (double)quality;
break;
}
//--
if (!got_picture)
return NULL; // skipped image
if (init_vo(sh, avctx->pix_fmt) < 0)
return NULL;
if (dr1 && pic->opaque)
mpi = (mp_image_t *)pic->opaque;
if (!mpi)
mpi = mpcodecs_get_image(sh, MP_IMGTYPE_EXPORT, MP_IMGFLAG_PRESERVE,
avctx->width, avctx->height);
if (!mpi) { // temporary error?
mp_tmsg(MSGT_DECVIDEO, MSGL_WARN,
"[VD_FFMPEG] Couldn't allocate image for codec.\n");
return NULL;
}
if (!dr1) {
mpi->planes[0] = pic->data[0];
mpi->planes[1] = pic->data[1];
mpi->planes[2] = pic->data[2];
mpi->planes[3] = pic->data[3];
mpi->stride[0] = pic->linesize[0];
mpi->stride[1] = pic->linesize[1];
mpi->stride[2] = pic->linesize[2];
mpi->stride[3] = pic->linesize[3];
}
if (!mpi->planes[0])
return NULL;
if (ctx->best_csp == IMGFMT_422P && mpi->chroma_y_shift == 1) {
// we have 422p but user wants 420p
mpi->stride[1] *= 2;
mpi->stride[2] *= 2;
}
#if HAVE_BIGENDIAN
// FIXME: this might cause problems for buffers with FF_BUFFER_HINTS_PRESERVE
if (mpi->bpp == 8)
swap_palette(mpi->planes[1]);
#endif
mpi->qscale = pic->qscale_table;
mpi->qstride = pic->qstride;
mpi->pict_type = pic->pict_type;
mpi->qscale_type = pic->qscale_type;
mpi->fields = MP_IMGFIELD_ORDERED;
if (pic->interlaced_frame)
mpi->fields |= MP_IMGFIELD_INTERLACED;
if (pic->top_field_first)
mpi->fields |= MP_IMGFIELD_TOP_FIRST;
if (pic->repeat_pict == 1)
mpi->fields |= MP_IMGFIELD_REPEAT_FIRST;
return mpi;
}
static enum PixelFormat get_format(struct AVCodecContext *avctx,
const enum PixelFormat *fmt)
{
sh_video_t *sh = avctx->opaque;
int i;
for (i = 0; fmt[i] != PIX_FMT_NONE; i++) {
int imgfmt = pixfmt2imgfmt(fmt[i]);
if (!IMGFMT_IS_HWACCEL(imgfmt))
continue;
mp_msg(MSGT_DECVIDEO, MSGL_V, "[VD_FFMPEG] Trying pixfmt=%d.\n", i);
if (init_vo(sh, fmt[i]) >= 0)
break;
}
return fmt[i];
}
static int control(sh_video_t *sh, int cmd, void *arg, ...)
{
vd_ffmpeg_ctx *ctx = sh->context;
AVCodecContext *avctx = ctx->avctx;
switch (cmd) {
case VDCTRL_QUERY_FORMAT: {
int format = (*((int *)arg));
if (format == ctx->best_csp)
return CONTROL_TRUE;
// possible conversions:
switch (format) {
case IMGFMT_YV12:
case IMGFMT_IYUV:
case IMGFMT_I420:
// "converted" using pointer/stride modification
if (ctx->best_csp == IMGFMT_YV12)
return CONTROL_TRUE; // u/v swap
if (ctx->best_csp == IMGFMT_422P && !ctx->do_dr1)
return CONTROL_TRUE; // half stride
break;
}
return CONTROL_FALSE;
}
case VDCTRL_RESYNC_STREAM:
avcodec_flush_buffers(avctx);
return CONTROL_TRUE;
case VDCTRL_QUERY_UNSEEN_FRAMES:;
int delay = avctx->has_b_frames;
if (avctx->active_thread_type & FF_THREAD_FRAME)
delay += avctx->thread_count - 1;
return delay + 10;
case VDCTRL_RESET_ASPECT:
if (ctx->vo_initialized)
ctx->vo_initialized = false;
init_vo(sh, avctx->pix_fmt);
return true;
}
return CONTROL_UNKNOWN;
}
const struct vd_functions mpcodecs_vd_ffmpeg = {
.info = &info,
.init = init,
.uninit = uninit,
.control = control,
.decode2 = decode
};