mirror of https://github.com/mpv-player/mpv
782 lines
24 KiB
C
782 lines
24 KiB
C
/*
|
|
* This file is part of MPlayer.
|
|
*
|
|
* MPlayer is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License as published by
|
|
* the Free Software Foundation; either version 2 of the License, or
|
|
* (at your option) any later version.
|
|
*
|
|
* MPlayer is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License along
|
|
* with MPlayer; if not, write to the Free Software Foundation, Inc.,
|
|
* 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
|
|
*/
|
|
|
|
#include <stdio.h>
|
|
#include <stdlib.h>
|
|
#include <assert.h>
|
|
#include <time.h>
|
|
#include <stdbool.h>
|
|
#include <sys/types.h>
|
|
|
|
#include <libavutil/common.h>
|
|
#include <libavutil/opt.h>
|
|
#include <libavutil/intreadwrite.h>
|
|
#include <libavutil/pixdesc.h>
|
|
|
|
#include "compat/libav.h"
|
|
|
|
#include "talloc.h"
|
|
#include "config.h"
|
|
#include "core/mp_msg.h"
|
|
#include "core/options.h"
|
|
#include "core/av_opts.h"
|
|
#include "core/av_common.h"
|
|
#include "core/codecs.h"
|
|
|
|
#include "compat/mpbswap.h"
|
|
#include "video/fmt-conversion.h"
|
|
|
|
#include "vd.h"
|
|
#include "video/img_format.h"
|
|
#include "video/mp_image_pool.h"
|
|
#include "video/filter/vf.h"
|
|
#include "demux/stheader.h"
|
|
#include "demux/demux_packet.h"
|
|
#include "osdep/numcores.h"
|
|
#include "video/csputils.h"
|
|
|
|
#include "libavcodec/avcodec.h"
|
|
#include "lavc.h"
|
|
|
|
#if AVPALETTE_SIZE != MP_PALETTE_SIZE
|
|
#error palette too large, adapt video/mp_image.h:MP_PALETTE_SIZE
|
|
#endif
|
|
|
|
#include "core/m_option.h"
|
|
|
|
static void init_avctx(sh_video_t *sh, const char *decoder, struct hwdec *hwdec);
|
|
static void uninit_avctx(sh_video_t *sh);
|
|
static void setup_refcounting_hw(struct AVCodecContext *s);
|
|
static void draw_slice_hwdec(struct AVCodecContext *s, const AVFrame *src,
|
|
int offset[4], int y, int type, int height);
|
|
|
|
static enum PixelFormat get_format_hwdec(struct AVCodecContext *avctx,
|
|
const enum PixelFormat *pix_fmt);
|
|
|
|
static void uninit(struct sh_video *sh);
|
|
|
|
#define OPT_BASE_STRUCT struct MPOpts
|
|
|
|
const m_option_t lavc_decode_opts_conf[] = {
|
|
OPT_FLAG_CONSTANTS("fast", lavc_param.fast, 0, 0, CODEC_FLAG2_FAST),
|
|
OPT_STRING("skiploopfilter", lavc_param.skip_loop_filter_str, 0),
|
|
OPT_STRING("skipidct", lavc_param.skip_idct_str, 0),
|
|
OPT_STRING("skipframe", lavc_param.skip_frame_str, 0),
|
|
OPT_INTRANGE("threads", lavc_param.threads, 0, 0, 16),
|
|
OPT_FLAG_CONSTANTS("bitexact", lavc_param.bitexact, 0, 0, CODEC_FLAG_BITEXACT),
|
|
OPT_STRING("o", lavc_param.avopt, 0),
|
|
{NULL, NULL, 0, 0, 0, 0, NULL}
|
|
};
|
|
|
|
// keep in sync with --hwdec option
|
|
enum hwdec_type {
|
|
HWDEC_NONE = 0,
|
|
HWDEC_VDPAU = 1,
|
|
HWDEC_VDA = 2,
|
|
HWDEC_CRYSTALHD = 3,
|
|
};
|
|
|
|
struct hwdec {
|
|
enum hwdec_type api;
|
|
const char *codec, *hw_codec;
|
|
};
|
|
|
|
static const struct hwdec hwdec[] = {
|
|
{HWDEC_VDPAU, "h264", "h264_vdpau"},
|
|
{HWDEC_VDPAU, "wmv3", "wmv3_vdpau"},
|
|
{HWDEC_VDPAU, "vc1", "vc1_vdpau"},
|
|
{HWDEC_VDPAU, "mpegvideo", "mpegvideo_vdpau"},
|
|
{HWDEC_VDPAU, "mpeg1video", "mpeg1video_vdpau"},
|
|
{HWDEC_VDPAU, "mpeg2video", "mpegvideo_vdpau"},
|
|
{HWDEC_VDPAU, "mpeg2", "mpeg2_vdpau"},
|
|
{HWDEC_VDPAU, "mpeg4", "mpeg4_vdpau"},
|
|
|
|
{HWDEC_VDA, "h264", "h264_vda"},
|
|
|
|
{HWDEC_CRYSTALHD, "mpeg2", "mpeg2_crystalhd"},
|
|
{HWDEC_CRYSTALHD, "msmpeg4", "msmpeg4_crystalhd"},
|
|
{HWDEC_CRYSTALHD, "wmv3", "wmv3_crystalhd"},
|
|
{HWDEC_CRYSTALHD, "vc1", "vc1_crystalhd"},
|
|
{HWDEC_CRYSTALHD, "h264", "h264_crystalhd"},
|
|
{HWDEC_CRYSTALHD, "mpeg4", "mpeg4_crystalhd"},
|
|
|
|
{0}
|
|
};
|
|
|
|
static struct hwdec *find_hwcodec(enum hwdec_type api, const char *codec)
|
|
{
|
|
for (int n = 0; hwdec[n].api; n++) {
|
|
if (hwdec[n].api == api && strcmp(hwdec[n].codec, codec) == 0)
|
|
return (struct hwdec *)&hwdec[n];
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
static bool hwdec_codec_allowed(sh_video_t *sh, struct hwdec *hwdec)
|
|
{
|
|
bstr s = bstr0(sh->opts->hwdec_codecs);
|
|
while (s.len) {
|
|
bstr item;
|
|
bstr_split_tok(s, ",", &item, &s);
|
|
if (bstr_equals0(item, "all") || bstr_equals0(item, hwdec->codec))
|
|
return true;
|
|
}
|
|
return false;
|
|
}
|
|
|
|
static enum AVDiscard str2AVDiscard(char *str)
|
|
{
|
|
if (!str) return AVDISCARD_DEFAULT;
|
|
if (strcasecmp(str, "none" ) == 0) return AVDISCARD_NONE;
|
|
if (strcasecmp(str, "default") == 0) return AVDISCARD_DEFAULT;
|
|
if (strcasecmp(str, "nonref" ) == 0) return AVDISCARD_NONREF;
|
|
if (strcasecmp(str, "bidir" ) == 0) return AVDISCARD_BIDIR;
|
|
if (strcasecmp(str, "nonkey" ) == 0) return AVDISCARD_NONKEY;
|
|
if (strcasecmp(str, "all" ) == 0) return AVDISCARD_ALL;
|
|
mp_msg(MSGT_DECVIDEO, MSGL_ERR, "Unknown discard value %s\n", str);
|
|
return AVDISCARD_DEFAULT;
|
|
}
|
|
|
|
static int init(sh_video_t *sh, const char *decoder)
|
|
{
|
|
vd_ffmpeg_ctx *ctx;
|
|
ctx = sh->context = talloc_zero(NULL, vd_ffmpeg_ctx);
|
|
ctx->non_dr1_pool = talloc_steal(ctx, mp_image_pool_new(16));
|
|
|
|
struct hwdec *hwdec = find_hwcodec(sh->opts->hwdec_api, decoder);
|
|
struct hwdec *use_hwdec = NULL;
|
|
if (hwdec && hwdec_codec_allowed(sh, hwdec)) {
|
|
AVCodec *lavc_hwcodec = avcodec_find_decoder_by_name(hwdec->hw_codec);
|
|
if (lavc_hwcodec) {
|
|
ctx->software_fallback_decoder = talloc_strdup(ctx, decoder);
|
|
decoder = lavc_hwcodec->name;
|
|
use_hwdec = hwdec;
|
|
} else {
|
|
mp_tmsg(MSGT_DECVIDEO, MSGL_WARN, "Decoder '%s' not found in "
|
|
"libavcodec, using software decoding.\n", hwdec->hw_codec);
|
|
}
|
|
}
|
|
|
|
init_avctx(sh, decoder, use_hwdec);
|
|
if (!ctx->avctx) {
|
|
if (ctx->software_fallback_decoder) {
|
|
mp_tmsg(MSGT_DECVIDEO, MSGL_ERR, "Error initializing hardware "
|
|
"decoding, falling back to software decoding.\n");
|
|
decoder = ctx->software_fallback_decoder;
|
|
ctx->software_fallback_decoder = NULL;
|
|
init_avctx(sh, decoder, NULL);
|
|
}
|
|
if (!ctx->avctx) {
|
|
uninit(sh);
|
|
return 0;
|
|
}
|
|
}
|
|
return 1;
|
|
}
|
|
|
|
static void set_from_bih(AVCodecContext *avctx, uint32_t format,
|
|
BITMAPINFOHEADER *bih)
|
|
{
|
|
|
|
switch (format) {
|
|
case mmioFOURCC('S','V','Q','3'):
|
|
case mmioFOURCC('A','V','R','n'):
|
|
case mmioFOURCC('M','J','P','G'):
|
|
/* AVRn stores huffman table in AVI header */
|
|
/* Pegasus MJPEG stores it also in AVI header, but it uses the common
|
|
* MJPG fourcc :( */
|
|
if (bih->biSize <= sizeof(*bih))
|
|
break;
|
|
av_opt_set_int(avctx, "extern_huff", 1, AV_OPT_SEARCH_CHILDREN);
|
|
avctx->extradata_size = bih->biSize - sizeof(*bih);
|
|
avctx->extradata = av_mallocz(avctx->extradata_size +
|
|
FF_INPUT_BUFFER_PADDING_SIZE);
|
|
memcpy(avctx->extradata, bih + 1, avctx->extradata_size);
|
|
break;
|
|
|
|
case mmioFOURCC('R','V','1','0'):
|
|
case mmioFOURCC('R','V','1','3'):
|
|
case mmioFOURCC('R','V','2','0'):
|
|
case mmioFOURCC('R','V','3','0'):
|
|
case mmioFOURCC('R','V','4','0'):
|
|
if (bih->biSize < sizeof(*bih) + 8) {
|
|
// only 1 packet per frame & sub_id from fourcc
|
|
avctx->extradata_size = 8;
|
|
avctx->extradata = av_mallocz(avctx->extradata_size +
|
|
FF_INPUT_BUFFER_PADDING_SIZE);
|
|
((uint32_t *)avctx->extradata)[0] = 0;
|
|
((uint32_t *)avctx->extradata)[1] =
|
|
format == mmioFOURCC('R','V','1','3') ?
|
|
0x10003001 : 0x10000000;
|
|
} else {
|
|
// has extra slice header (demux_rm or rm->avi streamcopy)
|
|
avctx->extradata_size = bih->biSize - sizeof(*bih);
|
|
avctx->extradata = av_mallocz(avctx->extradata_size +
|
|
FF_INPUT_BUFFER_PADDING_SIZE);
|
|
memcpy(avctx->extradata, bih + 1, avctx->extradata_size);
|
|
}
|
|
break;
|
|
|
|
default:
|
|
if (bih->biSize <= sizeof(*bih))
|
|
break;
|
|
avctx->extradata_size = bih->biSize - sizeof(*bih);
|
|
avctx->extradata = av_mallocz(avctx->extradata_size +
|
|
FF_INPUT_BUFFER_PADDING_SIZE);
|
|
memcpy(avctx->extradata, bih + 1, avctx->extradata_size);
|
|
break;
|
|
}
|
|
|
|
avctx->bits_per_coded_sample = bih->biBitCount;
|
|
avctx->coded_width = bih->biWidth;
|
|
avctx->coded_height = bih->biHeight;
|
|
}
|
|
|
|
static void init_avctx(sh_video_t *sh, const char *decoder, struct hwdec *hwdec)
|
|
{
|
|
vd_ffmpeg_ctx *ctx = sh->context;
|
|
struct lavc_param *lavc_param = &sh->opts->lavc_param;
|
|
bool mp_rawvideo = false;
|
|
|
|
assert(!ctx->avctx);
|
|
|
|
if (strcmp(decoder, "mp-rawvideo") == 0) {
|
|
mp_rawvideo = true;
|
|
decoder = "rawvideo";
|
|
}
|
|
|
|
AVCodec *lavc_codec = avcodec_find_decoder_by_name(decoder);
|
|
if (!lavc_codec)
|
|
return;
|
|
|
|
ctx->do_dr1 = ctx->do_hw_dr1 = 0;
|
|
ctx->pix_fmt = PIX_FMT_NONE;
|
|
ctx->vo_initialized = 0;
|
|
ctx->hwdec = hwdec;
|
|
ctx->pic = avcodec_alloc_frame();
|
|
ctx->avctx = avcodec_alloc_context3(lavc_codec);
|
|
AVCodecContext *avctx = ctx->avctx;
|
|
avctx->opaque = sh;
|
|
avctx->codec_type = AVMEDIA_TYPE_VIDEO;
|
|
avctx->codec_id = lavc_codec->id;
|
|
|
|
avctx->thread_count = lavc_param->threads;
|
|
|
|
// Hack to allow explicitly selecting vdpau hw decoders
|
|
if (!hwdec && (lavc_codec->capabilities & CODEC_CAP_HWACCEL_VDPAU)) {
|
|
ctx->hwdec = talloc(ctx, struct hwdec);
|
|
*ctx->hwdec = (struct hwdec) {
|
|
.api = HWDEC_VDPAU,
|
|
.codec = sh->gsh->codec,
|
|
.hw_codec = decoder,
|
|
};
|
|
}
|
|
|
|
if (ctx->hwdec && ctx->hwdec->api == HWDEC_VDPAU) {
|
|
assert(lavc_codec->capabilities & CODEC_CAP_HWACCEL_VDPAU);
|
|
ctx->do_hw_dr1 = true;
|
|
avctx->thread_count = 1;
|
|
avctx->get_format = get_format_hwdec;
|
|
setup_refcounting_hw(avctx);
|
|
if (ctx->hwdec->api == HWDEC_VDPAU) {
|
|
avctx->draw_horiz_band = draw_slice_hwdec;
|
|
avctx->slice_flags =
|
|
SLICE_FLAG_CODED_ORDER | SLICE_FLAG_ALLOW_FIELD;
|
|
}
|
|
} else {
|
|
#if HAVE_AVUTIL_REFCOUNTING
|
|
avctx->refcounted_frames = 1;
|
|
#else
|
|
if (lavc_codec->capabilities & CODEC_CAP_DR1) {
|
|
ctx->do_dr1 = true;
|
|
avctx->get_buffer = mp_codec_get_buffer;
|
|
avctx->release_buffer = mp_codec_release_buffer;
|
|
}
|
|
#endif
|
|
}
|
|
|
|
if (avctx->thread_count == 0) {
|
|
int threads = default_thread_count();
|
|
if (threads < 1) {
|
|
mp_msg(MSGT_DECVIDEO, MSGL_WARN, "[VD_FFMPEG] Could not determine "
|
|
"thread count to use, defaulting to 1.\n");
|
|
threads = 1;
|
|
}
|
|
threads = FFMIN(threads, 16);
|
|
avctx->thread_count = threads;
|
|
}
|
|
|
|
avctx->flags |= lavc_param->bitexact;
|
|
|
|
avctx->flags2 |= lavc_param->fast;
|
|
avctx->skip_loop_filter = str2AVDiscard(lavc_param->skip_loop_filter_str);
|
|
avctx->skip_idct = str2AVDiscard(lavc_param->skip_idct_str);
|
|
avctx->skip_frame = str2AVDiscard(lavc_param->skip_frame_str);
|
|
|
|
if (lavc_param->avopt) {
|
|
if (parse_avopts(avctx, lavc_param->avopt) < 0) {
|
|
mp_msg(MSGT_DECVIDEO, MSGL_ERR,
|
|
"Your options /%s/ look like gibberish to me pal\n",
|
|
lavc_param->avopt);
|
|
uninit_avctx(sh);
|
|
return;
|
|
}
|
|
}
|
|
|
|
// Do this after the above avopt handling in case it changes values
|
|
ctx->skip_frame = avctx->skip_frame;
|
|
|
|
avctx->codec_tag = sh->format;
|
|
avctx->coded_width = sh->disp_w;
|
|
avctx->coded_height = sh->disp_h;
|
|
|
|
// demux_mkv
|
|
if (sh->bih)
|
|
set_from_bih(avctx, sh->format, sh->bih);
|
|
|
|
if (mp_rawvideo && sh->format >= IMGFMT_START && sh->format < IMGFMT_END) {
|
|
avctx->pix_fmt = imgfmt2pixfmt(sh->format);
|
|
avctx->codec_tag = 0;
|
|
}
|
|
|
|
if (sh->gsh->lav_headers)
|
|
mp_copy_lav_codec_headers(avctx, sh->gsh->lav_headers);
|
|
|
|
/* open it */
|
|
if (avcodec_open2(avctx, lavc_codec, NULL) < 0) {
|
|
mp_tmsg(MSGT_DECVIDEO, MSGL_ERR, "Could not open codec.\n");
|
|
uninit_avctx(sh);
|
|
return;
|
|
}
|
|
}
|
|
|
|
static void uninit_avctx(sh_video_t *sh)
|
|
{
|
|
vd_ffmpeg_ctx *ctx = sh->context;
|
|
AVCodecContext *avctx = ctx->avctx;
|
|
|
|
if (avctx) {
|
|
if (avctx->codec && avcodec_close(avctx) < 0)
|
|
mp_tmsg(MSGT_DECVIDEO, MSGL_ERR, "Could not close codec.\n");
|
|
|
|
av_freep(&avctx->extradata);
|
|
av_freep(&avctx->slice_offset);
|
|
}
|
|
|
|
av_freep(&ctx->avctx);
|
|
avcodec_free_frame(&ctx->pic);
|
|
|
|
#if !HAVE_AVUTIL_REFCOUNTING
|
|
mp_buffer_pool_free(&ctx->dr1_buffer_pool);
|
|
#endif
|
|
}
|
|
|
|
static void uninit(sh_video_t *sh)
|
|
{
|
|
vd_ffmpeg_ctx *ctx = sh->context;
|
|
|
|
uninit_avctx(sh);
|
|
talloc_free(ctx);
|
|
}
|
|
|
|
static int init_vo(sh_video_t *sh, AVFrame *frame)
|
|
{
|
|
vd_ffmpeg_ctx *ctx = sh->context;
|
|
int width = frame->width;
|
|
int height = frame->height;
|
|
float aspect = av_q2d(frame->sample_aspect_ratio) * width / height;
|
|
int pix_fmt = frame->format;
|
|
|
|
#if LIBAVCODEC_VERSION_INT < AV_VERSION_INT(54, 40, 0)
|
|
pix_fmt = ctx->avctx->pix_fmt;
|
|
#endif
|
|
|
|
/* Reconfiguring filter/VO chain may invalidate direct rendering buffers
|
|
* we have allocated for libavcodec (including the VDPAU HW decoding
|
|
* case). Is it guaranteed that the code below only triggers in a situation
|
|
* with no busy direct rendering buffers for reference frames?
|
|
*/
|
|
if (av_cmp_q(frame->sample_aspect_ratio, ctx->last_sample_aspect_ratio) ||
|
|
width != sh->disp_w || height != sh->disp_h ||
|
|
pix_fmt != ctx->pix_fmt || !ctx->vo_initialized)
|
|
{
|
|
mp_image_pool_clear(ctx->non_dr1_pool);
|
|
ctx->vo_initialized = 0;
|
|
mp_msg(MSGT_DECVIDEO, MSGL_V, "[ffmpeg] aspect_ratio: %f\n", aspect);
|
|
|
|
// Do not overwrite s->aspect on the first call, so that a container
|
|
// aspect if available is preferred.
|
|
// But set it even if the sample aspect did not change, since a
|
|
// resolution change can cause an aspect change even if the
|
|
// _sample_ aspect is unchanged.
|
|
if (sh->aspect == 0 || ctx->last_sample_aspect_ratio.den)
|
|
sh->aspect = aspect;
|
|
ctx->last_sample_aspect_ratio = frame->sample_aspect_ratio;
|
|
sh->disp_w = width;
|
|
sh->disp_h = height;
|
|
|
|
ctx->pix_fmt = pix_fmt;
|
|
ctx->best_csp = pixfmt2imgfmt(pix_fmt);
|
|
|
|
ctx->image_params = (struct mp_image_params) {
|
|
.imgfmt = ctx->best_csp,
|
|
.w = width,
|
|
.h = height,
|
|
// Ideally, we should also set aspect ratio, but we aren't there yet
|
|
// - so vd.c calculates display size from sh->aspect.
|
|
.d_w = width,
|
|
.d_h = height,
|
|
.colorspace = avcol_spc_to_mp_csp(ctx->avctx->colorspace),
|
|
.colorlevels = avcol_range_to_mp_csp_levels(ctx->avctx->color_range),
|
|
.chroma_location =
|
|
avchroma_location_to_mp(ctx->avctx->chroma_sample_location),
|
|
};
|
|
|
|
if (mpcodecs_reconfig_vo(sh, &ctx->image_params) < 0)
|
|
return -1;
|
|
|
|
ctx->vo_initialized = 1;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static enum PixelFormat get_format_hwdec(struct AVCodecContext *avctx,
|
|
const enum PixelFormat *fmt)
|
|
{
|
|
sh_video_t *sh = avctx->opaque;
|
|
vd_ffmpeg_ctx *ctx = sh->context;
|
|
|
|
mp_msg(MSGT_DECVIDEO, MSGL_V, "Pixel formats supported by decoder:");
|
|
for (int i = 0; fmt[i] != PIX_FMT_NONE; i++)
|
|
mp_msg(MSGT_DECVIDEO, MSGL_V, " %s", av_get_pix_fmt_name(fmt[i]));
|
|
mp_msg(MSGT_DECVIDEO, MSGL_V, "\n");
|
|
|
|
assert(ctx->hwdec);
|
|
|
|
for (int i = 0; fmt[i] != PIX_FMT_NONE; i++) {
|
|
int imgfmt = pixfmt2imgfmt(fmt[i]);
|
|
if (ctx->hwdec->api == HWDEC_VDPAU && IMGFMT_IS_VDPAU(imgfmt))
|
|
return fmt[i];
|
|
}
|
|
|
|
return PIX_FMT_NONE;
|
|
}
|
|
|
|
static void draw_slice_hwdec(struct AVCodecContext *s,
|
|
const AVFrame *src, int offset[4],
|
|
int y, int type, int height)
|
|
{
|
|
sh_video_t *sh = s->opaque;
|
|
struct vf_instance *vf = sh->vfilter;
|
|
void *state_ptr = src->data[0];
|
|
vf->control(vf, VFCTRL_HWDEC_DECODER_RENDER, state_ptr);
|
|
}
|
|
|
|
static struct mp_image *get_surface_hwdec(struct sh_video *sh, AVFrame *pic)
|
|
{
|
|
vd_ffmpeg_ctx *ctx = sh->context;
|
|
|
|
/* Decoders using ffmpeg's hwaccel architecture (everything except vdpau)
|
|
* can fall back to software decoding automatically. However, we don't
|
|
* want that: multithreading was already disabled. ffmpeg's fallback
|
|
* isn't really useful, and causes more trouble than it helps.
|
|
*
|
|
* Instead of trying to "adjust" the thread_count fields in avctx, let
|
|
* decoding fail hard. Then decode_with_fallback() will do our own software
|
|
* fallback. Fully reinitializing the decoder is saner, and will probably
|
|
* save us from other weird corner cases, like having to "reroute" the
|
|
* get_buffer callback.
|
|
*/
|
|
int imgfmt = pixfmt2imgfmt(pic->format);
|
|
if (!IMGFMT_IS_HWACCEL(imgfmt))
|
|
return NULL;
|
|
|
|
// Video with non mod-16 width/height will have allocation sizes that are
|
|
// rounded up. This conflicts with our video size change detection and
|
|
// leads to an endless loop. On the other hand, vdpau seems to round up
|
|
// frame allocations internally. So use the original video resolution
|
|
// instead.
|
|
AVFrame pic_resized = *pic;
|
|
pic_resized.width = ctx->avctx->width;
|
|
pic_resized.height = ctx->avctx->height;
|
|
|
|
if (init_vo(sh, &pic_resized) < 0)
|
|
return NULL;
|
|
|
|
assert(IMGFMT_IS_HWACCEL(ctx->best_csp));
|
|
|
|
struct mp_image *mpi = NULL;
|
|
|
|
struct vf_instance *vf = sh->vfilter;
|
|
vf->control(vf, VFCTRL_HWDEC_ALLOC_SURFACE, &mpi);
|
|
|
|
if (mpi) {
|
|
for (int i = 0; i < 4; i++)
|
|
pic->data[i] = mpi->planes[i];
|
|
}
|
|
|
|
return mpi;
|
|
}
|
|
|
|
#if HAVE_AVUTIL_REFCOUNTING
|
|
|
|
static void free_mpi(void *opaque, uint8_t *data)
|
|
{
|
|
struct mp_image *mpi = opaque;
|
|
talloc_free(mpi);
|
|
}
|
|
|
|
static int get_buffer2_hwdec(AVCodecContext *avctx, AVFrame *pic, int flags)
|
|
{
|
|
sh_video_t *sh = avctx->opaque;
|
|
|
|
struct mp_image *mpi = get_surface_hwdec(sh, pic);
|
|
if (!mpi)
|
|
return -1;
|
|
|
|
pic->buf[0] = av_buffer_create(NULL, 0, free_mpi, mpi, 0);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void setup_refcounting_hw(AVCodecContext *avctx)
|
|
{
|
|
avctx->get_buffer2 = get_buffer2_hwdec;
|
|
avctx->refcounted_frames = 1;
|
|
}
|
|
|
|
#else /* HAVE_AVUTIL_REFCOUNTING */
|
|
|
|
static int get_buffer_hwdec(AVCodecContext *avctx, AVFrame *pic)
|
|
{
|
|
sh_video_t *sh = avctx->opaque;
|
|
|
|
struct mp_image *mpi = get_surface_hwdec(sh, pic);
|
|
if (!mpi)
|
|
return -1;
|
|
|
|
pic->opaque = mpi;
|
|
pic->type = FF_BUFFER_TYPE_USER;
|
|
|
|
/* The libavcodec reordered_opaque functionality is implemented by
|
|
* a similar copy in avcodec_default_get_buffer() and without a
|
|
* workaround like this it'd stop working when a custom buffer
|
|
* callback is used.
|
|
*/
|
|
pic->reordered_opaque = avctx->reordered_opaque;
|
|
return 0;
|
|
}
|
|
|
|
static void release_buffer_hwdec(AVCodecContext *avctx, AVFrame *pic)
|
|
{
|
|
mp_image_t *mpi = pic->opaque;
|
|
|
|
assert(pic->type == FF_BUFFER_TYPE_USER);
|
|
assert(mpi);
|
|
|
|
talloc_free(mpi);
|
|
|
|
for (int i = 0; i < 4; i++)
|
|
pic->data[i] = NULL;
|
|
}
|
|
|
|
static void setup_refcounting_hw(AVCodecContext *avctx)
|
|
{
|
|
avctx->get_buffer = get_buffer_hwdec;
|
|
avctx->release_buffer = release_buffer_hwdec;
|
|
}
|
|
|
|
#endif /* HAVE_AVUTIL_REFCOUNTING */
|
|
|
|
#if HAVE_AVUTIL_REFCOUNTING
|
|
|
|
static struct mp_image *image_from_decoder(struct sh_video *sh)
|
|
{
|
|
vd_ffmpeg_ctx *ctx = sh->context;
|
|
AVFrame *pic = ctx->pic;
|
|
|
|
struct mp_image *img = mp_image_from_av_frame(pic);
|
|
av_frame_unref(pic);
|
|
|
|
return img;
|
|
}
|
|
|
|
#else /* HAVE_AVUTIL_REFCOUNTING */
|
|
|
|
static void fb_ref(void *b)
|
|
{
|
|
mp_buffer_ref(b);
|
|
}
|
|
|
|
static void fb_unref(void *b)
|
|
{
|
|
mp_buffer_unref(b);
|
|
}
|
|
|
|
static bool fb_is_unique(void *b)
|
|
{
|
|
return mp_buffer_is_unique(b);
|
|
}
|
|
|
|
static struct mp_image *image_from_decoder(struct sh_video *sh)
|
|
{
|
|
vd_ffmpeg_ctx *ctx = sh->context;
|
|
AVFrame *pic = ctx->pic;
|
|
|
|
struct mp_image new = {0};
|
|
mp_image_copy_fields_from_av_frame(&new, pic);
|
|
|
|
struct mp_image *mpi;
|
|
if (ctx->do_hw_dr1 && pic->opaque) {
|
|
mpi = pic->opaque; // reordered frame
|
|
assert(mpi);
|
|
mpi = mp_image_new_ref(mpi);
|
|
mp_image_copy_attributes(mpi, &new);
|
|
} else if (ctx->do_dr1 && pic->opaque) {
|
|
struct FrameBuffer *fb = pic->opaque;
|
|
mp_buffer_ref(fb); // initial reference for mpi
|
|
mpi = mp_image_new_external_ref(&new, fb, fb_ref, fb_unref,
|
|
fb_is_unique, NULL);
|
|
} else {
|
|
mpi = mp_image_pool_new_copy(ctx->non_dr1_pool, &new);
|
|
}
|
|
return mpi;
|
|
}
|
|
|
|
#endif /* HAVE_AVUTIL_REFCOUNTING */
|
|
|
|
static int decode(struct sh_video *sh, struct demux_packet *packet,
|
|
int flags, double *reordered_pts, struct mp_image **out_image)
|
|
{
|
|
int got_picture = 0;
|
|
int ret;
|
|
vd_ffmpeg_ctx *ctx = sh->context;
|
|
AVFrame *pic = ctx->pic;
|
|
AVCodecContext *avctx = ctx->avctx;
|
|
AVPacket pkt;
|
|
|
|
if (flags & 2)
|
|
avctx->skip_frame = AVDISCARD_ALL;
|
|
else if (flags & 1)
|
|
avctx->skip_frame = AVDISCARD_NONREF;
|
|
else
|
|
avctx->skip_frame = ctx->skip_frame;
|
|
|
|
mp_set_av_packet(&pkt, packet);
|
|
|
|
// The avcodec opaque field stupidly supports only int64_t type
|
|
union pts { int64_t i; double d; };
|
|
avctx->reordered_opaque = (union pts){.d = *reordered_pts}.i;
|
|
ret = avcodec_decode_video2(avctx, pic, &got_picture, &pkt);
|
|
if (ret < 0) {
|
|
mp_msg(MSGT_DECVIDEO, MSGL_WARN, "Error while decoding frame!\n");
|
|
return -1;
|
|
}
|
|
*reordered_pts = (union pts){.i = pic->reordered_opaque}.d;
|
|
|
|
if (!got_picture)
|
|
return 0; // skipped image
|
|
|
|
if (init_vo(sh, pic) < 0)
|
|
return -1;
|
|
|
|
struct mp_image *mpi = image_from_decoder(sh);
|
|
assert(mpi->planes[0]);
|
|
|
|
mpi->colorspace = ctx->image_params.colorspace;
|
|
mpi->levels = ctx->image_params.colorlevels;
|
|
mpi->chroma_location = ctx->image_params.chroma_location;
|
|
|
|
*out_image = mpi;
|
|
return 1;
|
|
}
|
|
|
|
static struct mp_image *decode_with_fallback(struct sh_video *sh,
|
|
struct demux_packet *packet,
|
|
int flags, double *reordered_pts)
|
|
{
|
|
vd_ffmpeg_ctx *ctx = sh->context;
|
|
if (!ctx->avctx)
|
|
return NULL;
|
|
|
|
struct mp_image *mpi = NULL;
|
|
int res = decode(sh, packet, flags, reordered_pts, &mpi);
|
|
if (res >= 0)
|
|
return mpi;
|
|
|
|
// Failed hardware decoding? Try again in software.
|
|
if (ctx->software_fallback_decoder) {
|
|
uninit_avctx(sh);
|
|
sh->vf_initialized = 0;
|
|
mp_tmsg(MSGT_DECVIDEO, MSGL_ERR, "Error using hardware "
|
|
"decoding, falling back to software decoding.\n");
|
|
const char *decoder = ctx->software_fallback_decoder;
|
|
ctx->software_fallback_decoder = NULL;
|
|
init_avctx(sh, decoder, NULL);
|
|
if (ctx->avctx) {
|
|
mpi = NULL;
|
|
decode(sh, packet, flags, reordered_pts, &mpi);
|
|
return mpi;
|
|
}
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static int control(sh_video_t *sh, int cmd, void *arg)
|
|
{
|
|
vd_ffmpeg_ctx *ctx = sh->context;
|
|
AVCodecContext *avctx = ctx->avctx;
|
|
switch (cmd) {
|
|
case VDCTRL_RESYNC_STREAM:
|
|
avcodec_flush_buffers(avctx);
|
|
return CONTROL_TRUE;
|
|
case VDCTRL_QUERY_UNSEEN_FRAMES:;
|
|
int delay = avctx->has_b_frames;
|
|
assert(delay >= 0);
|
|
if (avctx->active_thread_type & FF_THREAD_FRAME)
|
|
delay += avctx->thread_count - 1;
|
|
*(int *)arg = delay;
|
|
return CONTROL_TRUE;
|
|
case VDCTRL_REINIT_VO:
|
|
if (ctx->vo_initialized)
|
|
mpcodecs_reconfig_vo(sh, &ctx->image_params);
|
|
return true;
|
|
case VDCTRL_GET_PARAMS:
|
|
*(struct mp_image_params *)arg = ctx->image_params;
|
|
return ctx->vo_initialized ? true : CONTROL_NA;
|
|
}
|
|
return CONTROL_UNKNOWN;
|
|
}
|
|
|
|
static void add_decoders(struct mp_decoder_list *list)
|
|
{
|
|
mp_add_lavc_decoders(list, AVMEDIA_TYPE_VIDEO);
|
|
mp_add_decoder(list, "lavc", "mp-rawvideo", "mp-rawvideo",
|
|
"raw video");
|
|
}
|
|
|
|
const struct vd_functions mpcodecs_vd_ffmpeg = {
|
|
.name = "lavc",
|
|
.add_decoders = add_decoders,
|
|
.init = init,
|
|
.uninit = uninit,
|
|
.control = control,
|
|
.decode = decode_with_fallback,
|
|
};
|