mirror of
https://git.ffmpeg.org/ffmpeg.git
synced 2024-12-25 16:52:31 +00:00
201cb35061
P012, Y212 and XV36 are used for 12bit content in FFmpeg VAAPI, so these formats should be used in FFmpeg QSV too, however the SDK only declares support for P016, Y216 and Y416. So this commit fudged mappings between AV_PIX_FMT_P012 and MFX_FOURCC_P016, AV_PIX_FMT_Y212 and MFX_FOURCC_Y216, AV_PIX_FMT_XV36 and MFX_FOURCC_Y416. Signed-off-by: Fei Wang <fei.w.wang@intel.com> Signed-off-by: Wenbin Chen <wenbin.chen@intel.com> Signed-off-by: Haihao Xiang <haihao.xiang@intel.com>
2227 lines
71 KiB
C
2227 lines
71 KiB
C
/*
|
|
* This file is part of FFmpeg.
|
|
*
|
|
* FFmpeg is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with FFmpeg; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
#include <stdatomic.h>
|
|
#include <stdint.h>
|
|
#include <string.h>
|
|
|
|
#include <mfxvideo.h>
|
|
|
|
#include "config.h"
|
|
|
|
#if HAVE_PTHREADS
|
|
#include <pthread.h>
|
|
#endif
|
|
|
|
#define COBJMACROS
|
|
#if CONFIG_VAAPI
|
|
#include "hwcontext_vaapi.h"
|
|
#endif
|
|
#if CONFIG_D3D11VA
|
|
#include "hwcontext_d3d11va.h"
|
|
#endif
|
|
#if CONFIG_DXVA2
|
|
#include "hwcontext_dxva2.h"
|
|
#endif
|
|
|
|
#include "buffer.h"
|
|
#include "common.h"
|
|
#include "hwcontext.h"
|
|
#include "hwcontext_internal.h"
|
|
#include "hwcontext_qsv.h"
|
|
#include "mem.h"
|
|
#include "pixfmt.h"
|
|
#include "pixdesc.h"
|
|
#include "time.h"
|
|
#include "imgutils.h"
|
|
#include "avassert.h"
|
|
|
|
#define QSV_VERSION_ATLEAST(MAJOR, MINOR) \
|
|
(MFX_VERSION_MAJOR > (MAJOR) || \
|
|
MFX_VERSION_MAJOR == (MAJOR) && MFX_VERSION_MINOR >= (MINOR))
|
|
|
|
#define MFX_IMPL_VIA_MASK(impl) (0x0f00 & (impl))
|
|
#define QSV_ONEVPL QSV_VERSION_ATLEAST(2, 0)
|
|
#define QSV_HAVE_OPAQUE !QSV_ONEVPL
|
|
|
|
#if QSV_ONEVPL
|
|
#include <mfxdispatcher.h>
|
|
#else
|
|
#define MFXUnload(a) do { } while(0)
|
|
#endif
|
|
|
|
typedef struct QSVDevicePriv {
|
|
AVBufferRef *child_device_ctx;
|
|
} QSVDevicePriv;
|
|
|
|
typedef struct QSVDeviceContext {
|
|
mfxHDL handle;
|
|
mfxHandleType handle_type;
|
|
mfxVersion ver;
|
|
mfxIMPL impl;
|
|
|
|
enum AVHWDeviceType child_device_type;
|
|
enum AVPixelFormat child_pix_fmt;
|
|
} QSVDeviceContext;
|
|
|
|
typedef struct QSVFramesContext {
|
|
mfxSession session_download;
|
|
atomic_int session_download_init;
|
|
mfxSession session_upload;
|
|
atomic_int session_upload_init;
|
|
#if HAVE_PTHREADS
|
|
pthread_mutex_t session_lock;
|
|
#endif
|
|
|
|
AVBufferRef *child_frames_ref;
|
|
mfxFrameSurface1 *surfaces_internal;
|
|
mfxHDLPair *handle_pairs_internal;
|
|
int nb_surfaces_used;
|
|
|
|
// used in the frame allocator for non-opaque surfaces
|
|
mfxMemId *mem_ids;
|
|
#if QSV_HAVE_OPAQUE
|
|
// used in the opaque alloc request for opaque surfaces
|
|
mfxFrameSurface1 **surface_ptrs;
|
|
|
|
mfxExtOpaqueSurfaceAlloc opaque_alloc;
|
|
mfxExtBuffer *ext_buffers[1];
|
|
#endif
|
|
AVFrame realigned_upload_frame;
|
|
AVFrame realigned_download_frame;
|
|
} QSVFramesContext;
|
|
|
|
static const struct {
|
|
enum AVPixelFormat pix_fmt;
|
|
uint32_t fourcc;
|
|
uint16_t mfx_shift;
|
|
} supported_pixel_formats[] = {
|
|
{ AV_PIX_FMT_NV12, MFX_FOURCC_NV12, 0 },
|
|
{ AV_PIX_FMT_BGRA, MFX_FOURCC_RGB4, 0 },
|
|
{ AV_PIX_FMT_P010, MFX_FOURCC_P010, 1 },
|
|
{ AV_PIX_FMT_PAL8, MFX_FOURCC_P8, 0 },
|
|
#if CONFIG_VAAPI
|
|
{ AV_PIX_FMT_YUYV422,
|
|
MFX_FOURCC_YUY2, 0 },
|
|
{ AV_PIX_FMT_Y210,
|
|
MFX_FOURCC_Y210, 1 },
|
|
// VUYX is used for VAAPI child device,
|
|
// the SDK only delares support for AYUV
|
|
{ AV_PIX_FMT_VUYX,
|
|
MFX_FOURCC_AYUV, 0 },
|
|
// XV30 is used for VAAPI child device,
|
|
// the SDK only delares support for Y410
|
|
{ AV_PIX_FMT_XV30,
|
|
MFX_FOURCC_Y410, 0 },
|
|
#if QSV_VERSION_ATLEAST(1, 31)
|
|
// P012 is used for VAAPI child device,
|
|
// the SDK only delares support for P016
|
|
{ AV_PIX_FMT_P012,
|
|
MFX_FOURCC_P016, 1 },
|
|
// Y212 is used for VAAPI child device,
|
|
// the SDK only delares support for Y216
|
|
{ AV_PIX_FMT_Y212,
|
|
MFX_FOURCC_Y216, 1 },
|
|
// XV36 is used for VAAPI child device,
|
|
// the SDK only delares support for Y416
|
|
{ AV_PIX_FMT_XV36,
|
|
MFX_FOURCC_Y416, 1 },
|
|
#endif
|
|
#endif
|
|
};
|
|
|
|
extern int ff_qsv_get_surface_base_handle(mfxFrameSurface1 *surf,
|
|
enum AVHWDeviceType base_dev_type,
|
|
void **base_handle);
|
|
|
|
/**
|
|
* Caller needs to allocate enough space for base_handle pointer.
|
|
**/
|
|
int ff_qsv_get_surface_base_handle(mfxFrameSurface1 *surf,
|
|
enum AVHWDeviceType base_dev_type,
|
|
void **base_handle)
|
|
{
|
|
mfxHDLPair *handle_pair;
|
|
handle_pair = surf->Data.MemId;
|
|
switch (base_dev_type) {
|
|
#if CONFIG_VAAPI
|
|
case AV_HWDEVICE_TYPE_VAAPI:
|
|
base_handle[0] = handle_pair->first;
|
|
return 0;
|
|
#endif
|
|
#if CONFIG_D3D11VA
|
|
case AV_HWDEVICE_TYPE_D3D11VA:
|
|
base_handle[0] = handle_pair->first;
|
|
base_handle[1] = handle_pair->second;
|
|
return 0;
|
|
#endif
|
|
#if CONFIG_DXVA2
|
|
case AV_HWDEVICE_TYPE_DXVA2:
|
|
base_handle[0] = handle_pair->first;
|
|
return 0;
|
|
#endif
|
|
}
|
|
return AVERROR(EINVAL);
|
|
}
|
|
|
|
static uint32_t qsv_fourcc_from_pix_fmt(enum AVPixelFormat pix_fmt)
|
|
{
|
|
int i;
|
|
for (i = 0; i < FF_ARRAY_ELEMS(supported_pixel_formats); i++) {
|
|
if (supported_pixel_formats[i].pix_fmt == pix_fmt)
|
|
return supported_pixel_formats[i].fourcc;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static uint16_t qsv_shift_from_pix_fmt(enum AVPixelFormat pix_fmt)
|
|
{
|
|
for (int i = 0; i < FF_ARRAY_ELEMS(supported_pixel_formats); i++) {
|
|
if (supported_pixel_formats[i].pix_fmt == pix_fmt)
|
|
return supported_pixel_formats[i].mfx_shift;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
#if CONFIG_D3D11VA
|
|
static uint32_t qsv_get_d3d11va_bind_flags(int mem_type)
|
|
{
|
|
uint32_t bind_flags = 0;
|
|
|
|
if ((mem_type & MFX_MEMTYPE_VIDEO_MEMORY_ENCODER_TARGET) && (mem_type & MFX_MEMTYPE_INTERNAL_FRAME))
|
|
bind_flags = D3D11_BIND_DECODER | D3D11_BIND_VIDEO_ENCODER;
|
|
else
|
|
bind_flags = D3D11_BIND_DECODER;
|
|
|
|
if ((MFX_MEMTYPE_FROM_VPPOUT & mem_type) || (MFX_MEMTYPE_VIDEO_MEMORY_PROCESSOR_TARGET & mem_type))
|
|
bind_flags = D3D11_BIND_RENDER_TARGET;
|
|
|
|
return bind_flags;
|
|
}
|
|
#endif
|
|
|
|
static int qsv_fill_border(AVFrame *dst, const AVFrame *src)
|
|
{
|
|
const AVPixFmtDescriptor *desc;
|
|
int i, planes_nb = 0;
|
|
if (dst->format != src->format)
|
|
return AVERROR(EINVAL);
|
|
|
|
desc = av_pix_fmt_desc_get(dst->format);
|
|
|
|
for (i = 0; i < desc->nb_components; i++)
|
|
planes_nb = FFMAX(planes_nb, desc->comp[i].plane + 1);
|
|
|
|
for (i = 0; i < planes_nb; i++) {
|
|
int sheight, dheight, y;
|
|
ptrdiff_t swidth = av_image_get_linesize(src->format,
|
|
src->width,
|
|
i);
|
|
ptrdiff_t dwidth = av_image_get_linesize(dst->format,
|
|
dst->width,
|
|
i);
|
|
const AVComponentDescriptor comp = desc->comp[i];
|
|
if (swidth < 0 || dwidth < 0) {
|
|
av_log(NULL, AV_LOG_ERROR, "av_image_get_linesize failed\n");
|
|
return AVERROR(EINVAL);
|
|
}
|
|
sheight = src->height;
|
|
dheight = dst->height;
|
|
if (i) {
|
|
sheight = AV_CEIL_RSHIFT(src->height, desc->log2_chroma_h);
|
|
dheight = AV_CEIL_RSHIFT(dst->height, desc->log2_chroma_h);
|
|
}
|
|
//fill right padding
|
|
for (y = 0; y < sheight; y++) {
|
|
void *line_ptr = dst->data[i] + y*dst->linesize[i] + swidth;
|
|
av_memcpy_backptr(line_ptr,
|
|
comp.depth > 8 ? 2 : 1,
|
|
dwidth - swidth);
|
|
}
|
|
//fill bottom padding
|
|
for (y = sheight; y < dheight; y++) {
|
|
memcpy(dst->data[i]+y*dst->linesize[i],
|
|
dst->data[i]+(sheight-1)*dst->linesize[i],
|
|
dwidth);
|
|
}
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int qsv_device_init(AVHWDeviceContext *ctx)
|
|
{
|
|
AVQSVDeviceContext *hwctx = ctx->hwctx;
|
|
QSVDeviceContext *s = ctx->internal->priv;
|
|
int hw_handle_supported = 0;
|
|
mfxHandleType handle_type;
|
|
enum AVHWDeviceType device_type;
|
|
enum AVPixelFormat pix_fmt;
|
|
mfxStatus err;
|
|
|
|
err = MFXQueryIMPL(hwctx->session, &s->impl);
|
|
if (err == MFX_ERR_NONE)
|
|
err = MFXQueryVersion(hwctx->session, &s->ver);
|
|
if (err != MFX_ERR_NONE) {
|
|
av_log(ctx, AV_LOG_ERROR, "Error querying the session attributes\n");
|
|
return AVERROR_UNKNOWN;
|
|
}
|
|
|
|
if (MFX_IMPL_VIA_VAAPI == MFX_IMPL_VIA_MASK(s->impl)) {
|
|
#if CONFIG_VAAPI
|
|
handle_type = MFX_HANDLE_VA_DISPLAY;
|
|
device_type = AV_HWDEVICE_TYPE_VAAPI;
|
|
pix_fmt = AV_PIX_FMT_VAAPI;
|
|
hw_handle_supported = 1;
|
|
#endif
|
|
} else if (MFX_IMPL_VIA_D3D11 == MFX_IMPL_VIA_MASK(s->impl)) {
|
|
#if CONFIG_D3D11VA
|
|
handle_type = MFX_HANDLE_D3D11_DEVICE;
|
|
device_type = AV_HWDEVICE_TYPE_D3D11VA;
|
|
pix_fmt = AV_PIX_FMT_D3D11;
|
|
hw_handle_supported = 1;
|
|
#endif
|
|
} else if (MFX_IMPL_VIA_D3D9 == MFX_IMPL_VIA_MASK(s->impl)) {
|
|
#if CONFIG_DXVA2
|
|
handle_type = MFX_HANDLE_D3D9_DEVICE_MANAGER;
|
|
device_type = AV_HWDEVICE_TYPE_DXVA2;
|
|
pix_fmt = AV_PIX_FMT_DXVA2_VLD;
|
|
hw_handle_supported = 1;
|
|
#endif
|
|
}
|
|
|
|
if (hw_handle_supported) {
|
|
err = MFXVideoCORE_GetHandle(hwctx->session, handle_type, &s->handle);
|
|
if (err == MFX_ERR_NONE) {
|
|
s->handle_type = handle_type;
|
|
s->child_device_type = device_type;
|
|
s->child_pix_fmt = pix_fmt;
|
|
}
|
|
}
|
|
if (!s->handle) {
|
|
av_log(ctx, AV_LOG_VERBOSE, "No supported hw handle could be retrieved "
|
|
"from the session\n");
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static void qsv_frames_uninit(AVHWFramesContext *ctx)
|
|
{
|
|
QSVFramesContext *s = ctx->internal->priv;
|
|
|
|
if (s->session_download) {
|
|
MFXVideoVPP_Close(s->session_download);
|
|
MFXClose(s->session_download);
|
|
}
|
|
s->session_download = NULL;
|
|
s->session_download_init = 0;
|
|
|
|
if (s->session_upload) {
|
|
MFXVideoVPP_Close(s->session_upload);
|
|
MFXClose(s->session_upload);
|
|
}
|
|
s->session_upload = NULL;
|
|
s->session_upload_init = 0;
|
|
|
|
#if HAVE_PTHREADS
|
|
pthread_mutex_destroy(&s->session_lock);
|
|
#endif
|
|
|
|
av_freep(&s->mem_ids);
|
|
#if QSV_HAVE_OPAQUE
|
|
av_freep(&s->surface_ptrs);
|
|
#endif
|
|
av_freep(&s->surfaces_internal);
|
|
av_freep(&s->handle_pairs_internal);
|
|
av_frame_unref(&s->realigned_upload_frame);
|
|
av_frame_unref(&s->realigned_download_frame);
|
|
av_buffer_unref(&s->child_frames_ref);
|
|
}
|
|
|
|
static void qsv_pool_release_dummy(void *opaque, uint8_t *data)
|
|
{
|
|
}
|
|
|
|
static AVBufferRef *qsv_pool_alloc(void *opaque, size_t size)
|
|
{
|
|
AVHWFramesContext *ctx = (AVHWFramesContext*)opaque;
|
|
QSVFramesContext *s = ctx->internal->priv;
|
|
AVQSVFramesContext *hwctx = ctx->hwctx;
|
|
|
|
if (s->nb_surfaces_used < hwctx->nb_surfaces) {
|
|
s->nb_surfaces_used++;
|
|
return av_buffer_create((uint8_t*)(s->surfaces_internal + s->nb_surfaces_used - 1),
|
|
sizeof(*hwctx->surfaces), qsv_pool_release_dummy, NULL, 0);
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static int qsv_init_child_ctx(AVHWFramesContext *ctx)
|
|
{
|
|
AVQSVFramesContext *hwctx = ctx->hwctx;
|
|
QSVFramesContext *s = ctx->internal->priv;
|
|
QSVDeviceContext *device_priv = ctx->device_ctx->internal->priv;
|
|
|
|
AVBufferRef *child_device_ref = NULL;
|
|
AVBufferRef *child_frames_ref = NULL;
|
|
|
|
AVHWDeviceContext *child_device_ctx;
|
|
AVHWFramesContext *child_frames_ctx;
|
|
|
|
int i, ret = 0;
|
|
|
|
if (!device_priv->handle) {
|
|
av_log(ctx, AV_LOG_ERROR,
|
|
"Cannot create a non-opaque internal surface pool without "
|
|
"a hardware handle\n");
|
|
return AVERROR(EINVAL);
|
|
}
|
|
|
|
child_device_ref = av_hwdevice_ctx_alloc(device_priv->child_device_type);
|
|
if (!child_device_ref)
|
|
return AVERROR(ENOMEM);
|
|
child_device_ctx = (AVHWDeviceContext*)child_device_ref->data;
|
|
|
|
#if CONFIG_VAAPI
|
|
if (child_device_ctx->type == AV_HWDEVICE_TYPE_VAAPI) {
|
|
AVVAAPIDeviceContext *child_device_hwctx = child_device_ctx->hwctx;
|
|
child_device_hwctx->display = (VADisplay)device_priv->handle;
|
|
}
|
|
#endif
|
|
#if CONFIG_D3D11VA
|
|
if (child_device_ctx->type == AV_HWDEVICE_TYPE_D3D11VA) {
|
|
AVD3D11VADeviceContext *child_device_hwctx = child_device_ctx->hwctx;
|
|
ID3D11Device_AddRef((ID3D11Device*)device_priv->handle);
|
|
child_device_hwctx->device = (ID3D11Device*)device_priv->handle;
|
|
}
|
|
#endif
|
|
#if CONFIG_DXVA2
|
|
if (child_device_ctx->type == AV_HWDEVICE_TYPE_DXVA2) {
|
|
AVDXVA2DeviceContext *child_device_hwctx = child_device_ctx->hwctx;
|
|
child_device_hwctx->devmgr = (IDirect3DDeviceManager9*)device_priv->handle;
|
|
}
|
|
#endif
|
|
|
|
ret = av_hwdevice_ctx_init(child_device_ref);
|
|
if (ret < 0) {
|
|
av_log(ctx, AV_LOG_ERROR, "Error initializing a child device context\n");
|
|
goto fail;
|
|
}
|
|
|
|
child_frames_ref = av_hwframe_ctx_alloc(child_device_ref);
|
|
if (!child_frames_ref) {
|
|
ret = AVERROR(ENOMEM);
|
|
goto fail;
|
|
}
|
|
child_frames_ctx = (AVHWFramesContext*)child_frames_ref->data;
|
|
|
|
child_frames_ctx->format = device_priv->child_pix_fmt;
|
|
child_frames_ctx->sw_format = ctx->sw_format;
|
|
child_frames_ctx->initial_pool_size = ctx->initial_pool_size;
|
|
child_frames_ctx->width = FFALIGN(ctx->width, 16);
|
|
child_frames_ctx->height = FFALIGN(ctx->height, 16);
|
|
|
|
#if CONFIG_D3D11VA
|
|
if (child_device_ctx->type == AV_HWDEVICE_TYPE_D3D11VA) {
|
|
AVD3D11VAFramesContext *child_frames_hwctx = child_frames_ctx->hwctx;
|
|
if (hwctx->frame_type == 0)
|
|
hwctx->frame_type = MFX_MEMTYPE_VIDEO_MEMORY_PROCESSOR_TARGET;
|
|
if (hwctx->frame_type & MFX_MEMTYPE_SHARED_RESOURCE)
|
|
child_frames_hwctx->MiscFlags = D3D11_RESOURCE_MISC_SHARED;
|
|
child_frames_hwctx->BindFlags = qsv_get_d3d11va_bind_flags(hwctx->frame_type);
|
|
}
|
|
#endif
|
|
#if CONFIG_DXVA2
|
|
if (child_device_ctx->type == AV_HWDEVICE_TYPE_DXVA2) {
|
|
AVDXVA2FramesContext *child_frames_hwctx = child_frames_ctx->hwctx;
|
|
if (hwctx->frame_type & MFX_MEMTYPE_VIDEO_MEMORY_PROCESSOR_TARGET)
|
|
child_frames_hwctx->surface_type = DXVA2_VideoProcessorRenderTarget;
|
|
else
|
|
child_frames_hwctx->surface_type = DXVA2_VideoDecoderRenderTarget;
|
|
}
|
|
#endif
|
|
|
|
ret = av_hwframe_ctx_init(child_frames_ref);
|
|
if (ret < 0) {
|
|
av_log(ctx, AV_LOG_ERROR, "Error initializing a child frames context\n");
|
|
goto fail;
|
|
}
|
|
|
|
#if CONFIG_VAAPI
|
|
if (child_device_ctx->type == AV_HWDEVICE_TYPE_VAAPI) {
|
|
AVVAAPIFramesContext *child_frames_hwctx = child_frames_ctx->hwctx;
|
|
for (i = 0; i < ctx->initial_pool_size; i++) {
|
|
s->handle_pairs_internal[i].first = child_frames_hwctx->surface_ids + i;
|
|
s->handle_pairs_internal[i].second = (mfxMemId)MFX_INFINITE;
|
|
s->surfaces_internal[i].Data.MemId = (mfxMemId)&s->handle_pairs_internal[i];
|
|
}
|
|
hwctx->frame_type = MFX_MEMTYPE_VIDEO_MEMORY_DECODER_TARGET;
|
|
}
|
|
#endif
|
|
#if CONFIG_D3D11VA
|
|
if (child_device_ctx->type == AV_HWDEVICE_TYPE_D3D11VA) {
|
|
AVD3D11VAFramesContext *child_frames_hwctx = child_frames_ctx->hwctx;
|
|
for (i = 0; i < ctx->initial_pool_size; i++) {
|
|
s->handle_pairs_internal[i].first = (mfxMemId)child_frames_hwctx->texture_infos[i].texture;
|
|
if(child_frames_hwctx->BindFlags & D3D11_BIND_RENDER_TARGET) {
|
|
s->handle_pairs_internal[i].second = (mfxMemId)MFX_INFINITE;
|
|
} else {
|
|
s->handle_pairs_internal[i].second = (mfxMemId)child_frames_hwctx->texture_infos[i].index;
|
|
}
|
|
s->surfaces_internal[i].Data.MemId = (mfxMemId)&s->handle_pairs_internal[i];
|
|
}
|
|
if (child_frames_hwctx->BindFlags & D3D11_BIND_RENDER_TARGET) {
|
|
hwctx->frame_type |= MFX_MEMTYPE_VIDEO_MEMORY_PROCESSOR_TARGET;
|
|
} else {
|
|
hwctx->frame_type |= MFX_MEMTYPE_VIDEO_MEMORY_DECODER_TARGET;
|
|
}
|
|
}
|
|
#endif
|
|
#if CONFIG_DXVA2
|
|
if (child_device_ctx->type == AV_HWDEVICE_TYPE_DXVA2) {
|
|
AVDXVA2FramesContext *child_frames_hwctx = child_frames_ctx->hwctx;
|
|
for (i = 0; i < ctx->initial_pool_size; i++) {
|
|
s->handle_pairs_internal[i].first = (mfxMemId)child_frames_hwctx->surfaces[i];
|
|
s->handle_pairs_internal[i].second = (mfxMemId)MFX_INFINITE;
|
|
s->surfaces_internal[i].Data.MemId = (mfxMemId)&s->handle_pairs_internal[i];
|
|
}
|
|
if (child_frames_hwctx->surface_type == DXVA2_VideoProcessorRenderTarget)
|
|
hwctx->frame_type = MFX_MEMTYPE_VIDEO_MEMORY_PROCESSOR_TARGET;
|
|
else
|
|
hwctx->frame_type = MFX_MEMTYPE_VIDEO_MEMORY_DECODER_TARGET;
|
|
}
|
|
#endif
|
|
|
|
s->child_frames_ref = child_frames_ref;
|
|
child_frames_ref = NULL;
|
|
|
|
fail:
|
|
av_buffer_unref(&child_device_ref);
|
|
av_buffer_unref(&child_frames_ref);
|
|
return ret;
|
|
}
|
|
|
|
static int qsv_init_surface(AVHWFramesContext *ctx, mfxFrameSurface1 *surf)
|
|
{
|
|
const AVPixFmtDescriptor *desc;
|
|
uint32_t fourcc;
|
|
|
|
desc = av_pix_fmt_desc_get(ctx->sw_format);
|
|
if (!desc)
|
|
return AVERROR(EINVAL);
|
|
|
|
fourcc = qsv_fourcc_from_pix_fmt(ctx->sw_format);
|
|
if (!fourcc)
|
|
return AVERROR(EINVAL);
|
|
|
|
surf->Info.BitDepthLuma = desc->comp[0].depth;
|
|
surf->Info.BitDepthChroma = desc->comp[0].depth;
|
|
surf->Info.Shift = qsv_shift_from_pix_fmt(ctx->sw_format);
|
|
|
|
if (desc->log2_chroma_w && desc->log2_chroma_h)
|
|
surf->Info.ChromaFormat = MFX_CHROMAFORMAT_YUV420;
|
|
else if (desc->log2_chroma_w)
|
|
surf->Info.ChromaFormat = MFX_CHROMAFORMAT_YUV422;
|
|
else
|
|
surf->Info.ChromaFormat = MFX_CHROMAFORMAT_YUV444;
|
|
|
|
surf->Info.FourCC = fourcc;
|
|
surf->Info.Width = FFALIGN(ctx->width, 16);
|
|
surf->Info.CropW = ctx->width;
|
|
surf->Info.Height = FFALIGN(ctx->height, 16);
|
|
surf->Info.CropH = ctx->height;
|
|
surf->Info.FrameRateExtN = 25;
|
|
surf->Info.FrameRateExtD = 1;
|
|
surf->Info.PicStruct = MFX_PICSTRUCT_PROGRESSIVE;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int qsv_init_pool(AVHWFramesContext *ctx, uint32_t fourcc)
|
|
{
|
|
QSVFramesContext *s = ctx->internal->priv;
|
|
AVQSVFramesContext *frames_hwctx = ctx->hwctx;
|
|
|
|
int i, ret = 0;
|
|
|
|
if (ctx->initial_pool_size <= 0) {
|
|
av_log(ctx, AV_LOG_ERROR, "QSV requires a fixed frame pool size\n");
|
|
return AVERROR(EINVAL);
|
|
}
|
|
|
|
s->handle_pairs_internal = av_calloc(ctx->initial_pool_size,
|
|
sizeof(*s->handle_pairs_internal));
|
|
if (!s->handle_pairs_internal)
|
|
return AVERROR(ENOMEM);
|
|
|
|
s->surfaces_internal = av_calloc(ctx->initial_pool_size,
|
|
sizeof(*s->surfaces_internal));
|
|
if (!s->surfaces_internal)
|
|
return AVERROR(ENOMEM);
|
|
|
|
for (i = 0; i < ctx->initial_pool_size; i++) {
|
|
ret = qsv_init_surface(ctx, &s->surfaces_internal[i]);
|
|
if (ret < 0)
|
|
return ret;
|
|
}
|
|
|
|
#if QSV_HAVE_OPAQUE
|
|
if (!(frames_hwctx->frame_type & MFX_MEMTYPE_OPAQUE_FRAME)) {
|
|
ret = qsv_init_child_ctx(ctx);
|
|
if (ret < 0)
|
|
return ret;
|
|
}
|
|
#else
|
|
ret = qsv_init_child_ctx(ctx);
|
|
if (ret < 0)
|
|
return ret;
|
|
#endif
|
|
|
|
ctx->internal->pool_internal = av_buffer_pool_init2(sizeof(mfxFrameSurface1),
|
|
ctx, qsv_pool_alloc, NULL);
|
|
if (!ctx->internal->pool_internal)
|
|
return AVERROR(ENOMEM);
|
|
|
|
frames_hwctx->surfaces = s->surfaces_internal;
|
|
frames_hwctx->nb_surfaces = ctx->initial_pool_size;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static mfxStatus frame_alloc(mfxHDL pthis, mfxFrameAllocRequest *req,
|
|
mfxFrameAllocResponse *resp)
|
|
{
|
|
AVHWFramesContext *ctx = pthis;
|
|
QSVFramesContext *s = ctx->internal->priv;
|
|
AVQSVFramesContext *hwctx = ctx->hwctx;
|
|
mfxFrameInfo *i = &req->Info;
|
|
mfxFrameInfo *i1 = &hwctx->surfaces[0].Info;
|
|
|
|
if (!(req->Type & MFX_MEMTYPE_VIDEO_MEMORY_PROCESSOR_TARGET) ||
|
|
!(req->Type & (MFX_MEMTYPE_FROM_VPPIN | MFX_MEMTYPE_FROM_VPPOUT)) ||
|
|
!(req->Type & MFX_MEMTYPE_EXTERNAL_FRAME))
|
|
return MFX_ERR_UNSUPPORTED;
|
|
if (i->Width > i1->Width || i->Height > i1->Height ||
|
|
i->FourCC != i1->FourCC || i->ChromaFormat != i1->ChromaFormat) {
|
|
av_log(ctx, AV_LOG_ERROR, "Mismatching surface properties in an "
|
|
"allocation request: %dx%d %d %d vs %dx%d %d %d\n",
|
|
i->Width, i->Height, i->FourCC, i->ChromaFormat,
|
|
i1->Width, i1->Height, i1->FourCC, i1->ChromaFormat);
|
|
return MFX_ERR_UNSUPPORTED;
|
|
}
|
|
|
|
resp->mids = s->mem_ids;
|
|
resp->NumFrameActual = hwctx->nb_surfaces;
|
|
|
|
return MFX_ERR_NONE;
|
|
}
|
|
|
|
static mfxStatus frame_free(mfxHDL pthis, mfxFrameAllocResponse *resp)
|
|
{
|
|
return MFX_ERR_NONE;
|
|
}
|
|
|
|
static mfxStatus frame_lock(mfxHDL pthis, mfxMemId mid, mfxFrameData *ptr)
|
|
{
|
|
return MFX_ERR_UNSUPPORTED;
|
|
}
|
|
|
|
static mfxStatus frame_unlock(mfxHDL pthis, mfxMemId mid, mfxFrameData *ptr)
|
|
{
|
|
return MFX_ERR_UNSUPPORTED;
|
|
}
|
|
|
|
static mfxStatus frame_get_hdl(mfxHDL pthis, mfxMemId mid, mfxHDL *hdl)
|
|
{
|
|
mfxHDLPair *pair_dst = (mfxHDLPair*)hdl;
|
|
mfxHDLPair *pair_src = (mfxHDLPair*)mid;
|
|
|
|
pair_dst->first = pair_src->first;
|
|
|
|
if (pair_src->second != (mfxMemId)MFX_INFINITE)
|
|
pair_dst->second = pair_src->second;
|
|
return MFX_ERR_NONE;
|
|
}
|
|
|
|
#if QSV_ONEVPL
|
|
|
|
static int qsv_d3d11_update_config(void *ctx, mfxHDL handle, mfxConfig cfg)
|
|
{
|
|
#if CONFIG_D3D11VA
|
|
mfxStatus sts;
|
|
IDXGIAdapter *pDXGIAdapter;
|
|
DXGI_ADAPTER_DESC adapterDesc;
|
|
IDXGIDevice *pDXGIDevice = NULL;
|
|
HRESULT hr;
|
|
ID3D11Device *device = handle;
|
|
mfxVariant impl_value;
|
|
|
|
hr = ID3D11Device_QueryInterface(device, &IID_IDXGIDevice, (void**)&pDXGIDevice);
|
|
if (SUCCEEDED(hr)) {
|
|
hr = IDXGIDevice_GetAdapter(pDXGIDevice, &pDXGIAdapter);
|
|
if (FAILED(hr)) {
|
|
av_log(ctx, AV_LOG_ERROR, "Error IDXGIDevice_GetAdapter %d\n", hr);
|
|
goto fail;
|
|
}
|
|
|
|
hr = IDXGIAdapter_GetDesc(pDXGIAdapter, &adapterDesc);
|
|
if (FAILED(hr)) {
|
|
av_log(ctx, AV_LOG_ERROR, "Error IDXGIAdapter_GetDesc %d\n", hr);
|
|
goto fail;
|
|
}
|
|
} else {
|
|
av_log(ctx, AV_LOG_ERROR, "Error ID3D11Device_QueryInterface %d\n", hr);
|
|
goto fail;
|
|
}
|
|
|
|
impl_value.Type = MFX_VARIANT_TYPE_U16;
|
|
impl_value.Data.U16 = adapterDesc.DeviceId;
|
|
sts = MFXSetConfigFilterProperty(cfg,
|
|
(const mfxU8 *)"mfxExtendedDeviceId.DeviceID", impl_value);
|
|
if (sts != MFX_ERR_NONE) {
|
|
av_log(ctx, AV_LOG_ERROR, "Error adding a MFX configuration"
|
|
"DeviceID property: %d.\n", sts);
|
|
goto fail;
|
|
}
|
|
|
|
impl_value.Type = MFX_VARIANT_TYPE_PTR;
|
|
impl_value.Data.Ptr = &adapterDesc.AdapterLuid;
|
|
sts = MFXSetConfigFilterProperty(cfg,
|
|
(const mfxU8 *)"mfxExtendedDeviceId.DeviceLUID", impl_value);
|
|
if (sts != MFX_ERR_NONE) {
|
|
av_log(ctx, AV_LOG_ERROR, "Error adding a MFX configuration"
|
|
"DeviceLUID property: %d.\n", sts);
|
|
goto fail;
|
|
}
|
|
|
|
impl_value.Type = MFX_VARIANT_TYPE_U32;
|
|
impl_value.Data.U32 = 0x0001;
|
|
sts = MFXSetConfigFilterProperty(cfg,
|
|
(const mfxU8 *)"mfxExtendedDeviceId.LUIDDeviceNodeMask", impl_value);
|
|
if (sts != MFX_ERR_NONE) {
|
|
av_log(ctx, AV_LOG_ERROR, "Error adding a MFX configuration"
|
|
"LUIDDeviceNodeMask property: %d.\n", sts);
|
|
goto fail;
|
|
}
|
|
|
|
return 0;
|
|
|
|
fail:
|
|
#endif
|
|
return AVERROR_UNKNOWN;
|
|
}
|
|
|
|
static int qsv_d3d9_update_config(void *ctx, mfxHDL handle, mfxConfig cfg)
|
|
{
|
|
int ret = AVERROR_UNKNOWN;
|
|
#if CONFIG_DXVA2
|
|
mfxStatus sts;
|
|
IDirect3DDeviceManager9* devmgr = handle;
|
|
IDirect3DDevice9Ex *device = NULL;
|
|
HANDLE device_handle = 0;
|
|
IDirect3D9Ex *d3d9ex = NULL;
|
|
LUID luid;
|
|
D3DDEVICE_CREATION_PARAMETERS params;
|
|
HRESULT hr;
|
|
mfxVariant impl_value;
|
|
|
|
hr = IDirect3DDeviceManager9_OpenDeviceHandle(devmgr, &device_handle);
|
|
if (FAILED(hr)) {
|
|
av_log(ctx, AV_LOG_ERROR, "Error OpenDeviceHandle %d\n", hr);
|
|
goto fail;
|
|
}
|
|
|
|
hr = IDirect3DDeviceManager9_LockDevice(devmgr, device_handle, &device, TRUE);
|
|
if (FAILED(hr)) {
|
|
av_log(ctx, AV_LOG_ERROR, "Error LockDevice %d\n", hr);
|
|
goto fail;
|
|
}
|
|
|
|
hr = IDirect3DDevice9Ex_GetCreationParameters(device, ¶ms);
|
|
if (FAILED(hr)) {
|
|
av_log(ctx, AV_LOG_ERROR, "Error IDirect3DDevice9_GetCreationParameters %d\n", hr);
|
|
goto unlock;
|
|
}
|
|
|
|
hr = IDirect3DDevice9Ex_GetDirect3D(device, &d3d9ex);
|
|
if (FAILED(hr)) {
|
|
av_log(ctx, AV_LOG_ERROR, "Error IDirect3DDevice9Ex_GetAdapterLUID %d\n", hr);
|
|
goto unlock;
|
|
}
|
|
|
|
hr = IDirect3D9Ex_GetAdapterLUID(d3d9ex, params.AdapterOrdinal, &luid);
|
|
if (FAILED(hr)) {
|
|
av_log(ctx, AV_LOG_ERROR, "Error IDirect3DDevice9Ex_GetAdapterLUID %d\n", hr);
|
|
goto unlock;
|
|
}
|
|
|
|
impl_value.Type = MFX_VARIANT_TYPE_PTR;
|
|
impl_value.Data.Ptr = &luid;
|
|
sts = MFXSetConfigFilterProperty(cfg,
|
|
(const mfxU8 *)"mfxExtendedDeviceId.DeviceLUID", impl_value);
|
|
if (sts != MFX_ERR_NONE) {
|
|
av_log(ctx, AV_LOG_ERROR, "Error adding a MFX configuration"
|
|
"DeviceLUID property: %d.\n", sts);
|
|
goto unlock;
|
|
}
|
|
|
|
ret = 0;
|
|
|
|
unlock:
|
|
IDirect3DDeviceManager9_UnlockDevice(devmgr, device_handle, FALSE);
|
|
fail:
|
|
#endif
|
|
return ret;
|
|
}
|
|
|
|
static int qsv_va_update_config(void *ctx, mfxHDL handle, mfxConfig cfg)
|
|
{
|
|
#if CONFIG_VAAPI
|
|
#if VA_CHECK_VERSION(1, 15, 0)
|
|
mfxStatus sts;
|
|
VADisplay dpy = handle;
|
|
VAStatus vas;
|
|
VADisplayAttribute attr = {
|
|
.type = VADisplayPCIID,
|
|
};
|
|
mfxVariant impl_value;
|
|
|
|
vas = vaGetDisplayAttributes(dpy, &attr, 1);
|
|
if (vas == VA_STATUS_SUCCESS && attr.flags != VA_DISPLAY_ATTRIB_NOT_SUPPORTED) {
|
|
impl_value.Type = MFX_VARIANT_TYPE_U16;
|
|
impl_value.Data.U16 = (attr.value & 0xFFFF);
|
|
sts = MFXSetConfigFilterProperty(cfg,
|
|
(const mfxU8 *)"mfxExtendedDeviceId.DeviceID", impl_value);
|
|
if (sts != MFX_ERR_NONE) {
|
|
av_log(ctx, AV_LOG_ERROR, "Error adding a MFX configuration"
|
|
"DeviceID property: %d.\n", sts);
|
|
goto fail;
|
|
}
|
|
} else {
|
|
av_log(ctx, AV_LOG_ERROR, "libva: Failed to get device id from the driver. Please "
|
|
"consider to upgrade the driver to support VA-API 1.15.0\n");
|
|
goto fail;
|
|
}
|
|
|
|
return 0;
|
|
|
|
fail:
|
|
#else
|
|
av_log(ctx, AV_LOG_ERROR, "libva: This version of libva doesn't support retrieving "
|
|
"the device information from the driver. Please consider to upgrade libva to "
|
|
"support VA-API 1.15.0\n");
|
|
#endif
|
|
#endif
|
|
return AVERROR_UNKNOWN;
|
|
}
|
|
|
|
static int qsv_new_mfx_loader(void *ctx,
|
|
mfxHDL handle,
|
|
mfxHandleType handle_type,
|
|
mfxIMPL implementation,
|
|
mfxVersion *pver,
|
|
void **ploader)
|
|
{
|
|
mfxStatus sts;
|
|
mfxLoader loader = NULL;
|
|
mfxConfig cfg;
|
|
mfxVariant impl_value;
|
|
|
|
*ploader = NULL;
|
|
loader = MFXLoad();
|
|
if (!loader) {
|
|
av_log(ctx, AV_LOG_ERROR, "Error creating a MFX loader\n");
|
|
goto fail;
|
|
}
|
|
|
|
/* Create configurations for implementation */
|
|
cfg = MFXCreateConfig(loader);
|
|
if (!cfg) {
|
|
av_log(ctx, AV_LOG_ERROR, "Error creating a MFX configuration\n");
|
|
goto fail;
|
|
}
|
|
|
|
impl_value.Type = MFX_VARIANT_TYPE_U32;
|
|
impl_value.Data.U32 = (implementation == MFX_IMPL_SOFTWARE) ?
|
|
MFX_IMPL_TYPE_SOFTWARE : MFX_IMPL_TYPE_HARDWARE;
|
|
sts = MFXSetConfigFilterProperty(cfg,
|
|
(const mfxU8 *)"mfxImplDescription.Impl", impl_value);
|
|
if (sts != MFX_ERR_NONE) {
|
|
av_log(ctx, AV_LOG_ERROR, "Error adding a MFX configuration "
|
|
"property: %d.\n", sts);
|
|
goto fail;
|
|
}
|
|
|
|
impl_value.Type = MFX_VARIANT_TYPE_U32;
|
|
impl_value.Data.U32 = pver->Version;
|
|
sts = MFXSetConfigFilterProperty(cfg,
|
|
(const mfxU8 *)"mfxImplDescription.ApiVersion.Version",
|
|
impl_value);
|
|
if (sts != MFX_ERR_NONE) {
|
|
av_log(ctx, AV_LOG_ERROR, "Error adding a MFX configuration "
|
|
"property: %d.\n", sts);
|
|
goto fail;
|
|
}
|
|
|
|
impl_value.Type = MFX_VARIANT_TYPE_U16;
|
|
impl_value.Data.U16 = 0x8086; // Intel device only
|
|
sts = MFXSetConfigFilterProperty(cfg,
|
|
(const mfxU8 *)"mfxExtendedDeviceId.VendorID", impl_value);
|
|
if (sts != MFX_ERR_NONE) {
|
|
av_log(ctx, AV_LOG_ERROR, "Error adding a MFX configuration"
|
|
"VendorID property: %d.\n", sts);
|
|
goto fail;
|
|
}
|
|
|
|
if (MFX_HANDLE_VA_DISPLAY == handle_type) {
|
|
if (handle && qsv_va_update_config(ctx, handle, cfg))
|
|
goto fail;
|
|
|
|
impl_value.Data.U32 = MFX_ACCEL_MODE_VIA_VAAPI;
|
|
} else if (MFX_HANDLE_D3D9_DEVICE_MANAGER == handle_type) {
|
|
if (handle && qsv_d3d9_update_config(ctx, handle, cfg))
|
|
goto fail;
|
|
|
|
impl_value.Data.U32 = MFX_ACCEL_MODE_VIA_D3D9;
|
|
} else {
|
|
if (handle && qsv_d3d11_update_config(ctx, handle, cfg))
|
|
goto fail;
|
|
|
|
impl_value.Data.U32 = MFX_ACCEL_MODE_VIA_D3D11;
|
|
}
|
|
|
|
impl_value.Type = MFX_VARIANT_TYPE_U32;
|
|
sts = MFXSetConfigFilterProperty(cfg,
|
|
(const mfxU8 *)"mfxImplDescription.AccelerationMode", impl_value);
|
|
if (sts != MFX_ERR_NONE) {
|
|
av_log(ctx, AV_LOG_ERROR, "Error adding a MFX configuration"
|
|
"AccelerationMode property: %d.\n", sts);
|
|
goto fail;
|
|
}
|
|
|
|
*ploader = loader;
|
|
|
|
return 0;
|
|
|
|
fail:
|
|
if (loader)
|
|
MFXUnload(loader);
|
|
|
|
return AVERROR_UNKNOWN;
|
|
}
|
|
|
|
static int qsv_create_mfx_session_from_loader(void *ctx, mfxLoader loader, mfxSession *psession)
|
|
{
|
|
mfxStatus sts;
|
|
mfxSession session = NULL;
|
|
uint32_t impl_idx = 0;
|
|
mfxVersion ver;
|
|
|
|
while (1) {
|
|
/* Enumerate all implementations */
|
|
mfxImplDescription *impl_desc;
|
|
|
|
sts = MFXEnumImplementations(loader, impl_idx,
|
|
MFX_IMPLCAPS_IMPLDESCSTRUCTURE,
|
|
(mfxHDL *)&impl_desc);
|
|
/* Failed to find an available implementation */
|
|
if (sts == MFX_ERR_NOT_FOUND)
|
|
break;
|
|
else if (sts != MFX_ERR_NONE) {
|
|
impl_idx++;
|
|
continue;
|
|
}
|
|
|
|
sts = MFXCreateSession(loader, impl_idx, &session);
|
|
MFXDispReleaseImplDescription(loader, impl_desc);
|
|
if (sts == MFX_ERR_NONE)
|
|
break;
|
|
|
|
impl_idx++;
|
|
}
|
|
|
|
if (sts != MFX_ERR_NONE) {
|
|
av_log(ctx, AV_LOG_ERROR, "Error creating a MFX session: %d.\n", sts);
|
|
goto fail;
|
|
}
|
|
|
|
sts = MFXQueryVersion(session, &ver);
|
|
if (sts != MFX_ERR_NONE) {
|
|
av_log(ctx, AV_LOG_ERROR, "Error querying a MFX session: %d.\n", sts);
|
|
goto fail;
|
|
}
|
|
|
|
av_log(ctx, AV_LOG_VERBOSE, "Initialize MFX session: implementation "
|
|
"version is %d.%d\n", ver.Major, ver.Minor);
|
|
|
|
*psession = session;
|
|
|
|
return 0;
|
|
|
|
fail:
|
|
if (session)
|
|
MFXClose(session);
|
|
|
|
return AVERROR_UNKNOWN;
|
|
}
|
|
|
|
static int qsv_create_mfx_session(void *ctx,
|
|
mfxHDL handle,
|
|
mfxHandleType handle_type,
|
|
mfxIMPL implementation,
|
|
mfxVersion *pver,
|
|
mfxSession *psession,
|
|
void **ploader)
|
|
{
|
|
mfxLoader loader = NULL;
|
|
|
|
av_log(ctx, AV_LOG_VERBOSE,
|
|
"Use Intel(R) oneVPL to create MFX session, API version is "
|
|
"%d.%d, the required implementation version is %d.%d\n",
|
|
MFX_VERSION_MAJOR, MFX_VERSION_MINOR, pver->Major, pver->Minor);
|
|
|
|
if (handle_type != MFX_HANDLE_VA_DISPLAY &&
|
|
handle_type != MFX_HANDLE_D3D9_DEVICE_MANAGER &&
|
|
handle_type != MFX_HANDLE_D3D11_DEVICE) {
|
|
av_log(ctx, AV_LOG_ERROR,
|
|
"Invalid MFX device handle type\n");
|
|
return AVERROR(EXDEV);
|
|
}
|
|
|
|
*psession = NULL;
|
|
|
|
if (!*ploader) {
|
|
if (qsv_new_mfx_loader(ctx, handle, handle_type, implementation, pver, (void **)&loader))
|
|
goto fail;
|
|
|
|
av_assert0(loader);
|
|
} else
|
|
loader = *ploader; // Use the input mfxLoader to create mfx session
|
|
|
|
if (qsv_create_mfx_session_from_loader(ctx, loader, psession))
|
|
goto fail;
|
|
|
|
if (!*ploader)
|
|
*ploader = loader;
|
|
|
|
return 0;
|
|
|
|
fail:
|
|
if (!*ploader && loader)
|
|
MFXUnload(loader);
|
|
|
|
return AVERROR_UNKNOWN;
|
|
}
|
|
|
|
#else
|
|
|
|
static int qsv_create_mfx_session(void *ctx,
|
|
mfxHDL handle,
|
|
mfxHandleType handle_type,
|
|
mfxIMPL implementation,
|
|
mfxVersion *pver,
|
|
mfxSession *psession,
|
|
void **ploader)
|
|
{
|
|
mfxVersion ver;
|
|
mfxStatus sts;
|
|
mfxSession session = NULL;
|
|
|
|
av_log(ctx, AV_LOG_VERBOSE,
|
|
"Use Intel(R) Media SDK to create MFX session, API version is "
|
|
"%d.%d, the required implementation version is %d.%d\n",
|
|
MFX_VERSION_MAJOR, MFX_VERSION_MINOR, pver->Major, pver->Minor);
|
|
|
|
*ploader = NULL;
|
|
*psession = NULL;
|
|
ver = *pver;
|
|
sts = MFXInit(implementation, &ver, &session);
|
|
if (sts != MFX_ERR_NONE) {
|
|
av_log(ctx, AV_LOG_ERROR, "Error initializing an MFX session: "
|
|
"%d.\n", sts);
|
|
goto fail;
|
|
}
|
|
|
|
sts = MFXQueryVersion(session, &ver);
|
|
if (sts != MFX_ERR_NONE) {
|
|
av_log(ctx, AV_LOG_ERROR, "Error querying an MFX session: "
|
|
"%d.\n", sts);
|
|
goto fail;
|
|
}
|
|
|
|
av_log(ctx, AV_LOG_VERBOSE, "Initialize MFX session: implementation "
|
|
"version is %d.%d\n", ver.Major, ver.Minor);
|
|
|
|
MFXClose(session);
|
|
|
|
sts = MFXInit(implementation, &ver, &session);
|
|
if (sts != MFX_ERR_NONE) {
|
|
av_log(ctx, AV_LOG_ERROR, "Error initializing an MFX session: "
|
|
"%d.\n", sts);
|
|
goto fail;
|
|
}
|
|
|
|
*psession = session;
|
|
|
|
return 0;
|
|
|
|
fail:
|
|
if (session)
|
|
MFXClose(session);
|
|
|
|
return AVERROR_UNKNOWN;
|
|
}
|
|
|
|
#endif
|
|
|
|
static int qsv_init_internal_session(AVHWFramesContext *ctx,
|
|
mfxSession *session, int upload)
|
|
{
|
|
AVQSVFramesContext *frames_hwctx = ctx->hwctx;
|
|
QSVDeviceContext *device_priv = ctx->device_ctx->internal->priv;
|
|
int opaque = 0;
|
|
|
|
mfxFrameAllocator frame_allocator = {
|
|
.pthis = ctx,
|
|
.Alloc = frame_alloc,
|
|
.Lock = frame_lock,
|
|
.Unlock = frame_unlock,
|
|
.GetHDL = frame_get_hdl,
|
|
.Free = frame_free,
|
|
};
|
|
|
|
mfxVideoParam par;
|
|
mfxStatus err;
|
|
int ret = AVERROR_UNKNOWN;
|
|
AVQSVDeviceContext *hwctx = ctx->device_ctx->hwctx;
|
|
/* hwctx->loader is non-NULL for oneVPL user and NULL for non-oneVPL user */
|
|
void **loader = &hwctx->loader;
|
|
|
|
#if QSV_HAVE_OPAQUE
|
|
QSVFramesContext *s = ctx->internal->priv;
|
|
opaque = !!(frames_hwctx->frame_type & MFX_MEMTYPE_OPAQUE_FRAME);
|
|
#endif
|
|
|
|
ret = qsv_create_mfx_session(ctx, device_priv->handle, device_priv->handle_type,
|
|
device_priv->impl, &device_priv->ver, session, loader);
|
|
if (ret)
|
|
goto fail;
|
|
|
|
if (device_priv->handle) {
|
|
err = MFXVideoCORE_SetHandle(*session, device_priv->handle_type,
|
|
device_priv->handle);
|
|
if (err != MFX_ERR_NONE) {
|
|
ret = AVERROR_UNKNOWN;
|
|
goto fail;
|
|
}
|
|
}
|
|
|
|
if (!opaque) {
|
|
err = MFXVideoCORE_SetFrameAllocator(*session, &frame_allocator);
|
|
if (err != MFX_ERR_NONE) {
|
|
ret = AVERROR_UNKNOWN;
|
|
goto fail;
|
|
}
|
|
}
|
|
|
|
memset(&par, 0, sizeof(par));
|
|
|
|
if (!opaque) {
|
|
par.IOPattern = upload ? MFX_IOPATTERN_OUT_VIDEO_MEMORY :
|
|
MFX_IOPATTERN_IN_VIDEO_MEMORY;
|
|
}
|
|
#if QSV_HAVE_OPAQUE
|
|
else {
|
|
par.ExtParam = s->ext_buffers;
|
|
par.NumExtParam = FF_ARRAY_ELEMS(s->ext_buffers);
|
|
par.IOPattern = upload ? MFX_IOPATTERN_OUT_OPAQUE_MEMORY :
|
|
MFX_IOPATTERN_IN_OPAQUE_MEMORY;
|
|
}
|
|
#endif
|
|
|
|
par.IOPattern |= upload ? MFX_IOPATTERN_IN_SYSTEM_MEMORY :
|
|
MFX_IOPATTERN_OUT_SYSTEM_MEMORY;
|
|
par.AsyncDepth = 1;
|
|
|
|
par.vpp.In = frames_hwctx->surfaces[0].Info;
|
|
|
|
/* Apparently VPP requires the frame rate to be set to some value, otherwise
|
|
* init will fail (probably for the framerate conversion filter). Since we
|
|
* are only doing data upload/download here, we just invent an arbitrary
|
|
* value */
|
|
par.vpp.In.FrameRateExtN = 25;
|
|
par.vpp.In.FrameRateExtD = 1;
|
|
par.vpp.Out = par.vpp.In;
|
|
|
|
err = MFXVideoVPP_Init(*session, &par);
|
|
if (err != MFX_ERR_NONE) {
|
|
av_log(ctx, AV_LOG_VERBOSE, "Error opening the internal VPP session."
|
|
"Surface upload/download will not be possible\n");
|
|
|
|
ret = AVERROR_UNKNOWN;
|
|
goto fail;
|
|
}
|
|
|
|
return 0;
|
|
|
|
fail:
|
|
if (*session)
|
|
MFXClose(*session);
|
|
|
|
*session = NULL;
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int qsv_frames_init(AVHWFramesContext *ctx)
|
|
{
|
|
QSVFramesContext *s = ctx->internal->priv;
|
|
AVQSVFramesContext *frames_hwctx = ctx->hwctx;
|
|
|
|
int opaque = 0;
|
|
|
|
uint32_t fourcc;
|
|
int i, ret;
|
|
|
|
#if QSV_HAVE_OPAQUE
|
|
opaque = !!(frames_hwctx->frame_type & MFX_MEMTYPE_OPAQUE_FRAME);
|
|
#endif
|
|
|
|
fourcc = qsv_fourcc_from_pix_fmt(ctx->sw_format);
|
|
if (!fourcc) {
|
|
av_log(ctx, AV_LOG_ERROR, "Unsupported pixel format\n");
|
|
return AVERROR(ENOSYS);
|
|
}
|
|
|
|
if (!ctx->pool) {
|
|
ret = qsv_init_pool(ctx, fourcc);
|
|
if (ret < 0) {
|
|
av_log(ctx, AV_LOG_ERROR, "Error creating an internal frame pool\n");
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
if (!opaque) {
|
|
s->mem_ids = av_calloc(frames_hwctx->nb_surfaces, sizeof(*s->mem_ids));
|
|
if (!s->mem_ids)
|
|
return AVERROR(ENOMEM);
|
|
|
|
for (i = 0; i < frames_hwctx->nb_surfaces; i++)
|
|
s->mem_ids[i] = frames_hwctx->surfaces[i].Data.MemId;
|
|
}
|
|
#if QSV_HAVE_OPAQUE
|
|
else {
|
|
s->surface_ptrs = av_calloc(frames_hwctx->nb_surfaces,
|
|
sizeof(*s->surface_ptrs));
|
|
if (!s->surface_ptrs)
|
|
return AVERROR(ENOMEM);
|
|
|
|
for (i = 0; i < frames_hwctx->nb_surfaces; i++)
|
|
s->surface_ptrs[i] = frames_hwctx->surfaces + i;
|
|
|
|
s->opaque_alloc.In.Surfaces = s->surface_ptrs;
|
|
s->opaque_alloc.In.NumSurface = frames_hwctx->nb_surfaces;
|
|
s->opaque_alloc.In.Type = frames_hwctx->frame_type;
|
|
|
|
s->opaque_alloc.Out = s->opaque_alloc.In;
|
|
|
|
s->opaque_alloc.Header.BufferId = MFX_EXTBUFF_OPAQUE_SURFACE_ALLOCATION;
|
|
s->opaque_alloc.Header.BufferSz = sizeof(s->opaque_alloc);
|
|
|
|
s->ext_buffers[0] = (mfxExtBuffer*)&s->opaque_alloc;
|
|
}
|
|
#endif
|
|
|
|
s->session_download = NULL;
|
|
s->session_upload = NULL;
|
|
|
|
s->session_download_init = 0;
|
|
s->session_upload_init = 0;
|
|
|
|
#if HAVE_PTHREADS
|
|
pthread_mutex_init(&s->session_lock, NULL);
|
|
#endif
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int qsv_get_buffer(AVHWFramesContext *ctx, AVFrame *frame)
|
|
{
|
|
frame->buf[0] = av_buffer_pool_get(ctx->pool);
|
|
if (!frame->buf[0])
|
|
return AVERROR(ENOMEM);
|
|
|
|
frame->data[3] = frame->buf[0]->data;
|
|
frame->format = AV_PIX_FMT_QSV;
|
|
frame->width = ctx->width;
|
|
frame->height = ctx->height;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int qsv_transfer_get_formats(AVHWFramesContext *ctx,
|
|
enum AVHWFrameTransferDirection dir,
|
|
enum AVPixelFormat **formats)
|
|
{
|
|
enum AVPixelFormat *fmts;
|
|
|
|
fmts = av_malloc_array(2, sizeof(*fmts));
|
|
if (!fmts)
|
|
return AVERROR(ENOMEM);
|
|
|
|
fmts[0] = ctx->sw_format;
|
|
fmts[1] = AV_PIX_FMT_NONE;
|
|
|
|
*formats = fmts;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int qsv_frames_derive_from(AVHWFramesContext *dst_ctx,
|
|
AVHWFramesContext *src_ctx, int flags)
|
|
{
|
|
AVQSVFramesContext *src_hwctx = src_ctx->hwctx;
|
|
int i;
|
|
|
|
switch (dst_ctx->device_ctx->type) {
|
|
#if CONFIG_VAAPI
|
|
case AV_HWDEVICE_TYPE_VAAPI:
|
|
{
|
|
AVVAAPIFramesContext *dst_hwctx = dst_ctx->hwctx;
|
|
dst_hwctx->surface_ids = av_calloc(src_hwctx->nb_surfaces,
|
|
sizeof(*dst_hwctx->surface_ids));
|
|
if (!dst_hwctx->surface_ids)
|
|
return AVERROR(ENOMEM);
|
|
for (i = 0; i < src_hwctx->nb_surfaces; i++) {
|
|
mfxHDLPair *pair = (mfxHDLPair*)src_hwctx->surfaces[i].Data.MemId;
|
|
dst_hwctx->surface_ids[i] = *(VASurfaceID*)pair->first;
|
|
}
|
|
dst_hwctx->nb_surfaces = src_hwctx->nb_surfaces;
|
|
}
|
|
break;
|
|
#endif
|
|
#if CONFIG_D3D11VA
|
|
case AV_HWDEVICE_TYPE_D3D11VA:
|
|
{
|
|
D3D11_TEXTURE2D_DESC texDesc;
|
|
dst_ctx->initial_pool_size = src_ctx->initial_pool_size;
|
|
AVD3D11VAFramesContext *dst_hwctx = dst_ctx->hwctx;
|
|
dst_hwctx->texture_infos = av_calloc(src_hwctx->nb_surfaces,
|
|
sizeof(*dst_hwctx->texture_infos));
|
|
if (!dst_hwctx->texture_infos)
|
|
return AVERROR(ENOMEM);
|
|
if (src_hwctx->frame_type & MFX_MEMTYPE_SHARED_RESOURCE)
|
|
dst_hwctx->MiscFlags = D3D11_RESOURCE_MISC_SHARED;
|
|
for (i = 0; i < src_hwctx->nb_surfaces; i++) {
|
|
mfxHDLPair *pair = (mfxHDLPair*)src_hwctx->surfaces[i].Data.MemId;
|
|
dst_hwctx->texture_infos[i].texture = (ID3D11Texture2D*)pair->first;
|
|
dst_hwctx->texture_infos[i].index = pair->second == (mfxMemId)MFX_INFINITE ? (intptr_t)0 : (intptr_t)pair->second;
|
|
}
|
|
ID3D11Texture2D_GetDesc(dst_hwctx->texture_infos[0].texture, &texDesc);
|
|
dst_hwctx->BindFlags = texDesc.BindFlags;
|
|
}
|
|
break;
|
|
#endif
|
|
#if CONFIG_DXVA2
|
|
case AV_HWDEVICE_TYPE_DXVA2:
|
|
{
|
|
AVDXVA2FramesContext *dst_hwctx = dst_ctx->hwctx;
|
|
dst_hwctx->surfaces = av_calloc(src_hwctx->nb_surfaces,
|
|
sizeof(*dst_hwctx->surfaces));
|
|
if (!dst_hwctx->surfaces)
|
|
return AVERROR(ENOMEM);
|
|
for (i = 0; i < src_hwctx->nb_surfaces; i++) {
|
|
mfxHDLPair *pair = (mfxHDLPair*)src_hwctx->surfaces[i].Data.MemId;
|
|
dst_hwctx->surfaces[i] = (IDirect3DSurface9*)pair->first;
|
|
}
|
|
dst_hwctx->nb_surfaces = src_hwctx->nb_surfaces;
|
|
if (src_hwctx->frame_type == MFX_MEMTYPE_VIDEO_MEMORY_DECODER_TARGET)
|
|
dst_hwctx->surface_type = DXVA2_VideoDecoderRenderTarget;
|
|
else
|
|
dst_hwctx->surface_type = DXVA2_VideoProcessorRenderTarget;
|
|
}
|
|
break;
|
|
#endif
|
|
default:
|
|
return AVERROR(ENOSYS);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int qsv_map_from(AVHWFramesContext *ctx,
|
|
AVFrame *dst, const AVFrame *src, int flags)
|
|
{
|
|
QSVFramesContext *s = ctx->internal->priv;
|
|
mfxFrameSurface1 *surf = (mfxFrameSurface1*)src->data[3];
|
|
AVHWFramesContext *child_frames_ctx;
|
|
const AVPixFmtDescriptor *desc;
|
|
uint8_t *child_data;
|
|
AVFrame *dummy;
|
|
int ret = 0;
|
|
|
|
if (!s->child_frames_ref)
|
|
return AVERROR(ENOSYS);
|
|
child_frames_ctx = (AVHWFramesContext*)s->child_frames_ref->data;
|
|
|
|
switch (child_frames_ctx->device_ctx->type) {
|
|
#if CONFIG_VAAPI
|
|
case AV_HWDEVICE_TYPE_VAAPI:
|
|
{
|
|
mfxHDLPair *pair = (mfxHDLPair*)surf->Data.MemId;
|
|
/* pair->first is *VASurfaceID while data[3] in vaapi frame is VASurfaceID, so
|
|
* we need this casting for vaapi.
|
|
* Add intptr_t to force cast from VASurfaceID(uint) type to pointer(long) type
|
|
* to avoid compile warning */
|
|
child_data = (uint8_t*)(intptr_t)*(VASurfaceID*)pair->first;
|
|
break;
|
|
}
|
|
#endif
|
|
#if CONFIG_D3D11VA
|
|
case AV_HWDEVICE_TYPE_D3D11VA:
|
|
{
|
|
mfxHDLPair *pair = (mfxHDLPair*)surf->Data.MemId;
|
|
child_data = pair->first;
|
|
break;
|
|
}
|
|
#endif
|
|
#if CONFIG_DXVA2
|
|
case AV_HWDEVICE_TYPE_DXVA2:
|
|
{
|
|
mfxHDLPair *pair = (mfxHDLPair*)surf->Data.MemId;
|
|
child_data = pair->first;
|
|
break;
|
|
}
|
|
#endif
|
|
default:
|
|
return AVERROR(ENOSYS);
|
|
}
|
|
|
|
if (dst->format == child_frames_ctx->format) {
|
|
ret = ff_hwframe_map_create(s->child_frames_ref,
|
|
dst, src, NULL, NULL);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
dst->width = src->width;
|
|
dst->height = src->height;
|
|
|
|
if (child_frames_ctx->device_ctx->type == AV_HWDEVICE_TYPE_D3D11VA) {
|
|
mfxHDLPair *pair = (mfxHDLPair*)surf->Data.MemId;
|
|
dst->data[0] = pair->first;
|
|
dst->data[1] = pair->second == (mfxMemId)MFX_INFINITE ? (uint8_t *)0 : pair->second;
|
|
} else {
|
|
dst->data[3] = child_data;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
desc = av_pix_fmt_desc_get(dst->format);
|
|
if (desc && desc->flags & AV_PIX_FMT_FLAG_HWACCEL) {
|
|
// This only supports mapping to software.
|
|
return AVERROR(ENOSYS);
|
|
}
|
|
|
|
dummy = av_frame_alloc();
|
|
if (!dummy)
|
|
return AVERROR(ENOMEM);
|
|
|
|
dummy->buf[0] = av_buffer_ref(src->buf[0]);
|
|
dummy->hw_frames_ctx = av_buffer_ref(s->child_frames_ref);
|
|
if (!dummy->buf[0] || !dummy->hw_frames_ctx)
|
|
goto fail;
|
|
|
|
dummy->format = child_frames_ctx->format;
|
|
dummy->width = src->width;
|
|
dummy->height = src->height;
|
|
|
|
if (child_frames_ctx->device_ctx->type == AV_HWDEVICE_TYPE_D3D11VA) {
|
|
mfxHDLPair *pair = (mfxHDLPair*)surf->Data.MemId;
|
|
dummy->data[0] = pair->first;
|
|
dummy->data[1] = pair->second == (mfxMemId)MFX_INFINITE ? (uint8_t *)0 : pair->second;
|
|
} else {
|
|
dummy->data[3] = child_data;
|
|
}
|
|
|
|
ret = av_hwframe_map(dst, dummy, flags);
|
|
|
|
fail:
|
|
av_frame_free(&dummy);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int qsv_transfer_data_child(AVHWFramesContext *ctx, AVFrame *dst,
|
|
const AVFrame *src)
|
|
{
|
|
QSVFramesContext *s = ctx->internal->priv;
|
|
AVHWFramesContext *child_frames_ctx = (AVHWFramesContext*)s->child_frames_ref->data;
|
|
int download = !!src->hw_frames_ctx;
|
|
mfxFrameSurface1 *surf = (mfxFrameSurface1*)(download ? src->data[3] : dst->data[3]);
|
|
|
|
AVFrame *dummy;
|
|
int ret;
|
|
|
|
dummy = av_frame_alloc();
|
|
if (!dummy)
|
|
return AVERROR(ENOMEM);
|
|
|
|
dummy->format = child_frames_ctx->format;
|
|
dummy->width = src->width;
|
|
dummy->height = src->height;
|
|
dummy->buf[0] = download ? src->buf[0] : dst->buf[0];
|
|
dummy->data[3] = surf->Data.MemId;
|
|
dummy->hw_frames_ctx = s->child_frames_ref;
|
|
|
|
ret = download ? av_hwframe_transfer_data(dst, dummy, 0) :
|
|
av_hwframe_transfer_data(dummy, src, 0);
|
|
|
|
dummy->buf[0] = NULL;
|
|
dummy->data[3] = NULL;
|
|
dummy->hw_frames_ctx = NULL;
|
|
|
|
av_frame_free(&dummy);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int map_frame_to_surface(const AVFrame *frame, mfxFrameSurface1 *surface)
|
|
{
|
|
switch (frame->format) {
|
|
case AV_PIX_FMT_NV12:
|
|
case AV_PIX_FMT_P010:
|
|
case AV_PIX_FMT_P012:
|
|
surface->Data.Y = frame->data[0];
|
|
surface->Data.UV = frame->data[1];
|
|
break;
|
|
|
|
case AV_PIX_FMT_YUV420P:
|
|
surface->Data.Y = frame->data[0];
|
|
surface->Data.U = frame->data[1];
|
|
surface->Data.V = frame->data[2];
|
|
break;
|
|
|
|
case AV_PIX_FMT_BGRA:
|
|
surface->Data.B = frame->data[0];
|
|
surface->Data.G = frame->data[0] + 1;
|
|
surface->Data.R = frame->data[0] + 2;
|
|
surface->Data.A = frame->data[0] + 3;
|
|
break;
|
|
#if CONFIG_VAAPI
|
|
case AV_PIX_FMT_YUYV422:
|
|
surface->Data.Y = frame->data[0];
|
|
surface->Data.U = frame->data[0] + 1;
|
|
surface->Data.V = frame->data[0] + 3;
|
|
break;
|
|
|
|
case AV_PIX_FMT_Y210:
|
|
case AV_PIX_FMT_Y212:
|
|
surface->Data.Y16 = (mfxU16 *)frame->data[0];
|
|
surface->Data.U16 = (mfxU16 *)frame->data[0] + 1;
|
|
surface->Data.V16 = (mfxU16 *)frame->data[0] + 3;
|
|
break;
|
|
case AV_PIX_FMT_VUYX:
|
|
surface->Data.V = frame->data[0];
|
|
surface->Data.U = frame->data[0] + 1;
|
|
surface->Data.Y = frame->data[0] + 2;
|
|
// Only set Data.A to a valid address, the SDK doesn't
|
|
// use the value from the frame.
|
|
surface->Data.A = frame->data[0] + 3;
|
|
break;
|
|
case AV_PIX_FMT_XV30:
|
|
surface->Data.U = frame->data[0];
|
|
break;
|
|
case AV_PIX_FMT_XV36:
|
|
surface->Data.U = frame->data[0];
|
|
surface->Data.Y = frame->data[0] + 2;
|
|
surface->Data.V = frame->data[0] + 4;
|
|
// Only set Data.A to a valid address, the SDK doesn't
|
|
// use the value from the frame.
|
|
surface->Data.A = frame->data[0] + 6;
|
|
break;
|
|
#endif
|
|
default:
|
|
return MFX_ERR_UNSUPPORTED;
|
|
}
|
|
surface->Data.Pitch = frame->linesize[0];
|
|
surface->Data.TimeStamp = frame->pts;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int qsv_internal_session_check_init(AVHWFramesContext *ctx, int upload)
|
|
{
|
|
QSVFramesContext *s = ctx->internal->priv;
|
|
atomic_int *inited = upload ? &s->session_upload_init : &s->session_download_init;
|
|
mfxSession *session = upload ? &s->session_upload : &s->session_download;
|
|
int ret = 0;
|
|
|
|
if (atomic_load(inited))
|
|
return 0;
|
|
|
|
#if HAVE_PTHREADS
|
|
pthread_mutex_lock(&s->session_lock);
|
|
#endif
|
|
|
|
if (!atomic_load(inited)) {
|
|
ret = qsv_init_internal_session(ctx, session, upload);
|
|
atomic_store(inited, 1);
|
|
}
|
|
|
|
#if HAVE_PTHREADS
|
|
pthread_mutex_unlock(&s->session_lock);
|
|
#endif
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int qsv_transfer_data_from(AVHWFramesContext *ctx, AVFrame *dst,
|
|
const AVFrame *src)
|
|
{
|
|
QSVFramesContext *s = ctx->internal->priv;
|
|
mfxFrameSurface1 out = {{ 0 }};
|
|
mfxFrameSurface1 *in = (mfxFrameSurface1*)src->data[3];
|
|
|
|
mfxSyncPoint sync = NULL;
|
|
mfxStatus err;
|
|
int ret = 0;
|
|
/* download to temp frame if the output is not padded as libmfx requires */
|
|
AVFrame *tmp_frame = &s->realigned_download_frame;
|
|
AVFrame *dst_frame;
|
|
int realigned = 0;
|
|
|
|
ret = qsv_internal_session_check_init(ctx, 0);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
/* According to MSDK spec for mfxframeinfo, "Width must be a multiple of 16.
|
|
* Height must be a multiple of 16 for progressive frame sequence and a
|
|
* multiple of 32 otherwise.", so allign all frames to 16 before downloading. */
|
|
if (dst->height & 15 || dst->linesize[0] & 15) {
|
|
realigned = 1;
|
|
if (tmp_frame->format != dst->format ||
|
|
tmp_frame->width != FFALIGN(dst->linesize[0], 16) ||
|
|
tmp_frame->height != FFALIGN(dst->height, 16)) {
|
|
av_frame_unref(tmp_frame);
|
|
|
|
tmp_frame->format = dst->format;
|
|
tmp_frame->width = FFALIGN(dst->linesize[0], 16);
|
|
tmp_frame->height = FFALIGN(dst->height, 16);
|
|
ret = av_frame_get_buffer(tmp_frame, 0);
|
|
if (ret < 0)
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
dst_frame = realigned ? tmp_frame : dst;
|
|
|
|
if (!s->session_download) {
|
|
if (s->child_frames_ref)
|
|
return qsv_transfer_data_child(ctx, dst_frame, src);
|
|
|
|
av_log(ctx, AV_LOG_ERROR, "Surface download not possible\n");
|
|
return AVERROR(ENOSYS);
|
|
}
|
|
|
|
out.Info = in->Info;
|
|
map_frame_to_surface(dst_frame, &out);
|
|
|
|
do {
|
|
err = MFXVideoVPP_RunFrameVPPAsync(s->session_download, in, &out, NULL, &sync);
|
|
if (err == MFX_WRN_DEVICE_BUSY)
|
|
av_usleep(1);
|
|
} while (err == MFX_WRN_DEVICE_BUSY);
|
|
|
|
if (err < 0 || !sync) {
|
|
av_log(ctx, AV_LOG_ERROR, "Error downloading the surface\n");
|
|
return AVERROR_UNKNOWN;
|
|
}
|
|
|
|
do {
|
|
err = MFXVideoCORE_SyncOperation(s->session_download, sync, 1000);
|
|
} while (err == MFX_WRN_IN_EXECUTION);
|
|
if (err < 0) {
|
|
av_log(ctx, AV_LOG_ERROR, "Error synchronizing the operation: %d\n", err);
|
|
return AVERROR_UNKNOWN;
|
|
}
|
|
|
|
if (realigned) {
|
|
tmp_frame->width = dst->width;
|
|
tmp_frame->height = dst->height;
|
|
ret = av_frame_copy(dst, tmp_frame);
|
|
tmp_frame->width = FFALIGN(dst->linesize[0], 16);
|
|
tmp_frame->height = FFALIGN(dst->height, 16);
|
|
if (ret < 0)
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int qsv_transfer_data_to(AVHWFramesContext *ctx, AVFrame *dst,
|
|
const AVFrame *src)
|
|
{
|
|
QSVFramesContext *s = ctx->internal->priv;
|
|
mfxFrameSurface1 in = {{ 0 }};
|
|
mfxFrameSurface1 *out = (mfxFrameSurface1*)dst->data[3];
|
|
mfxFrameInfo tmp_info;
|
|
|
|
mfxSyncPoint sync = NULL;
|
|
mfxStatus err;
|
|
int ret = 0;
|
|
/* make a copy if the input is not padded as libmfx requires */
|
|
AVFrame *tmp_frame = &s->realigned_upload_frame;
|
|
const AVFrame *src_frame;
|
|
int realigned = 0;
|
|
|
|
ret = qsv_internal_session_check_init(ctx, 1);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
/* According to MSDK spec for mfxframeinfo, "Width must be a multiple of 16.
|
|
* Height must be a multiple of 16 for progressive frame sequence and a
|
|
* multiple of 32 otherwise.", so allign all frames to 16 before uploading. */
|
|
if (src->height & 15 || src->linesize[0] & 15) {
|
|
realigned = 1;
|
|
if (tmp_frame->format != src->format ||
|
|
tmp_frame->width != FFALIGN(src->width, 16) ||
|
|
tmp_frame->height != FFALIGN(src->height, 16)) {
|
|
av_frame_unref(tmp_frame);
|
|
|
|
tmp_frame->format = src->format;
|
|
tmp_frame->width = FFALIGN(src->width, 16);
|
|
tmp_frame->height = FFALIGN(src->height, 16);
|
|
ret = av_frame_get_buffer(tmp_frame, 0);
|
|
if (ret < 0)
|
|
return ret;
|
|
}
|
|
ret = av_frame_copy(tmp_frame, src);
|
|
if (ret < 0) {
|
|
av_frame_unref(tmp_frame);
|
|
return ret;
|
|
}
|
|
ret = qsv_fill_border(tmp_frame, src);
|
|
if (ret < 0) {
|
|
av_frame_unref(tmp_frame);
|
|
return ret;
|
|
}
|
|
|
|
tmp_info = out->Info;
|
|
out->Info.CropW = FFMIN(out->Info.Width, tmp_frame->width);
|
|
out->Info.CropH = FFMIN(out->Info.Height, tmp_frame->height);
|
|
}
|
|
|
|
src_frame = realigned ? tmp_frame : src;
|
|
|
|
if (!s->session_upload) {
|
|
if (s->child_frames_ref)
|
|
return qsv_transfer_data_child(ctx, dst, src_frame);
|
|
|
|
av_log(ctx, AV_LOG_ERROR, "Surface upload not possible\n");
|
|
return AVERROR(ENOSYS);
|
|
}
|
|
|
|
in.Info = out->Info;
|
|
map_frame_to_surface(src_frame, &in);
|
|
|
|
do {
|
|
err = MFXVideoVPP_RunFrameVPPAsync(s->session_upload, &in, out, NULL, &sync);
|
|
if (err == MFX_WRN_DEVICE_BUSY)
|
|
av_usleep(1);
|
|
} while (err == MFX_WRN_DEVICE_BUSY);
|
|
|
|
if (err < 0 || !sync) {
|
|
av_log(ctx, AV_LOG_ERROR, "Error uploading the surface\n");
|
|
return AVERROR_UNKNOWN;
|
|
}
|
|
|
|
do {
|
|
err = MFXVideoCORE_SyncOperation(s->session_upload, sync, 1000);
|
|
} while (err == MFX_WRN_IN_EXECUTION);
|
|
if (err < 0) {
|
|
av_log(ctx, AV_LOG_ERROR, "Error synchronizing the operation\n");
|
|
return AVERROR_UNKNOWN;
|
|
}
|
|
|
|
if (realigned) {
|
|
out->Info.CropW = tmp_info.CropW;
|
|
out->Info.CropH = tmp_info.CropH;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int qsv_frames_derive_to(AVHWFramesContext *dst_ctx,
|
|
AVHWFramesContext *src_ctx, int flags)
|
|
{
|
|
QSVFramesContext *s = dst_ctx->internal->priv;
|
|
AVQSVFramesContext *dst_hwctx = dst_ctx->hwctx;
|
|
int i;
|
|
|
|
if (src_ctx->initial_pool_size == 0) {
|
|
av_log(dst_ctx, AV_LOG_ERROR, "Only fixed-size pools can be "
|
|
"mapped to QSV frames.\n");
|
|
return AVERROR(EINVAL);
|
|
}
|
|
|
|
switch (src_ctx->device_ctx->type) {
|
|
#if CONFIG_VAAPI
|
|
case AV_HWDEVICE_TYPE_VAAPI:
|
|
{
|
|
AVVAAPIFramesContext *src_hwctx = src_ctx->hwctx;
|
|
s->handle_pairs_internal = av_calloc(src_ctx->initial_pool_size,
|
|
sizeof(*s->handle_pairs_internal));
|
|
if (!s->handle_pairs_internal)
|
|
return AVERROR(ENOMEM);
|
|
s->surfaces_internal = av_calloc(src_hwctx->nb_surfaces,
|
|
sizeof(*s->surfaces_internal));
|
|
if (!s->surfaces_internal)
|
|
return AVERROR(ENOMEM);
|
|
for (i = 0; i < src_hwctx->nb_surfaces; i++) {
|
|
qsv_init_surface(dst_ctx, &s->surfaces_internal[i]);
|
|
s->handle_pairs_internal[i].first = src_hwctx->surface_ids + i;
|
|
s->handle_pairs_internal[i].second = (mfxMemId)MFX_INFINITE;
|
|
s->surfaces_internal[i].Data.MemId = (mfxMemId)&s->handle_pairs_internal[i];
|
|
}
|
|
dst_hwctx->nb_surfaces = src_hwctx->nb_surfaces;
|
|
dst_hwctx->frame_type = MFX_MEMTYPE_VIDEO_MEMORY_DECODER_TARGET;
|
|
}
|
|
break;
|
|
#endif
|
|
#if CONFIG_D3D11VA
|
|
case AV_HWDEVICE_TYPE_D3D11VA:
|
|
{
|
|
AVD3D11VAFramesContext *src_hwctx = src_ctx->hwctx;
|
|
s->handle_pairs_internal = av_calloc(src_ctx->initial_pool_size,
|
|
sizeof(*s->handle_pairs_internal));
|
|
if (!s->handle_pairs_internal)
|
|
return AVERROR(ENOMEM);
|
|
s->surfaces_internal = av_calloc(src_ctx->initial_pool_size,
|
|
sizeof(*s->surfaces_internal));
|
|
if (!s->surfaces_internal)
|
|
return AVERROR(ENOMEM);
|
|
for (i = 0; i < src_ctx->initial_pool_size; i++) {
|
|
qsv_init_surface(dst_ctx, &s->surfaces_internal[i]);
|
|
s->handle_pairs_internal[i].first = (mfxMemId)src_hwctx->texture_infos[i].texture;
|
|
if (src_hwctx->BindFlags & D3D11_BIND_RENDER_TARGET) {
|
|
s->handle_pairs_internal[i].second = (mfxMemId)MFX_INFINITE;
|
|
} else {
|
|
s->handle_pairs_internal[i].second = (mfxMemId)src_hwctx->texture_infos[i].index;
|
|
}
|
|
s->surfaces_internal[i].Data.MemId = (mfxMemId)&s->handle_pairs_internal[i];
|
|
}
|
|
dst_hwctx->nb_surfaces = src_ctx->initial_pool_size;
|
|
if (src_hwctx->BindFlags & D3D11_BIND_RENDER_TARGET) {
|
|
dst_hwctx->frame_type |= MFX_MEMTYPE_VIDEO_MEMORY_PROCESSOR_TARGET;
|
|
} else {
|
|
dst_hwctx->frame_type |= MFX_MEMTYPE_VIDEO_MEMORY_DECODER_TARGET;
|
|
}
|
|
}
|
|
break;
|
|
#endif
|
|
#if CONFIG_DXVA2
|
|
case AV_HWDEVICE_TYPE_DXVA2:
|
|
{
|
|
AVDXVA2FramesContext *src_hwctx = src_ctx->hwctx;
|
|
s->handle_pairs_internal = av_calloc(src_ctx->initial_pool_size,
|
|
sizeof(*s->handle_pairs_internal));
|
|
if (!s->handle_pairs_internal)
|
|
return AVERROR(ENOMEM);
|
|
s->surfaces_internal = av_calloc(src_hwctx->nb_surfaces,
|
|
sizeof(*s->surfaces_internal));
|
|
if (!s->surfaces_internal)
|
|
return AVERROR(ENOMEM);
|
|
for (i = 0; i < src_hwctx->nb_surfaces; i++) {
|
|
qsv_init_surface(dst_ctx, &s->surfaces_internal[i]);
|
|
s->handle_pairs_internal[i].first = (mfxMemId)src_hwctx->surfaces[i];
|
|
s->handle_pairs_internal[i].second = (mfxMemId)MFX_INFINITE;
|
|
s->surfaces_internal[i].Data.MemId = (mfxMemId)&s->handle_pairs_internal[i];
|
|
}
|
|
dst_hwctx->nb_surfaces = src_hwctx->nb_surfaces;
|
|
if (src_hwctx->surface_type == DXVA2_VideoProcessorRenderTarget)
|
|
dst_hwctx->frame_type = MFX_MEMTYPE_VIDEO_MEMORY_PROCESSOR_TARGET;
|
|
else
|
|
dst_hwctx->frame_type = MFX_MEMTYPE_VIDEO_MEMORY_DECODER_TARGET;
|
|
}
|
|
break;
|
|
#endif
|
|
default:
|
|
return AVERROR(ENOSYS);
|
|
}
|
|
|
|
dst_hwctx->surfaces = s->surfaces_internal;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int qsv_map_to(AVHWFramesContext *dst_ctx,
|
|
AVFrame *dst, const AVFrame *src, int flags)
|
|
{
|
|
AVQSVFramesContext *hwctx = dst_ctx->hwctx;
|
|
int i, err, index = -1;
|
|
|
|
for (i = 0; i < hwctx->nb_surfaces && index < 0; i++) {
|
|
switch(src->format) {
|
|
#if CONFIG_VAAPI
|
|
case AV_PIX_FMT_VAAPI:
|
|
{
|
|
mfxHDLPair *pair = (mfxHDLPair*)hwctx->surfaces[i].Data.MemId;
|
|
if (*(VASurfaceID*)pair->first == (VASurfaceID)src->data[3]) {
|
|
index = i;
|
|
break;
|
|
}
|
|
}
|
|
#endif
|
|
#if CONFIG_D3D11VA
|
|
case AV_PIX_FMT_D3D11:
|
|
{
|
|
mfxHDLPair *pair = (mfxHDLPair*)hwctx->surfaces[i].Data.MemId;
|
|
if (pair->first == src->data[0]
|
|
&& (pair->second == src->data[1]
|
|
|| (pair->second == (mfxMemId)MFX_INFINITE && src->data[1] == (uint8_t *)0))) {
|
|
index = i;
|
|
break;
|
|
}
|
|
}
|
|
#endif
|
|
#if CONFIG_DXVA2
|
|
case AV_PIX_FMT_DXVA2_VLD:
|
|
{
|
|
mfxHDLPair *pair = (mfxHDLPair*)hwctx->surfaces[i].Data.MemId;
|
|
if (pair->first == src->data[3]) {
|
|
index = i;
|
|
break;
|
|
}
|
|
}
|
|
#endif
|
|
}
|
|
}
|
|
if (index < 0) {
|
|
av_log(dst_ctx, AV_LOG_ERROR, "Trying to map from a surface which "
|
|
"is not in the mapped frames context.\n");
|
|
return AVERROR(EINVAL);
|
|
}
|
|
|
|
err = ff_hwframe_map_create(dst->hw_frames_ctx,
|
|
dst, src, NULL, NULL);
|
|
if (err)
|
|
return err;
|
|
|
|
dst->width = src->width;
|
|
dst->height = src->height;
|
|
dst->data[3] = (uint8_t*)&hwctx->surfaces[index];
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int qsv_frames_get_constraints(AVHWDeviceContext *ctx,
|
|
const void *hwconfig,
|
|
AVHWFramesConstraints *constraints)
|
|
{
|
|
int i;
|
|
|
|
constraints->valid_sw_formats = av_malloc_array(FF_ARRAY_ELEMS(supported_pixel_formats) + 1,
|
|
sizeof(*constraints->valid_sw_formats));
|
|
if (!constraints->valid_sw_formats)
|
|
return AVERROR(ENOMEM);
|
|
|
|
for (i = 0; i < FF_ARRAY_ELEMS(supported_pixel_formats); i++)
|
|
constraints->valid_sw_formats[i] = supported_pixel_formats[i].pix_fmt;
|
|
constraints->valid_sw_formats[FF_ARRAY_ELEMS(supported_pixel_formats)] = AV_PIX_FMT_NONE;
|
|
|
|
constraints->valid_hw_formats = av_malloc_array(2, sizeof(*constraints->valid_hw_formats));
|
|
if (!constraints->valid_hw_formats)
|
|
return AVERROR(ENOMEM);
|
|
|
|
constraints->valid_hw_formats[0] = AV_PIX_FMT_QSV;
|
|
constraints->valid_hw_formats[1] = AV_PIX_FMT_NONE;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void qsv_device_free(AVHWDeviceContext *ctx)
|
|
{
|
|
AVQSVDeviceContext *hwctx = ctx->hwctx;
|
|
QSVDevicePriv *priv = ctx->user_opaque;
|
|
|
|
if (hwctx->session)
|
|
MFXClose(hwctx->session);
|
|
|
|
if (hwctx->loader)
|
|
MFXUnload(hwctx->loader);
|
|
av_buffer_unref(&priv->child_device_ctx);
|
|
av_freep(&priv);
|
|
}
|
|
|
|
static mfxIMPL choose_implementation(const char *device, enum AVHWDeviceType child_device_type)
|
|
{
|
|
static const struct {
|
|
const char *name;
|
|
mfxIMPL impl;
|
|
} impl_map[] = {
|
|
{ "auto", MFX_IMPL_AUTO },
|
|
{ "sw", MFX_IMPL_SOFTWARE },
|
|
{ "hw", MFX_IMPL_HARDWARE },
|
|
{ "auto_any", MFX_IMPL_AUTO_ANY },
|
|
{ "hw_any", MFX_IMPL_HARDWARE_ANY },
|
|
{ "hw2", MFX_IMPL_HARDWARE2 },
|
|
{ "hw3", MFX_IMPL_HARDWARE3 },
|
|
{ "hw4", MFX_IMPL_HARDWARE4 },
|
|
};
|
|
|
|
mfxIMPL impl = MFX_IMPL_AUTO_ANY;
|
|
int i;
|
|
|
|
if (device) {
|
|
for (i = 0; i < FF_ARRAY_ELEMS(impl_map); i++)
|
|
if (!strcmp(device, impl_map[i].name)) {
|
|
impl = impl_map[i].impl;
|
|
break;
|
|
}
|
|
if (i == FF_ARRAY_ELEMS(impl_map))
|
|
impl = strtol(device, NULL, 0);
|
|
}
|
|
|
|
if (impl != MFX_IMPL_SOFTWARE) {
|
|
if (child_device_type == AV_HWDEVICE_TYPE_D3D11VA)
|
|
impl |= MFX_IMPL_VIA_D3D11;
|
|
else if (child_device_type == AV_HWDEVICE_TYPE_DXVA2)
|
|
impl |= MFX_IMPL_VIA_D3D9;
|
|
}
|
|
|
|
return impl;
|
|
}
|
|
|
|
static int qsv_device_derive_from_child(AVHWDeviceContext *ctx,
|
|
mfxIMPL implementation,
|
|
AVHWDeviceContext *child_device_ctx,
|
|
int flags)
|
|
{
|
|
AVQSVDeviceContext *hwctx = ctx->hwctx;
|
|
|
|
mfxVersion ver = { { 3, 1 } };
|
|
mfxHDL handle;
|
|
mfxHandleType handle_type;
|
|
mfxStatus err;
|
|
int ret;
|
|
|
|
switch (child_device_ctx->type) {
|
|
#if CONFIG_VAAPI
|
|
case AV_HWDEVICE_TYPE_VAAPI:
|
|
{
|
|
AVVAAPIDeviceContext *child_device_hwctx = child_device_ctx->hwctx;
|
|
handle_type = MFX_HANDLE_VA_DISPLAY;
|
|
handle = (mfxHDL)child_device_hwctx->display;
|
|
}
|
|
break;
|
|
#endif
|
|
#if CONFIG_D3D11VA
|
|
case AV_HWDEVICE_TYPE_D3D11VA:
|
|
{
|
|
AVD3D11VADeviceContext *child_device_hwctx = child_device_ctx->hwctx;
|
|
handle_type = MFX_HANDLE_D3D11_DEVICE;
|
|
handle = (mfxHDL)child_device_hwctx->device;
|
|
}
|
|
break;
|
|
#endif
|
|
#if CONFIG_DXVA2
|
|
case AV_HWDEVICE_TYPE_DXVA2:
|
|
{
|
|
AVDXVA2DeviceContext *child_device_hwctx = child_device_ctx->hwctx;
|
|
handle_type = MFX_HANDLE_D3D9_DEVICE_MANAGER;
|
|
handle = (mfxHDL)child_device_hwctx->devmgr;
|
|
}
|
|
break;
|
|
#endif
|
|
default:
|
|
ret = AVERROR(ENOSYS);
|
|
goto fail;
|
|
}
|
|
|
|
ret = qsv_create_mfx_session(ctx, handle, handle_type, implementation, &ver,
|
|
&hwctx->session, &hwctx->loader);
|
|
if (ret)
|
|
goto fail;
|
|
|
|
err = MFXVideoCORE_SetHandle(hwctx->session, handle_type, handle);
|
|
if (err != MFX_ERR_NONE) {
|
|
av_log(ctx, AV_LOG_ERROR, "Error setting child device handle: "
|
|
"%d\n", err);
|
|
ret = AVERROR_UNKNOWN;
|
|
goto fail;
|
|
}
|
|
|
|
return 0;
|
|
|
|
fail:
|
|
if (hwctx->session)
|
|
MFXClose(hwctx->session);
|
|
|
|
if (hwctx->loader)
|
|
MFXUnload(hwctx->loader);
|
|
|
|
hwctx->session = NULL;
|
|
hwctx->loader = NULL;
|
|
return ret;
|
|
}
|
|
|
|
static int qsv_device_derive(AVHWDeviceContext *ctx,
|
|
AVHWDeviceContext *child_device_ctx,
|
|
AVDictionary *opts, int flags)
|
|
{
|
|
mfxIMPL impl;
|
|
impl = choose_implementation("hw_any", child_device_ctx->type);
|
|
return qsv_device_derive_from_child(ctx, impl,
|
|
child_device_ctx, flags);
|
|
}
|
|
|
|
static int qsv_device_create(AVHWDeviceContext *ctx, const char *device,
|
|
AVDictionary *opts, int flags)
|
|
{
|
|
QSVDevicePriv *priv;
|
|
enum AVHWDeviceType child_device_type;
|
|
AVHWDeviceContext *child_device;
|
|
AVDictionary *child_device_opts;
|
|
AVDictionaryEntry *e;
|
|
|
|
mfxIMPL impl;
|
|
int ret;
|
|
|
|
priv = av_mallocz(sizeof(*priv));
|
|
if (!priv)
|
|
return AVERROR(ENOMEM);
|
|
|
|
ctx->user_opaque = priv;
|
|
ctx->free = qsv_device_free;
|
|
|
|
e = av_dict_get(opts, "child_device_type", NULL, 0);
|
|
if (e) {
|
|
child_device_type = av_hwdevice_find_type_by_name(e->value);
|
|
if (child_device_type == AV_HWDEVICE_TYPE_NONE) {
|
|
av_log(ctx, AV_LOG_ERROR, "Unknown child device type "
|
|
"\"%s\".\n", e->value);
|
|
return AVERROR(EINVAL);
|
|
}
|
|
} else if (CONFIG_VAAPI) {
|
|
child_device_type = AV_HWDEVICE_TYPE_VAAPI;
|
|
#if QSV_ONEVPL
|
|
} else if (CONFIG_D3D11VA) { // Use D3D11 by default if d3d11va is enabled
|
|
av_log(ctx, AV_LOG_VERBOSE,
|
|
"Defaulting child_device_type to AV_HWDEVICE_TYPE_D3D11VA for oneVPL."
|
|
"Please explicitly set child device type via \"-init_hw_device\" "
|
|
"option if needed.\n");
|
|
child_device_type = AV_HWDEVICE_TYPE_D3D11VA;
|
|
} else if (CONFIG_DXVA2) {
|
|
child_device_type = AV_HWDEVICE_TYPE_DXVA2;
|
|
#else
|
|
} else if (CONFIG_DXVA2) {
|
|
av_log(NULL, AV_LOG_WARNING,
|
|
"WARNING: defaulting child_device_type to AV_HWDEVICE_TYPE_DXVA2 for compatibility "
|
|
"with old commandlines. This behaviour will be removed "
|
|
"in the future. Please explicitly set device type via \"-init_hw_device\" option.\n");
|
|
child_device_type = AV_HWDEVICE_TYPE_DXVA2;
|
|
} else if (CONFIG_D3D11VA) {
|
|
child_device_type = AV_HWDEVICE_TYPE_D3D11VA;
|
|
#endif
|
|
} else {
|
|
av_log(ctx, AV_LOG_ERROR, "No supported child device type is enabled\n");
|
|
return AVERROR(ENOSYS);
|
|
}
|
|
|
|
child_device_opts = NULL;
|
|
switch (child_device_type) {
|
|
#if CONFIG_VAAPI
|
|
case AV_HWDEVICE_TYPE_VAAPI:
|
|
{
|
|
// libmfx does not actually implement VAAPI properly, rather it
|
|
// depends on the specific behaviour of a matching iHD driver when
|
|
// used on recent Intel hardware. Set options to the VAAPI device
|
|
// creation so that we should pick a usable setup by default if
|
|
// possible, even when multiple devices and drivers are available.
|
|
av_dict_set(&child_device_opts, "kernel_driver", "i915", 0);
|
|
av_dict_set(&child_device_opts, "driver", "iHD", 0);
|
|
}
|
|
break;
|
|
#endif
|
|
#if CONFIG_D3D11VA
|
|
case AV_HWDEVICE_TYPE_D3D11VA:
|
|
break;
|
|
#endif
|
|
#if CONFIG_DXVA2
|
|
case AV_HWDEVICE_TYPE_DXVA2:
|
|
#if QSV_ONEVPL
|
|
{
|
|
av_log(ctx, AV_LOG_VERBOSE,
|
|
"d3d11va is not available or child device type is set to dxva2 "
|
|
"explicitly for oneVPL.\n");
|
|
}
|
|
#endif
|
|
break;
|
|
#endif
|
|
default:
|
|
{
|
|
av_log(ctx, AV_LOG_ERROR, "No supported child device type is enabled\n");
|
|
return AVERROR(ENOSYS);
|
|
}
|
|
break;
|
|
}
|
|
|
|
e = av_dict_get(opts, "child_device", NULL, 0);
|
|
ret = av_hwdevice_ctx_create(&priv->child_device_ctx, child_device_type,
|
|
e ? e->value : NULL, child_device_opts, 0);
|
|
|
|
av_dict_free(&child_device_opts);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
child_device = (AVHWDeviceContext*)priv->child_device_ctx->data;
|
|
|
|
impl = choose_implementation(device, child_device_type);
|
|
|
|
return qsv_device_derive_from_child(ctx, impl, child_device, 0);
|
|
}
|
|
|
|
const HWContextType ff_hwcontext_type_qsv = {
|
|
.type = AV_HWDEVICE_TYPE_QSV,
|
|
.name = "QSV",
|
|
|
|
.device_hwctx_size = sizeof(AVQSVDeviceContext),
|
|
.device_priv_size = sizeof(QSVDeviceContext),
|
|
.frames_hwctx_size = sizeof(AVQSVFramesContext),
|
|
.frames_priv_size = sizeof(QSVFramesContext),
|
|
|
|
.device_create = qsv_device_create,
|
|
.device_derive = qsv_device_derive,
|
|
.device_init = qsv_device_init,
|
|
.frames_get_constraints = qsv_frames_get_constraints,
|
|
.frames_init = qsv_frames_init,
|
|
.frames_uninit = qsv_frames_uninit,
|
|
.frames_get_buffer = qsv_get_buffer,
|
|
.transfer_get_formats = qsv_transfer_get_formats,
|
|
.transfer_data_to = qsv_transfer_data_to,
|
|
.transfer_data_from = qsv_transfer_data_from,
|
|
.map_to = qsv_map_to,
|
|
.map_from = qsv_map_from,
|
|
.frames_derive_to = qsv_frames_derive_to,
|
|
.frames_derive_from = qsv_frames_derive_from,
|
|
|
|
.pix_fmts = (const enum AVPixelFormat[]){ AV_PIX_FMT_QSV, AV_PIX_FMT_NONE },
|
|
};
|