1
0
mirror of https://github.com/mpv-player/mpv synced 2024-12-26 00:42:57 +00:00
mpv/video/vaapi.c

556 lines
16 KiB
C
Raw Normal View History

/*
* This file is part of mpv.
*
* mpv is free software; you can redistribute it and/or modify
* it under the terms of the GNU General Public License as published by
* the Free Software Foundation; either version 2 of the License, or
* (at your option) any later version.
*
* mpv is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*
* You should have received a copy of the GNU General Public License along
* with mpv. If not, see <http://www.gnu.org/licenses/>.
*/
#include <assert.h>
#include "vaapi.h"
#include "common/common.h"
#include "common/msg.h"
#include "osdep/threads.h"
#include "mp_image.h"
#include "img_format.h"
#include "mp_image_pool.h"
bool check_va_status(struct mp_log *log, VAStatus status, const char *msg)
{
if (status != VA_STATUS_SUCCESS) {
mp_err(log, "%s: %s\n", msg, vaErrorStr(status));
return false;
}
return true;
}
int va_get_colorspace_flag(enum mp_csp csp)
{
switch (csp) {
case MP_CSP_BT_601: return VA_SRC_BT601;
case MP_CSP_BT_709: return VA_SRC_BT709;
case MP_CSP_SMPTE_240M: return VA_SRC_SMPTE_240;
}
return 0;
}
struct fmtentry {
uint32_t va;
enum mp_imgfmt mp;
};
static const struct fmtentry va_to_imgfmt[] = {
{VA_FOURCC_NV12, IMGFMT_NV12},
{VA_FOURCC_YV12, IMGFMT_420P},
{VA_FOURCC_IYUV, IMGFMT_420P},
{VA_FOURCC_UYVY, IMGFMT_UYVY},
{VA_FOURCC_YUY2, IMGFMT_YUYV},
// Note: not sure about endian issues (the mp formats are byte-addressed)
{VA_FOURCC_RGBA, IMGFMT_RGBA},
{VA_FOURCC_RGBX, IMGFMT_RGBA},
{VA_FOURCC_BGRA, IMGFMT_BGRA},
{VA_FOURCC_BGRX, IMGFMT_BGRA},
{0 , IMGFMT_NONE}
};
enum mp_imgfmt va_fourcc_to_imgfmt(uint32_t fourcc)
{
for (const struct fmtentry *entry = va_to_imgfmt; entry->va; ++entry) {
if (entry->va == fourcc)
return entry->mp;
}
return IMGFMT_NONE;
}
uint32_t va_fourcc_from_imgfmt(int imgfmt)
{
for (const struct fmtentry *entry = va_to_imgfmt; entry->va; ++entry) {
if (entry->mp == imgfmt)
return entry->va;
}
return 0;
}
static struct mp_image *ctx_download_image(struct mp_hwdec_ctx *ctx,
struct mp_image *mpi,
struct mp_image_pool *swpool)
{
return va_surface_download(mpi, swpool);
}
struct va_image_formats {
VAImageFormat *entries;
int num;
};
static void va_get_formats(struct mp_vaapi_ctx *ctx)
{
struct va_image_formats *formats = talloc_ptrtype(ctx, formats);
formats->num = vaMaxNumImageFormats(ctx->display);
formats->entries = talloc_array(formats, VAImageFormat, formats->num);
VAStatus status = vaQueryImageFormats(ctx->display, formats->entries,
&formats->num);
if (!CHECK_VA_STATUS(ctx, "vaQueryImageFormats()"))
return;
MP_VERBOSE(ctx, "%d image formats available:\n", formats->num);
for (int i = 0; i < formats->num; i++)
2016-01-11 19:30:36 +00:00
MP_VERBOSE(ctx, " %s\n", mp_tag_str(formats->entries[i].fourcc));
ctx->image_formats = formats;
}
struct mp_vaapi_ctx *va_initialize(VADisplay *display, struct mp_log *plog,
bool probing)
{
struct mp_vaapi_ctx *res = NULL;
struct mp_log *log = mp_log_new(NULL, plog, "/vaapi");
int major_version, minor_version;
int status = vaInitialize(display, &major_version, &minor_version);
if (status != VA_STATUS_SUCCESS && probing)
goto error;
if (!check_va_status(log, status, "vaInitialize()"))
goto error;
mp_verbose(log, "VA API version %d.%d\n", major_version, minor_version);
res = talloc_ptrtype(NULL, res);
*res = (struct mp_vaapi_ctx) {
.log = talloc_steal(res, log),
.display = display,
.hwctx = {
.type = HWDEC_VAAPI,
.ctx = res,
.download_image = ctx_download_image,
},
};
mpthread_mutex_init_recursive(&res->lock);
va_get_formats(res);
if (!res->image_formats)
goto error;
return res;
error:
if (res && res->display)
vaTerminate(res->display);
talloc_free(log);
talloc_free(res);
return NULL;
}
// Undo va_initialize, and close the VADisplay.
void va_destroy(struct mp_vaapi_ctx *ctx)
{
if (ctx) {
if (ctx->display)
vaTerminate(ctx->display);
pthread_mutex_destroy(&ctx->lock);
talloc_free(ctx);
}
}
2015-01-21 21:12:30 +00:00
VAImageFormat *va_image_format_from_imgfmt(struct mp_vaapi_ctx *ctx, int imgfmt)
{
2015-01-21 21:12:30 +00:00
struct va_image_formats *formats = ctx->image_formats;
const int fourcc = va_fourcc_from_imgfmt(imgfmt);
if (!formats || !formats->num || !fourcc)
return NULL;
for (int i = 0; i < formats->num; i++) {
if (formats->entries[i].fourcc == fourcc)
return &formats->entries[i];
}
return NULL;
}
struct va_surface {
struct mp_vaapi_ctx *ctx;
VADisplay display;
VASurfaceID id;
int rt_format;
// The actually allocated surface size (needed for cropping).
// mp_images can have a smaller size than this, which means they are
// cropped down to a smaller size by removing right/bottom pixels.
int w, h;
VAImage image; // used for software decoding case
bool is_derived; // is image derived by vaDeriveImage()?
};
VASurfaceID va_surface_id(struct mp_image *mpi)
{
return mpi && mpi->imgfmt == IMGFMT_VAAPI ?
(VASurfaceID)(uintptr_t)mpi->planes[3] : VA_INVALID_ID;
}
static struct va_surface *va_surface_in_mp_image(struct mp_image *mpi)
{
return mpi && mpi->imgfmt == IMGFMT_VAAPI ?
(struct va_surface*)mpi->planes[0] : NULL;
}
int va_surface_rt_format(struct mp_image *mpi)
{
struct va_surface *surface = va_surface_in_mp_image(mpi);
return surface ? surface->rt_format : 0;
}
// Return the real size of the underlying surface. (HW decoding might allocate
// padded surfaces for example.)
void va_surface_get_uncropped_size(struct mp_image *mpi, int *out_w, int *out_h)
{
struct va_surface *s = va_surface_in_mp_image(mpi);
*out_w = s ? s->w : 0;
*out_h = s ? s->h : 0;
}
static void release_va_surface(void *arg)
{
struct va_surface *surface = arg;
va_lock(surface->ctx);
if (surface->id != VA_INVALID_ID) {
if (surface->image.image_id != VA_INVALID_ID)
vaDestroyImage(surface->display, surface->image.image_id);
vaDestroySurfaces(surface->display, &surface->id, 1);
}
va_unlock(surface->ctx);
talloc_free(surface);
}
static struct mp_image *alloc_surface(struct mp_vaapi_ctx *ctx, int rt_format,
int w, int h)
{
VASurfaceID id = VA_INVALID_ID;
VAStatus status;
va_lock(ctx);
status = vaCreateSurfaces(ctx->display, rt_format, w, h, &id, 1, NULL, 0);
va_unlock(ctx);
if (!CHECK_VA_STATUS(ctx, "vaCreateSurfaces()"))
return NULL;
struct va_surface *surface = talloc_ptrtype(NULL, surface);
if (!surface)
return NULL;
*surface = (struct va_surface){
.ctx = ctx,
.id = id,
.rt_format = rt_format,
.w = w,
.h = h,
.display = ctx->display,
.image = { .image_id = VA_INVALID_ID, .buf = VA_INVALID_ID },
};
struct mp_image img = {0};
mp_image_setfmt(&img, IMGFMT_VAAPI);
mp_image_set_size(&img, w, h);
img.planes[0] = (uint8_t*)surface;
img.planes[3] = (uint8_t*)(uintptr_t)surface->id;
return mp_image_new_custom_ref(&img, surface, release_va_surface);
}
static void va_surface_image_destroy(struct va_surface *surface)
{
if (!surface || surface->image.image_id == VA_INVALID_ID)
return;
vaDestroyImage(surface->display, surface->image.image_id);
surface->image.image_id = VA_INVALID_ID;
surface->is_derived = false;
}
static int va_surface_image_alloc(struct mp_image *img, VAImageFormat *format)
{
struct va_surface *p = va_surface_in_mp_image(img);
if (!format || !p)
return -1;
VADisplay *display = p->display;
if (p->image.image_id != VA_INVALID_ID &&
p->image.format.fourcc == format->fourcc)
return 0;
int r = 0;
va_lock(p->ctx);
va_surface_image_destroy(p);
VAStatus status = vaDeriveImage(display, p->id, &p->image);
if (status == VA_STATUS_SUCCESS) {
/* vaDeriveImage() is supported, check format */
if (p->image.format.fourcc == format->fourcc &&
p->image.width == p->w && p->image.height == p->h)
{
p->is_derived = true;
MP_TRACE(p->ctx, "Using vaDeriveImage()\n");
} else {
vaDestroyImage(p->display, p->image.image_id);
status = VA_STATUS_ERROR_OPERATION_FAILED;
}
}
if (status != VA_STATUS_SUCCESS) {
p->image.image_id = VA_INVALID_ID;
status = vaCreateImage(p->display, format, p->w, p->h, &p->image);
if (!CHECK_VA_STATUS(p->ctx, "vaCreateImage()")) {
p->image.image_id = VA_INVALID_ID;
r = -1;
}
}
va_unlock(p->ctx);
return r;
}
// img must be a VAAPI surface; make sure its internal VAImage is allocated
// to a format corresponding to imgfmt (or return an error).
int va_surface_alloc_imgfmt(struct mp_image *img, int imgfmt)
{
struct va_surface *p = va_surface_in_mp_image(img);
if (!p)
return -1;
// Multiple FourCCs can refer to the same imgfmt, so check by doing the
// surjective conversion first.
if (p->image.image_id != VA_INVALID_ID &&
va_fourcc_to_imgfmt(p->image.format.fourcc) == imgfmt)
return 0;
2015-01-21 21:12:30 +00:00
VAImageFormat *format = va_image_format_from_imgfmt(p->ctx, imgfmt);
if (!format)
return -1;
if (va_surface_image_alloc(img, format) < 0)
return -1;
return 0;
}
bool va_image_map(struct mp_vaapi_ctx *ctx, VAImage *image, struct mp_image *mpi)
{
int imgfmt = va_fourcc_to_imgfmt(image->format.fourcc);
if (imgfmt == IMGFMT_NONE)
return false;
void *data = NULL;
va_lock(ctx);
const VAStatus status = vaMapBuffer(ctx->display, image->buf, &data);
va_unlock(ctx);
if (!CHECK_VA_STATUS(ctx, "vaMapBuffer()"))
return false;
*mpi = (struct mp_image) {0};
mp_image_setfmt(mpi, imgfmt);
mp_image_set_size(mpi, image->width, image->height);
for (int p = 0; p < image->num_planes; p++) {
mpi->stride[p] = image->pitches[p];
mpi->planes[p] = (uint8_t *)data + image->offsets[p];
}
if (image->format.fourcc == VA_FOURCC_YV12) {
MPSWAP(int, mpi->stride[1], mpi->stride[2]);
MPSWAP(uint8_t *, mpi->planes[1], mpi->planes[2]);
}
return true;
}
bool va_image_unmap(struct mp_vaapi_ctx *ctx, VAImage *image)
{
va_lock(ctx);
const VAStatus status = vaUnmapBuffer(ctx->display, image->buf);
va_unlock(ctx);
return CHECK_VA_STATUS(ctx, "vaUnmapBuffer()");
}
// va_dst: copy destination, must be IMGFMT_VAAPI
// sw_src: copy source, must be a software pixel format
int va_surface_upload(struct mp_image *va_dst, struct mp_image *sw_src)
{
struct va_surface *p = va_surface_in_mp_image(va_dst);
if (!p)
return -1;
if (va_surface_alloc_imgfmt(va_dst, sw_src->imgfmt) < 0)
return -1;
struct mp_image img;
if (!va_image_map(p->ctx, &p->image, &img))
return -1;
assert(sw_src->w <= img.w && sw_src->h <= img.h);
mp_image_set_size(&img, sw_src->w, sw_src->h); // copy only visible part
mp_image_copy(&img, sw_src);
va_image_unmap(p->ctx, &p->image);
if (!p->is_derived) {
va_lock(p->ctx);
VAStatus status = vaPutImage(p->display, p->id,
p->image.image_id,
0, 0, sw_src->w, sw_src->h,
0, 0, sw_src->w, sw_src->h);
va_unlock(p->ctx);
if (!CHECK_VA_STATUS(p->ctx, "vaPutImage()"))
return -1;
}
if (p->is_derived)
va_surface_image_destroy(p);
return 0;
}
static struct mp_image *try_download(struct mp_image *src,
struct mp_image_pool *pool)
{
VAStatus status;
struct va_surface *p = va_surface_in_mp_image(src);
if (!p)
return NULL;
VAImage *image = &p->image;
if (image->image_id == VA_INVALID_ID ||
!va_fourcc_to_imgfmt(image->format.fourcc))
return NULL;
if (!p->is_derived) {
va_lock(p->ctx);
status = vaGetImage(p->display, p->id, 0, 0,
p->w, p->h, image->image_id);
va_unlock(p->ctx);
if (status != VA_STATUS_SUCCESS)
return NULL;
}
struct mp_image *dst = NULL;
struct mp_image tmp;
if (va_image_map(p->ctx, image, &tmp)) {
assert(src->w <= tmp.w && src->h <= tmp.h);
mp_image_set_size(&tmp, src->w, src->h); // copy only visible part
dst = mp_image_pool_get(pool, tmp.imgfmt, tmp.w, tmp.h);
if (dst) {
va_lock(p->ctx);
mp_check_gpu_memcpy(p->ctx->log, &p->ctx->gpu_memcpy_message);
va_unlock(p->ctx);
mp_image_copy_gpu(dst, &tmp);
mp_image_copy_attributes(dst, src);
}
va_image_unmap(p->ctx, image);
}
if (p->is_derived)
va_surface_image_destroy(p);
return dst;
}
// Return a software copy of the IMGFMT_VAAPI src image.
// pool is optional (used for allocating returned images).
struct mp_image *va_surface_download(struct mp_image *src,
struct mp_image_pool *pool)
{
struct va_surface *p = va_surface_in_mp_image(src);
if (!p)
return NULL;
struct mp_vaapi_ctx *ctx = p->ctx;
va_lock(ctx);
VAStatus status = vaSyncSurface(p->display, p->id);
va_unlock(ctx);
if (!CHECK_VA_STATUS(ctx, "vaSyncSurface()"))
return NULL;
struct mp_image *mpi = try_download(src, pool);
if (mpi)
return mpi;
// We have no clue which format will work, so try them all.
// Make sure to start with the most preferred format (nv12), to avoid
// slower code paths.
for (int n = 0; va_to_imgfmt[n].mp; n++) {
VAImageFormat *format =
va_image_format_from_imgfmt(ctx, va_to_imgfmt[n].mp);
if (format) {
if (va_surface_image_alloc(src, format) < 0)
continue;
mpi = try_download(src, pool);
if (mpi)
return mpi;
}
}
MP_ERR(ctx, "failed to get surface data.\n");
return NULL;
}
vaapi: determine surface format in decoder, not in renderer Until now, we have made the assumption that a driver will use only 1 hardware surface format. the format is dictated by the driver (you don't create surfaces with a specific format - you just pass a rt_format and get a surface that will be in a specific driver-chosen format). In particular, the renderer created a dummy surface to probe the format, and hoped the decoder would produce the same format. Due to a driver bug this required a workaround to actually get the same format as the driver did. Change this so that the format is determined in the decoder. The format is then passed down as hw_subfmt, which allows the renderer to configure itself with the correct format. If the hardware surface changes its format midstream, the renderer can be reconfigured using the normal mechanisms. This calls va_surface_init_subformat() each time after the decoder returns a surface. Since libavcodec/AVFrame has no concept of sub- formats, this is unavoidable. It creates and destroys a derived VAImage, but this shouldn't have any bad performance effects (at least I didn't notice any measurable effects). Note that vaDeriveImage() failures are silently ignored as some drivers (the vdpau wrapper) support neither vaDeriveImage, nor EGL interop. In addition, we still probe whether we can map an image in the EGL interop code. This is important as it's the only way to determine whether EGL interop is supported at all. With respect to the driver bug mentioned above, it doesn't matter which format the test surface has. In vf_vavpp, also remove the rt_format guessing business. I think the existing logic was a bit meaningless anyway. It's not even a given that vavpp produces the same rt_format for output.
2016-04-11 18:46:05 +00:00
// Set the hw_subfmt from the surface's real format. Because of this bug:
// https://bugs.freedesktop.org/show_bug.cgi?id=79848
// it should be assumed that the real format is only known after an arbitrary
// vaCreateContext() call has been made, or even better, after the surface
// has been rendered to.
// If the hw_subfmt is already set, this is a NOP.
void va_surface_init_subformat(struct mp_image *mpi)
{
VAStatus status;
if (mpi->params.hw_subfmt)
return;
struct va_surface *p = va_surface_in_mp_image(mpi);
if (!p)
return;
VAImage va_image = { .image_id = VA_INVALID_ID };
va_lock(p->ctx);
status = vaDeriveImage(p->display, va_surface_id(mpi), &va_image);
if (status != VA_STATUS_SUCCESS)
goto err;
mpi->params.hw_subfmt = va_fourcc_to_imgfmt(va_image.format.fourcc);
vaapi: determine surface format in decoder, not in renderer Until now, we have made the assumption that a driver will use only 1 hardware surface format. the format is dictated by the driver (you don't create surfaces with a specific format - you just pass a rt_format and get a surface that will be in a specific driver-chosen format). In particular, the renderer created a dummy surface to probe the format, and hoped the decoder would produce the same format. Due to a driver bug this required a workaround to actually get the same format as the driver did. Change this so that the format is determined in the decoder. The format is then passed down as hw_subfmt, which allows the renderer to configure itself with the correct format. If the hardware surface changes its format midstream, the renderer can be reconfigured using the normal mechanisms. This calls va_surface_init_subformat() each time after the decoder returns a surface. Since libavcodec/AVFrame has no concept of sub- formats, this is unavoidable. It creates and destroys a derived VAImage, but this shouldn't have any bad performance effects (at least I didn't notice any measurable effects). Note that vaDeriveImage() failures are silently ignored as some drivers (the vdpau wrapper) support neither vaDeriveImage, nor EGL interop. In addition, we still probe whether we can map an image in the EGL interop code. This is important as it's the only way to determine whether EGL interop is supported at all. With respect to the driver bug mentioned above, it doesn't matter which format the test surface has. In vf_vavpp, also remove the rt_format guessing business. I think the existing logic was a bit meaningless anyway. It's not even a given that vavpp produces the same rt_format for output.
2016-04-11 18:46:05 +00:00
status = vaDestroyImage(p->display, va_image.image_id);
CHECK_VA_STATUS(p->ctx, "vaDestroyImage()");
err:
va_unlock(p->ctx);
}
struct pool_alloc_ctx {
struct mp_vaapi_ctx *vaapi;
int rt_format;
};
static struct mp_image *alloc_pool(void *pctx, int fmt, int w, int h)
{
struct pool_alloc_ctx *alloc_ctx = pctx;
if (fmt != IMGFMT_VAAPI)
return NULL;
return alloc_surface(alloc_ctx->vaapi, alloc_ctx->rt_format, w, h);
}
// The allocator of the given image pool to allocate VAAPI surfaces, using
// the given rt_format.
void va_pool_set_allocator(struct mp_image_pool *pool, struct mp_vaapi_ctx *ctx,
int rt_format)
{
struct pool_alloc_ctx *alloc_ctx = talloc_ptrtype(pool, alloc_ctx);
*alloc_ctx = (struct pool_alloc_ctx){
.vaapi = ctx,
.rt_format = rt_format,
};
mp_image_pool_set_allocator(pool, alloc_pool, alloc_ctx);
mp_image_pool_set_lru(pool);
}
bool va_guess_if_emulated(struct mp_vaapi_ctx *ctx)
{
va_lock(ctx);
const char *s = vaQueryVendorString(ctx->display);
va_unlock(ctx);
return s && strstr(s, "VDPAU backend");
}