ffmpeg/libavcodec/vdpau_vp9.c
Andreas Rheinhardt 8238bc0b5e avcodec/defs: Add AV_PROFILE_* defines, deprecate FF_PROFILE_* defines
These defines are also used in other contexts than just AVCodecContext
ones, e.g. in libavformat. Furthermore, given that these defines are
public, the AV-prefix is the right one, so deprecate (and not just move)
the FF-macros.

Signed-off-by: Andreas Rheinhardt <andreas.rheinhardt@outlook.com>
2023-09-07 00:39:02 +02:00

240 lines
8.2 KiB
C

/*
* VP9 HW decode acceleration through VDPAU
*
* Copyright (c) 2019 Manoj Gupta Bonda
*
* This file is part of FFmpeg.
*
* FFmpeg is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2.1 of the License, or (at your option) any later version.
*
* FFmpeg is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with FFmpeg; if not, write to the Free Software Foundation,
* Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
#include <vdpau/vdpau.h>
#include "libavutil/pixdesc.h"
#include "avcodec.h"
#include "hwaccel_internal.h"
#include "vp9dec.h"
#include "vdpau.h"
#include "vdpau_internal.h"
static int vdpau_vp9_start_frame(AVCodecContext *avctx,
const uint8_t *buffer, uint32_t size)
{
VP9Context *s = avctx->priv_data;
VP9SharedContext *h = &(s->s);
VP9Frame pic = h->frames[CUR_FRAME];
struct vdpau_picture_context *pic_ctx = pic.hwaccel_picture_private;
int i;
VdpPictureInfoVP9 *info = &pic_ctx->info.vp9;
const AVPixFmtDescriptor *pixdesc = av_pix_fmt_desc_get(avctx->sw_pix_fmt);
if (!pixdesc) {
return AV_PIX_FMT_NONE;
}
info->width = avctx->width;
info->height = avctx->height;
/* fill LvPictureInfoVP9 struct */
info->lastReference = VDP_INVALID_HANDLE;
info->goldenReference = VDP_INVALID_HANDLE;
info->altReference = VDP_INVALID_HANDLE;
if (h->refs[h->h.refidx[0]].f && h->refs[h->h.refidx[0]].f->private_ref) {
info->lastReference = ff_vdpau_get_surface_id(h->refs[h->h.refidx[0]].f);
}
if (h->refs[h->h.refidx[1]].f && h->refs[h->h.refidx[1]].f->private_ref) {
info->goldenReference = ff_vdpau_get_surface_id(h->refs[h->h.refidx[1]].f);
}
if (h->refs[h->h.refidx[2]].f && h->refs[h->h.refidx[2]].f->private_ref) {
info->altReference = ff_vdpau_get_surface_id(h->refs[h->h.refidx[2]].f);
}
info->profile = h->h.profile;
info->frameContextIdx = h->h.framectxid;
info->keyFrame = h->h.keyframe;
info->showFrame = !h->h.invisible;
info->errorResilient = h->h.errorres;
info->frameParallelDecoding = h->h.parallelmode;
info->subSamplingX = pixdesc->log2_chroma_w;
info->subSamplingY = pixdesc->log2_chroma_h;
info->intraOnly = h->h.intraonly;
info->allowHighPrecisionMv = h->h.keyframe ? 0 : h->h.highprecisionmvs;
info->refreshEntropyProbs = h->h.refreshctx;
info->bitDepthMinus8Luma = pixdesc->comp[0].depth - 8;
info->bitDepthMinus8Chroma = pixdesc->comp[1].depth - 8;
info->loopFilterLevel = h->h.filter.level;
info->loopFilterSharpness = h->h.filter.sharpness;
info->modeRefLfEnabled = h->h.lf_delta.enabled;
info->log2TileColumns = h->h.tiling.log2_tile_cols;
info->log2TileRows = h->h.tiling.log2_tile_rows;
info->segmentEnabled = h->h.segmentation.enabled;
info->segmentMapUpdate = h->h.segmentation.update_map;
info->segmentMapTemporalUpdate = h->h.segmentation.temporal;
info->segmentFeatureMode = h->h.segmentation.absolute_vals;
info->qpYAc = h->h.yac_qi;
info->qpYDc = h->h.ydc_qdelta;
info->qpChDc = h->h.uvdc_qdelta;
info->qpChAc = h->h.uvac_qdelta;
info->resetFrameContext = h->h.resetctx;
info->mcompFilterType = h->h.filtermode ^ (h->h.filtermode <= 1);
info->uncompressedHeaderSize = h->h.uncompressed_header_size;
info->compressedHeaderSize = h->h.compressed_header_size;
info->refFrameSignBias[0] = 0;
for (i = 0; i < FF_ARRAY_ELEMS(info->mbModeLfDelta); i++)
info->mbModeLfDelta[i] = h->h.lf_delta.mode[i];
for (i = 0; i < FF_ARRAY_ELEMS(info->mbRefLfDelta); i++)
info->mbRefLfDelta[i] = h->h.lf_delta.ref[i];
for (i = 0; i < FF_ARRAY_ELEMS(info->mbSegmentTreeProbs); i++)
info->mbSegmentTreeProbs[i] = h->h.segmentation.prob[i];
for (i = 0; i < FF_ARRAY_ELEMS(info->activeRefIdx); i++) {
info->activeRefIdx[i] = h->h.refidx[i];
info->segmentPredProbs[i] = h->h.segmentation.pred_prob[i];
info->refFrameSignBias[i + 1] = h->h.signbias[i];
}
for (i = 0; i < FF_ARRAY_ELEMS(info->segmentFeatureEnable); i++) {
info->segmentFeatureEnable[i][0] = h->h.segmentation.feat[i].q_enabled;
info->segmentFeatureEnable[i][1] = h->h.segmentation.feat[i].lf_enabled;
info->segmentFeatureEnable[i][2] = h->h.segmentation.feat[i].ref_enabled;
info->segmentFeatureEnable[i][3] = h->h.segmentation.feat[i].skip_enabled;
info->segmentFeatureData[i][0] = h->h.segmentation.feat[i].q_val;
info->segmentFeatureData[i][1] = h->h.segmentation.feat[i].lf_val;
info->segmentFeatureData[i][2] = h->h.segmentation.feat[i].ref_val;
info->segmentFeatureData[i][3] = 0;
}
switch (avctx->colorspace) {
default:
case AVCOL_SPC_UNSPECIFIED:
info->colorSpace = 0;
break;
case AVCOL_SPC_BT470BG:
info->colorSpace = 1;
break;
case AVCOL_SPC_BT709:
info->colorSpace = 2;
break;
case AVCOL_SPC_SMPTE170M:
info->colorSpace = 3;
break;
case AVCOL_SPC_SMPTE240M:
info->colorSpace = 4;
break;
case AVCOL_SPC_BT2020_NCL:
info->colorSpace = 5;
break;
case AVCOL_SPC_RESERVED:
info->colorSpace = 6;
break;
case AVCOL_SPC_RGB:
info->colorSpace = 7;
break;
}
return ff_vdpau_common_start_frame(pic_ctx, buffer, size);
}
static const uint8_t start_code_prefix[3] = { 0x00, 0x00, 0x01 };
static int vdpau_vp9_decode_slice(AVCodecContext *avctx,
const uint8_t *buffer, uint32_t size)
{
VP9SharedContext *h = avctx->priv_data;
VP9Frame pic = h->frames[CUR_FRAME];
struct vdpau_picture_context *pic_ctx = pic.hwaccel_picture_private;
int val;
val = ff_vdpau_add_buffer(pic_ctx, start_code_prefix, 3);
if (val)
return val;
val = ff_vdpau_add_buffer(pic_ctx, buffer, size);
if (val)
return val;
return 0;
}
static int vdpau_vp9_end_frame(AVCodecContext *avctx)
{
VP9SharedContext *h = avctx->priv_data;
VP9Frame pic = h->frames[CUR_FRAME];
struct vdpau_picture_context *pic_ctx = pic.hwaccel_picture_private;
int val;
val = ff_vdpau_common_end_frame(avctx, pic.tf.f, pic_ctx);
if (val < 0)
return val;
return 0;
}
static int vdpau_vp9_init(AVCodecContext *avctx)
{
VdpDecoderProfile profile;
uint32_t level = avctx->level;
switch (avctx->profile) {
case AV_PROFILE_VP9_0:
profile = VDP_DECODER_PROFILE_VP9_PROFILE_0;
break;
case AV_PROFILE_VP9_1:
profile = VDP_DECODER_PROFILE_VP9_PROFILE_1;
break;
case AV_PROFILE_VP9_2:
profile = VDP_DECODER_PROFILE_VP9_PROFILE_2;
break;
case AV_PROFILE_VP9_3:
profile = VDP_DECODER_PROFILE_VP9_PROFILE_3;
break;
default:
return AVERROR(ENOTSUP);
}
return ff_vdpau_common_init(avctx, profile, level);
}
const FFHWAccel ff_vp9_vdpau_hwaccel = {
.p.name = "vp9_vdpau",
.p.type = AVMEDIA_TYPE_VIDEO,
.p.id = AV_CODEC_ID_VP9,
.p.pix_fmt = AV_PIX_FMT_VDPAU,
.start_frame = vdpau_vp9_start_frame,
.end_frame = vdpau_vp9_end_frame,
.decode_slice = vdpau_vp9_decode_slice,
.frame_priv_data_size = sizeof(struct vdpau_picture_context),
.init = vdpau_vp9_init,
.uninit = ff_vdpau_common_uninit,
.frame_params = ff_vdpau_common_frame_params,
.priv_data_size = sizeof(VDPAUContext),
.caps_internal = HWACCEL_CAP_ASYNC_SAFE,
};