mirror of https://git.ffmpeg.org/ffmpeg.git
185 lines
6.7 KiB
C
185 lines
6.7 KiB
C
/*
|
|
* VP9 HW decode acceleration through NVDEC
|
|
*
|
|
* Copyright (c) 2016 Timo Rothenpieler
|
|
*
|
|
* This file is part of FFmpeg.
|
|
*
|
|
* FFmpeg is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with FFmpeg; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
#include "libavutil/pixdesc.h"
|
|
|
|
#include "avcodec.h"
|
|
#include "nvdec.h"
|
|
#include "decode.h"
|
|
#include "internal.h"
|
|
#include "vp9shared.h"
|
|
|
|
static int nvdec_vp9_start_frame(AVCodecContext *avctx, const uint8_t *buffer, uint32_t size)
|
|
{
|
|
VP9SharedContext *h = avctx->priv_data;
|
|
const AVPixFmtDescriptor *pixdesc = av_pix_fmt_desc_get(avctx->sw_pix_fmt);
|
|
|
|
NVDECContext *ctx = avctx->internal->hwaccel_priv_data;
|
|
CUVIDPICPARAMS *pp = &ctx->pic_params;
|
|
CUVIDVP9PICPARAMS *ppc = &pp->CodecSpecific.vp9;
|
|
FrameDecodeData *fdd;
|
|
NVDECFrame *cf;
|
|
AVFrame *cur_frame = h->frames[CUR_FRAME].tf.f;
|
|
|
|
int ret, i;
|
|
|
|
ret = ff_nvdec_start_frame(avctx, cur_frame);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
fdd = (FrameDecodeData*)cur_frame->private_ref->data;
|
|
cf = (NVDECFrame*)fdd->hwaccel_priv;
|
|
|
|
*pp = (CUVIDPICPARAMS) {
|
|
.PicWidthInMbs = (cur_frame->width + 15) / 16,
|
|
.FrameHeightInMbs = (cur_frame->height + 15) / 16,
|
|
.CurrPicIdx = cf->idx,
|
|
|
|
.CodecSpecific.vp9 = {
|
|
.width = cur_frame->width,
|
|
.height = cur_frame->height,
|
|
|
|
.LastRefIdx = ff_nvdec_get_ref_idx(h->refs[h->h.refidx[0]].f),
|
|
.GoldenRefIdx = ff_nvdec_get_ref_idx(h->refs[h->h.refidx[1]].f),
|
|
.AltRefIdx = ff_nvdec_get_ref_idx(h->refs[h->h.refidx[2]].f),
|
|
|
|
.profile = h->h.profile,
|
|
.frameContextIdx = h->h.framectxid,
|
|
.frameType = !h->h.keyframe,
|
|
.showFrame = !h->h.invisible,
|
|
.errorResilient = h->h.errorres,
|
|
.frameParallelDecoding = h->h.parallelmode,
|
|
.subSamplingX = pixdesc->log2_chroma_w,
|
|
.subSamplingY = pixdesc->log2_chroma_h,
|
|
.intraOnly = h->h.intraonly,
|
|
.allow_high_precision_mv = h->h.keyframe ? 0 : h->h.highprecisionmvs,
|
|
.refreshEntropyProbs = h->h.refreshctx,
|
|
|
|
.bitDepthMinus8Luma = pixdesc->comp[0].depth - 8,
|
|
.bitDepthMinus8Chroma = pixdesc->comp[1].depth - 8,
|
|
|
|
.loopFilterLevel = h->h.filter.level,
|
|
.loopFilterSharpness = h->h.filter.sharpness,
|
|
.modeRefLfEnabled = h->h.lf_delta.enabled,
|
|
|
|
.log2_tile_columns = h->h.tiling.log2_tile_cols,
|
|
.log2_tile_rows = h->h.tiling.log2_tile_rows,
|
|
|
|
.segmentEnabled = h->h.segmentation.enabled,
|
|
.segmentMapUpdate = h->h.segmentation.update_map,
|
|
.segmentMapTemporalUpdate = h->h.segmentation.temporal,
|
|
.segmentFeatureMode = h->h.segmentation.absolute_vals,
|
|
|
|
.qpYAc = h->h.yac_qi,
|
|
.qpYDc = h->h.ydc_qdelta,
|
|
.qpChDc = h->h.uvdc_qdelta,
|
|
.qpChAc = h->h.uvac_qdelta,
|
|
|
|
.resetFrameContext = h->h.resetctx,
|
|
.mcomp_filter_type = h->h.filtermode ^ (h->h.filtermode <= 1),
|
|
|
|
.frameTagSize = h->h.uncompressed_header_size,
|
|
.offsetToDctParts = h->h.compressed_header_size,
|
|
|
|
.refFrameSignBias[0] = 0,
|
|
}
|
|
};
|
|
|
|
for (i = 0; i < 2; i++)
|
|
ppc->mbModeLfDelta[i] = h->h.lf_delta.mode[i];
|
|
|
|
for (i = 0; i < 4; i++)
|
|
ppc->mbRefLfDelta[i] = h->h.lf_delta.ref[i];
|
|
|
|
for (i = 0; i < 7; i++)
|
|
ppc->mb_segment_tree_probs[i] = h->h.segmentation.prob[i];
|
|
|
|
for (i = 0; i < 3; i++) {
|
|
ppc->activeRefIdx[i] = h->h.refidx[i];
|
|
ppc->segment_pred_probs[i] = h->h.segmentation.pred_prob[i];
|
|
ppc->refFrameSignBias[i + 1] = h->h.signbias[i];
|
|
}
|
|
|
|
for (i = 0; i < 8; i++) {
|
|
ppc->segmentFeatureEnable[i][0] = h->h.segmentation.feat[i].q_enabled;
|
|
ppc->segmentFeatureEnable[i][1] = h->h.segmentation.feat[i].lf_enabled;
|
|
ppc->segmentFeatureEnable[i][2] = h->h.segmentation.feat[i].ref_enabled;
|
|
ppc->segmentFeatureEnable[i][3] = h->h.segmentation.feat[i].skip_enabled;
|
|
|
|
ppc->segmentFeatureData[i][0] = h->h.segmentation.feat[i].q_val;
|
|
ppc->segmentFeatureData[i][1] = h->h.segmentation.feat[i].lf_val;
|
|
ppc->segmentFeatureData[i][2] = h->h.segmentation.feat[i].ref_val;
|
|
ppc->segmentFeatureData[i][3] = 0;
|
|
}
|
|
|
|
switch (avctx->colorspace) {
|
|
default:
|
|
case AVCOL_SPC_UNSPECIFIED:
|
|
ppc->colorSpace = 0;
|
|
break;
|
|
case AVCOL_SPC_BT470BG:
|
|
ppc->colorSpace = 1;
|
|
break;
|
|
case AVCOL_SPC_BT709:
|
|
ppc->colorSpace = 2;
|
|
break;
|
|
case AVCOL_SPC_SMPTE170M:
|
|
ppc->colorSpace = 3;
|
|
break;
|
|
case AVCOL_SPC_SMPTE240M:
|
|
ppc->colorSpace = 4;
|
|
break;
|
|
case AVCOL_SPC_BT2020_NCL:
|
|
ppc->colorSpace = 5;
|
|
break;
|
|
case AVCOL_SPC_RESERVED:
|
|
ppc->colorSpace = 6;
|
|
break;
|
|
case AVCOL_SPC_RGB:
|
|
ppc->colorSpace = 7;
|
|
break;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int nvdec_vp9_frame_params(AVCodecContext *avctx,
|
|
AVBufferRef *hw_frames_ctx)
|
|
{
|
|
// VP9 uses a fixed size pool of 8 possible reference frames
|
|
return ff_nvdec_frame_params(avctx, hw_frames_ctx, 8, 0);
|
|
}
|
|
|
|
const AVHWAccel ff_vp9_nvdec_hwaccel = {
|
|
.name = "vp9_nvdec",
|
|
.type = AVMEDIA_TYPE_VIDEO,
|
|
.id = AV_CODEC_ID_VP9,
|
|
.pix_fmt = AV_PIX_FMT_CUDA,
|
|
.start_frame = nvdec_vp9_start_frame,
|
|
.end_frame = ff_nvdec_simple_end_frame,
|
|
.decode_slice = ff_nvdec_simple_decode_slice,
|
|
.frame_params = nvdec_vp9_frame_params,
|
|
.init = ff_nvdec_decode_init,
|
|
.uninit = ff_nvdec_decode_uninit,
|
|
.priv_data_size = sizeof(NVDECContext),
|
|
};
|