ffmpeg/libavcodec/vaapi_vc1.c
Andreas Rheinhardt fe6037fd04 avcodec/mpegpicture: Split MPVPicture into WorkPicture and ordinary Pic
There are two types of MPVPictures: Three (cur_pic, last_pic, next_pic)
that are directly part of MpegEncContext and an array of MPVPictures
that are separately allocated and are mostly accessed via pointers
(cur|last|next)_pic_ptr; they are also used to store AVFrames in the
encoder (necessary due to B-frames). As the name implies, each of the
former is directly associated with one of the _ptr pointers:
They actually share the same underlying buffers, but the ones
that are part of the context can have their data pointers offset
and their linesize doubled for field pictures.

Up until now, each of these had their own references; in particular,
there was an underlying av_frame_ref() to sync cur_pic and cur_pic_ptr
etc. This is wasteful.

This commit changes this relationship: cur_pic, last_pic and next_pic
now become MPVWorkPictures; this structure does not have an AVFrame
at all any more, but only the cached values of data and linesize.
It also contains a pointer to the corresponding MPVPicture, establishing
a more natural relationsship between the two.
This already means that creating the context-pictures from the pointers
can no longer fail.

What has not been changed is the fact that the MPVPicture* pointers
are not ownership pointers and that the MPVPictures are part of an
array of MPVPictures that is owned by a single AVCodecContext.
Doing so will be done in a latter commit.

Signed-off-by: Andreas Rheinhardt <andreas.rheinhardt@outlook.com>
2024-06-12 11:38:13 +02:00

538 lines
20 KiB
C

/*
* VC-1 HW decode acceleration through VA API
*
* Copyright (C) 2008-2009 Splitted-Desktop Systems
*
* This file is part of FFmpeg.
*
* FFmpeg is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2.1 of the License, or (at your option) any later version.
*
* FFmpeg is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with FFmpeg; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
#include "config_components.h"
#include "libavutil/mem.h"
#include "hwaccel_internal.h"
#include "mpegvideodec.h"
#include "vaapi_decode.h"
#include "vc1.h"
/** Translate FFmpeg MV modes to VA API */
static int get_VAMvModeVC1(enum MVModes mv_mode)
{
switch (mv_mode) {
case MV_PMODE_1MV_HPEL_BILIN: return VAMvMode1MvHalfPelBilinear;
case MV_PMODE_1MV: return VAMvMode1Mv;
case MV_PMODE_1MV_HPEL: return VAMvMode1MvHalfPel;
case MV_PMODE_MIXED_MV: return VAMvModeMixedMv;
case MV_PMODE_INTENSITY_COMP: return VAMvModeIntensityCompensation;
}
return 0;
}
/** Check whether the MVTYPEMB bitplane is present */
static inline int vc1_has_MVTYPEMB_bitplane(const VC1Context *v)
{
if (v->mv_type_is_raw)
return 0;
return v->fcm == PROGRESSIVE &&
(v->s.pict_type == AV_PICTURE_TYPE_P && !v->p_frame_skipped) &&
(v->mv_mode == MV_PMODE_MIXED_MV ||
(v->mv_mode == MV_PMODE_INTENSITY_COMP &&
v->mv_mode2 == MV_PMODE_MIXED_MV));
}
/** Check whether the SKIPMB bitplane is present */
static inline int vc1_has_SKIPMB_bitplane(const VC1Context *v)
{
if (v->skip_is_raw)
return 0;
return (v->fcm == PROGRESSIVE || v->fcm == ILACE_FRAME) &&
((v->s.pict_type == AV_PICTURE_TYPE_P && !v->p_frame_skipped) ||
(v->s.pict_type == AV_PICTURE_TYPE_B && !v->bi_type));
}
/** Check whether the DIRECTMB bitplane is present */
static inline int vc1_has_DIRECTMB_bitplane(const VC1Context *v)
{
if (v->dmb_is_raw)
return 0;
return (v->fcm == PROGRESSIVE || v->fcm == ILACE_FRAME) &&
(v->s.pict_type == AV_PICTURE_TYPE_B && !v->bi_type);
}
/** Check whether the ACPRED bitplane is present */
static inline int vc1_has_ACPRED_bitplane(const VC1Context *v)
{
if (v->acpred_is_raw)
return 0;
return v->profile == PROFILE_ADVANCED &&
(v->s.pict_type == AV_PICTURE_TYPE_I ||
(v->s.pict_type == AV_PICTURE_TYPE_B && v->bi_type));
}
/** Check whether the OVERFLAGS bitplane is present */
static inline int vc1_has_OVERFLAGS_bitplane(const VC1Context *v)
{
if (v->overflg_is_raw)
return 0;
return v->profile == PROFILE_ADVANCED &&
(v->s.pict_type == AV_PICTURE_TYPE_I ||
(v->s.pict_type == AV_PICTURE_TYPE_B && v->bi_type)) &&
(v->overlap && v->pq <= 8) &&
v->condover == CONDOVER_SELECT;
}
/** Check whether the FIELDTX bitplane is present */
static inline int vc1_has_FIELDTX_bitplane(const VC1Context *v)
{
if (v->fieldtx_is_raw)
return 0;
return v->fcm == ILACE_FRAME &&
(v->s.pict_type == AV_PICTURE_TYPE_I ||
(v->s.pict_type == AV_PICTURE_TYPE_B && v->bi_type));
}
/** Check whether the FORWARDMB bitplane is present */
static inline int vc1_has_FORWARDMB_bitplane(const VC1Context *v)
{
if (v->fmb_is_raw)
return 0;
return v->fcm == ILACE_FIELD &&
(v->s.pict_type == AV_PICTURE_TYPE_B && !v->bi_type);
}
/** Reconstruct bitstream PTYPE (7.1.1.4, index into Table-35) */
static int vc1_get_PTYPE(const VC1Context *v)
{
const MpegEncContext *s = &v->s;
switch (s->pict_type) {
case AV_PICTURE_TYPE_I: return 0;
case AV_PICTURE_TYPE_P: return v->p_frame_skipped ? 4 : 1;
case AV_PICTURE_TYPE_B: return v->bi_type ? 3 : 2;
}
return 0;
}
/** Reconstruct bitstream FPTYPE (9.1.1.42, index into Table-105) */
static int vc1_get_FPTYPE(const VC1Context *v)
{
const MpegEncContext *s = &v->s;
switch (s->pict_type) {
case AV_PICTURE_TYPE_I: return 0;
case AV_PICTURE_TYPE_P: return 3;
case AV_PICTURE_TYPE_B: return v->bi_type ? 7 : 4;
}
return 0;
}
/** Reconstruct bitstream MVMODE (7.1.1.32) */
static inline VAMvModeVC1 vc1_get_MVMODE(const VC1Context *v)
{
if ((v->fcm == PROGRESSIVE || v->fcm == ILACE_FIELD) &&
((v->s.pict_type == AV_PICTURE_TYPE_P && !v->p_frame_skipped) ||
(v->s.pict_type == AV_PICTURE_TYPE_B && !v->bi_type)))
return get_VAMvModeVC1(v->mv_mode);
return 0;
}
/** Reconstruct bitstream MVMODE2 (7.1.1.33) */
static inline VAMvModeVC1 vc1_get_MVMODE2(const VC1Context *v)
{
if ((v->fcm == PROGRESSIVE || v->fcm == ILACE_FIELD) &&
(v->s.pict_type == AV_PICTURE_TYPE_P && !v->p_frame_skipped) &&
v->mv_mode == MV_PMODE_INTENSITY_COMP)
return get_VAMvModeVC1(v->mv_mode2);
return 0;
}
av_unused static inline int vc1_get_INTCOMPFIELD(const VC1Context *v)
{
if ((v->s.pict_type == AV_PICTURE_TYPE_P && !v->p_frame_skipped) &&
v->fcm == ILACE_FIELD &&
v->mv_mode == MV_PMODE_INTENSITY_COMP)
switch (v->intcompfield) {
case 1: return 1;
case 2: return 2;
case 3: return 0;
}
return 0;
}
static inline int vc1_get_LUMSCALE(const VC1Context *v)
{
if (v->s.pict_type == AV_PICTURE_TYPE_P && !v->p_frame_skipped) {
if ((v->fcm == PROGRESSIVE && v->mv_mode == MV_PMODE_INTENSITY_COMP) ||
(v->fcm == ILACE_FRAME && v->intcomp))
return v->lumscale;
else if (v->fcm == ILACE_FIELD && v->mv_mode == MV_PMODE_INTENSITY_COMP)
switch (v->intcompfield) {
case 1: return v->lumscale;
case 2: return v->lumscale2;
case 3: return v->lumscale;
}
}
return 0;
}
static inline int vc1_get_LUMSHIFT(const VC1Context *v)
{
if (v->s.pict_type == AV_PICTURE_TYPE_P && !v->p_frame_skipped) {
if ((v->fcm == PROGRESSIVE && v->mv_mode == MV_PMODE_INTENSITY_COMP) ||
(v->fcm == ILACE_FRAME && v->intcomp))
return v->lumshift;
else if (v->fcm == ILACE_FIELD && v->mv_mode == MV_PMODE_INTENSITY_COMP)
switch (v->intcompfield) {
case 1: return v->lumshift;
case 2: return v->lumshift2;
case 3: return v->lumshift;
}
}
return 0;
}
av_unused static inline int vc1_get_LUMSCALE2(const VC1Context *v)
{
if ((v->s.pict_type == AV_PICTURE_TYPE_P && !v->p_frame_skipped) &&
v->fcm == ILACE_FIELD &&
v->mv_mode == MV_PMODE_INTENSITY_COMP &&
v->intcompfield == 3)
return v->lumscale2;
return 0;
}
av_unused static inline int vc1_get_LUMSHIFT2(const VC1Context *v)
{
if ((v->s.pict_type == AV_PICTURE_TYPE_P && !v->p_frame_skipped) &&
v->fcm == ILACE_FIELD &&
v->mv_mode == MV_PMODE_INTENSITY_COMP &&
v->intcompfield == 3)
return v->lumshift2;
return 0;
}
/** Reconstruct bitstream TTFRM (7.1.1.41, Table-53) */
static inline int vc1_get_TTFRM(const VC1Context *v)
{
switch (v->ttfrm) {
case TT_8X8: return 0;
case TT_8X4: return 1;
case TT_4X8: return 2;
case TT_4X4: return 3;
}
return 0;
}
/** Pack FFmpeg bitplanes into a VABitPlaneBuffer element */
static inline void vc1_pack_bitplanes(uint8_t *bitplane, int n, const uint8_t *ff_bp[3], int x, int y, int stride)
{
const int bitplane_index = n / 2;
const int ff_bp_index = y * stride + x;
uint8_t v = 0;
if (ff_bp[0])
v = ff_bp[0][ff_bp_index];
if (ff_bp[1])
v |= ff_bp[1][ff_bp_index] << 1;
if (ff_bp[2])
v |= ff_bp[2][ff_bp_index] << 2;
bitplane[bitplane_index] = (bitplane[bitplane_index] << 4) | v;
}
static int vaapi_vc1_start_frame(AVCodecContext *avctx, av_unused const uint8_t *buffer, av_unused uint32_t size)
{
const VC1Context *v = avctx->priv_data;
const MpegEncContext *s = &v->s;
VAAPIDecodePicture *pic = s->cur_pic.ptr->hwaccel_picture_private;
VAPictureParameterBufferVC1 pic_param;
int err;
pic->output_surface = ff_vaapi_get_surface_id(s->cur_pic.ptr->f);
pic_param = (VAPictureParameterBufferVC1) {
.forward_reference_picture = VA_INVALID_ID,
.backward_reference_picture = VA_INVALID_ID,
.inloop_decoded_picture = VA_INVALID_ID,
.sequence_fields.bits = {
.pulldown = v->broadcast,
.interlace = v->interlace,
.tfcntrflag = v->tfcntrflag,
.finterpflag = v->finterpflag,
.psf = v->psf,
.multires = v->multires,
.overlap = v->overlap,
.syncmarker = v->resync_marker,
.rangered = v->rangered,
.max_b_frames = s->avctx->max_b_frames,
.profile = v->profile,
},
.coded_width = s->avctx->coded_width,
.coded_height = s->avctx->coded_height,
.entrypoint_fields.bits = {
.broken_link = v->broken_link,
.closed_entry = v->closed_entry,
.panscan_flag = v->panscanflag,
.loopfilter = s->loop_filter,
},
.conditional_overlap_flag = v->condover,
.fast_uvmc_flag = v->fastuvmc,
.range_mapping_fields.bits = {
.luma_flag = v->range_mapy_flag,
.luma = v->range_mapy,
.chroma_flag = v->range_mapuv_flag,
.chroma = v->range_mapuv,
},
.b_picture_fraction = v->bfraction_lut_index,
.cbp_table = (v->fcm == PROGRESSIVE ? v->cbptab : v->icbptab),
.mb_mode_table = v->mbmodetab,
.range_reduction_frame = v->rangeredfrm,
.rounding_control = v->rnd,
.post_processing = v->postproc,
.picture_resolution_index = v->respic,
.picture_fields.bits = {
.picture_type = (v->fcm == ILACE_FIELD ? vc1_get_FPTYPE(v) : vc1_get_PTYPE(v)),
.frame_coding_mode = v->fcm,
.top_field_first = v->tff,
.is_first_field = !v->second_field,
.intensity_compensation = v->intcomp,
},
.luma_scale = vc1_get_LUMSCALE(v),
.luma_shift = vc1_get_LUMSHIFT(v),
#if VA_CHECK_VERSION(1, 1, 0)
.luma_scale2 = vc1_get_LUMSCALE2(v),
.luma_shift2 = vc1_get_LUMSHIFT2(v),
.intensity_compensation_field = vc1_get_INTCOMPFIELD(v),
#endif
.raw_coding.flags = {
.mv_type_mb = v->mv_type_is_raw,
.direct_mb = v->dmb_is_raw,
.skip_mb = v->skip_is_raw,
.field_tx = v->fieldtx_is_raw,
.forward_mb = v->fmb_is_raw,
.ac_pred = v->acpred_is_raw,
.overflags = v->overflg_is_raw,
},
.bitplane_present.flags = {
.bp_mv_type_mb = vc1_has_MVTYPEMB_bitplane(v),
.bp_direct_mb = vc1_has_DIRECTMB_bitplane(v),
.bp_skip_mb = vc1_has_SKIPMB_bitplane(v),
.bp_field_tx = vc1_has_FIELDTX_bitplane(v),
.bp_forward_mb = vc1_has_FORWARDMB_bitplane(v),
.bp_ac_pred = vc1_has_ACPRED_bitplane(v),
.bp_overflags = vc1_has_OVERFLAGS_bitplane(v),
},
.reference_fields.bits = {
.reference_distance_flag = v->refdist_flag,
.reference_distance = v->refdist,
.num_reference_pictures = v->numref,
.reference_field_pic_indicator = v->reffield,
},
.mv_fields.bits = {
.mv_mode = vc1_get_MVMODE(v),
.mv_mode2 = vc1_get_MVMODE2(v),
.mv_table = (v->fcm == PROGRESSIVE ? s->mv_table_index : v->imvtab),
.two_mv_block_pattern_table = v->twomvbptab,
.four_mv_switch = v->fourmvswitch,
.four_mv_block_pattern_table = v->fourmvbptab,
.extended_mv_flag = v->extended_mv,
.extended_mv_range = v->mvrange,
.extended_dmv_flag = v->extended_dmv,
.extended_dmv_range = v->dmvrange,
},
.pic_quantizer_fields.bits = {
.dquant = v->dquant,
.quantizer = v->quantizer_mode,
.half_qp = v->halfpq,
.pic_quantizer_scale = v->pq,
.pic_quantizer_type = v->pquantizer,
.dq_frame = v->dquantfrm,
.dq_profile = v->dqprofile,
.dq_sb_edge = v->dqprofile == DQPROFILE_SINGLE_EDGE ? v->dqsbedge : 0,
.dq_db_edge = v->dqprofile == DQPROFILE_DOUBLE_EDGES ? v->dqsbedge : 0,
.dq_binary_level = v->dqbilevel,
.alt_pic_quantizer = v->altpq,
},
.transform_fields.bits = {
.variable_sized_transform_flag = v->vstransform,
.mb_level_transform_type_flag = v->ttmbf,
.frame_level_transform_type = vc1_get_TTFRM(v),
.transform_ac_codingset_idx1 = v->c_ac_table_index,
.transform_ac_codingset_idx2 = v->y_ac_table_index,
.intra_transform_dc_table = v->s.dc_table_index,
},
};
switch (s->pict_type) {
case AV_PICTURE_TYPE_B:
if (s->next_pic.ptr)
pic_param.backward_reference_picture = ff_vaapi_get_surface_id(s->next_pic.ptr->f);
// fall-through
case AV_PICTURE_TYPE_P:
if (s->last_pic.ptr)
pic_param.forward_reference_picture = ff_vaapi_get_surface_id(s->last_pic.ptr->f);
break;
}
err = ff_vaapi_decode_make_param_buffer(avctx, pic,
VAPictureParameterBufferType,
&pic_param, sizeof(pic_param));
if (err)
goto fail;
if (pic_param.bitplane_present.value & 0x7f) {
uint8_t *bitplane;
const uint8_t *ff_bp[3];
int x, y, n;
size_t size = (s->mb_width * s->mb_height + 1) / 2;
bitplane = av_mallocz(size);
if (!bitplane) {
err = AVERROR(ENOMEM);
goto fail;
}
switch (s->pict_type) {
case AV_PICTURE_TYPE_P:
ff_bp[0] = pic_param.bitplane_present.flags.bp_direct_mb ? v->direct_mb_plane : NULL;
ff_bp[1] = pic_param.bitplane_present.flags.bp_skip_mb ? s->mbskip_table : NULL;
ff_bp[2] = pic_param.bitplane_present.flags.bp_mv_type_mb ? v->mv_type_mb_plane : NULL;
break;
case AV_PICTURE_TYPE_B:
if (!v->bi_type) {
ff_bp[0] = pic_param.bitplane_present.flags.bp_direct_mb ? v->direct_mb_plane : NULL;
ff_bp[1] = pic_param.bitplane_present.flags.bp_skip_mb ? s->mbskip_table : NULL;
ff_bp[2] = pic_param.bitplane_present.flags.bp_forward_mb ? v->forward_mb_plane : NULL;
break;
}
/* fall-through (BI-type) */
case AV_PICTURE_TYPE_I:
ff_bp[0] = pic_param.bitplane_present.flags.bp_field_tx ? v->fieldtx_plane : NULL;
ff_bp[1] = pic_param.bitplane_present.flags.bp_ac_pred ? v->acpred_plane : NULL;
ff_bp[2] = pic_param.bitplane_present.flags.bp_overflags ? v->over_flags_plane : NULL;
break;
default:
ff_bp[0] = NULL;
ff_bp[1] = NULL;
ff_bp[2] = NULL;
break;
}
n = 0;
for (y = 0; y < s->mb_height; y++)
for (x = 0; x < s->mb_width; x++, n++)
vc1_pack_bitplanes(bitplane, n, ff_bp, x, y, s->mb_stride);
if (n & 1) /* move last nibble to the high order */
bitplane[n/2] <<= 4;
err = ff_vaapi_decode_make_param_buffer(avctx, pic,
VABitPlaneBufferType,
bitplane, size);
av_free(bitplane);
if (err)
goto fail;
}
return 0;
fail:
ff_vaapi_decode_cancel(avctx, pic);
return err;
}
static int vaapi_vc1_end_frame(AVCodecContext *avctx)
{
VC1Context *v = avctx->priv_data;
MpegEncContext *s = &v->s;
VAAPIDecodePicture *pic = s->cur_pic.ptr->hwaccel_picture_private;
int ret;
ret = ff_vaapi_decode_issue(avctx, pic);
if (ret < 0)
goto fail;
fail:
return ret;
}
static int vaapi_vc1_decode_slice(AVCodecContext *avctx, const uint8_t *buffer, uint32_t size)
{
const VC1Context *v = avctx->priv_data;
const MpegEncContext *s = &v->s;
VAAPIDecodePicture *pic = s->cur_pic.ptr->hwaccel_picture_private;
VASliceParameterBufferVC1 slice_param;
int mb_height;
int err;
/* Current bit buffer is beyond any marker for VC-1, so skip it */
if (avctx->codec_id == AV_CODEC_ID_VC1 && IS_MARKER(AV_RB32(buffer))) {
buffer += 4;
size -= 4;
}
if (v->fcm == ILACE_FIELD)
mb_height = avctx->coded_height + 31 >> 5;
else
mb_height = avctx->coded_height + 15 >> 4;
slice_param = (VASliceParameterBufferVC1) {
.slice_data_size = size,
.slice_data_offset = 0,
.slice_data_flag = VA_SLICE_DATA_FLAG_ALL,
.macroblock_offset = get_bits_count(&s->gb),
.slice_vertical_position = s->mb_y % mb_height,
};
err = ff_vaapi_decode_make_slice_buffer(avctx, pic,
&slice_param, 1, sizeof(slice_param),
buffer, size);
if (err < 0) {
ff_vaapi_decode_cancel(avctx, pic);
return err;
}
return 0;
}
#if CONFIG_WMV3_VAAPI_HWACCEL
const FFHWAccel ff_wmv3_vaapi_hwaccel = {
.p.name = "wmv3_vaapi",
.p.type = AVMEDIA_TYPE_VIDEO,
.p.id = AV_CODEC_ID_WMV3,
.p.pix_fmt = AV_PIX_FMT_VAAPI,
.start_frame = &vaapi_vc1_start_frame,
.end_frame = &vaapi_vc1_end_frame,
.decode_slice = &vaapi_vc1_decode_slice,
.frame_priv_data_size = sizeof(VAAPIDecodePicture),
.init = &ff_vaapi_decode_init,
.uninit = &ff_vaapi_decode_uninit,
.frame_params = &ff_vaapi_common_frame_params,
.priv_data_size = sizeof(VAAPIDecodeContext),
.caps_internal = HWACCEL_CAP_ASYNC_SAFE,
};
#endif
const FFHWAccel ff_vc1_vaapi_hwaccel = {
.p.name = "vc1_vaapi",
.p.type = AVMEDIA_TYPE_VIDEO,
.p.id = AV_CODEC_ID_VC1,
.p.pix_fmt = AV_PIX_FMT_VAAPI,
.start_frame = &vaapi_vc1_start_frame,
.end_frame = &vaapi_vc1_end_frame,
.decode_slice = &vaapi_vc1_decode_slice,
.frame_priv_data_size = sizeof(VAAPIDecodePicture),
.init = &ff_vaapi_decode_init,
.uninit = &ff_vaapi_decode_uninit,
.frame_params = &ff_vaapi_common_frame_params,
.priv_data_size = sizeof(VAAPIDecodeContext),
.caps_internal = HWACCEL_CAP_ASYNC_SAFE,
};