diff --git a/libavcodec/vaapi_vp9.c b/libavcodec/vaapi_vp9.c index 9b3e81a388..a656ffb13c 100644 --- a/libavcodec/vaapi_vp9.c +++ b/libavcodec/vaapi_vp9.c @@ -21,72 +21,15 @@ */ #include "libavutil/pixdesc.h" -#include "vaapi_internal.h" +#include "vaapi_decode.h" #include "vp9.h" -static void fill_picture_parameters(AVCodecContext *avctx, - const VP9SharedContext *h, - VADecPictureParameterBufferVP9 *pp) +static VASurfaceID vaapi_vp9_surface_id(const VP9Frame *vf) { - const AVPixFmtDescriptor *pixdesc = av_pix_fmt_desc_get(avctx->sw_pix_fmt); - int i; - - pp->frame_width = avctx->width; - pp->frame_height = avctx->height; - - pp->frame_header_length_in_bytes = h->h.uncompressed_header_size; - pp->first_partition_size = h->h.compressed_header_size; - - pp->profile = h->h.profile; - pp->bit_depth = h->h.bpp; - - pp->filter_level = h->h.filter.level; - pp->sharpness_level = h->h.filter.sharpness; - pp->log2_tile_rows = h->h.tiling.log2_tile_rows; - pp->log2_tile_columns = h->h.tiling.log2_tile_cols; - - pp->pic_fields.bits.subsampling_x = pixdesc->log2_chroma_w; - pp->pic_fields.bits.subsampling_y = pixdesc->log2_chroma_h; - pp->pic_fields.bits.frame_type = !h->h.keyframe; - pp->pic_fields.bits.show_frame = !h->h.invisible; - pp->pic_fields.bits.error_resilient_mode = h->h.errorres; - pp->pic_fields.bits.intra_only = h->h.intraonly; - pp->pic_fields.bits.allow_high_precision_mv = h->h.keyframe ? 0 : h->h.highprecisionmvs; - pp->pic_fields.bits.mcomp_filter_type = h->h.filtermode ^ (h->h.filtermode <= 1); - pp->pic_fields.bits.frame_parallel_decoding_mode = h->h.parallelmode; - pp->pic_fields.bits.reset_frame_context = h->h.resetctx; - pp->pic_fields.bits.refresh_frame_context = h->h.refreshctx; - pp->pic_fields.bits.frame_context_idx = h->h.framectxid; - - pp->pic_fields.bits.segmentation_enabled = h->h.segmentation.enabled; - pp->pic_fields.bits.segmentation_temporal_update = h->h.segmentation.temporal; - pp->pic_fields.bits.segmentation_update_map = h->h.segmentation.update_map; - - pp->pic_fields.bits.last_ref_frame = h->h.refidx[0]; - pp->pic_fields.bits.last_ref_frame_sign_bias = h->h.signbias[0]; - pp->pic_fields.bits.golden_ref_frame = h->h.refidx[1]; - pp->pic_fields.bits.golden_ref_frame_sign_bias = h->h.signbias[1]; - pp->pic_fields.bits.alt_ref_frame = h->h.refidx[2]; - pp->pic_fields.bits.alt_ref_frame_sign_bias = h->h.signbias[2]; - pp->pic_fields.bits.lossless_flag = h->h.lossless; - - for (i = 0; i < 7; i++) - pp->mb_segment_tree_probs[i] = h->h.segmentation.prob[i]; - - if (h->h.segmentation.temporal) { - for (i = 0; i < 3; i++) - pp->segment_pred_probs[i] = h->h.segmentation.pred_prob[i]; - } else { - memset(pp->segment_pred_probs, 255, sizeof(pp->segment_pred_probs)); - } - - for (i = 0; i < 8; i++) { - if (h->refs[i].f->buf[0]) { - pp->reference_frames[i] = ff_vaapi_get_surface_id(h->refs[i].f); - } else { - pp->reference_frames[i] = VA_INVALID_ID; - } - } + if (vf) + return ff_vaapi_get_surface_id(vf->tf.f); + else + return VA_INVALID_SURFACE; } static int vaapi_vp9_start_frame(AVCodecContext *avctx, @@ -94,62 +37,130 @@ static int vaapi_vp9_start_frame(AVCodecContext *avctx, av_unused uint32_t size) { const VP9SharedContext *h = avctx->priv_data; - FFVAContext * const vactx = ff_vaapi_get_context(avctx); - VADecPictureParameterBufferVP9 *pic_param; + VAAPIDecodePicture *pic = h->frames[CUR_FRAME].hwaccel_picture_private; + VADecPictureParameterBufferVP9 pic_param; + const AVPixFmtDescriptor *pixdesc = av_pix_fmt_desc_get(avctx->sw_pix_fmt); + int err, i; - vactx->slice_param_size = sizeof(VASliceParameterBufferVP9); + pic->output_surface = vaapi_vp9_surface_id(&h->frames[CUR_FRAME]); - pic_param = ff_vaapi_alloc_pic_param(vactx, sizeof(VADecPictureParameterBufferVP9)); - if (!pic_param) - return -1; - fill_picture_parameters(avctx, h, pic_param); + pic_param = (VADecPictureParameterBufferVP9) { + .frame_width = avctx->width, + .frame_height = avctx->height, + + .pic_fields.bits = { + .subsampling_x = pixdesc->log2_chroma_w, + .subsampling_y = pixdesc->log2_chroma_h, + .frame_type = !h->h.keyframe, + .show_frame = !h->h.invisible, + .error_resilient_mode = h->h.errorres, + .intra_only = h->h.intraonly, + .allow_high_precision_mv = h->h.keyframe ? 0 : h->h.highprecisionmvs, + .mcomp_filter_type = h->h.filtermode ^ (h->h.filtermode <= 1), + .frame_parallel_decoding_mode = h->h.parallelmode, + .reset_frame_context = h->h.resetctx, + .refresh_frame_context = h->h.refreshctx, + .frame_context_idx = h->h.framectxid, + + .segmentation_enabled = h->h.segmentation.enabled, + .segmentation_temporal_update = h->h.segmentation.temporal, + .segmentation_update_map = h->h.segmentation.update_map, + + .last_ref_frame = h->h.refidx[0], + .last_ref_frame_sign_bias = h->h.signbias[0], + .golden_ref_frame = h->h.refidx[1], + .golden_ref_frame_sign_bias = h->h.signbias[1], + .alt_ref_frame = h->h.refidx[2], + .alt_ref_frame_sign_bias = h->h.signbias[2], + .lossless_flag = h->h.lossless, + }, + + .filter_level = h->h.filter.level, + .sharpness_level = h->h.filter.sharpness, + .log2_tile_rows = h->h.tiling.log2_tile_rows, + .log2_tile_columns = h->h.tiling.log2_tile_cols, + + .frame_header_length_in_bytes = h->h.uncompressed_header_size, + .first_partition_size = h->h.compressed_header_size, + + .profile = h->h.profile, + .bit_depth = h->h.bpp, + }; + + for (i = 0; i < 7; i++) + pic_param.mb_segment_tree_probs[i] = h->h.segmentation.prob[i]; + + if (h->h.segmentation.temporal) { + for (i = 0; i < 3; i++) + pic_param.segment_pred_probs[i] = h->h.segmentation.pred_prob[i]; + } else { + memset(pic_param.segment_pred_probs, 255, sizeof(pic_param.segment_pred_probs)); + } + + for (i = 0; i < 8; i++) { + if (h->refs[i].f->buf[0]) + pic_param.reference_frames[i] = ff_vaapi_get_surface_id(h->refs[i].f); + else + pic_param.reference_frames[i] = VA_INVALID_ID; + } + + err = ff_vaapi_decode_make_param_buffer(avctx, pic, + VAPictureParameterBufferType, + &pic_param, sizeof(pic_param)); + if (err < 0) { + ff_vaapi_decode_cancel(avctx, pic); + return err; + } return 0; } static int vaapi_vp9_end_frame(AVCodecContext *avctx) { - FFVAContext * const vactx = ff_vaapi_get_context(avctx); const VP9SharedContext *h = avctx->priv_data; - int ret; + VAAPIDecodePicture *pic = h->frames[CUR_FRAME].hwaccel_picture_private; - ret = ff_vaapi_commit_slices(vactx); - if (ret < 0) - goto finish; - - ret = ff_vaapi_render_picture(vactx, ff_vaapi_get_surface_id(h->frames[CUR_FRAME].tf.f)); - if (ret < 0) - goto finish; - -finish: - ff_vaapi_common_end_frame(avctx); - return ret; + return ff_vaapi_decode_issue(avctx, pic); } static int vaapi_vp9_decode_slice(AVCodecContext *avctx, const uint8_t *buffer, uint32_t size) { - FFVAContext * const vactx = ff_vaapi_get_context(avctx); const VP9SharedContext *h = avctx->priv_data; - VASliceParameterBufferVP9 *slice_param; - int i; + VAAPIDecodePicture *pic = h->frames[CUR_FRAME].hwaccel_picture_private; + VASliceParameterBufferVP9 slice_param; + int err, i; - slice_param = (VASliceParameterBufferVP9*)ff_vaapi_alloc_slice(vactx, buffer, size); - if (!slice_param) - return -1; + slice_param = (VASliceParameterBufferVP9) { + .slice_data_size = size, + .slice_data_offset = 0, + .slice_data_flag = VA_SLICE_DATA_FLAG_ALL, + }; for (i = 0; i < 8; i++) { - slice_param->seg_param[i].segment_flags.fields.segment_reference_enabled = h->h.segmentation.feat[i].ref_enabled; - slice_param->seg_param[i].segment_flags.fields.segment_reference = h->h.segmentation.feat[i].ref_val; - slice_param->seg_param[i].segment_flags.fields.segment_reference_skipped = h->h.segmentation.feat[i].skip_enabled; + slice_param.seg_param[i] = (VASegmentParameterVP9) { + .segment_flags.fields = { + .segment_reference_enabled = h->h.segmentation.feat[i].ref_enabled, + .segment_reference = h->h.segmentation.feat[i].ref_val, + .segment_reference_skipped = h->h.segmentation.feat[i].skip_enabled, + }, - memcpy(slice_param->seg_param[i].filter_level, h->h.segmentation.feat[i].lflvl, sizeof(slice_param->seg_param[i].filter_level)); + .luma_dc_quant_scale = h->h.segmentation.feat[i].qmul[0][0], + .luma_ac_quant_scale = h->h.segmentation.feat[i].qmul[0][1], + .chroma_dc_quant_scale = h->h.segmentation.feat[i].qmul[1][0], + .chroma_ac_quant_scale = h->h.segmentation.feat[i].qmul[1][1], + }; - slice_param->seg_param[i].luma_dc_quant_scale = h->h.segmentation.feat[i].qmul[0][0]; - slice_param->seg_param[i].luma_ac_quant_scale = h->h.segmentation.feat[i].qmul[0][1]; - slice_param->seg_param[i].chroma_dc_quant_scale = h->h.segmentation.feat[i].qmul[1][0]; - slice_param->seg_param[i].chroma_ac_quant_scale = h->h.segmentation.feat[i].qmul[1][1]; + memcpy(slice_param.seg_param[i].filter_level, h->h.segmentation.feat[i].lflvl, sizeof(slice_param.seg_param[i].filter_level)); + } + + err = ff_vaapi_decode_make_slice_buffer(avctx, pic, + &slice_param, sizeof(slice_param), + buffer, size); + if (err) { + ff_vaapi_decode_cancel(avctx, pic); + return err; } return 0; @@ -163,7 +174,8 @@ AVHWAccel ff_vp9_vaapi_hwaccel = { .start_frame = vaapi_vp9_start_frame, .end_frame = vaapi_vp9_end_frame, .decode_slice = vaapi_vp9_decode_slice, - .init = ff_vaapi_context_init, - .uninit = ff_vaapi_context_fini, - .priv_data_size = sizeof(FFVAContext), + .frame_priv_data_size = sizeof(VAAPIDecodePicture), + .init = ff_vaapi_decode_init, + .uninit = ff_vaapi_decode_uninit, + .priv_data_size = sizeof(VAAPIDecodeContext), };