mirror of https://git.ffmpeg.org/ffmpeg.git
583 lines
26 KiB
C
583 lines
26 KiB
C
/*
|
|
* This file is part of FFmpeg.
|
|
*
|
|
* FFmpeg is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with FFmpeg; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
#include "av1dec.h"
|
|
|
|
#include "vulkan_decode.h"
|
|
|
|
/* Maximum number of tiles specified by any defined level */
|
|
#define MAX_TILES 256
|
|
|
|
const VkExtensionProperties ff_vk_dec_av1_ext = {
|
|
.extensionName = VK_STD_VULKAN_VIDEO_CODEC_AV1_DECODE_EXTENSION_NAME,
|
|
.specVersion = VK_STD_VULKAN_VIDEO_CODEC_AV1_DECODE_SPEC_VERSION,
|
|
};
|
|
|
|
typedef struct AV1VulkanDecodePicture {
|
|
FFVulkanDecodePicture vp;
|
|
|
|
/* Workaround for a spec issue.
|
|
*Can be removed once no longer needed, and threading can be enabled. */
|
|
FFVulkanDecodeContext *dec;
|
|
|
|
StdVideoAV1MESATile tiles[MAX_TILES];
|
|
StdVideoAV1MESATileList tile_list;
|
|
const uint32_t *tile_offsets;
|
|
|
|
/* Current picture */
|
|
VkVideoDecodeAV1DpbSlotInfoMESA vkav1_ref;
|
|
StdVideoAV1MESAFrameHeader av1_frame_header;
|
|
VkVideoDecodeAV1PictureInfoMESA av1_pic_info;
|
|
|
|
/* Picture refs */
|
|
const AV1Frame *ref_src [AV1_NUM_REF_FRAMES];
|
|
VkVideoDecodeAV1DpbSlotInfoMESA vkav1_refs[AV1_NUM_REF_FRAMES];
|
|
|
|
uint8_t frame_id_set;
|
|
uint8_t frame_id;
|
|
} AV1VulkanDecodePicture;
|
|
|
|
static int vk_av1_fill_pict(AVCodecContext *avctx, const AV1Frame **ref_src,
|
|
VkVideoReferenceSlotInfoKHR *ref_slot, /* Main structure */
|
|
VkVideoPictureResourceInfoKHR *ref, /* Goes in ^ */
|
|
VkVideoDecodeAV1DpbSlotInfoMESA *vkav1_ref, /* Goes in ^ */
|
|
const AV1Frame *pic, int is_current, int has_grain,
|
|
int dpb_slot_index)
|
|
{
|
|
FFVulkanDecodeContext *dec = avctx->internal->hwaccel_priv_data;
|
|
AV1VulkanDecodePicture *hp = pic->hwaccel_picture_private;
|
|
FFVulkanDecodePicture *vkpic = &hp->vp;
|
|
|
|
int err = ff_vk_decode_prepare_frame(dec, pic->f, vkpic, is_current,
|
|
has_grain || dec->dedicated_dpb);
|
|
if (err < 0)
|
|
return err;
|
|
|
|
*vkav1_ref = (VkVideoDecodeAV1DpbSlotInfoMESA) {
|
|
.sType = VK_STRUCTURE_TYPE_VIDEO_DECODE_AV1_DPB_SLOT_INFO_MESA,
|
|
.frameIdx = hp->frame_id,
|
|
};
|
|
|
|
for (unsigned i = 0; i < 7; i++) {
|
|
const int idx = pic->raw_frame_header->ref_frame_idx[i];
|
|
vkav1_ref->ref_order_hint[i] = pic->raw_frame_header->ref_order_hint[idx];
|
|
}
|
|
|
|
vkav1_ref->disable_frame_end_update_cdf = pic->raw_frame_header->disable_frame_end_update_cdf;
|
|
|
|
*ref = (VkVideoPictureResourceInfoKHR) {
|
|
.sType = VK_STRUCTURE_TYPE_VIDEO_PICTURE_RESOURCE_INFO_KHR,
|
|
.codedOffset = (VkOffset2D){ 0, 0 },
|
|
.codedExtent = (VkExtent2D){ pic->f->width, pic->f->height },
|
|
.baseArrayLayer = ((has_grain || dec->dedicated_dpb) && dec->layered_dpb) ?
|
|
dpb_slot_index : 0,
|
|
.imageViewBinding = vkpic->img_view_ref,
|
|
};
|
|
|
|
*ref_slot = (VkVideoReferenceSlotInfoKHR) {
|
|
.sType = VK_STRUCTURE_TYPE_VIDEO_REFERENCE_SLOT_INFO_KHR,
|
|
.pNext = vkav1_ref,
|
|
.slotIndex = dpb_slot_index,
|
|
.pPictureResource = ref,
|
|
};
|
|
|
|
if (ref_src)
|
|
*ref_src = pic;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int vk_av1_create_params(AVCodecContext *avctx, AVBufferRef **buf)
|
|
{
|
|
const AV1DecContext *s = avctx->priv_data;
|
|
FFVulkanDecodeContext *dec = avctx->internal->hwaccel_priv_data;
|
|
FFVulkanDecodeShared *ctx = dec->shared_ctx;
|
|
|
|
const AV1RawSequenceHeader *seq = s->raw_seq;
|
|
|
|
StdVideoAV1MESASequenceHeader av1_sequence_header;
|
|
VkVideoDecodeAV1SessionParametersAddInfoMESA av1_params_info;
|
|
VkVideoDecodeAV1SessionParametersCreateInfoMESA av1_params;
|
|
VkVideoSessionParametersCreateInfoKHR session_params_create;
|
|
|
|
int err;
|
|
|
|
av1_sequence_header = (StdVideoAV1MESASequenceHeader) {
|
|
.flags = (StdVideoAV1MESASequenceHeaderFlags) {
|
|
.still_picture = seq->still_picture,
|
|
.reduced_still_picture_header = seq->reduced_still_picture_header,
|
|
.use_128x128_superblock = seq->use_128x128_superblock,
|
|
.enable_filter_intra = seq->enable_filter_intra,
|
|
.enable_intra_edge_filter = seq->enable_intra_edge_filter,
|
|
.enable_interintra_compound = seq->enable_interintra_compound,
|
|
.enable_masked_compound = seq->enable_masked_compound,
|
|
.enable_warped_motion = seq->enable_warped_motion,
|
|
.enable_dual_filter = seq->enable_dual_filter,
|
|
.enable_order_hint = seq->enable_order_hint,
|
|
.enable_jnt_comp = seq->enable_jnt_comp,
|
|
.enable_ref_frame_mvs = seq->enable_ref_frame_mvs,
|
|
.frame_id_numbers_present_flag = seq->frame_id_numbers_present_flag,
|
|
.enable_superres = seq->enable_superres,
|
|
.enable_cdef = seq->enable_cdef,
|
|
.enable_restoration = seq->enable_restoration,
|
|
.film_grain_params_present = seq->film_grain_params_present,
|
|
.timing_info_present_flag = seq->timing_info_present_flag,
|
|
.initial_display_delay_present_flag = seq->initial_display_delay_present_flag,
|
|
},
|
|
.seq_profile = seq->seq_profile,
|
|
.frame_width_bits_minus_1 = seq->frame_width_bits_minus_1,
|
|
.frame_height_bits_minus_1 = seq->frame_height_bits_minus_1,
|
|
.max_frame_width_minus_1 = seq->max_frame_width_minus_1,
|
|
.max_frame_height_minus_1 = seq->max_frame_height_minus_1,
|
|
.delta_frame_id_length_minus_2 = seq->delta_frame_id_length_minus_2,
|
|
.additional_frame_id_length_minus_1 = seq->additional_frame_id_length_minus_1,
|
|
.order_hint_bits_minus_1 = seq->order_hint_bits_minus_1,
|
|
.timing_info = (StdVideoAV1MESATimingInfo) {
|
|
.flags = (StdVideoAV1MESATimingInfoFlags) {
|
|
.equal_picture_interval = seq->timing_info.equal_picture_interval,
|
|
},
|
|
.num_units_in_display_tick = seq->timing_info.num_units_in_display_tick,
|
|
.time_scale = seq->timing_info.time_scale,
|
|
.num_ticks_per_picture_minus_1 = seq->timing_info.num_ticks_per_picture_minus_1,
|
|
},
|
|
.color_config = (StdVideoAV1MESAColorConfig) {
|
|
.flags = (StdVideoAV1MESAColorConfigFlags) {
|
|
.mono_chrome = seq->color_config.mono_chrome,
|
|
.color_range = seq->color_config.color_range,
|
|
.separate_uv_delta_q = seq->color_config.separate_uv_delta_q,
|
|
},
|
|
.bit_depth = seq->color_config.twelve_bit ? 12 :
|
|
seq->color_config.high_bitdepth ? 10 : 8,
|
|
.subsampling_x = seq->color_config.subsampling_x,
|
|
.subsampling_y = seq->color_config.subsampling_y,
|
|
},
|
|
};
|
|
|
|
av1_params_info = (VkVideoDecodeAV1SessionParametersAddInfoMESA) {
|
|
.sType = VK_STRUCTURE_TYPE_VIDEO_DECODE_AV1_SESSION_PARAMETERS_ADD_INFO_MESA,
|
|
.sequence_header = &av1_sequence_header,
|
|
};
|
|
av1_params = (VkVideoDecodeAV1SessionParametersCreateInfoMESA) {
|
|
.sType = VK_STRUCTURE_TYPE_VIDEO_DECODE_AV1_SESSION_PARAMETERS_CREATE_INFO_MESA,
|
|
.pParametersAddInfo = &av1_params_info,
|
|
};
|
|
session_params_create = (VkVideoSessionParametersCreateInfoKHR) {
|
|
.sType = VK_STRUCTURE_TYPE_VIDEO_SESSION_PARAMETERS_CREATE_INFO_KHR,
|
|
.pNext = &av1_params,
|
|
.videoSession = ctx->common.session,
|
|
.videoSessionParametersTemplate = VK_NULL_HANDLE,
|
|
};
|
|
|
|
err = ff_vk_decode_create_params(buf, avctx, ctx, &session_params_create);
|
|
if (err < 0)
|
|
return err;
|
|
|
|
av_log(avctx, AV_LOG_DEBUG, "Created frame parameters\n");
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int vk_av1_start_frame(AVCodecContext *avctx,
|
|
av_unused const uint8_t *buffer,
|
|
av_unused uint32_t size)
|
|
{
|
|
int err;
|
|
int ref_count = 0;
|
|
AV1DecContext *s = avctx->priv_data;
|
|
const AV1Frame *pic = &s->cur_frame;
|
|
FFVulkanDecodeContext *dec = avctx->internal->hwaccel_priv_data;
|
|
AV1VulkanDecodePicture *ap = pic->hwaccel_picture_private;
|
|
FFVulkanDecodePicture *vp = &ap->vp;
|
|
|
|
const AV1RawFrameHeader *frame_header = s->raw_frame_header;
|
|
const AV1RawFilmGrainParams *film_grain = &s->cur_frame.film_grain;
|
|
const int apply_grain = !(avctx->export_side_data & AV_CODEC_EXPORT_DATA_FILM_GRAIN) &&
|
|
film_grain->apply_grain;
|
|
|
|
if (!dec->session_params) {
|
|
err = vk_av1_create_params(avctx, &dec->session_params);
|
|
if (err < 0)
|
|
return err;
|
|
}
|
|
|
|
if (!ap->frame_id_set) {
|
|
unsigned slot_idx = 0;
|
|
for (unsigned i = 0; i < 32; i++) {
|
|
if (!(dec->frame_id_alloc_mask & (1 << i))) {
|
|
slot_idx = i;
|
|
break;
|
|
}
|
|
}
|
|
ap->frame_id = slot_idx;
|
|
ap->frame_id_set = 1;
|
|
dec->frame_id_alloc_mask |= (1 << slot_idx);
|
|
}
|
|
|
|
/* Fill in references */
|
|
for (int i = 0; i < AV1_NUM_REF_FRAMES; i++) {
|
|
const AV1Frame *ref_frame = &s->ref[i];
|
|
if (s->ref[i].f->pict_type == AV_PICTURE_TYPE_NONE)
|
|
continue;
|
|
|
|
err = vk_av1_fill_pict(avctx, &ap->ref_src[i], &vp->ref_slots[i],
|
|
&vp->refs[i], &ap->vkav1_refs[i],
|
|
ref_frame, 0, 0, i);
|
|
if (err < 0)
|
|
return err;
|
|
|
|
ref_count++;
|
|
}
|
|
|
|
err = vk_av1_fill_pict(avctx, NULL, &vp->ref_slot, &vp->ref,
|
|
&ap->vkav1_ref,
|
|
pic, 1, apply_grain, 8);
|
|
if (err < 0)
|
|
return err;
|
|
|
|
ap->tile_list.nb_tiles = 0;
|
|
ap->tile_list.tile_list = ap->tiles;
|
|
|
|
ap->av1_pic_info = (VkVideoDecodeAV1PictureInfoMESA) {
|
|
.sType = VK_STRUCTURE_TYPE_VIDEO_DECODE_AV1_PICTURE_INFO_MESA,
|
|
.frame_header = &ap->av1_frame_header,
|
|
.tile_list = &ap->tile_list,
|
|
};
|
|
|
|
vp->decode_info = (VkVideoDecodeInfoKHR) {
|
|
.sType = VK_STRUCTURE_TYPE_VIDEO_DECODE_INFO_KHR,
|
|
.pNext = &ap->av1_pic_info,
|
|
.flags = 0x0,
|
|
.pSetupReferenceSlot = &vp->ref_slot,
|
|
.referenceSlotCount = ref_count,
|
|
.pReferenceSlots = vp->ref_slots,
|
|
.dstPictureResource = (VkVideoPictureResourceInfoKHR) {
|
|
.sType = VK_STRUCTURE_TYPE_VIDEO_PICTURE_RESOURCE_INFO_KHR,
|
|
.codedOffset = (VkOffset2D){ 0, 0 },
|
|
.codedExtent = (VkExtent2D){ pic->f->width, pic->f->height },
|
|
.baseArrayLayer = 0,
|
|
.imageViewBinding = vp->img_view_out,
|
|
},
|
|
};
|
|
|
|
/* Setup frame header */
|
|
ap->av1_frame_header = (StdVideoAV1MESAFrameHeader) {
|
|
.flags = (StdVideoAV1MESAFrameHeaderFlags) {
|
|
.error_resilient_mode = frame_header->error_resilient_mode,
|
|
.disable_cdf_update = frame_header->disable_cdf_update,
|
|
.use_superres = frame_header->use_superres,
|
|
.render_and_frame_size_different = frame_header->render_and_frame_size_different,
|
|
.allow_screen_content_tools = frame_header->allow_screen_content_tools,
|
|
.is_filter_switchable = frame_header->is_filter_switchable,
|
|
.force_integer_mv = frame_header->force_integer_mv,
|
|
.frame_size_override_flag = frame_header->frame_size_override_flag,
|
|
.buffer_removal_time_present_flag = frame_header->buffer_removal_time_present_flag,
|
|
.allow_intrabc = frame_header->allow_intrabc,
|
|
.frame_refs_short_signaling = frame_header->frame_refs_short_signaling,
|
|
.allow_high_precision_mv = frame_header->allow_high_precision_mv,
|
|
.is_motion_mode_switchable = frame_header->is_motion_mode_switchable,
|
|
.use_ref_frame_mvs = frame_header->use_ref_frame_mvs,
|
|
.disable_frame_end_update_cdf = frame_header->disable_frame_end_update_cdf,
|
|
.allow_warped_motion = frame_header->allow_warped_motion,
|
|
.reduced_tx_set = frame_header->reduced_tx_set,
|
|
.reference_select = frame_header->reference_select,
|
|
.skip_mode_present = frame_header->skip_mode_present,
|
|
.delta_q_present = frame_header->delta_q_present,
|
|
},
|
|
.frame_to_show_map_idx = frame_header->frame_to_show_map_idx,
|
|
.frame_presentation_time = frame_header->frame_presentation_time,
|
|
.display_frame_id = frame_header->display_frame_id,
|
|
.frame_type = frame_header->frame_type,
|
|
.current_frame_id = frame_header->current_frame_id,
|
|
.order_hint = frame_header->order_hint,
|
|
.primary_ref_frame = frame_header->primary_ref_frame,
|
|
.frame_width_minus_1 = frame_header->frame_width_minus_1,
|
|
.frame_height_minus_1 = frame_header->frame_height_minus_1,
|
|
.coded_denom = frame_header->coded_denom,
|
|
.render_width_minus_1 = frame_header->render_width_minus_1,
|
|
.render_height_minus_1 = frame_header->render_height_minus_1,
|
|
.refresh_frame_flags = frame_header->refresh_frame_flags,
|
|
.interpolation_filter = frame_header->interpolation_filter,
|
|
.tx_mode = frame_header->tx_mode,
|
|
.tiling = (StdVideoAV1MESATileInfo) {
|
|
.flags = (StdVideoAV1MESATileInfoFlags) {
|
|
.uniform_tile_spacing_flag = frame_header->uniform_tile_spacing_flag,
|
|
},
|
|
.tile_cols = frame_header->tile_cols,
|
|
.tile_rows = frame_header->tile_rows,
|
|
.context_update_tile_id = frame_header->context_update_tile_id,
|
|
.tile_size_bytes_minus1 = frame_header->tile_size_bytes_minus1,
|
|
},
|
|
.quantization = (StdVideoAV1MESAQuantization) {
|
|
.flags.using_qmatrix = frame_header->using_qmatrix,
|
|
.base_q_idx = frame_header->base_q_idx,
|
|
.delta_q_y_dc = frame_header->delta_q_y_dc,
|
|
.diff_uv_delta = frame_header->diff_uv_delta,
|
|
.delta_q_u_dc = frame_header->delta_q_u_dc,
|
|
.delta_q_u_ac = frame_header->delta_q_u_ac,
|
|
.delta_q_v_dc = frame_header->delta_q_v_dc,
|
|
.delta_q_v_ac = frame_header->delta_q_v_ac,
|
|
.qm_y = frame_header->qm_y,
|
|
.qm_u = frame_header->qm_u,
|
|
.qm_v = frame_header->qm_v,
|
|
},
|
|
.delta_q = (StdVideoAV1MESADeltaQ) {
|
|
.flags = (StdVideoAV1MESADeltaQFlags) {
|
|
.delta_lf_present = frame_header->delta_lf_present,
|
|
.delta_lf_multi = frame_header->delta_lf_multi,
|
|
},
|
|
.delta_q_res = frame_header->delta_q_res,
|
|
.delta_lf_res = frame_header->delta_lf_res,
|
|
},
|
|
.loop_filter = (StdVideoAV1MESALoopFilter) {
|
|
.flags = (StdVideoAV1MESALoopFilterFlags) {
|
|
.delta_enabled = frame_header->loop_filter_delta_enabled,
|
|
.delta_update = frame_header->loop_filter_delta_update,
|
|
},
|
|
.level = {
|
|
frame_header->loop_filter_level[0], frame_header->loop_filter_level[1],
|
|
frame_header->loop_filter_level[2], frame_header->loop_filter_level[3],
|
|
},
|
|
.sharpness = frame_header->loop_filter_sharpness,
|
|
.mode_deltas = {
|
|
frame_header->loop_filter_mode_deltas[0], frame_header->loop_filter_mode_deltas[1],
|
|
},
|
|
},
|
|
.cdef = (StdVideoAV1MESACDEF) {
|
|
.damping_minus_3 = frame_header->cdef_damping_minus_3,
|
|
.bits = frame_header->cdef_bits,
|
|
},
|
|
.lr = (StdVideoAV1MESALoopRestoration) {
|
|
.lr_unit_shift = frame_header->lr_unit_shift,
|
|
.lr_uv_shift = frame_header->lr_uv_shift,
|
|
.lr_type = { frame_header->lr_type[0], frame_header->lr_type[1], frame_header->lr_type[2] },
|
|
},
|
|
.segmentation = (StdVideoAV1MESASegmentation) {
|
|
.flags = (StdVideoAV1MESASegmentationFlags) {
|
|
.enabled = frame_header->segmentation_enabled,
|
|
.update_map = frame_header->segmentation_update_map,
|
|
.temporal_update = frame_header->segmentation_temporal_update,
|
|
.update_data = frame_header->segmentation_update_data,
|
|
},
|
|
},
|
|
.film_grain = (StdVideoAV1MESAFilmGrainParameters) {
|
|
.flags = (StdVideoAV1MESAFilmGrainFlags) {
|
|
.apply_grain = apply_grain,
|
|
.chroma_scaling_from_luma = film_grain->chroma_scaling_from_luma,
|
|
.overlap_flag = film_grain->overlap_flag,
|
|
.clip_to_restricted_range = film_grain->clip_to_restricted_range,
|
|
},
|
|
.grain_scaling_minus_8 = film_grain->grain_scaling_minus_8,
|
|
.ar_coeff_lag = film_grain->ar_coeff_lag,
|
|
.ar_coeff_shift_minus_6 = film_grain->ar_coeff_shift_minus_6,
|
|
.grain_scale_shift = film_grain->grain_scale_shift,
|
|
.grain_seed = film_grain->grain_seed,
|
|
.num_y_points = film_grain->num_y_points,
|
|
.num_cb_points = film_grain->num_cb_points,
|
|
.num_cr_points = film_grain->num_cr_points,
|
|
.cb_mult = film_grain->cb_mult,
|
|
.cb_luma_mult = film_grain->cb_luma_mult,
|
|
.cb_offset = film_grain->cb_offset,
|
|
.cr_mult = film_grain->cr_mult,
|
|
.cr_luma_mult = film_grain->cr_luma_mult,
|
|
.cr_offset = film_grain->cr_offset,
|
|
},
|
|
};
|
|
|
|
for (int i = 0; i < 64; i++) {
|
|
ap->av1_frame_header.tiling.width_in_sbs_minus_1[i] = frame_header->width_in_sbs_minus_1[i];
|
|
ap->av1_frame_header.tiling.height_in_sbs_minus_1[i] = frame_header->height_in_sbs_minus_1[i];
|
|
ap->av1_frame_header.tiling.tile_start_col_sb[i] = frame_header->tile_start_col_sb[i];
|
|
ap->av1_frame_header.tiling.tile_start_row_sb[i] = frame_header->tile_start_row_sb[i];
|
|
}
|
|
|
|
for (int i = 0; i < 8; i++) {
|
|
ap->av1_frame_header.segmentation.feature_enabled_bits[i] = 0;
|
|
for (int j = 0; j < 8; j++) {
|
|
ap->av1_frame_header.segmentation.feature_enabled_bits[i] |= (frame_header->feature_enabled[i][j] << j);
|
|
ap->av1_frame_header.segmentation.feature_data[i][j] = frame_header->feature_value[i][j];
|
|
}
|
|
|
|
ap->av1_frame_header.loop_filter.ref_deltas[i] = frame_header->loop_filter_ref_deltas[i];
|
|
|
|
ap->av1_frame_header.cdef.y_pri_strength[i] = frame_header->cdef_y_pri_strength[i];
|
|
ap->av1_frame_header.cdef.y_sec_strength[i] = frame_header->cdef_y_sec_strength[i];
|
|
ap->av1_frame_header.cdef.uv_pri_strength[i] = frame_header->cdef_uv_pri_strength[i];
|
|
ap->av1_frame_header.cdef.uv_sec_strength[i] = frame_header->cdef_uv_sec_strength[i];
|
|
|
|
ap->av1_frame_header.ref_order_hint[i] = frame_header->ref_order_hint[i];
|
|
ap->av1_frame_header.global_motion[i] = (StdVideoAV1MESAGlobalMotion) {
|
|
.flags = (StdVideoAV1MESAGlobalMotionFlags) {
|
|
.gm_invalid = s->cur_frame.gm_invalid[i],
|
|
},
|
|
.gm_type = s->cur_frame.gm_type[i],
|
|
.gm_params = {
|
|
s->cur_frame.gm_params[i][0], s->cur_frame.gm_params[i][1],
|
|
s->cur_frame.gm_params[i][2], s->cur_frame.gm_params[i][3],
|
|
s->cur_frame.gm_params[i][4], s->cur_frame.gm_params[i][5],
|
|
},
|
|
};
|
|
}
|
|
|
|
for (int i = 0; i < 7; i++) {
|
|
ap->av1_frame_header.ref_frame_idx[i] = frame_header->ref_frame_idx[i];
|
|
ap->av1_frame_header.delta_frame_id_minus1[i] = frame_header->delta_frame_id_minus1[i];
|
|
}
|
|
|
|
ap->av1_pic_info.skip_mode_frame_idx[0] = s->cur_frame.skip_mode_frame_idx[0];
|
|
ap->av1_pic_info.skip_mode_frame_idx[1] = s->cur_frame.skip_mode_frame_idx[1];
|
|
|
|
if (apply_grain) {
|
|
for (int i = 0; i < 14; i++) {
|
|
ap->av1_frame_header.film_grain.point_y_value[i] = film_grain->point_y_value[i];
|
|
ap->av1_frame_header.film_grain.point_y_scaling[i] = film_grain->point_y_scaling[i];
|
|
}
|
|
|
|
for (int i = 0; i < 10; i++) {
|
|
ap->av1_frame_header.film_grain.point_cb_value[i] = film_grain->point_cb_value[i];
|
|
ap->av1_frame_header.film_grain.point_cb_scaling[i] = film_grain->point_cb_scaling[i];
|
|
ap->av1_frame_header.film_grain.point_cr_value[i] = film_grain->point_cr_value[i];
|
|
ap->av1_frame_header.film_grain.point_cr_scaling[i] = film_grain->point_cr_scaling[i];
|
|
}
|
|
|
|
for (int i = 0; i < 24; i++) {
|
|
ap->av1_frame_header.film_grain.ar_coeffs_y_plus_128[i] = film_grain->ar_coeffs_y_plus_128[i];
|
|
ap->av1_frame_header.film_grain.ar_coeffs_cb_plus_128[i] = film_grain->ar_coeffs_cb_plus_128[i];
|
|
ap->av1_frame_header.film_grain.ar_coeffs_cr_plus_128[i] = film_grain->ar_coeffs_cr_plus_128[i];
|
|
}
|
|
|
|
ap->av1_frame_header.film_grain.ar_coeffs_cb_plus_128[24] = film_grain->ar_coeffs_cb_plus_128[24];
|
|
ap->av1_frame_header.film_grain.ar_coeffs_cr_plus_128[24] = film_grain->ar_coeffs_cr_plus_128[24];
|
|
}
|
|
|
|
/* Workaround for a spec issue. */
|
|
ap->dec = dec;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int vk_av1_decode_slice(AVCodecContext *avctx,
|
|
const uint8_t *data,
|
|
uint32_t size)
|
|
{
|
|
int err;
|
|
const AV1DecContext *s = avctx->priv_data;
|
|
AV1VulkanDecodePicture *ap = s->cur_frame.hwaccel_picture_private;
|
|
FFVulkanDecodePicture *vp = &ap->vp;
|
|
|
|
for (int i = s->tg_start; i <= s->tg_end; i++) {
|
|
ap->tiles[ap->tile_list.nb_tiles] = (StdVideoAV1MESATile) {
|
|
.size = s->tile_group_info[i].tile_size,
|
|
.offset = s->tile_group_info[i].tile_offset,
|
|
.row = s->tile_group_info[i].tile_row,
|
|
.column = s->tile_group_info[i].tile_column,
|
|
.tg_start = s->tg_start,
|
|
.tg_end = s->tg_end,
|
|
};
|
|
|
|
err = ff_vk_decode_add_slice(avctx, vp,
|
|
data + s->tile_group_info[i].tile_offset,
|
|
s->tile_group_info[i].tile_size, 0,
|
|
&ap->tile_list.nb_tiles,
|
|
&ap->tile_offsets);
|
|
if (err < 0)
|
|
return err;
|
|
|
|
ap->tiles[ap->tile_list.nb_tiles - 1].offset = ap->tile_offsets[ap->tile_list.nb_tiles - 1];
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int vk_av1_end_frame(AVCodecContext *avctx)
|
|
{
|
|
const AV1DecContext *s = avctx->priv_data;
|
|
FFVulkanDecodeContext *dec = avctx->internal->hwaccel_priv_data;
|
|
const AV1Frame *pic = &s->cur_frame;
|
|
AV1VulkanDecodePicture *ap = pic->hwaccel_picture_private;
|
|
FFVulkanDecodePicture *vp = &ap->vp;
|
|
FFVulkanDecodePicture *rvp[AV1_NUM_REF_FRAMES] = { 0 };
|
|
AVFrame *rav[AV1_NUM_REF_FRAMES] = { 0 };
|
|
|
|
if (!ap->tile_list.nb_tiles)
|
|
return 0;
|
|
|
|
if (!dec->session_params) {
|
|
int err = vk_av1_create_params(avctx, &dec->session_params);
|
|
if (err < 0)
|
|
return err;
|
|
}
|
|
|
|
for (int i = 0; i < vp->decode_info.referenceSlotCount; i++) {
|
|
const AV1Frame *rp = ap->ref_src[i];
|
|
AV1VulkanDecodePicture *rhp = rp->hwaccel_picture_private;
|
|
|
|
rvp[i] = &rhp->vp;
|
|
rav[i] = ap->ref_src[i]->f;
|
|
}
|
|
|
|
av_log(avctx, AV_LOG_VERBOSE, "Decoding frame, %"SIZE_SPECIFIER" bytes, %i tiles\n",
|
|
vp->slices_size, ap->tile_list.nb_tiles);
|
|
|
|
return ff_vk_decode_frame(avctx, pic->f, vp, rav, rvp);
|
|
}
|
|
|
|
static void vk_av1_free_frame_priv(FFRefStructOpaque _hwctx, void *data)
|
|
{
|
|
AVHWDeviceContext *hwctx = _hwctx.nc;
|
|
AV1VulkanDecodePicture *ap = data;
|
|
|
|
/* Workaround for a spec issue. */
|
|
if (ap->frame_id_set)
|
|
ap->dec->frame_id_alloc_mask &= ~(1 << ap->frame_id);
|
|
|
|
/* Free frame resources, this also destroys the session parameters. */
|
|
ff_vk_decode_free_frame(hwctx, &ap->vp);
|
|
}
|
|
|
|
const FFHWAccel ff_av1_vulkan_hwaccel = {
|
|
.p.name = "av1_vulkan",
|
|
.p.type = AVMEDIA_TYPE_VIDEO,
|
|
.p.id = AV_CODEC_ID_AV1,
|
|
.p.pix_fmt = AV_PIX_FMT_VULKAN,
|
|
.start_frame = &vk_av1_start_frame,
|
|
.decode_slice = &vk_av1_decode_slice,
|
|
.end_frame = &vk_av1_end_frame,
|
|
.free_frame_priv = &vk_av1_free_frame_priv,
|
|
.frame_priv_data_size = sizeof(AV1VulkanDecodePicture),
|
|
.init = &ff_vk_decode_init,
|
|
.update_thread_context = &ff_vk_update_thread_context,
|
|
.decode_params = &ff_vk_params_invalidate,
|
|
.flush = &ff_vk_decode_flush,
|
|
.uninit = &ff_vk_decode_uninit,
|
|
.frame_params = &ff_vk_frame_params,
|
|
.priv_data_size = sizeof(FFVulkanDecodeContext),
|
|
|
|
/* NOTE: Threading is intentionally disabled here. Due to the design of Vulkan,
|
|
* where frames are opaque to users, and mostly opaque for driver developers,
|
|
* there's an issue with current hardware accelerator implementations of AV1,
|
|
* where they require an internal index. With regular hwaccel APIs, this index
|
|
* is given to users as an opaque handle directly. With Vulkan, due to increased
|
|
* flexibility, this index cannot be present anywhere.
|
|
* The current implementation tracks the index for the driver and submits it
|
|
* as necessary information. Due to needing to modify the decoding context,
|
|
* which is not thread-safe, on frame free, threading is disabled.
|
|
* In the future, once this is fixed in the spec, the workarounds may be removed
|
|
* and threading enabled. */
|
|
.caps_internal = HWACCEL_CAP_ASYNC_SAFE,
|
|
};
|