ffmpeg/libavcodec/libaomenc.c
Andreas Rheinhardt 790f793844 avutil/common: Don't auto-include mem.h
There are lots of files that don't need it: The number of object
files that actually need it went down from 2011 to 884 here.

Keep it for external users in order to not cause breakages.

Also improve the other headers a bit while just at it.

Signed-off-by: Andreas Rheinhardt <andreas.rheinhardt@outlook.com>
2024-03-31 00:08:43 +01:00

1576 lines
65 KiB
C

/*
* Copyright (c) 2010, Google, Inc.
*
* This file is part of FFmpeg.
*
* FFmpeg is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2.1 of the License, or (at your option) any later version.
*
* FFmpeg is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with FFmpeg; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
/**
* @file
* AV1 encoder support via libaom
*/
#include <limits.h>
#define AOM_DISABLE_CTRL_TYPECHECKS 1
#include <aom/aom_encoder.h>
#include <aom/aomcx.h>
#include "libavutil/avassert.h"
#include "libavutil/base64.h"
#include "libavutil/common.h"
#include "libavutil/cpu.h"
#include "libavutil/imgutils.h"
#include "libavutil/mathematics.h"
#include "libavutil/mem.h"
#include "libavutil/opt.h"
#include "libavutil/pixdesc.h"
#include "av1.h"
#include "avcodec.h"
#include "bsf.h"
#include "codec_internal.h"
#include "encode.h"
#include "internal.h"
#include "libaom.h"
#include "packet_internal.h"
#include "profiles.h"
/*
* Portion of struct aom_codec_cx_pkt from aom_encoder.h.
* One encoded frame returned from the library.
*/
struct FrameListData {
void *buf; /**< compressed data buffer */
size_t sz; /**< length of compressed data */
int64_t pts; /**< time stamp to show frame
(in timebase units) */
unsigned long duration; /**< duration to show frame
(in timebase units) */
uint32_t flags; /**< flags for this frame */
uint64_t sse[4];
int have_sse; /**< true if we have pending sse[] */
uint64_t frame_number;
struct FrameListData *next;
};
typedef struct AOMEncoderContext {
AVClass *class;
AVBSFContext *bsf;
struct aom_codec_ctx encoder;
struct aom_image rawimg;
struct aom_fixed_buf twopass_stats;
unsigned twopass_stats_size;
struct FrameListData *coded_frame_list;
int cpu_used;
int auto_alt_ref;
int arnr_max_frames;
int arnr_strength;
int aq_mode;
int lag_in_frames;
int error_resilient;
int crf;
int static_thresh;
int drop_threshold;
int denoise_noise_level;
int denoise_block_size;
uint64_t sse[4];
int have_sse; /**< true if we have pending sse[] */
uint64_t frame_number;
int rc_undershoot_pct;
int rc_overshoot_pct;
int minsection_pct;
int maxsection_pct;
int frame_parallel;
int tile_cols, tile_rows;
int tile_cols_log2, tile_rows_log2;
aom_superblock_size_t superblock_size;
int uniform_tiles;
int row_mt;
int enable_cdef;
int enable_global_motion;
int enable_intrabc;
int enable_restoration;
int usage;
int tune;
int still_picture;
int enable_rect_partitions;
int enable_1to4_partitions;
int enable_ab_partitions;
int enable_angle_delta;
int enable_cfl_intra;
int enable_paeth_intra;
int enable_smooth_intra;
int enable_intra_edge_filter;
int enable_palette;
int enable_filter_intra;
int enable_flip_idtx;
int enable_tx64;
int reduced_tx_type_set;
int use_intra_dct_only;
int use_inter_dct_only;
int use_intra_default_tx_only;
int enable_ref_frame_mvs;
int enable_interinter_wedge;
int enable_interintra_wedge;
int enable_interintra_comp;
int enable_masked_comp;
int enable_obmc;
int enable_onesided_comp;
int enable_reduced_reference_set;
int enable_smooth_interintra;
int enable_diff_wtd_comp;
int enable_dist_wtd_comp;
int enable_dual_filter;
AVDictionary *aom_params;
} AOMContext;
static const char *const ctlidstr[] = {
[AOME_SET_CPUUSED] = "AOME_SET_CPUUSED",
[AOME_SET_CQ_LEVEL] = "AOME_SET_CQ_LEVEL",
[AOME_SET_ENABLEAUTOALTREF] = "AOME_SET_ENABLEAUTOALTREF",
[AOME_SET_ARNR_MAXFRAMES] = "AOME_SET_ARNR_MAXFRAMES",
[AOME_SET_ARNR_STRENGTH] = "AOME_SET_ARNR_STRENGTH",
[AOME_SET_STATIC_THRESHOLD] = "AOME_SET_STATIC_THRESHOLD",
[AV1E_SET_COLOR_RANGE] = "AV1E_SET_COLOR_RANGE",
[AV1E_SET_COLOR_PRIMARIES] = "AV1E_SET_COLOR_PRIMARIES",
[AV1E_SET_MATRIX_COEFFICIENTS] = "AV1E_SET_MATRIX_COEFFICIENTS",
[AV1E_SET_TRANSFER_CHARACTERISTICS] = "AV1E_SET_TRANSFER_CHARACTERISTICS",
[AV1E_SET_AQ_MODE] = "AV1E_SET_AQ_MODE",
[AV1E_SET_FRAME_PARALLEL_DECODING] = "AV1E_SET_FRAME_PARALLEL_DECODING",
[AV1E_SET_SUPERBLOCK_SIZE] = "AV1E_SET_SUPERBLOCK_SIZE",
[AV1E_SET_TILE_COLUMNS] = "AV1E_SET_TILE_COLUMNS",
[AV1E_SET_TILE_ROWS] = "AV1E_SET_TILE_ROWS",
[AV1E_SET_ENABLE_RESTORATION] = "AV1E_SET_ENABLE_RESTORATION",
#ifdef AOM_CTRL_AV1E_SET_ROW_MT
[AV1E_SET_ROW_MT] = "AV1E_SET_ROW_MT",
#endif
#ifdef AOM_CTRL_AV1E_SET_DENOISE_NOISE_LEVEL
[AV1E_SET_DENOISE_NOISE_LEVEL] = "AV1E_SET_DENOISE_NOISE_LEVEL",
#endif
#ifdef AOM_CTRL_AV1E_SET_DENOISE_BLOCK_SIZE
[AV1E_SET_DENOISE_BLOCK_SIZE] = "AV1E_SET_DENOISE_BLOCK_SIZE",
#endif
#ifdef AOM_CTRL_AV1E_SET_MAX_REFERENCE_FRAMES
[AV1E_SET_MAX_REFERENCE_FRAMES] = "AV1E_SET_MAX_REFERENCE_FRAMES",
#endif
#ifdef AOM_CTRL_AV1E_SET_ENABLE_GLOBAL_MOTION
[AV1E_SET_ENABLE_GLOBAL_MOTION] = "AV1E_SET_ENABLE_GLOBAL_MOTION",
#endif
#ifdef AOM_CTRL_AV1E_SET_ENABLE_INTRABC
[AV1E_SET_ENABLE_INTRABC] = "AV1E_SET_ENABLE_INTRABC",
#endif
[AV1E_SET_ENABLE_CDEF] = "AV1E_SET_ENABLE_CDEF",
[AOME_SET_TUNING] = "AOME_SET_TUNING",
#if AOM_ENCODER_ABI_VERSION >= 22
[AV1E_SET_ENABLE_1TO4_PARTITIONS] = "AV1E_SET_ENABLE_1TO4_PARTITIONS",
[AV1E_SET_ENABLE_AB_PARTITIONS] = "AV1E_SET_ENABLE_AB_PARTITIONS",
[AV1E_SET_ENABLE_RECT_PARTITIONS] = "AV1E_SET_ENABLE_RECT_PARTITIONS",
[AV1E_SET_ENABLE_ANGLE_DELTA] = "AV1E_SET_ENABLE_ANGLE_DELTA",
[AV1E_SET_ENABLE_CFL_INTRA] = "AV1E_SET_ENABLE_CFL_INTRA",
[AV1E_SET_ENABLE_FILTER_INTRA] = "AV1E_SET_ENABLE_FILTER_INTRA",
[AV1E_SET_ENABLE_INTRA_EDGE_FILTER] = "AV1E_SET_ENABLE_INTRA_EDGE_FILTER",
[AV1E_SET_ENABLE_PAETH_INTRA] = "AV1E_SET_ENABLE_PAETH_INTRA",
[AV1E_SET_ENABLE_SMOOTH_INTRA] = "AV1E_SET_ENABLE_SMOOTH_INTRA",
[AV1E_SET_ENABLE_PALETTE] = "AV1E_SET_ENABLE_PALETTE",
[AV1E_SET_ENABLE_FLIP_IDTX] = "AV1E_SET_ENABLE_FLIP_IDTX",
[AV1E_SET_ENABLE_TX64] = "AV1E_SET_ENABLE_TX64",
[AV1E_SET_INTRA_DCT_ONLY] = "AV1E_SET_INTRA_DCT_ONLY",
[AV1E_SET_INTER_DCT_ONLY] = "AV1E_SET_INTER_DCT_ONLY",
[AV1E_SET_INTRA_DEFAULT_TX_ONLY] = "AV1E_SET_INTRA_DEFAULT_TX_ONLY",
[AV1E_SET_REDUCED_TX_TYPE_SET] = "AV1E_SET_REDUCED_TX_TYPE_SET",
[AV1E_SET_ENABLE_DIFF_WTD_COMP] = "AV1E_SET_ENABLE_DIFF_WTD_COMP",
[AV1E_SET_ENABLE_DIST_WTD_COMP] = "AV1E_SET_ENABLE_DIST_WTD_COMP",
[AV1E_SET_ENABLE_DUAL_FILTER] = "AV1E_SET_ENABLE_DUAL_FILTER",
[AV1E_SET_ENABLE_INTERINTER_WEDGE] = "AV1E_SET_ENABLE_INTERINTER_WEDGE",
[AV1E_SET_ENABLE_INTERINTRA_WEDGE] = "AV1E_SET_ENABLE_INTERINTRA_WEDGE",
[AV1E_SET_ENABLE_MASKED_COMP] = "AV1E_SET_ENABLE_MASKED_COMP",
[AV1E_SET_ENABLE_INTERINTRA_COMP] = "AV1E_SET_ENABLE_INTERINTRA_COMP",
[AV1E_SET_ENABLE_OBMC] = "AV1E_SET_ENABLE_OBMC",
[AV1E_SET_ENABLE_ONESIDED_COMP] = "AV1E_SET_ENABLE_ONESIDED_COMP",
[AV1E_SET_REDUCED_REFERENCE_SET] = "AV1E_SET_REDUCED_REFERENCE_SET",
[AV1E_SET_ENABLE_SMOOTH_INTERINTRA] = "AV1E_SET_ENABLE_SMOOTH_INTERINTRA",
[AV1E_SET_ENABLE_REF_FRAME_MVS] = "AV1E_SET_ENABLE_REF_FRAME_MVS",
#endif
#ifdef AOM_CTRL_AV1E_GET_NUM_OPERATING_POINTS
[AV1E_GET_NUM_OPERATING_POINTS] = "AV1E_GET_NUM_OPERATING_POINTS",
#endif
#ifdef AOM_CTRL_AV1E_GET_SEQ_LEVEL_IDX
[AV1E_GET_SEQ_LEVEL_IDX] = "AV1E_GET_SEQ_LEVEL_IDX",
#endif
#ifdef AOM_CTRL_AV1E_GET_TARGET_SEQ_LEVEL_IDX
[AV1E_GET_TARGET_SEQ_LEVEL_IDX] = "AV1E_GET_TARGET_SEQ_LEVEL_IDX",
#endif
[AV1_GET_NEW_FRAME_IMAGE] = "AV1_GET_NEW_FRAME_IMAGE",
};
static av_cold void log_encoder_error(AVCodecContext *avctx, const char *desc)
{
AOMContext *ctx = avctx->priv_data;
const char *error = aom_codec_error(&ctx->encoder);
const char *detail = aom_codec_error_detail(&ctx->encoder);
av_log(avctx, AV_LOG_ERROR, "%s: %s\n", desc, error);
if (detail)
av_log(avctx, AV_LOG_ERROR, " Additional information: %s\n", detail);
}
static av_cold void dump_enc_cfg(AVCodecContext *avctx,
const struct aom_codec_enc_cfg *cfg,
int level)
{
int width = -30;
av_log(avctx, level, "aom_codec_enc_cfg\n");
av_log(avctx, level, "generic settings\n"
" %*s%u\n %*s%u\n %*s%u\n %*s%u\n %*s%u\n"
" %*s%u\n %*s%u\n"
" %*s{%u/%u}\n %*s%u\n %*s%d\n %*s%u\n",
width, "g_usage:", cfg->g_usage,
width, "g_threads:", cfg->g_threads,
width, "g_profile:", cfg->g_profile,
width, "g_w:", cfg->g_w,
width, "g_h:", cfg->g_h,
width, "g_bit_depth:", cfg->g_bit_depth,
width, "g_input_bit_depth:", cfg->g_input_bit_depth,
width, "g_timebase:", cfg->g_timebase.num, cfg->g_timebase.den,
width, "g_error_resilient:", cfg->g_error_resilient,
width, "g_pass:", cfg->g_pass,
width, "g_lag_in_frames:", cfg->g_lag_in_frames);
av_log(avctx, level, "rate control settings\n"
" %*s%u\n %*s%d\n %*s%p(%"SIZE_SPECIFIER")\n %*s%u\n",
width, "rc_dropframe_thresh:", cfg->rc_dropframe_thresh,
width, "rc_end_usage:", cfg->rc_end_usage,
width, "rc_twopass_stats_in:", cfg->rc_twopass_stats_in.buf, cfg->rc_twopass_stats_in.sz,
width, "rc_target_bitrate:", cfg->rc_target_bitrate);
av_log(avctx, level, "quantizer settings\n"
" %*s%u\n %*s%u\n",
width, "rc_min_quantizer:", cfg->rc_min_quantizer,
width, "rc_max_quantizer:", cfg->rc_max_quantizer);
av_log(avctx, level, "bitrate tolerance\n"
" %*s%u\n %*s%u\n",
width, "rc_undershoot_pct:", cfg->rc_undershoot_pct,
width, "rc_overshoot_pct:", cfg->rc_overshoot_pct);
av_log(avctx, level, "decoder buffer model\n"
" %*s%u\n %*s%u\n %*s%u\n",
width, "rc_buf_sz:", cfg->rc_buf_sz,
width, "rc_buf_initial_sz:", cfg->rc_buf_initial_sz,
width, "rc_buf_optimal_sz:", cfg->rc_buf_optimal_sz);
av_log(avctx, level, "2 pass rate control settings\n"
" %*s%u\n %*s%u\n %*s%u\n",
width, "rc_2pass_vbr_bias_pct:", cfg->rc_2pass_vbr_bias_pct,
width, "rc_2pass_vbr_minsection_pct:", cfg->rc_2pass_vbr_minsection_pct,
width, "rc_2pass_vbr_maxsection_pct:", cfg->rc_2pass_vbr_maxsection_pct);
av_log(avctx, level, "keyframing settings\n"
" %*s%d\n %*s%u\n %*s%u\n",
width, "kf_mode:", cfg->kf_mode,
width, "kf_min_dist:", cfg->kf_min_dist,
width, "kf_max_dist:", cfg->kf_max_dist);
av_log(avctx, level, "tile settings\n"
" %*s%d\n %*s%d\n",
width, "tile_width_count:", cfg->tile_width_count,
width, "tile_height_count:", cfg->tile_height_count);
av_log(avctx, level, "\n");
}
static void coded_frame_add(void *list, struct FrameListData *cx_frame)
{
struct FrameListData **p = list;
while (*p)
p = &(*p)->next;
*p = cx_frame;
cx_frame->next = NULL;
}
static av_cold void free_coded_frame(struct FrameListData *cx_frame)
{
av_freep(&cx_frame->buf);
av_freep(&cx_frame);
}
static av_cold void free_frame_list(struct FrameListData *list)
{
struct FrameListData *p = list;
while (p) {
list = list->next;
free_coded_frame(p);
p = list;
}
}
static av_cold int codecctl_int(AVCodecContext *avctx,
#ifdef UENUM1BYTE
aome_enc_control_id id,
#else
enum aome_enc_control_id id,
#endif
int val)
{
AOMContext *ctx = avctx->priv_data;
char buf[80];
int width = -30;
int res;
snprintf(buf, sizeof(buf), "%s:", ctlidstr[id]);
av_log(avctx, AV_LOG_DEBUG, " %*s%d\n", width, buf, val);
res = aom_codec_control(&ctx->encoder, id, val);
if (res != AOM_CODEC_OK) {
snprintf(buf, sizeof(buf), "Failed to set %s codec control",
ctlidstr[id]);
log_encoder_error(avctx, buf);
return AVERROR(EINVAL);
}
return 0;
}
#if defined(AOM_CTRL_AV1E_GET_NUM_OPERATING_POINTS) && \
defined(AOM_CTRL_AV1E_GET_SEQ_LEVEL_IDX) && \
defined(AOM_CTRL_AV1E_GET_TARGET_SEQ_LEVEL_IDX)
static av_cold int codecctl_intp(AVCodecContext *avctx,
#ifdef UENUM1BYTE
aome_enc_control_id id,
#else
enum aome_enc_control_id id,
#endif
int* ptr)
{
AOMContext *ctx = avctx->priv_data;
char buf[80];
int width = -30;
int res;
snprintf(buf, sizeof(buf), "%s:", ctlidstr[id]);
av_log(avctx, AV_LOG_DEBUG, " %*s%d\n", width, buf, *ptr);
res = aom_codec_control(&ctx->encoder, id, ptr);
if (res != AOM_CODEC_OK) {
snprintf(buf, sizeof(buf), "Failed to set %s codec control",
ctlidstr[id]);
log_encoder_error(avctx, buf);
return AVERROR(EINVAL);
}
return 0;
}
#endif
static av_cold int codecctl_imgp(AVCodecContext *avctx,
#ifdef UENUM1BYTE
aome_enc_control_id id,
#else
enum aome_enc_control_id id,
#endif
struct aom_image *img)
{
AOMContext *ctx = avctx->priv_data;
char buf[80];
int res;
snprintf(buf, sizeof(buf), "%s:", ctlidstr[id]);
res = aom_codec_control(&ctx->encoder, id, img);
if (res != AOM_CODEC_OK) {
snprintf(buf, sizeof(buf), "Failed to get %s codec control",
ctlidstr[id]);
log_encoder_error(avctx, buf);
return AVERROR(EINVAL);
}
return 0;
}
static av_cold int aom_free(AVCodecContext *avctx)
{
AOMContext *ctx = avctx->priv_data;
#if defined(AOM_CTRL_AV1E_GET_NUM_OPERATING_POINTS) && \
defined(AOM_CTRL_AV1E_GET_SEQ_LEVEL_IDX) && \
defined(AOM_CTRL_AV1E_GET_TARGET_SEQ_LEVEL_IDX)
if (ctx->encoder.iface && !(avctx->flags & AV_CODEC_FLAG_PASS1)) {
int num_operating_points;
int levels[32];
int target_levels[32];
if (!codecctl_intp(avctx, AV1E_GET_NUM_OPERATING_POINTS,
&num_operating_points) &&
!codecctl_intp(avctx, AV1E_GET_SEQ_LEVEL_IDX, levels) &&
!codecctl_intp(avctx, AV1E_GET_TARGET_SEQ_LEVEL_IDX,
target_levels)) {
for (int i = 0; i < num_operating_points; i++) {
if (levels[i] > target_levels[i]) {
// Warn when the target level was not met
av_log(avctx, AV_LOG_WARNING,
"Could not encode to target level %d.%d for "
"operating point %d. The output level is %d.%d.\n",
2 + (target_levels[i] >> 2), target_levels[i] & 3,
i, 2 + (levels[i] >> 2), levels[i] & 3);
} else if (target_levels[i] < 31) {
// Log the encoded level if a target level was given
av_log(avctx, AV_LOG_INFO,
"Output level for operating point %d is %d.%d.\n",
i, 2 + (levels[i] >> 2), levels[i] & 3);
}
}
}
}
#endif
aom_codec_destroy(&ctx->encoder);
av_freep(&ctx->twopass_stats.buf);
av_freep(&avctx->stats_out);
free_frame_list(ctx->coded_frame_list);
av_bsf_free(&ctx->bsf);
return 0;
}
static int set_pix_fmt(AVCodecContext *avctx, aom_codec_caps_t codec_caps,
struct aom_codec_enc_cfg *enccfg, aom_codec_flags_t *flags,
aom_img_fmt_t *img_fmt)
{
AOMContext av_unused *ctx = avctx->priv_data;
const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(avctx->pix_fmt);
enccfg->g_bit_depth = enccfg->g_input_bit_depth = desc->comp[0].depth;
switch (avctx->pix_fmt) {
case AV_PIX_FMT_GRAY8:
enccfg->monochrome = 1;
/* Fall-through */
case AV_PIX_FMT_YUV420P:
enccfg->g_profile = AV_PROFILE_AV1_MAIN;
*img_fmt = AOM_IMG_FMT_I420;
return 0;
case AV_PIX_FMT_YUV422P:
enccfg->g_profile = AV_PROFILE_AV1_PROFESSIONAL;
*img_fmt = AOM_IMG_FMT_I422;
return 0;
case AV_PIX_FMT_YUV444P:
case AV_PIX_FMT_GBRP:
enccfg->g_profile = AV_PROFILE_AV1_HIGH;
*img_fmt = AOM_IMG_FMT_I444;
return 0;
case AV_PIX_FMT_GRAY10:
case AV_PIX_FMT_GRAY12:
enccfg->monochrome = 1;
/* Fall-through */
case AV_PIX_FMT_YUV420P10:
case AV_PIX_FMT_YUV420P12:
if (codec_caps & AOM_CODEC_CAP_HIGHBITDEPTH) {
enccfg->g_profile =
enccfg->g_bit_depth == 10 ? AV_PROFILE_AV1_MAIN : AV_PROFILE_AV1_PROFESSIONAL;
*img_fmt = AOM_IMG_FMT_I42016;
*flags |= AOM_CODEC_USE_HIGHBITDEPTH;
return 0;
}
break;
case AV_PIX_FMT_YUV422P10:
case AV_PIX_FMT_YUV422P12:
if (codec_caps & AOM_CODEC_CAP_HIGHBITDEPTH) {
enccfg->g_profile = AV_PROFILE_AV1_PROFESSIONAL;
*img_fmt = AOM_IMG_FMT_I42216;
*flags |= AOM_CODEC_USE_HIGHBITDEPTH;
return 0;
}
break;
case AV_PIX_FMT_YUV444P10:
case AV_PIX_FMT_YUV444P12:
case AV_PIX_FMT_GBRP10:
case AV_PIX_FMT_GBRP12:
if (codec_caps & AOM_CODEC_CAP_HIGHBITDEPTH) {
enccfg->g_profile =
enccfg->g_bit_depth == 10 ? AV_PROFILE_AV1_HIGH : AV_PROFILE_AV1_PROFESSIONAL;
*img_fmt = AOM_IMG_FMT_I44416;
*flags |= AOM_CODEC_USE_HIGHBITDEPTH;
return 0;
}
break;
default:
break;
}
av_log(avctx, AV_LOG_ERROR, "Unsupported pixel format.\n");
return AVERROR_INVALIDDATA;
}
static void set_color_range(AVCodecContext *avctx)
{
aom_color_range_t aom_cr;
switch (avctx->color_range) {
case AVCOL_RANGE_UNSPECIFIED:
case AVCOL_RANGE_MPEG: aom_cr = AOM_CR_STUDIO_RANGE; break;
case AVCOL_RANGE_JPEG: aom_cr = AOM_CR_FULL_RANGE; break;
default:
av_log(avctx, AV_LOG_WARNING, "Unsupported color range (%d)\n",
avctx->color_range);
return;
}
codecctl_int(avctx, AV1E_SET_COLOR_RANGE, aom_cr);
}
static int count_uniform_tiling(int dim, int sb_size, int tiles_log2)
{
int sb_dim = (dim + sb_size - 1) / sb_size;
int tile_dim = (sb_dim + (1 << tiles_log2) - 1) >> tiles_log2;
av_assert0(tile_dim > 0);
return (sb_dim + tile_dim - 1) / tile_dim;
}
static int choose_tiling(AVCodecContext *avctx,
struct aom_codec_enc_cfg *enccfg)
{
AOMContext *ctx = avctx->priv_data;
int sb_128x128_possible, sb_size, sb_width, sb_height;
int uniform_rows, uniform_cols;
int uniform_64x64_possible, uniform_128x128_possible;
int tile_size, rounding, i;
if (ctx->tile_cols_log2 >= 0)
ctx->tile_cols = 1 << ctx->tile_cols_log2;
if (ctx->tile_rows_log2 >= 0)
ctx->tile_rows = 1 << ctx->tile_rows_log2;
if (ctx->tile_cols == 0) {
ctx->tile_cols = (avctx->width + AV1_MAX_TILE_WIDTH - 1) /
AV1_MAX_TILE_WIDTH;
if (ctx->tile_cols > 1) {
av_log(avctx, AV_LOG_DEBUG, "Automatically using %d tile "
"columns to fill width.\n", ctx->tile_cols);
}
}
av_assert0(ctx->tile_cols > 0);
if (ctx->tile_rows == 0) {
int max_tile_width =
FFALIGN((FFALIGN(avctx->width, 128) +
ctx->tile_cols - 1) / ctx->tile_cols, 128);
ctx->tile_rows =
(max_tile_width * FFALIGN(avctx->height, 128) +
AV1_MAX_TILE_AREA - 1) / AV1_MAX_TILE_AREA;
if (ctx->tile_rows > 1) {
av_log(avctx, AV_LOG_DEBUG, "Automatically using %d tile "
"rows to fill area.\n", ctx->tile_rows);
}
}
av_assert0(ctx->tile_rows > 0);
if ((avctx->width + 63) / 64 < ctx->tile_cols ||
(avctx->height + 63) / 64 < ctx->tile_rows) {
av_log(avctx, AV_LOG_ERROR, "Invalid tile sizing: frame not "
"large enough to fit specified tile arrangement.\n");
return AVERROR(EINVAL);
}
if (ctx->tile_cols > AV1_MAX_TILE_COLS ||
ctx->tile_rows > AV1_MAX_TILE_ROWS) {
av_log(avctx, AV_LOG_ERROR, "Invalid tile sizing: AV1 does "
"not allow more than %dx%d tiles.\n",
AV1_MAX_TILE_COLS, AV1_MAX_TILE_ROWS);
return AVERROR(EINVAL);
}
if (avctx->width / ctx->tile_cols > AV1_MAX_TILE_WIDTH) {
av_log(avctx, AV_LOG_ERROR, "Invalid tile sizing: AV1 does "
"not allow tiles of width greater than %d.\n",
AV1_MAX_TILE_WIDTH);
return AVERROR(EINVAL);
}
ctx->superblock_size = AOM_SUPERBLOCK_SIZE_DYNAMIC;
if (ctx->tile_cols == 1 && ctx->tile_rows == 1) {
av_log(avctx, AV_LOG_DEBUG, "Using a single tile.\n");
return 0;
}
sb_128x128_possible =
(avctx->width + 127) / 128 >= ctx->tile_cols &&
(avctx->height + 127) / 128 >= ctx->tile_rows;
ctx->tile_cols_log2 = ctx->tile_cols == 1 ? 0 :
av_log2(ctx->tile_cols - 1) + 1;
ctx->tile_rows_log2 = ctx->tile_rows == 1 ? 0 :
av_log2(ctx->tile_rows - 1) + 1;
uniform_cols = count_uniform_tiling(avctx->width,
64, ctx->tile_cols_log2);
uniform_rows = count_uniform_tiling(avctx->height,
64, ctx->tile_rows_log2);
av_log(avctx, AV_LOG_DEBUG, "Uniform with 64x64 superblocks "
"-> %dx%d tiles.\n", uniform_cols, uniform_rows);
uniform_64x64_possible = uniform_cols == ctx->tile_cols &&
uniform_rows == ctx->tile_rows;
if (sb_128x128_possible) {
uniform_cols = count_uniform_tiling(avctx->width,
128, ctx->tile_cols_log2);
uniform_rows = count_uniform_tiling(avctx->height,
128, ctx->tile_rows_log2);
av_log(avctx, AV_LOG_DEBUG, "Uniform with 128x128 superblocks "
"-> %dx%d tiles.\n", uniform_cols, uniform_rows);
uniform_128x128_possible = uniform_cols == ctx->tile_cols &&
uniform_rows == ctx->tile_rows;
} else {
av_log(avctx, AV_LOG_DEBUG, "128x128 superblocks not possible.\n");
uniform_128x128_possible = 0;
}
ctx->uniform_tiles = 1;
if (uniform_64x64_possible && uniform_128x128_possible) {
av_log(avctx, AV_LOG_DEBUG, "Using uniform tiling with dynamic "
"superblocks (tile_cols_log2 = %d, tile_rows_log2 = %d).\n",
ctx->tile_cols_log2, ctx->tile_rows_log2);
return 0;
}
if (uniform_64x64_possible && !sb_128x128_possible) {
av_log(avctx, AV_LOG_DEBUG, "Using uniform tiling with 64x64 "
"superblocks (tile_cols_log2 = %d, tile_rows_log2 = %d).\n",
ctx->tile_cols_log2, ctx->tile_rows_log2);
ctx->superblock_size = AOM_SUPERBLOCK_SIZE_64X64;
return 0;
}
if (uniform_128x128_possible) {
av_log(avctx, AV_LOG_DEBUG, "Using uniform tiling with 128x128 "
"superblocks (tile_cols_log2 = %d, tile_rows_log2 = %d).\n",
ctx->tile_cols_log2, ctx->tile_rows_log2);
ctx->superblock_size = AOM_SUPERBLOCK_SIZE_128X128;
return 0;
}
ctx->uniform_tiles = 0;
if (sb_128x128_possible) {
sb_size = 128;
ctx->superblock_size = AOM_SUPERBLOCK_SIZE_128X128;
} else {
sb_size = 64;
ctx->superblock_size = AOM_SUPERBLOCK_SIZE_64X64;
}
av_log(avctx, AV_LOG_DEBUG, "Using fixed tiling with %dx%d "
"superblocks (tile_cols = %d, tile_rows = %d).\n",
sb_size, sb_size, ctx->tile_cols, ctx->tile_rows);
enccfg->tile_width_count = ctx->tile_cols;
enccfg->tile_height_count = ctx->tile_rows;
sb_width = (avctx->width + sb_size - 1) / sb_size;
sb_height = (avctx->height + sb_size - 1) / sb_size;
tile_size = sb_width / ctx->tile_cols;
rounding = sb_width % ctx->tile_cols;
for (i = 0; i < ctx->tile_cols; i++) {
enccfg->tile_widths[i] = tile_size +
(i < rounding / 2 ||
i > ctx->tile_cols - 1 - (rounding + 1) / 2);
}
tile_size = sb_height / ctx->tile_rows;
rounding = sb_height % ctx->tile_rows;
for (i = 0; i < ctx->tile_rows; i++) {
enccfg->tile_heights[i] = tile_size +
(i < rounding / 2 ||
i > ctx->tile_rows - 1 - (rounding + 1) / 2);
}
return 0;
}
static av_cold int aom_init(AVCodecContext *avctx,
const struct aom_codec_iface *iface)
{
AOMContext *ctx = avctx->priv_data;
const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(avctx->pix_fmt);
struct aom_codec_enc_cfg enccfg = { 0 };
#ifdef AOM_FRAME_IS_INTRAONLY
aom_codec_flags_t flags =
(avctx->flags & AV_CODEC_FLAG_PSNR) ? AOM_CODEC_USE_PSNR : 0;
#else
aom_codec_flags_t flags = 0;
#endif
AVCPBProperties *cpb_props;
int res;
aom_img_fmt_t img_fmt;
aom_codec_caps_t codec_caps = aom_codec_get_caps(iface);
av_log(avctx, AV_LOG_INFO, "%s\n", aom_codec_version_str());
av_log(avctx, AV_LOG_VERBOSE, "%s\n", aom_codec_build_config());
if ((res = aom_codec_enc_config_default(iface, &enccfg, ctx->usage)) != AOM_CODEC_OK) {
av_log(avctx, AV_LOG_ERROR, "Failed to get config: %s\n",
aom_codec_err_to_string(res));
return AVERROR(EINVAL);
}
if (set_pix_fmt(avctx, codec_caps, &enccfg, &flags, &img_fmt))
return AVERROR(EINVAL);
if(!avctx->bit_rate)
if(avctx->rc_max_rate || avctx->rc_buffer_size || avctx->rc_initial_buffer_occupancy) {
av_log( avctx, AV_LOG_ERROR, "Rate control parameters set without a bitrate\n");
return AVERROR(EINVAL);
}
dump_enc_cfg(avctx, &enccfg, AV_LOG_DEBUG);
enccfg.g_w = avctx->width;
enccfg.g_h = avctx->height;
enccfg.g_timebase.num = avctx->time_base.num;
enccfg.g_timebase.den = avctx->time_base.den;
enccfg.g_threads =
FFMIN(avctx->thread_count ? avctx->thread_count : av_cpu_count(), 64);
if (ctx->lag_in_frames >= 0)
enccfg.g_lag_in_frames = ctx->lag_in_frames;
if (avctx->flags & AV_CODEC_FLAG_PASS1)
enccfg.g_pass = AOM_RC_FIRST_PASS;
else if (avctx->flags & AV_CODEC_FLAG_PASS2)
enccfg.g_pass = AOM_RC_LAST_PASS;
else
enccfg.g_pass = AOM_RC_ONE_PASS;
if (avctx->rc_min_rate == avctx->rc_max_rate &&
avctx->rc_min_rate == avctx->bit_rate && avctx->bit_rate) {
enccfg.rc_end_usage = AOM_CBR;
} else if (ctx->crf >= 0) {
enccfg.rc_end_usage = AOM_CQ;
if (!avctx->bit_rate)
enccfg.rc_end_usage = AOM_Q;
}
if (avctx->bit_rate) {
enccfg.rc_target_bitrate = av_rescale_rnd(avctx->bit_rate, 1, 1000,
AV_ROUND_NEAR_INF);
} else if (enccfg.rc_end_usage != AOM_Q) {
enccfg.rc_end_usage = AOM_Q;
ctx->crf = 32;
av_log(avctx, AV_LOG_WARNING,
"Neither bitrate nor constrained quality specified, using default CRF of %d\n",
ctx->crf);
}
if (avctx->qmin >= 0)
enccfg.rc_min_quantizer = avctx->qmin;
if (avctx->qmax >= 0) {
enccfg.rc_max_quantizer = avctx->qmax;
} else if (!ctx->crf) {
enccfg.rc_max_quantizer = 0;
}
if (enccfg.rc_end_usage == AOM_CQ || enccfg.rc_end_usage == AOM_Q) {
if (ctx->crf < enccfg.rc_min_quantizer || ctx->crf > enccfg.rc_max_quantizer) {
av_log(avctx, AV_LOG_ERROR,
"CQ level %d must be between minimum and maximum quantizer value (%d-%d)\n",
ctx->crf, enccfg.rc_min_quantizer, enccfg.rc_max_quantizer);
return AVERROR(EINVAL);
}
}
enccfg.rc_dropframe_thresh = ctx->drop_threshold;
// 0-100 (0 => CBR, 100 => VBR)
enccfg.rc_2pass_vbr_bias_pct = round(avctx->qcompress * 100);
if (ctx->minsection_pct >= 0)
enccfg.rc_2pass_vbr_minsection_pct = ctx->minsection_pct;
else if (avctx->bit_rate)
enccfg.rc_2pass_vbr_minsection_pct =
avctx->rc_min_rate * 100LL / avctx->bit_rate;
if (ctx->maxsection_pct >= 0)
enccfg.rc_2pass_vbr_maxsection_pct = ctx->maxsection_pct;
else if (avctx->rc_max_rate)
enccfg.rc_2pass_vbr_maxsection_pct =
avctx->rc_max_rate * 100LL / avctx->bit_rate;
if (avctx->rc_buffer_size)
enccfg.rc_buf_sz =
avctx->rc_buffer_size * 1000LL / avctx->bit_rate;
if (avctx->rc_initial_buffer_occupancy)
enccfg.rc_buf_initial_sz =
avctx->rc_initial_buffer_occupancy * 1000LL / avctx->bit_rate;
enccfg.rc_buf_optimal_sz = enccfg.rc_buf_sz * 5 / 6;
if (ctx->rc_undershoot_pct >= 0)
enccfg.rc_undershoot_pct = ctx->rc_undershoot_pct;
if (ctx->rc_overshoot_pct >= 0)
enccfg.rc_overshoot_pct = ctx->rc_overshoot_pct;
// _enc_init() will balk if kf_min_dist differs from max w/AOM_KF_AUTO
if (avctx->keyint_min >= 0 && avctx->keyint_min == avctx->gop_size)
enccfg.kf_min_dist = avctx->keyint_min;
if (avctx->gop_size >= 0)
enccfg.kf_max_dist = avctx->gop_size;
if (enccfg.g_pass == AOM_RC_FIRST_PASS)
enccfg.g_lag_in_frames = 0;
else if (enccfg.g_pass == AOM_RC_LAST_PASS) {
int decode_size, ret;
if (!avctx->stats_in) {
av_log(avctx, AV_LOG_ERROR, "No stats file for second pass\n");
return AVERROR_INVALIDDATA;
}
ctx->twopass_stats.sz = strlen(avctx->stats_in) * 3 / 4;
ret = av_reallocp(&ctx->twopass_stats.buf, ctx->twopass_stats.sz);
if (ret < 0) {
av_log(avctx, AV_LOG_ERROR,
"Stat buffer alloc (%"SIZE_SPECIFIER" bytes) failed\n",
ctx->twopass_stats.sz);
ctx->twopass_stats.sz = 0;
return ret;
}
decode_size = av_base64_decode(ctx->twopass_stats.buf, avctx->stats_in,
ctx->twopass_stats.sz);
if (decode_size < 0) {
av_log(avctx, AV_LOG_ERROR, "Stat buffer decode failed\n");
return AVERROR_INVALIDDATA;
}
ctx->twopass_stats.sz = decode_size;
enccfg.rc_twopass_stats_in = ctx->twopass_stats;
}
/* 0-3: For non-zero values the encoder increasingly optimizes for reduced
* complexity playback on low powered devices at the expense of encode
* quality. */
if (avctx->profile != AV_PROFILE_UNKNOWN)
enccfg.g_profile = avctx->profile;
enccfg.g_error_resilient = ctx->error_resilient;
res = choose_tiling(avctx, &enccfg);
if (res < 0)
return res;
if (ctx->still_picture) {
// Set the maximum number of frames to 1. This will let libaom set
// still_picture and reduced_still_picture_header to 1 in the Sequence
// Header as required by AVIF still images.
enccfg.g_limit = 1;
// Reduce memory usage for still images.
enccfg.g_lag_in_frames = 0;
// All frames will be key frames.
enccfg.kf_max_dist = 0;
enccfg.kf_mode = AOM_KF_DISABLED;
}
/* Construct Encoder Context */
res = aom_codec_enc_init(&ctx->encoder, iface, &enccfg, flags);
if (res != AOM_CODEC_OK) {
dump_enc_cfg(avctx, &enccfg, AV_LOG_WARNING);
log_encoder_error(avctx, "Failed to initialize encoder");
return AVERROR(EINVAL);
}
dump_enc_cfg(avctx, &enccfg, AV_LOG_DEBUG);
// codec control failures are currently treated only as warnings
av_log(avctx, AV_LOG_DEBUG, "aom_codec_control\n");
codecctl_int(avctx, AOME_SET_CPUUSED, ctx->cpu_used);
if (ctx->auto_alt_ref >= 0)
codecctl_int(avctx, AOME_SET_ENABLEAUTOALTREF, ctx->auto_alt_ref);
if (ctx->arnr_max_frames >= 0)
codecctl_int(avctx, AOME_SET_ARNR_MAXFRAMES, ctx->arnr_max_frames);
if (ctx->arnr_strength >= 0)
codecctl_int(avctx, AOME_SET_ARNR_STRENGTH, ctx->arnr_strength);
if (ctx->enable_cdef >= 0)
codecctl_int(avctx, AV1E_SET_ENABLE_CDEF, ctx->enable_cdef);
if (ctx->enable_restoration >= 0)
codecctl_int(avctx, AV1E_SET_ENABLE_RESTORATION, ctx->enable_restoration);
#if AOM_ENCODER_ABI_VERSION >= 22
if (ctx->enable_rect_partitions >= 0)
codecctl_int(avctx, AV1E_SET_ENABLE_RECT_PARTITIONS, ctx->enable_rect_partitions);
if (ctx->enable_1to4_partitions >= 0)
codecctl_int(avctx, AV1E_SET_ENABLE_1TO4_PARTITIONS, ctx->enable_1to4_partitions);
if (ctx->enable_ab_partitions >= 0)
codecctl_int(avctx, AV1E_SET_ENABLE_AB_PARTITIONS, ctx->enable_ab_partitions);
if (ctx->enable_angle_delta >= 0)
codecctl_int(avctx, AV1E_SET_ENABLE_ANGLE_DELTA, ctx->enable_angle_delta);
if (ctx->enable_cfl_intra >= 0)
codecctl_int(avctx, AV1E_SET_ENABLE_CFL_INTRA, ctx->enable_cfl_intra);
if (ctx->enable_filter_intra >= 0)
codecctl_int(avctx, AV1E_SET_ENABLE_FILTER_INTRA, ctx->enable_filter_intra);
if (ctx->enable_intra_edge_filter >= 0)
codecctl_int(avctx, AV1E_SET_ENABLE_INTRA_EDGE_FILTER, ctx->enable_intra_edge_filter);
if (ctx->enable_paeth_intra >= 0)
codecctl_int(avctx, AV1E_SET_ENABLE_PAETH_INTRA, ctx->enable_paeth_intra);
if (ctx->enable_smooth_intra >= 0)
codecctl_int(avctx, AV1E_SET_ENABLE_SMOOTH_INTRA, ctx->enable_smooth_intra);
if (ctx->enable_palette >= 0)
codecctl_int(avctx, AV1E_SET_ENABLE_PALETTE, ctx->enable_palette);
if (ctx->enable_tx64 >= 0)
codecctl_int(avctx, AV1E_SET_ENABLE_TX64, ctx->enable_tx64);
if (ctx->enable_flip_idtx >= 0)
codecctl_int(avctx, AV1E_SET_ENABLE_FLIP_IDTX, ctx->enable_flip_idtx);
if (ctx->use_intra_dct_only >= 0)
codecctl_int(avctx, AV1E_SET_INTRA_DCT_ONLY, ctx->use_intra_dct_only);
if (ctx->use_inter_dct_only >= 0)
codecctl_int(avctx, AV1E_SET_INTER_DCT_ONLY, ctx->use_inter_dct_only);
if (ctx->use_intra_default_tx_only >= 0)
codecctl_int(avctx, AV1E_SET_INTRA_DEFAULT_TX_ONLY, ctx->use_intra_default_tx_only);
if (ctx->reduced_tx_type_set >= 0)
codecctl_int(avctx, AV1E_SET_REDUCED_TX_TYPE_SET, ctx->reduced_tx_type_set);
if (ctx->enable_ref_frame_mvs >= 0)
codecctl_int(avctx, AV1E_SET_ENABLE_REF_FRAME_MVS, ctx->enable_ref_frame_mvs);
if (ctx->enable_reduced_reference_set >= 0)
codecctl_int(avctx, AV1E_SET_REDUCED_REFERENCE_SET, ctx->enable_reduced_reference_set);
if (ctx->enable_diff_wtd_comp >= 0)
codecctl_int(avctx, AV1E_SET_ENABLE_DIFF_WTD_COMP, ctx->enable_diff_wtd_comp);
if (ctx->enable_dist_wtd_comp >= 0)
codecctl_int(avctx, AV1E_SET_ENABLE_DIST_WTD_COMP, ctx->enable_dist_wtd_comp);
if (ctx->enable_dual_filter >= 0)
codecctl_int(avctx, AV1E_SET_ENABLE_DUAL_FILTER, ctx->enable_dual_filter);
if (ctx->enable_interinter_wedge >= 0)
codecctl_int(avctx, AV1E_SET_ENABLE_INTERINTER_WEDGE, ctx->enable_interinter_wedge);
if (ctx->enable_masked_comp >= 0)
codecctl_int(avctx, AV1E_SET_ENABLE_MASKED_COMP, ctx->enable_masked_comp);
if (ctx->enable_interintra_comp >= 0)
codecctl_int(avctx, AV1E_SET_ENABLE_INTERINTRA_COMP, ctx->enable_interintra_comp);
if (ctx->enable_interintra_wedge >= 0)
codecctl_int(avctx, AV1E_SET_ENABLE_INTERINTRA_WEDGE, ctx->enable_interintra_wedge);
if (ctx->enable_obmc >= 0)
codecctl_int(avctx, AV1E_SET_ENABLE_OBMC, ctx->enable_obmc);
if (ctx->enable_onesided_comp >= 0)
codecctl_int(avctx, AV1E_SET_ENABLE_ONESIDED_COMP, ctx->enable_onesided_comp);
if (ctx->enable_smooth_interintra >= 0)
codecctl_int(avctx, AV1E_SET_ENABLE_SMOOTH_INTERINTRA, ctx->enable_smooth_interintra);
#endif
codecctl_int(avctx, AOME_SET_STATIC_THRESHOLD, ctx->static_thresh);
if (ctx->crf >= 0)
codecctl_int(avctx, AOME_SET_CQ_LEVEL, ctx->crf);
if (ctx->tune >= 0)
codecctl_int(avctx, AOME_SET_TUNING, ctx->tune);
if (desc->flags & AV_PIX_FMT_FLAG_RGB) {
codecctl_int(avctx, AV1E_SET_COLOR_PRIMARIES, AVCOL_PRI_BT709);
codecctl_int(avctx, AV1E_SET_MATRIX_COEFFICIENTS, AVCOL_SPC_RGB);
codecctl_int(avctx, AV1E_SET_TRANSFER_CHARACTERISTICS, AVCOL_TRC_IEC61966_2_1);
} else {
codecctl_int(avctx, AV1E_SET_COLOR_PRIMARIES, avctx->color_primaries);
codecctl_int(avctx, AV1E_SET_MATRIX_COEFFICIENTS, avctx->colorspace);
codecctl_int(avctx, AV1E_SET_TRANSFER_CHARACTERISTICS, avctx->color_trc);
}
if (ctx->aq_mode >= 0)
codecctl_int(avctx, AV1E_SET_AQ_MODE, ctx->aq_mode);
if (ctx->frame_parallel >= 0)
codecctl_int(avctx, AV1E_SET_FRAME_PARALLEL_DECODING, ctx->frame_parallel);
set_color_range(avctx);
codecctl_int(avctx, AV1E_SET_SUPERBLOCK_SIZE, ctx->superblock_size);
if (ctx->uniform_tiles) {
codecctl_int(avctx, AV1E_SET_TILE_COLUMNS, ctx->tile_cols_log2);
codecctl_int(avctx, AV1E_SET_TILE_ROWS, ctx->tile_rows_log2);
}
#ifdef AOM_CTRL_AV1E_SET_DENOISE_NOISE_LEVEL
if (ctx->denoise_noise_level >= 0)
codecctl_int(avctx, AV1E_SET_DENOISE_NOISE_LEVEL, ctx->denoise_noise_level);
#endif
#ifdef AOM_CTRL_AV1E_SET_DENOISE_BLOCK_SIZE
if (ctx->denoise_block_size >= 0)
codecctl_int(avctx, AV1E_SET_DENOISE_BLOCK_SIZE, ctx->denoise_block_size);
#endif
#ifdef AOM_CTRL_AV1E_SET_ENABLE_GLOBAL_MOTION
if (ctx->enable_global_motion >= 0)
codecctl_int(avctx, AV1E_SET_ENABLE_GLOBAL_MOTION, ctx->enable_global_motion);
#endif
#ifdef AOM_CTRL_AV1E_SET_MAX_REFERENCE_FRAMES
if (avctx->refs >= 3) {
codecctl_int(avctx, AV1E_SET_MAX_REFERENCE_FRAMES, avctx->refs);
}
#endif
#ifdef AOM_CTRL_AV1E_SET_ROW_MT
if (ctx->row_mt >= 0)
codecctl_int(avctx, AV1E_SET_ROW_MT, ctx->row_mt);
#endif
#ifdef AOM_CTRL_AV1E_SET_ENABLE_INTRABC
if (ctx->enable_intrabc >= 0)
codecctl_int(avctx, AV1E_SET_ENABLE_INTRABC, ctx->enable_intrabc);
#endif
#if AOM_ENCODER_ABI_VERSION >= 23
{
AVDictionaryEntry *en = NULL;
while ((en = av_dict_get(ctx->aom_params, "", en, AV_DICT_IGNORE_SUFFIX))) {
int ret = aom_codec_set_option(&ctx->encoder, en->key, en->value);
if (ret != AOM_CODEC_OK) {
log_encoder_error(avctx, en->key);
return AVERROR_EXTERNAL;
}
}
}
#endif
// provide dummy value to initialize wrapper, values will be updated each _encode()
aom_img_wrap(&ctx->rawimg, img_fmt, avctx->width, avctx->height, 1,
(unsigned char*)1);
if (codec_caps & AOM_CODEC_CAP_HIGHBITDEPTH)
ctx->rawimg.bit_depth = enccfg.g_bit_depth;
cpb_props = ff_encode_add_cpb_side_data(avctx);
if (!cpb_props)
return AVERROR(ENOMEM);
if (avctx->flags & AV_CODEC_FLAG_GLOBAL_HEADER) {
const AVBitStreamFilter *filter = av_bsf_get_by_name("extract_extradata");
int ret;
if (!filter) {
av_log(avctx, AV_LOG_ERROR, "extract_extradata bitstream filter "
"not found. This is a bug, please report it.\n");
return AVERROR_BUG;
}
ret = av_bsf_alloc(filter, &ctx->bsf);
if (ret < 0)
return ret;
ret = avcodec_parameters_from_context(ctx->bsf->par_in, avctx);
if (ret < 0)
return ret;
ret = av_bsf_init(ctx->bsf);
if (ret < 0)
return ret;
}
if (enccfg.rc_end_usage == AOM_CBR ||
enccfg.g_pass != AOM_RC_ONE_PASS) {
cpb_props->max_bitrate = avctx->rc_max_rate;
cpb_props->min_bitrate = avctx->rc_min_rate;
cpb_props->avg_bitrate = avctx->bit_rate;
}
cpb_props->buffer_size = avctx->rc_buffer_size;
return 0;
}
static inline void cx_pktcpy(AOMContext *ctx,
struct FrameListData *dst,
const struct aom_codec_cx_pkt *src)
{
dst->pts = src->data.frame.pts;
dst->duration = src->data.frame.duration;
dst->flags = src->data.frame.flags;
dst->sz = src->data.frame.sz;
dst->buf = src->data.frame.buf;
#ifdef AOM_FRAME_IS_INTRAONLY
dst->frame_number = ++ctx->frame_number;
dst->have_sse = ctx->have_sse;
if (ctx->have_sse) {
/* associate last-seen SSE to the frame. */
/* Transfers ownership from ctx to dst. */
memcpy(dst->sse, ctx->sse, sizeof(dst->sse));
ctx->have_sse = 0;
}
#endif
}
/**
* Store coded frame information in format suitable for return from encode2().
*
* Write information from @a cx_frame to @a pkt
* @return packet data size on success
* @return a negative AVERROR on error
*/
static int storeframe(AVCodecContext *avctx, struct FrameListData *cx_frame,
AVPacket *pkt)
{
AOMContext *ctx = avctx->priv_data;
int av_unused pict_type;
int ret = ff_get_encode_buffer(avctx, pkt, cx_frame->sz, 0);
if (ret < 0) {
av_log(avctx, AV_LOG_ERROR,
"Error getting output packet of size %"SIZE_SPECIFIER".\n", cx_frame->sz);
return ret;
}
memcpy(pkt->data, cx_frame->buf, pkt->size);
pkt->pts = pkt->dts = cx_frame->pts;
pkt->duration = cx_frame->duration;
if (!!(cx_frame->flags & AOM_FRAME_IS_KEY)) {
pkt->flags |= AV_PKT_FLAG_KEY;
#ifdef AOM_FRAME_IS_INTRAONLY
pict_type = AV_PICTURE_TYPE_I;
} else if (cx_frame->flags & AOM_FRAME_IS_INTRAONLY) {
pict_type = AV_PICTURE_TYPE_I;
} else {
pict_type = AV_PICTURE_TYPE_P;
}
ff_side_data_set_encoder_stats(pkt, 0, cx_frame->sse + 1,
cx_frame->have_sse ? 3 : 0, pict_type);
if (cx_frame->have_sse) {
int i;
for (i = 0; i < 3; ++i) {
avctx->error[i] += cx_frame->sse[i + 1];
}
cx_frame->have_sse = 0;
#endif
}
if (avctx->flags & AV_CODEC_FLAG_GLOBAL_HEADER) {
ret = av_bsf_send_packet(ctx->bsf, pkt);
if (ret < 0) {
av_log(avctx, AV_LOG_ERROR, "extract_extradata filter "
"failed to send input packet\n");
return ret;
}
ret = av_bsf_receive_packet(ctx->bsf, pkt);
if (ret < 0) {
av_log(avctx, AV_LOG_ERROR, "extract_extradata filter "
"failed to receive output packet\n");
return ret;
}
}
return pkt->size;
}
/**
* Queue multiple output frames from the encoder, returning the front-most.
* In cases where aom_codec_get_cx_data() returns more than 1 frame append
* the frame queue. Return the head frame if available.
* @return Stored frame size
* @return AVERROR(EINVAL) on output size error
* @return AVERROR(ENOMEM) on coded frame queue data allocation error
*/
static int queue_frames(AVCodecContext *avctx, AVPacket *pkt_out)
{
AOMContext *ctx = avctx->priv_data;
const struct aom_codec_cx_pkt *pkt;
const void *iter = NULL;
int size = 0;
if (ctx->coded_frame_list) {
struct FrameListData *cx_frame = ctx->coded_frame_list;
/* return the leading frame if we've already begun queueing */
size = storeframe(avctx, cx_frame, pkt_out);
if (size < 0)
return size;
ctx->coded_frame_list = cx_frame->next;
free_coded_frame(cx_frame);
}
/* consume all available output from the encoder before returning. buffers
* are only good through the next aom_codec call */
while ((pkt = aom_codec_get_cx_data(&ctx->encoder, &iter))) {
switch (pkt->kind) {
case AOM_CODEC_CX_FRAME_PKT:
if (!size) {
struct FrameListData cx_frame;
/* avoid storing the frame when the list is empty and we haven't yet
* provided a frame for output */
av_assert0(!ctx->coded_frame_list);
cx_pktcpy(ctx, &cx_frame, pkt);
size = storeframe(avctx, &cx_frame, pkt_out);
if (size < 0)
return size;
} else {
struct FrameListData *cx_frame =
av_malloc(sizeof(struct FrameListData));
if (!cx_frame) {
av_log(avctx, AV_LOG_ERROR,
"Frame queue element alloc failed\n");
return AVERROR(ENOMEM);
}
cx_pktcpy(ctx, cx_frame, pkt);
cx_frame->buf = av_malloc(cx_frame->sz);
if (!cx_frame->buf) {
av_log(avctx, AV_LOG_ERROR,
"Data buffer alloc (%"SIZE_SPECIFIER" bytes) failed\n",
cx_frame->sz);
av_freep(&cx_frame);
return AVERROR(ENOMEM);
}
memcpy(cx_frame->buf, pkt->data.frame.buf, pkt->data.frame.sz);
coded_frame_add(&ctx->coded_frame_list, cx_frame);
}
break;
case AOM_CODEC_STATS_PKT:
{
struct aom_fixed_buf *stats = &ctx->twopass_stats;
uint8_t *tmp = av_fast_realloc(stats->buf,
&ctx->twopass_stats_size,
stats->sz +
pkt->data.twopass_stats.sz);
if (!tmp) {
av_freep(&stats->buf);
stats->sz = 0;
av_log(avctx, AV_LOG_ERROR, "Stat buffer realloc failed\n");
return AVERROR(ENOMEM);
}
stats->buf = tmp;
memcpy((uint8_t *)stats->buf + stats->sz,
pkt->data.twopass_stats.buf, pkt->data.twopass_stats.sz);
stats->sz += pkt->data.twopass_stats.sz;
break;
}
#ifdef AOM_FRAME_IS_INTRAONLY
case AOM_CODEC_PSNR_PKT:
{
av_assert0(!ctx->have_sse);
ctx->sse[0] = pkt->data.psnr.sse[0];
ctx->sse[1] = pkt->data.psnr.sse[1];
ctx->sse[2] = pkt->data.psnr.sse[2];
ctx->sse[3] = pkt->data.psnr.sse[3];
ctx->have_sse = 1;
break;
}
#endif
case AOM_CODEC_CUSTOM_PKT:
// ignore unsupported/unrecognized packet types
break;
}
}
return size;
}
static enum AVPixelFormat aomfmt_to_pixfmt(struct aom_image *img)
{
switch (img->fmt) {
case AOM_IMG_FMT_I420:
case AOM_IMG_FMT_I42016:
if (img->bit_depth == 8)
return img->monochrome ? AV_PIX_FMT_GRAY8 : AV_PIX_FMT_YUV420P;
else if (img->bit_depth == 10)
return img->monochrome ? AV_PIX_FMT_GRAY10 : AV_PIX_FMT_YUV420P10;
else
return img->monochrome ? AV_PIX_FMT_GRAY12 : AV_PIX_FMT_YUV420P12;
case AOM_IMG_FMT_I422:
case AOM_IMG_FMT_I42216:
if (img->bit_depth == 8)
return AV_PIX_FMT_YUV422P;
else if (img->bit_depth == 10)
return AV_PIX_FMT_YUV422P10;
else
return AV_PIX_FMT_YUV422P12;
case AOM_IMG_FMT_I444:
case AOM_IMG_FMT_I44416:
if (img->bit_depth == 8)
return AV_PIX_FMT_YUV444P;
else if (img->bit_depth == 10)
return AV_PIX_FMT_YUV444P10;
else
return AV_PIX_FMT_YUV444P12;
};
return AV_PIX_FMT_NONE;
}
static int aom_encode(AVCodecContext *avctx, AVPacket *pkt,
const AVFrame *frame, int *got_packet)
{
AOMContext *ctx = avctx->priv_data;
struct aom_image *rawimg = NULL;
int64_t timestamp = 0;
unsigned long duration = 0;
int res, coded_size;
aom_enc_frame_flags_t flags = 0;
if (frame) {
rawimg = &ctx->rawimg;
rawimg->planes[AOM_PLANE_Y] = frame->data[0];
rawimg->planes[AOM_PLANE_U] = frame->data[1];
rawimg->planes[AOM_PLANE_V] = frame->data[2];
rawimg->stride[AOM_PLANE_Y] = frame->linesize[0];
rawimg->stride[AOM_PLANE_U] = frame->linesize[1];
rawimg->stride[AOM_PLANE_V] = frame->linesize[2];
timestamp = frame->pts;
if (frame->duration > ULONG_MAX) {
av_log(avctx, AV_LOG_WARNING,
"Frame duration too large: %"PRId64"\n", frame->duration);
} else if (frame->duration)
duration = frame->duration;
else if (avctx->framerate.num > 0 && avctx->framerate.den > 0)
duration = av_rescale_q(1, av_inv_q(avctx->framerate), avctx->time_base);
else {
FF_DISABLE_DEPRECATION_WARNINGS
duration =
#if FF_API_TICKS_PER_FRAME
avctx->ticks_per_frame ? avctx->ticks_per_frame :
#endif
1;
FF_ENABLE_DEPRECATION_WARNINGS
}
switch (frame->color_range) {
case AVCOL_RANGE_MPEG:
rawimg->range = AOM_CR_STUDIO_RANGE;
break;
case AVCOL_RANGE_JPEG:
rawimg->range = AOM_CR_FULL_RANGE;
break;
}
if (frame->pict_type == AV_PICTURE_TYPE_I)
flags |= AOM_EFLAG_FORCE_KF;
}
res = aom_codec_encode(&ctx->encoder, rawimg, timestamp, duration, flags);
if (res != AOM_CODEC_OK) {
log_encoder_error(avctx, "Error encoding frame");
return AVERROR_INVALIDDATA;
}
coded_size = queue_frames(avctx, pkt);
if (coded_size < 0)
return coded_size;
if (!frame && avctx->flags & AV_CODEC_FLAG_PASS1) {
size_t b64_size = AV_BASE64_SIZE(ctx->twopass_stats.sz);
avctx->stats_out = av_malloc(b64_size);
if (!avctx->stats_out) {
av_log(avctx, AV_LOG_ERROR, "Stat buffer alloc (%"SIZE_SPECIFIER" bytes) failed\n",
b64_size);
return AVERROR(ENOMEM);
}
av_base64_encode(avctx->stats_out, b64_size, ctx->twopass_stats.buf,
ctx->twopass_stats.sz);
}
*got_packet = !!coded_size;
if (*got_packet && avctx->flags & AV_CODEC_FLAG_RECON_FRAME) {
AVCodecInternal *avci = avctx->internal;
struct aom_image img;
av_frame_unref(avci->recon_frame);
res = codecctl_imgp(avctx, AV1_GET_NEW_FRAME_IMAGE, &img);
if (res < 0)
return res;
avci->recon_frame->format = aomfmt_to_pixfmt(&img);
if (avci->recon_frame->format == AV_PIX_FMT_NONE) {
av_log(ctx, AV_LOG_ERROR,
"Unhandled reconstructed frame colorspace: %d\n",
img.fmt);
return AVERROR(ENOSYS);
}
avci->recon_frame->width = img.d_w;
avci->recon_frame->height = img.d_h;
res = av_frame_get_buffer(avci->recon_frame, 0);
if (res < 0)
return res;
if ((img.fmt & AOM_IMG_FMT_HIGHBITDEPTH) && img.bit_depth == 8)
ff_aom_image_copy_16_to_8(avci->recon_frame, &img);
else {
const uint8_t *planes[4] = { img.planes[0], img.planes[1], img.planes[2] };
const int stride[4] = { img.stride[0], img.stride[1], img.stride[2] };
av_image_copy(avci->recon_frame->data, avci->recon_frame->linesize, planes,
stride, avci->recon_frame->format, img.d_w, img.d_h);
}
}
return 0;
}
static const enum AVPixelFormat av1_pix_fmts[] = {
AV_PIX_FMT_YUV420P,
AV_PIX_FMT_YUV422P,
AV_PIX_FMT_YUV444P,
AV_PIX_FMT_GBRP,
AV_PIX_FMT_NONE
};
static const enum AVPixelFormat av1_pix_fmts_with_gray[] = {
AV_PIX_FMT_YUV420P,
AV_PIX_FMT_YUV422P,
AV_PIX_FMT_YUV444P,
AV_PIX_FMT_GBRP,
AV_PIX_FMT_GRAY8,
AV_PIX_FMT_NONE
};
static const enum AVPixelFormat av1_pix_fmts_highbd[] = {
AV_PIX_FMT_YUV420P,
AV_PIX_FMT_YUV422P,
AV_PIX_FMT_YUV444P,
AV_PIX_FMT_GBRP,
AV_PIX_FMT_YUV420P10,
AV_PIX_FMT_YUV422P10,
AV_PIX_FMT_YUV444P10,
AV_PIX_FMT_YUV420P12,
AV_PIX_FMT_YUV422P12,
AV_PIX_FMT_YUV444P12,
AV_PIX_FMT_GBRP10,
AV_PIX_FMT_GBRP12,
AV_PIX_FMT_NONE
};
static const enum AVPixelFormat av1_pix_fmts_highbd_with_gray[] = {
AV_PIX_FMT_YUV420P,
AV_PIX_FMT_YUV422P,
AV_PIX_FMT_YUV444P,
AV_PIX_FMT_GBRP,
AV_PIX_FMT_YUV420P10,
AV_PIX_FMT_YUV422P10,
AV_PIX_FMT_YUV444P10,
AV_PIX_FMT_YUV420P12,
AV_PIX_FMT_YUV422P12,
AV_PIX_FMT_YUV444P12,
AV_PIX_FMT_GBRP10,
AV_PIX_FMT_GBRP12,
AV_PIX_FMT_GRAY8,
AV_PIX_FMT_GRAY10,
AV_PIX_FMT_GRAY12,
AV_PIX_FMT_NONE
};
static av_cold void av1_init_static(FFCodec *codec)
{
int supports_monochrome = aom_codec_version() >= 20001;
aom_codec_caps_t codec_caps = aom_codec_get_caps(aom_codec_av1_cx());
if (codec_caps & AOM_CODEC_CAP_HIGHBITDEPTH)
codec->p.pix_fmts = supports_monochrome ? av1_pix_fmts_highbd_with_gray :
av1_pix_fmts_highbd;
else
codec->p.pix_fmts = supports_monochrome ? av1_pix_fmts_with_gray :
av1_pix_fmts;
if (aom_codec_version_major() < 2)
codec->p.capabilities |= AV_CODEC_CAP_EXPERIMENTAL;
}
static av_cold int av1_init(AVCodecContext *avctx)
{
return aom_init(avctx, aom_codec_av1_cx());
}
#define OFFSET(x) offsetof(AOMContext, x)
#define VE AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_ENCODING_PARAM
static const AVOption options[] = {
{ "cpu-used", "Quality/Speed ratio modifier", OFFSET(cpu_used), AV_OPT_TYPE_INT, {.i64 = 1}, 0, 8, VE},
{ "auto-alt-ref", "Enable use of alternate reference "
"frames (2-pass only)", OFFSET(auto_alt_ref), AV_OPT_TYPE_INT, {.i64 = -1}, -1, 2, VE},
{ "lag-in-frames", "Number of frames to look ahead at for "
"alternate reference frame selection", OFFSET(lag_in_frames), AV_OPT_TYPE_INT, {.i64 = -1}, -1, INT_MAX, VE},
{ "arnr-max-frames", "altref noise reduction max frame count", OFFSET(arnr_max_frames), AV_OPT_TYPE_INT, {.i64 = -1}, -1, INT_MAX, VE},
{ "arnr-strength", "altref noise reduction filter strength", OFFSET(arnr_strength), AV_OPT_TYPE_INT, {.i64 = -1}, -1, 6, VE},
{ "aq-mode", "adaptive quantization mode", OFFSET(aq_mode), AV_OPT_TYPE_INT, {.i64 = -1}, -1, 4, VE, .unit = "aq_mode"},
{ "none", "Aq not used", 0, AV_OPT_TYPE_CONST, {.i64 = 0}, 0, 0, VE, .unit = "aq_mode"},
{ "variance", "Variance based Aq", 0, AV_OPT_TYPE_CONST, {.i64 = 1}, 0, 0, VE, .unit = "aq_mode"},
{ "complexity", "Complexity based Aq", 0, AV_OPT_TYPE_CONST, {.i64 = 2}, 0, 0, VE, .unit = "aq_mode"},
{ "cyclic", "Cyclic Refresh Aq", 0, AV_OPT_TYPE_CONST, {.i64 = 3}, 0, 0, VE, .unit = "aq_mode"},
{ "error-resilience", "Error resilience configuration", OFFSET(error_resilient), AV_OPT_TYPE_FLAGS, {.i64 = 0}, INT_MIN, INT_MAX, VE, .unit = "er"},
{ "default", "Improve resiliency against losses of whole frames", 0, AV_OPT_TYPE_CONST, {.i64 = AOM_ERROR_RESILIENT_DEFAULT}, 0, 0, VE, .unit = "er"},
{ "crf", "Select the quality for constant quality mode", offsetof(AOMContext, crf), AV_OPT_TYPE_INT, {.i64 = -1}, -1, 63, VE },
{ "static-thresh", "A change threshold on blocks below which they will be skipped by the encoder", OFFSET(static_thresh), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, VE },
{ "drop-threshold", "Frame drop threshold", offsetof(AOMContext, drop_threshold), AV_OPT_TYPE_INT, {.i64 = 0 }, INT_MIN, INT_MAX, VE },
{ "denoise-noise-level", "Amount of noise to be removed", OFFSET(denoise_noise_level), AV_OPT_TYPE_INT, {.i64 = -1}, -1, INT_MAX, VE},
{ "denoise-block-size", "Denoise block size ", OFFSET(denoise_block_size), AV_OPT_TYPE_INT, {.i64 = -1}, -1, INT_MAX, VE},
{ "undershoot-pct", "Datarate undershoot (min) target (%)", OFFSET(rc_undershoot_pct), AV_OPT_TYPE_INT, {.i64 = -1}, -1, 100, VE},
{ "overshoot-pct", "Datarate overshoot (max) target (%)", OFFSET(rc_overshoot_pct), AV_OPT_TYPE_INT, {.i64 = -1}, -1, 1000, VE},
{ "minsection-pct", "GOP min bitrate (% of target)", OFFSET(minsection_pct), AV_OPT_TYPE_INT, {.i64 = -1}, -1, 100, VE},
{ "maxsection-pct", "GOP max bitrate (% of target)", OFFSET(maxsection_pct), AV_OPT_TYPE_INT, {.i64 = -1}, -1, 5000, VE},
{ "frame-parallel", "Enable frame parallel decodability features", OFFSET(frame_parallel), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
{ "tiles", "Tile columns x rows", OFFSET(tile_cols), AV_OPT_TYPE_IMAGE_SIZE, { .str = NULL }, 0, 0, VE },
{ "tile-columns", "Log2 of number of tile columns to use", OFFSET(tile_cols_log2), AV_OPT_TYPE_INT, {.i64 = -1}, -1, 6, VE},
{ "tile-rows", "Log2 of number of tile rows to use", OFFSET(tile_rows_log2), AV_OPT_TYPE_INT, {.i64 = -1}, -1, 6, VE},
{ "row-mt", "Enable row based multi-threading", OFFSET(row_mt), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
{ "enable-cdef", "Enable CDEF filtering", OFFSET(enable_cdef), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
{ "enable-global-motion", "Enable global motion", OFFSET(enable_global_motion), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
{ "enable-intrabc", "Enable intra block copy prediction mode", OFFSET(enable_intrabc), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
{ "enable-restoration", "Enable Loop Restoration filtering", OFFSET(enable_restoration), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
{ "usage", "Quality and compression efficiency vs speed trade-off", OFFSET(usage), AV_OPT_TYPE_INT, {.i64 = 0}, 0, INT_MAX, VE, .unit = "usage"},
{ "good", "Good quality", 0, AV_OPT_TYPE_CONST, {.i64 = 0 /* AOM_USAGE_GOOD_QUALITY */}, 0, 0, VE, .unit = "usage"},
{ "realtime", "Realtime encoding", 0, AV_OPT_TYPE_CONST, {.i64 = 1 /* AOM_USAGE_REALTIME */}, 0, 0, VE, .unit = "usage"},
{ "allintra", "All Intra encoding", 0, AV_OPT_TYPE_CONST, {.i64 = 2 /* AOM_USAGE_ALL_INTRA */}, 0, 0, VE, .unit = "usage"},
{ "tune", "The metric that the encoder tunes for. Automatically chosen by the encoder by default", OFFSET(tune), AV_OPT_TYPE_INT, {.i64 = -1}, -1, AOM_TUNE_SSIM, VE, .unit = "tune"},
{ "psnr", NULL, 0, AV_OPT_TYPE_CONST, {.i64 = AOM_TUNE_PSNR}, 0, 0, VE, .unit = "tune"},
{ "ssim", NULL, 0, AV_OPT_TYPE_CONST, {.i64 = AOM_TUNE_SSIM}, 0, 0, VE, .unit = "tune"},
FF_AV1_PROFILE_OPTS
{ "still-picture", "Encode in single frame mode (typically used for still AVIF images).", OFFSET(still_picture), AV_OPT_TYPE_BOOL, {.i64 = 0}, -1, 1, VE },
{ "enable-rect-partitions", "Enable rectangular partitions", OFFSET(enable_rect_partitions), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
{ "enable-1to4-partitions", "Enable 1:4/4:1 partitions", OFFSET(enable_1to4_partitions), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
{ "enable-ab-partitions", "Enable ab shape partitions", OFFSET(enable_ab_partitions), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
{ "enable-angle-delta", "Enable angle delta intra prediction", OFFSET(enable_angle_delta), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
{ "enable-cfl-intra", "Enable chroma predicted from luma intra prediction", OFFSET(enable_cfl_intra), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
{ "enable-filter-intra", "Enable filter intra predictor", OFFSET(enable_filter_intra), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
{ "enable-intra-edge-filter", "Enable intra edge filter", OFFSET(enable_intra_edge_filter), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
{ "enable-smooth-intra", "Enable smooth intra prediction mode", OFFSET(enable_smooth_intra), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
{ "enable-paeth-intra", "Enable paeth predictor in intra prediction", OFFSET(enable_paeth_intra), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
{ "enable-palette", "Enable palette prediction mode", OFFSET(enable_palette), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
{ "enable-flip-idtx", "Enable extended transform type", OFFSET(enable_flip_idtx), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
{ "enable-tx64", "Enable 64-pt transform", OFFSET(enable_tx64), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
{ "reduced-tx-type-set", "Use reduced set of transform types", OFFSET(reduced_tx_type_set), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
{ "use-intra-dct-only", "Use DCT only for INTRA modes", OFFSET(use_intra_dct_only), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
{ "use-inter-dct-only", "Use DCT only for INTER modes", OFFSET(use_inter_dct_only), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
{ "use-intra-default-tx-only", "Use default-transform only for INTRA modes", OFFSET(use_intra_default_tx_only), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
{ "enable-ref-frame-mvs", "Enable temporal mv prediction", OFFSET(enable_ref_frame_mvs), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
{ "enable-reduced-reference-set", "Use reduced set of single and compound references", OFFSET(enable_reduced_reference_set), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
{ "enable-obmc", "Enable obmc", OFFSET(enable_obmc), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
{ "enable-dual-filter", "Enable dual filter", OFFSET(enable_dual_filter), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
{ "enable-diff-wtd-comp", "Enable difference-weighted compound", OFFSET(enable_diff_wtd_comp), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
{ "enable-dist-wtd-comp", "Enable distance-weighted compound", OFFSET(enable_dist_wtd_comp), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
{ "enable-onesided-comp", "Enable one sided compound", OFFSET(enable_onesided_comp), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
{ "enable-interinter-wedge", "Enable interinter wedge compound", OFFSET(enable_interinter_wedge), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
{ "enable-interintra-wedge", "Enable interintra wedge compound", OFFSET(enable_interintra_wedge), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
{ "enable-masked-comp", "Enable masked compound", OFFSET(enable_masked_comp), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
{ "enable-interintra-comp", "Enable interintra compound", OFFSET(enable_interintra_comp), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
{ "enable-smooth-interintra", "Enable smooth interintra mode", OFFSET(enable_smooth_interintra), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
#if AOM_ENCODER_ABI_VERSION >= 23
{ "aom-params", "Set libaom options using a :-separated list of key=value pairs", OFFSET(aom_params), AV_OPT_TYPE_DICT, { 0 }, 0, 0, VE },
#endif
{ NULL },
};
static const FFCodecDefault defaults[] = {
{ "b", "0" },
{ "qmin", "-1" },
{ "qmax", "-1" },
{ "g", "-1" },
{ "keyint_min", "-1" },
{ NULL },
};
static const AVClass class_aom = {
.class_name = "libaom-av1 encoder",
.item_name = av_default_item_name,
.option = options,
.version = LIBAVUTIL_VERSION_INT,
};
FFCodec ff_libaom_av1_encoder = {
.p.name = "libaom-av1",
CODEC_LONG_NAME("libaom AV1"),
.p.type = AVMEDIA_TYPE_VIDEO,
.p.id = AV_CODEC_ID_AV1,
.p.capabilities = AV_CODEC_CAP_DR1 | AV_CODEC_CAP_DELAY |
AV_CODEC_CAP_ENCODER_RECON_FRAME |
AV_CODEC_CAP_OTHER_THREADS,
.p.profiles = NULL_IF_CONFIG_SMALL(ff_av1_profiles),
.p.priv_class = &class_aom,
.p.wrapper_name = "libaom",
.priv_data_size = sizeof(AOMContext),
.init = av1_init,
FF_CODEC_ENCODE_CB(aom_encode),
.close = aom_free,
.caps_internal = FF_CODEC_CAP_NOT_INIT_THREADSAFE |
FF_CODEC_CAP_INIT_CLEANUP |
FF_CODEC_CAP_AUTO_THREADS,
.defaults = defaults,
.init_static_data = av1_init_static,
};