mirror of https://github.com/mpv-player/mpv
cosmetics: vd_ffmpeg.c: reformat
This commit is contained in:
parent
3595dcc089
commit
3dd8b7326c
|
@ -84,7 +84,7 @@ static enum PixelFormat get_format(struct AVCodecContext *avctx,
|
|||
const enum PixelFormat *pix_fmt);
|
||||
static void uninit(struct sh_video *sh);
|
||||
|
||||
const m_option_t lavc_decode_opts_conf[]={
|
||||
const m_option_t lavc_decode_opts_conf[] = {
|
||||
OPT_INTRANGE("bug", lavc_param.workaround_bugs, 0, -1, 999999),
|
||||
OPT_INTRANGE("er", lavc_param.error_resilience, 0, 0, 99),
|
||||
OPT_FLAG_ON("gray", lavc_param.gray, 0),
|
||||
|
@ -106,7 +106,8 @@ const m_option_t lavc_decode_opts_conf[]={
|
|||
{NULL, NULL, 0, 0, 0, 0, NULL}
|
||||
};
|
||||
|
||||
static enum AVDiscard str2AVDiscard(char *str) {
|
||||
static enum AVDiscard str2AVDiscard(char *str)
|
||||
{
|
||||
if (!str) return AVDISCARD_DEFAULT;
|
||||
if (strcasecmp(str, "none" ) == 0) return AVDISCARD_NONE;
|
||||
if (strcasecmp(str, "default") == 0) return AVDISCARD_DEFAULT;
|
||||
|
@ -118,23 +119,26 @@ static enum AVDiscard str2AVDiscard(char *str) {
|
|||
return AVDISCARD_DEFAULT;
|
||||
}
|
||||
|
||||
// to set/get/query special features/parameters
|
||||
static int control(sh_video_t *sh, int cmd, void *arg, ...){
|
||||
static int control(sh_video_t *sh, int cmd, void *arg, ...)
|
||||
{
|
||||
vd_ffmpeg_ctx *ctx = sh->context;
|
||||
AVCodecContext *avctx = ctx->avctx;
|
||||
switch(cmd){
|
||||
switch (cmd) {
|
||||
case VDCTRL_QUERY_FORMAT:
|
||||
{
|
||||
int format =(*((int *)arg));
|
||||
if(format == ctx->best_csp) return CONTROL_TRUE;//supported
|
||||
int format = (*((int *)arg));
|
||||
if (format == ctx->best_csp)
|
||||
return CONTROL_TRUE;
|
||||
// possible conversions:
|
||||
switch(format){
|
||||
switch (format) {
|
||||
case IMGFMT_YV12:
|
||||
case IMGFMT_IYUV:
|
||||
case IMGFMT_I420:
|
||||
// "converted" using pointer/stride modification
|
||||
if(ctx->best_csp == IMGFMT_YV12) return CONTROL_TRUE;// u/v swap
|
||||
if(ctx->best_csp == IMGFMT_422P && !ctx->do_dr1) return CONTROL_TRUE;// half stride
|
||||
if (ctx->best_csp == IMGFMT_YV12)
|
||||
return CONTROL_TRUE; // u/v swap
|
||||
if (ctx->best_csp == IMGFMT_422P && !ctx->do_dr1)
|
||||
return CONTROL_TRUE; // half stride
|
||||
break;
|
||||
}
|
||||
return CONTROL_FALSE;
|
||||
|
@ -149,37 +153,41 @@ static int control(sh_video_t *sh, int cmd, void *arg, ...){
|
|||
return CONTROL_UNKNOWN;
|
||||
}
|
||||
|
||||
// init driver
|
||||
static int init(sh_video_t *sh){
|
||||
static int init(sh_video_t *sh)
|
||||
{
|
||||
struct lavc_param *lavc_param = &sh->opts->lavc_param;
|
||||
AVCodecContext *avctx;
|
||||
vd_ffmpeg_ctx *ctx;
|
||||
AVCodec *lavc_codec;
|
||||
int do_vis_debug= lavc_param->vismv || (lavc_param->debug&(FF_DEBUG_VIS_MB_TYPE|FF_DEBUG_VIS_QP));
|
||||
int do_vis_debug = lavc_param->vismv ||
|
||||
(lavc_param->debug & (FF_DEBUG_VIS_MB_TYPE | FF_DEBUG_VIS_QP));
|
||||
|
||||
ctx = sh->context = talloc_zero(NULL, vd_ffmpeg_ctx);
|
||||
|
||||
lavc_codec = avcodec_find_decoder_by_name(sh->codec->dll);
|
||||
if(!lavc_codec){
|
||||
mp_tmsg(MSGT_DECVIDEO, MSGL_ERR, "Cannot find codec '%s' in libavcodec...\n", sh->codec->dll);
|
||||
if (!lavc_codec) {
|
||||
mp_tmsg(MSGT_DECVIDEO, MSGL_ERR,
|
||||
"Cannot find codec '%s' in libavcodec...\n", sh->codec->dll);
|
||||
uninit(sh);
|
||||
return 0;
|
||||
}
|
||||
|
||||
if(sh->opts->vd_use_slices && (lavc_codec->capabilities&CODEC_CAP_DRAW_HORIZ_BAND) && !do_vis_debug)
|
||||
ctx->do_slices=1;
|
||||
if (sh->opts->vd_use_slices
|
||||
&& (lavc_codec->capabilities & CODEC_CAP_DRAW_HORIZ_BAND)
|
||||
&& !do_vis_debug)
|
||||
ctx->do_slices = 1;
|
||||
|
||||
if(lavc_codec->capabilities&CODEC_CAP_DR1 && !do_vis_debug
|
||||
&& lavc_codec->id != CODEC_ID_H264
|
||||
&& lavc_codec->id != CODEC_ID_INTERPLAY_VIDEO
|
||||
&& lavc_codec->id != CODEC_ID_ROQ && lavc_codec->id != CODEC_ID_VP8
|
||||
if (lavc_codec->capabilities & CODEC_CAP_DR1 && !do_vis_debug
|
||||
&& lavc_codec->id != CODEC_ID_H264
|
||||
&& lavc_codec->id != CODEC_ID_INTERPLAY_VIDEO
|
||||
&& lavc_codec->id != CODEC_ID_ROQ && lavc_codec->id != CODEC_ID_VP8
|
||||
#if LIBAVCODEC_VERSION_INT >= AV_VERSION_INT(52, 108, 0)
|
||||
&& lavc_codec->id != CODEC_ID_LAGARITH
|
||||
&& lavc_codec->id != CODEC_ID_LAGARITH
|
||||
#endif
|
||||
)
|
||||
ctx->do_dr1=1;
|
||||
ctx->b_age= ctx->ip_age[0]= ctx->ip_age[1]= 256*256*256*64;
|
||||
ctx->ip_count= ctx->b_count= 0;
|
||||
)
|
||||
ctx->do_dr1 = 1;
|
||||
ctx->b_age = ctx->ip_age[0] = ctx->ip_age[1] = 256 * 256 * 256 * 64;
|
||||
ctx->ip_count = ctx->b_count = 0;
|
||||
|
||||
ctx->pic = avcodec_alloc_frame();
|
||||
ctx->avctx = avcodec_alloc_context();
|
||||
|
@ -204,7 +212,7 @@ static int init(sh_video_t *sh){
|
|||
if (lavc_codec->capabilities & CODEC_CAP_HWACCEL_VDPAU)
|
||||
mp_msg(MSGT_DECVIDEO, MSGL_V, "[VD_FFMPEG] VDPAU hardware "
|
||||
"decoding.\n");
|
||||
avctx->slice_flags = SLICE_FLAG_CODED_ORDER|SLICE_FLAG_ALLOW_FIELD;
|
||||
avctx->slice_flags = SLICE_FLAG_CODED_ORDER | SLICE_FLAG_ALLOW_FIELD;
|
||||
}
|
||||
|
||||
if (lavc_param->threads == 0) {
|
||||
|
@ -226,33 +234,33 @@ static int init(sh_video_t *sh){
|
|||
"%d threads if supported.\n", lavc_param->threads);
|
||||
}
|
||||
|
||||
if(ctx->do_dr1){
|
||||
avctx->flags|= CODEC_FLAG_EMU_EDGE;
|
||||
avctx->get_buffer= get_buffer;
|
||||
avctx->release_buffer= release_buffer;
|
||||
avctx->reget_buffer= get_buffer;
|
||||
if (ctx->do_dr1) {
|
||||
avctx->flags |= CODEC_FLAG_EMU_EDGE;
|
||||
avctx->get_buffer = get_buffer;
|
||||
avctx->release_buffer = release_buffer;
|
||||
avctx->reget_buffer = get_buffer;
|
||||
}
|
||||
|
||||
avctx->flags|= lavc_param->bitexact;
|
||||
avctx->flags |= lavc_param->bitexact;
|
||||
|
||||
avctx->coded_width = sh->disp_w;
|
||||
avctx->coded_height= sh->disp_h;
|
||||
avctx->workaround_bugs= lavc_param->workaround_bugs;
|
||||
avctx->error_recognition= lavc_param->error_resilience;
|
||||
if(lavc_param->gray) avctx->flags|= CODEC_FLAG_GRAY;
|
||||
avctx->flags2|= lavc_param->fast;
|
||||
avctx->codec_tag= sh->format;
|
||||
avctx->stream_codec_tag= sh->video.fccHandler;
|
||||
avctx->idct_algo= lavc_param->idct_algo;
|
||||
avctx->error_concealment= lavc_param->error_concealment;
|
||||
avctx->debug= lavc_param->debug;
|
||||
avctx->coded_height = sh->disp_h;
|
||||
avctx->workaround_bugs = lavc_param->workaround_bugs;
|
||||
avctx->error_recognition = lavc_param->error_resilience;
|
||||
if (lavc_param->gray)
|
||||
avctx->flags |= CODEC_FLAG_GRAY;
|
||||
avctx->flags2 |= lavc_param->fast;
|
||||
avctx->codec_tag = sh->format;
|
||||
avctx->stream_codec_tag = sh->video.fccHandler;
|
||||
avctx->idct_algo = lavc_param->idct_algo;
|
||||
avctx->error_concealment = lavc_param->error_concealment;
|
||||
avctx->debug = lavc_param->debug;
|
||||
if (lavc_param->debug)
|
||||
av_log_set_level(AV_LOG_DEBUG);
|
||||
avctx->debug_mv= lavc_param->vismv;
|
||||
avctx->debug_mv = lavc_param->vismv;
|
||||
avctx->skip_top = lavc_param->skip_top;
|
||||
avctx->skip_bottom= lavc_param->skip_bottom;
|
||||
if(lavc_param->lowres_str != NULL)
|
||||
{
|
||||
avctx->skip_bottom = lavc_param->skip_bottom;
|
||||
if (lavc_param->lowres_str != NULL) {
|
||||
int lowres, lowres_w;
|
||||
sscanf(lavc_param->lowres_str, "%d,%d", &lowres, &lowres_w);
|
||||
if (lowres < 1 || lowres > 16 ||
|
||||
|
@ -264,9 +272,11 @@ static int init(sh_video_t *sh){
|
|||
avctx->skip_idct = str2AVDiscard(lavc_param->skip_idct_str);
|
||||
avctx->skip_frame = str2AVDiscard(lavc_param->skip_frame_str);
|
||||
|
||||
if(lavc_param->avopt){
|
||||
if(parse_avopts(avctx, lavc_param->avopt) < 0){
|
||||
mp_msg(MSGT_DECVIDEO, MSGL_ERR, "Your options /%s/ look like gibberish to me pal\n", lavc_param->avopt);
|
||||
if (lavc_param->avopt) {
|
||||
if (parse_avopts(avctx, lavc_param->avopt) < 0) {
|
||||
mp_msg(MSGT_DECVIDEO, MSGL_ERR,
|
||||
"Your options /%s/ look like gibberish to me pal\n",
|
||||
lavc_param->avopt);
|
||||
uninit(sh);
|
||||
return 0;
|
||||
}
|
||||
|
@ -275,93 +285,88 @@ static int init(sh_video_t *sh){
|
|||
// Do this after the above avopt handling in case it changes values
|
||||
ctx->skip_frame = avctx->skip_frame;
|
||||
|
||||
mp_dbg(MSGT_DECVIDEO, MSGL_DBG2, "libavcodec.size: %d x %d\n", avctx->width, avctx->height);
|
||||
mp_dbg(MSGT_DECVIDEO, MSGL_DBG2,
|
||||
"libavcodec.size: %d x %d\n", avctx->width, avctx->height);
|
||||
switch (sh->format) {
|
||||
case mmioFOURCC('S','V','Q','3'):
|
||||
/* SVQ3 extradata can show up as sh->ImageDesc if demux_mov is used, or
|
||||
in the phony AVI header if demux_lavf is used. The first case is
|
||||
handled here; the second case falls through to the next section. */
|
||||
/* SVQ3 extradata can show up as sh->ImageDesc if demux_mov is used, or
|
||||
* in the phony AVI header if demux_lavf is used. The first case is
|
||||
* handled here; the second case falls through to the next section. */
|
||||
if (sh->ImageDesc) {
|
||||
avctx->extradata_size = (*(int *)sh->ImageDesc) - sizeof(int);
|
||||
avctx->extradata = av_mallocz(avctx->extradata_size + FF_INPUT_BUFFER_PADDING_SIZE);
|
||||
memcpy(avctx->extradata, ((int *)sh->ImageDesc)+1, avctx->extradata_size);
|
||||
avctx->extradata = av_mallocz(avctx->extradata_size +
|
||||
FF_INPUT_BUFFER_PADDING_SIZE);
|
||||
memcpy(avctx->extradata, ((int *)sh->ImageDesc) + 1,
|
||||
avctx->extradata_size);
|
||||
break;
|
||||
}
|
||||
/* fallthrough */
|
||||
/* fallthrough */
|
||||
|
||||
case mmioFOURCC('A','V','R','n'):
|
||||
case mmioFOURCC('M','J','P','G'):
|
||||
/* AVRn stores huffman table in AVI header */
|
||||
/* Pegasus MJPEG stores it also in AVI header, but it uses the common
|
||||
MJPG fourcc :( */
|
||||
/* AVRn stores huffman table in AVI header */
|
||||
/* Pegasus MJPEG stores it also in AVI header, but it uses the common
|
||||
* MJPG fourcc :( */
|
||||
if (!sh->bih || sh->bih->biSize <= sizeof(*sh->bih))
|
||||
break;
|
||||
avctx->flags |= CODEC_FLAG_EXTERN_HUFF;
|
||||
avctx->extradata_size = sh->bih->biSize-sizeof(*sh->bih);
|
||||
avctx->extradata = av_mallocz(avctx->extradata_size + FF_INPUT_BUFFER_PADDING_SIZE);
|
||||
memcpy(avctx->extradata, sh->bih+1, avctx->extradata_size);
|
||||
|
||||
#if 0
|
||||
{
|
||||
int x;
|
||||
uint8_t *p = avctx->extradata;
|
||||
|
||||
for (x=0; x<avctx->extradata_size; x++)
|
||||
mp_msg(MSGT_DECVIDEO, MSGL_INFO, "[%x] ", p[x]);
|
||||
mp_msg(MSGT_DECVIDEO, MSGL_INFO, "\n");
|
||||
}
|
||||
#endif
|
||||
avctx->extradata_size = sh->bih->biSize - sizeof(*sh->bih);
|
||||
avctx->extradata = av_mallocz(avctx->extradata_size +
|
||||
FF_INPUT_BUFFER_PADDING_SIZE);
|
||||
memcpy(avctx->extradata, sh->bih + 1, avctx->extradata_size);
|
||||
break;
|
||||
|
||||
case mmioFOURCC('R', 'V', '1', '0'):
|
||||
case mmioFOURCC('R', 'V', '1', '3'):
|
||||
case mmioFOURCC('R', 'V', '2', '0'):
|
||||
case mmioFOURCC('R', 'V', '3', '0'):
|
||||
case mmioFOURCC('R', 'V', '4', '0'):
|
||||
if(sh->bih->biSize<sizeof(*sh->bih)+8){
|
||||
/* only 1 packet per frame & sub_id from fourcc */
|
||||
avctx->extradata_size= 8;
|
||||
avctx->extradata = av_mallocz(avctx->extradata_size + FF_INPUT_BUFFER_PADDING_SIZE);
|
||||
case mmioFOURCC('R','V','1','0'):
|
||||
case mmioFOURCC('R','V','1','3'):
|
||||
case mmioFOURCC('R','V','2','0'):
|
||||
case mmioFOURCC('R','V','3','0'):
|
||||
case mmioFOURCC('R','V','4','0'):
|
||||
if (sh->bih->biSize < sizeof(*sh->bih) + 8) {
|
||||
// only 1 packet per frame & sub_id from fourcc
|
||||
avctx->extradata_size = 8;
|
||||
avctx->extradata = av_mallocz(avctx->extradata_size +
|
||||
FF_INPUT_BUFFER_PADDING_SIZE);
|
||||
((uint32_t *)avctx->extradata)[0] = 0;
|
||||
((uint32_t *)avctx->extradata)[1] =
|
||||
(sh->format == mmioFOURCC('R', 'V', '1', '3')) ? 0x10003001 : 0x10000000;
|
||||
sh->format == mmioFOURCC('R','V','1','3') ?
|
||||
0x10003001 : 0x10000000;
|
||||
} else {
|
||||
/* has extra slice header (demux_rm or rm->avi streamcopy) */
|
||||
avctx->extradata_size = sh->bih->biSize-sizeof(*sh->bih);
|
||||
avctx->extradata = av_mallocz(avctx->extradata_size + FF_INPUT_BUFFER_PADDING_SIZE);
|
||||
memcpy(avctx->extradata, sh->bih+1, avctx->extradata_size);
|
||||
// has extra slice header (demux_rm or rm->avi streamcopy)
|
||||
avctx->extradata_size = sh->bih->biSize - sizeof(*sh->bih);
|
||||
avctx->extradata = av_mallocz(avctx->extradata_size +
|
||||
FF_INPUT_BUFFER_PADDING_SIZE);
|
||||
memcpy(avctx->extradata, sh->bih + 1, avctx->extradata_size);
|
||||
}
|
||||
avctx->sub_id= AV_RB32(avctx->extradata+4);
|
||||
|
||||
// printf("%X %X %d %d\n", extrahdr[0], extrahdr[1]);
|
||||
avctx->sub_id = AV_RB32(avctx->extradata + 4);
|
||||
break;
|
||||
|
||||
default:
|
||||
if (!sh->bih || sh->bih->biSize <= sizeof(*sh->bih))
|
||||
break;
|
||||
avctx->extradata_size = sh->bih->biSize-sizeof(*sh->bih);
|
||||
avctx->extradata = av_mallocz(avctx->extradata_size + FF_INPUT_BUFFER_PADDING_SIZE);
|
||||
memcpy(avctx->extradata, sh->bih+1, avctx->extradata_size);
|
||||
avctx->extradata_size = sh->bih->biSize - sizeof(*sh->bih);
|
||||
avctx->extradata = av_mallocz(avctx->extradata_size +
|
||||
FF_INPUT_BUFFER_PADDING_SIZE);
|
||||
memcpy(avctx->extradata, sh->bih + 1, avctx->extradata_size);
|
||||
break;
|
||||
}
|
||||
/* Pass palette to codec */
|
||||
if (sh->bih && (sh->bih->biBitCount <= 8)) {
|
||||
avctx->palctrl = calloc(1, sizeof(AVPaletteControl));
|
||||
avctx->palctrl->palette_changed = 1;
|
||||
if (sh->bih->biSize-sizeof(*sh->bih))
|
||||
if (sh->bih->biSize - sizeof(*sh->bih))
|
||||
/* Palette size in biSize */
|
||||
memcpy(avctx->palctrl->palette, sh->bih+1,
|
||||
FFMIN(sh->bih->biSize-sizeof(*sh->bih), AVPALETTE_SIZE));
|
||||
memcpy(avctx->palctrl->palette, sh->bih + 1,
|
||||
FFMIN(sh->bih->biSize - sizeof(*sh->bih), AVPALETTE_SIZE));
|
||||
else
|
||||
/* Palette size in biClrUsed */
|
||||
memcpy(avctx->palctrl->palette, sh->bih+1,
|
||||
memcpy(avctx->palctrl->palette, sh->bih + 1,
|
||||
FFMIN(sh->bih->biClrUsed * 4, AVPALETTE_SIZE));
|
||||
}
|
||||
}
|
||||
|
||||
if(sh->bih)
|
||||
avctx->bits_per_coded_sample= sh->bih->biBitCount;
|
||||
if (sh->bih)
|
||||
avctx->bits_per_coded_sample = sh->bih->biBitCount;
|
||||
|
||||
if(lavc_param->threads > 1) {
|
||||
if (lavc_param->threads > 1) {
|
||||
avctx->thread_count = lavc_param->threads;
|
||||
avcodec_thread_init(avctx, lavc_param->threads);
|
||||
}
|
||||
|
@ -375,20 +380,19 @@ static int init(sh_video_t *sh){
|
|||
return 1; //mpcodecs_config_vo(sh, sh->disp_w, sh->disp_h, IMGFMT_YV12);
|
||||
}
|
||||
|
||||
// uninit driver
|
||||
static void uninit(sh_video_t *sh){
|
||||
static void uninit(sh_video_t *sh)
|
||||
{
|
||||
vd_ffmpeg_ctx *ctx = sh->context;
|
||||
AVCodecContext *avctx = ctx->avctx;
|
||||
|
||||
if (sh->opts->lavc_param.vstats && avctx->coded_frame) {
|
||||
int i;
|
||||
for(i=1; i<32; i++){
|
||||
mp_msg(MSGT_DECVIDEO, MSGL_INFO, "QP: %d, count: %d\n", i, ctx->qp_stat[i]);
|
||||
}
|
||||
mp_tmsg(MSGT_DECVIDEO, MSGL_INFO, "[VD_FFMPEG] Arithmetic mean of QP: %2.4f, Harmonic mean of QP: %2.4f\n",
|
||||
for (int i = 1; i < 32; i++)
|
||||
mp_msg(MSGT_DECVIDEO, MSGL_INFO,
|
||||
"QP: %d, count: %d\n", i, ctx->qp_stat[i]);
|
||||
mp_tmsg(MSGT_DECVIDEO, MSGL_INFO, "[VD_FFMPEG] Arithmetic mean of QP: "
|
||||
"%2.4f, Harmonic mean of QP: %2.4f\n",
|
||||
ctx->qp_sum / avctx->coded_frame->coded_picture_number,
|
||||
1.0/(ctx->inv_qp_sum / avctx->coded_frame->coded_picture_number)
|
||||
);
|
||||
1.0 / (ctx->inv_qp_sum / avctx->coded_frame->coded_picture_number));
|
||||
}
|
||||
|
||||
if (avctx) {
|
||||
|
@ -407,55 +411,38 @@ static void uninit(sh_video_t *sh){
|
|||
|
||||
static void draw_slice(struct AVCodecContext *s,
|
||||
const AVFrame *src, int offset[4],
|
||||
int y, int type, int height){
|
||||
int y, int type, int height)
|
||||
{
|
||||
sh_video_t *sh = s->opaque;
|
||||
uint8_t *source[MP_MAX_PLANES]= {src->data[0] + offset[0], src->data[1] + offset[1], src->data[2] + offset[2]};
|
||||
int strides[MP_MAX_PLANES] = {src->linesize[0], src->linesize[1], src->linesize[2]};
|
||||
#if 0
|
||||
int start=0, i;
|
||||
int width= s->width;
|
||||
vd_ffmpeg_ctx *ctx = sh->context;
|
||||
int skip_stride = ((width << s->lowres)+15) >> 4;
|
||||
uint8_t *skip= &s->coded_frame->mbskip_table[(y>>4)*skip_stride];
|
||||
int threshold= s->coded_frame->age;
|
||||
if(s->pict_type!=B_TYPE){
|
||||
for(i=0; i*16<width+16; i++){
|
||||
if(i*16>=width || skip[i]>=threshold){
|
||||
if(start==i) start++;
|
||||
else{
|
||||
uint8_t *src2[3]= {src[0] + start*16,
|
||||
src[1] + start*8,
|
||||
src[2] + start*8};
|
||||
//printf("%2d-%2d x %d\n", start, i, y);
|
||||
mpcodecs_draw_slice (sh, src2, stride, (i-start)*16, height, start*16, y);
|
||||
start= i+1;
|
||||
}
|
||||
}
|
||||
}
|
||||
}else
|
||||
#endif
|
||||
if (height < 0)
|
||||
{
|
||||
uint8_t *source[MP_MAX_PLANES] = {
|
||||
src->data[0] + offset[0], src->data[1] + offset[1],
|
||||
src->data[2] + offset[2]
|
||||
};
|
||||
int strides[MP_MAX_PLANES] = {
|
||||
src->linesize[0], src->linesize[1], src->linesize[2]
|
||||
};
|
||||
if (height < 0) {
|
||||
int i;
|
||||
height = -height;
|
||||
y -= height;
|
||||
for (i = 0; i < MP_MAX_PLANES; i++)
|
||||
{
|
||||
for (i = 0; i < MP_MAX_PLANES; i++) {
|
||||
strides[i] = -strides[i];
|
||||
source[i] -= strides[i];
|
||||
}
|
||||
}
|
||||
if (y < sh->disp_h) {
|
||||
height = FFMIN(height, sh->disp_h-y);
|
||||
mpcodecs_draw_slice (sh, source, strides, sh->disp_w, height, 0, y);
|
||||
height = FFMIN(height, sh->disp_h - y);
|
||||
mpcodecs_draw_slice(sh, source, strides, sh->disp_w, height, 0, y);
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
static int init_vo(sh_video_t *sh, enum PixelFormat pix_fmt){
|
||||
static int init_vo(sh_video_t *sh, enum PixelFormat pix_fmt)
|
||||
{
|
||||
vd_ffmpeg_ctx *ctx = sh->context;
|
||||
AVCodecContext *avctx = ctx->avctx;
|
||||
float aspect= av_q2d(avctx->sample_aspect_ratio) * avctx->width / avctx->height;
|
||||
float aspect = av_q2d(avctx->sample_aspect_ratio) *
|
||||
avctx->width / avctx->height;
|
||||
int width, height;
|
||||
|
||||
width = avctx->width;
|
||||
|
@ -469,14 +456,14 @@ static int init_vo(sh_video_t *sh, enum PixelFormat pix_fmt){
|
|||
height = sh->bih->biHeight >> avctx->lowres;
|
||||
}
|
||||
|
||||
// it is possible another vo buffers to be used after vo config()
|
||||
// lavc reset its buffers on width/heigh change but not on aspect change!!!
|
||||
/* Reconfiguring filter/VO chain may invalidate direct rendering buffers
|
||||
* we have allocated for libavcodec (including the VDPAU HW decoding
|
||||
* case). Is it guaranteed that the code below only triggers in a situation
|
||||
* with no busy direct rendering buffers for reference frames?
|
||||
*/
|
||||
if (av_cmp_q(avctx->sample_aspect_ratio, ctx->last_sample_aspect_ratio) ||
|
||||
width != sh->disp_w ||
|
||||
height != sh->disp_h ||
|
||||
pix_fmt != ctx->pix_fmt ||
|
||||
!ctx->vo_initialized)
|
||||
{
|
||||
width != sh->disp_w || height != sh->disp_h ||
|
||||
pix_fmt != ctx->pix_fmt || !ctx->vo_initialized) {
|
||||
ctx->vo_initialized = 0;
|
||||
mp_msg(MSGT_DECVIDEO, MSGL_V, "[ffmpeg] aspect_ratio: %f\n", aspect);
|
||||
|
||||
|
@ -494,12 +481,13 @@ static int init_vo(sh_video_t *sh, enum PixelFormat pix_fmt){
|
|||
ctx->best_csp = pixfmt2imgfmt(pix_fmt);
|
||||
const unsigned int *supported_fmts;
|
||||
if (ctx->best_csp == IMGFMT_YV12)
|
||||
supported_fmts = (const unsigned int[])
|
||||
{IMGFMT_YV12, IMGFMT_I420, IMGFMT_IYUV, 0xffffffff};
|
||||
supported_fmts = (const unsigned int[]){
|
||||
IMGFMT_YV12, IMGFMT_I420, IMGFMT_IYUV, 0xffffffff
|
||||
};
|
||||
else if (ctx->best_csp == IMGFMT_422P)
|
||||
supported_fmts = (const unsigned int[])
|
||||
{IMGFMT_422P, IMGFMT_YV12, IMGFMT_I420, IMGFMT_IYUV,
|
||||
0xffffffff};
|
||||
supported_fmts = (const unsigned int[]){
|
||||
IMGFMT_422P, IMGFMT_YV12, IMGFMT_I420, IMGFMT_IYUV, 0xffffffff
|
||||
};
|
||||
else
|
||||
supported_fmts = (const unsigned int[]){ctx->best_csp, 0xffffffff};
|
||||
if (!mpcodecs_config_vo2(sh, sh->disp_w, sh->disp_h, supported_fmts,
|
||||
|
@ -510,22 +498,24 @@ static int init_vo(sh_video_t *sh, enum PixelFormat pix_fmt){
|
|||
return 0;
|
||||
}
|
||||
|
||||
static int get_buffer(AVCodecContext *avctx, AVFrame *pic){
|
||||
static int get_buffer(AVCodecContext *avctx, AVFrame *pic)
|
||||
{
|
||||
sh_video_t *sh = avctx->opaque;
|
||||
vd_ffmpeg_ctx *ctx = sh->context;
|
||||
mp_image_t *mpi=NULL;
|
||||
int flags= MP_IMGFLAG_ACCEPT_ALIGNED_STRIDE | MP_IMGFLAG_PREFER_ALIGNED_STRIDE;
|
||||
int type= MP_IMGTYPE_IPB;
|
||||
int width= avctx->width;
|
||||
int height= avctx->height;
|
||||
mp_image_t *mpi = NULL;
|
||||
int flags = MP_IMGFLAG_ACCEPT_ALIGNED_STRIDE |
|
||||
MP_IMGFLAG_PREFER_ALIGNED_STRIDE;
|
||||
int type = MP_IMGTYPE_IPB;
|
||||
int width = avctx->width;
|
||||
int height = avctx->height;
|
||||
// special case to handle reget_buffer without buffer hints
|
||||
if (pic->opaque && pic->data[0] && !pic->buffer_hints)
|
||||
return 0;
|
||||
avcodec_align_dimensions(avctx, &width, &height);
|
||||
//printf("get_buffer %d %d %d\n", pic->reference, ctx->ip_count, ctx->b_count);
|
||||
|
||||
if (pic->buffer_hints) {
|
||||
mp_msg(MSGT_DECVIDEO, MSGL_DBG2, "Buffer hints: %u\n", pic->buffer_hints);
|
||||
mp_msg(MSGT_DECVIDEO, MSGL_DBG2, "Buffer hints: %u\n",
|
||||
pic->buffer_hints);
|
||||
type = MP_IMGTYPE_TEMP;
|
||||
if (pic->buffer_hints & FF_BUFFER_HINTS_READABLE)
|
||||
flags |= MP_IMGFLAG_READABLE;
|
||||
|
@ -538,120 +528,96 @@ static int get_buffer(AVCodecContext *avctx, AVFrame *pic){
|
|||
flags |= MP_IMGFLAG_PRESERVE;
|
||||
}
|
||||
flags |= ctx->do_slices ? MP_IMGFLAG_DRAW_CALLBACK : 0;
|
||||
mp_msg(MSGT_DECVIDEO, MSGL_DBG2, type == MP_IMGTYPE_STATIC ? "using STATIC\n" : "using TEMP\n");
|
||||
mp_msg(MSGT_DECVIDEO, MSGL_DBG2,
|
||||
type == MP_IMGTYPE_STATIC ? "using STATIC\n" : "using TEMP\n");
|
||||
} else {
|
||||
if(!pic->reference){
|
||||
if (!pic->reference) {
|
||||
ctx->b_count++;
|
||||
flags |= ctx->do_slices ? MP_IMGFLAG_DRAW_CALLBACK:0;
|
||||
}else{
|
||||
flags |= ctx->do_slices ? MP_IMGFLAG_DRAW_CALLBACK : 0;
|
||||
} else {
|
||||
ctx->ip_count++;
|
||||
flags|= MP_IMGFLAG_PRESERVE|MP_IMGFLAG_READABLE
|
||||
| (ctx->do_slices ? MP_IMGFLAG_DRAW_CALLBACK : 0);
|
||||
flags |= MP_IMGFLAG_PRESERVE | MP_IMGFLAG_READABLE
|
||||
| (ctx->do_slices ? MP_IMGFLAG_DRAW_CALLBACK : 0);
|
||||
}
|
||||
}
|
||||
|
||||
if(init_vo(sh, avctx->pix_fmt) < 0){
|
||||
avctx->release_buffer= avcodec_default_release_buffer;
|
||||
avctx->get_buffer= avcodec_default_get_buffer;
|
||||
avctx->reget_buffer= avcodec_default_reget_buffer;
|
||||
if (init_vo(sh, avctx->pix_fmt) < 0) {
|
||||
avctx->release_buffer = avcodec_default_release_buffer;
|
||||
avctx->get_buffer = avcodec_default_get_buffer;
|
||||
avctx->reget_buffer = avcodec_default_reget_buffer;
|
||||
if (pic->data[0])
|
||||
release_buffer(avctx, pic);
|
||||
return avctx->get_buffer(avctx, pic);
|
||||
}
|
||||
|
||||
if (IMGFMT_IS_HWACCEL(ctx->best_csp)) {
|
||||
if (IMGFMT_IS_HWACCEL(ctx->best_csp))
|
||||
type = MP_IMGTYPE_NUMBERED | (0xffff << 16);
|
||||
} else
|
||||
if (!pic->buffer_hints) {
|
||||
if(ctx->b_count>1 || ctx->ip_count>2){
|
||||
else if (!pic->buffer_hints) {
|
||||
if (ctx->b_count > 1 || ctx->ip_count > 2) {
|
||||
mp_tmsg(MSGT_DECVIDEO, MSGL_WARN, "[VD_FFMPEG] DRI failure.\n");
|
||||
|
||||
ctx->do_dr1=0; //FIXME
|
||||
avctx->get_buffer= avcodec_default_get_buffer;
|
||||
avctx->reget_buffer= avcodec_default_reget_buffer;
|
||||
ctx->do_dr1 = 0; //FIXME
|
||||
avctx->get_buffer = avcodec_default_get_buffer;
|
||||
avctx->reget_buffer = avcodec_default_reget_buffer;
|
||||
if (pic->data[0])
|
||||
release_buffer(avctx, pic);
|
||||
return avctx->get_buffer(avctx, pic);
|
||||
}
|
||||
|
||||
if(avctx->has_b_frames || ctx->b_count){
|
||||
type= MP_IMGTYPE_IPB;
|
||||
}else{
|
||||
type= MP_IMGTYPE_IP;
|
||||
}
|
||||
mp_msg(MSGT_DECVIDEO, MSGL_DBG2, type== MP_IMGTYPE_IPB ? "using IPB\n" : "using IP\n");
|
||||
if (avctx->has_b_frames || ctx->b_count)
|
||||
type = MP_IMGTYPE_IPB;
|
||||
else
|
||||
type = MP_IMGTYPE_IP;
|
||||
mp_msg(MSGT_DECVIDEO, MSGL_DBG2,
|
||||
type == MP_IMGTYPE_IPB ? "using IPB\n" : "using IP\n");
|
||||
}
|
||||
|
||||
if (ctx->best_csp == IMGFMT_RGB8 || ctx->best_csp == IMGFMT_BGR8)
|
||||
flags |= MP_IMGFLAG_RGB_PALETTE;
|
||||
mpi= mpcodecs_get_image(sh, type, flags, width, height);
|
||||
if (!mpi) return -1;
|
||||
mpi = mpcodecs_get_image(sh, type, flags, width, height);
|
||||
if (!mpi)
|
||||
return -1;
|
||||
|
||||
// ok, let's see what did we get:
|
||||
if(mpi->flags&MP_IMGFLAG_DRAW_CALLBACK &&
|
||||
!(mpi->flags&MP_IMGFLAG_DIRECT)){
|
||||
if (mpi->flags & MP_IMGFLAG_DRAW_CALLBACK &&
|
||||
!(mpi->flags & MP_IMGFLAG_DIRECT)) {
|
||||
// nice, filter/vo likes draw_callback :)
|
||||
avctx->draw_horiz_band= draw_slice;
|
||||
avctx->draw_horiz_band = draw_slice;
|
||||
} else
|
||||
avctx->draw_horiz_band= NULL;
|
||||
if(IMGFMT_IS_HWACCEL(mpi->imgfmt)) {
|
||||
avctx->draw_horiz_band= draw_slice;
|
||||
}
|
||||
avctx->draw_horiz_band = NULL;
|
||||
if (IMGFMT_IS_HWACCEL(mpi->imgfmt))
|
||||
avctx->draw_horiz_band = draw_slice;
|
||||
|
||||
pic->data[0]= mpi->planes[0];
|
||||
pic->data[1]= mpi->planes[1];
|
||||
pic->data[2]= mpi->planes[2];
|
||||
pic->data[3]= mpi->planes[3];
|
||||
pic->data[0] = mpi->planes[0];
|
||||
pic->data[1] = mpi->planes[1];
|
||||
pic->data[2] = mpi->planes[2];
|
||||
pic->data[3] = mpi->planes[3];
|
||||
|
||||
#if 0
|
||||
assert(mpi->width >= ((width +align)&(~align)));
|
||||
assert(mpi->height >= ((height+align)&(~align)));
|
||||
assert(mpi->stride[0] >= mpi->width);
|
||||
if(mpi->imgfmt==IMGFMT_I420 || mpi->imgfmt==IMGFMT_YV12 || mpi->imgfmt==IMGFMT_IYUV){
|
||||
const int y_size= mpi->stride[0] * (mpi->h-1) + mpi->w;
|
||||
const int c_size= mpi->stride[1] * ((mpi->h>>1)-1) + (mpi->w>>1);
|
||||
|
||||
assert(mpi->planes[0] > mpi->planes[1] || mpi->planes[0] + y_size <= mpi->planes[1]);
|
||||
assert(mpi->planes[0] > mpi->planes[2] || mpi->planes[0] + y_size <= mpi->planes[2]);
|
||||
assert(mpi->planes[1] > mpi->planes[0] || mpi->planes[1] + c_size <= mpi->planes[0]);
|
||||
assert(mpi->planes[1] > mpi->planes[2] || mpi->planes[1] + c_size <= mpi->planes[2]);
|
||||
assert(mpi->planes[2] > mpi->planes[0] || mpi->planes[2] + c_size <= mpi->planes[0]);
|
||||
assert(mpi->planes[2] > mpi->planes[1] || mpi->planes[2] + c_size <= mpi->planes[1]);
|
||||
}
|
||||
#endif
|
||||
|
||||
/* Note, some (many) codecs in libavcodec must have stride1==stride2 && no changes between frames
|
||||
* lavc will check that and die with an error message, if its not true
|
||||
/* Note: some (many) codecs in libavcodec require
|
||||
* linesize[1] == linesize[2] and no changes between frames.
|
||||
* Lavc will check that and die with an error message if it's not true.
|
||||
*/
|
||||
pic->linesize[0]= mpi->stride[0];
|
||||
pic->linesize[1]= mpi->stride[1];
|
||||
pic->linesize[2]= mpi->stride[2];
|
||||
pic->linesize[3]= mpi->stride[3];
|
||||
pic->linesize[0] = mpi->stride[0];
|
||||
pic->linesize[1] = mpi->stride[1];
|
||||
pic->linesize[2] = mpi->stride[2];
|
||||
pic->linesize[3] = mpi->stride[3];
|
||||
|
||||
pic->opaque = mpi;
|
||||
//printf("%X\n", (int)mpi->planes[0]);
|
||||
#if 0
|
||||
if(mpi->flags&MP_IMGFLAG_DIRECT)
|
||||
printf("D");
|
||||
else if(mpi->flags&MP_IMGFLAG_DRAW_CALLBACK)
|
||||
printf("S");
|
||||
else
|
||||
printf(".");
|
||||
#endif
|
||||
if(pic->reference){
|
||||
pic->age= ctx->ip_age[0];
|
||||
|
||||
ctx->ip_age[0]= ctx->ip_age[1]+1;
|
||||
ctx->ip_age[1]= 1;
|
||||
if (pic->reference) {
|
||||
pic->age = ctx->ip_age[0];
|
||||
|
||||
ctx->ip_age[0] = ctx->ip_age[1] + 1;
|
||||
ctx->ip_age[1] = 1;
|
||||
ctx->b_age++;
|
||||
}else{
|
||||
pic->age= ctx->b_age;
|
||||
} else {
|
||||
pic->age = ctx->b_age;
|
||||
|
||||
ctx->ip_age[0]++;
|
||||
ctx->ip_age[1]++;
|
||||
ctx->b_age=1;
|
||||
ctx->b_age = 1;
|
||||
}
|
||||
pic->type= FF_BUFFER_TYPE_USER;
|
||||
pic->type = FF_BUFFER_TYPE_USER;
|
||||
|
||||
/* The libavcodec reordered_opaque functionality is implemented by
|
||||
* a similar copy in avcodec_default_get_buffer() and without a
|
||||
|
@ -662,16 +628,14 @@ else
|
|||
return 0;
|
||||
}
|
||||
|
||||
static void release_buffer(struct AVCodecContext *avctx, AVFrame *pic){
|
||||
mp_image_t *mpi= pic->opaque;
|
||||
static void release_buffer(struct AVCodecContext *avctx, AVFrame *pic)
|
||||
{
|
||||
mp_image_t *mpi = pic->opaque;
|
||||
sh_video_t *sh = avctx->opaque;
|
||||
vd_ffmpeg_ctx *ctx = sh->context;
|
||||
int i;
|
||||
|
||||
//printf("release buffer %d %d %d\n", mpi ? mpi->flags&MP_IMGFLAG_PRESERVE : -99, ctx->ip_count, ctx->b_count);
|
||||
|
||||
if(ctx->ip_count <= 2 && ctx->b_count<=1){
|
||||
if(mpi->flags&MP_IMGFLAG_PRESERVE)
|
||||
if (ctx->ip_count <= 2 && ctx->b_count <= 1) {
|
||||
if (mpi->flags & MP_IMGFLAG_PRESERVE)
|
||||
ctx->ip_count--;
|
||||
else
|
||||
ctx->b_count--;
|
||||
|
@ -685,26 +649,15 @@ static void release_buffer(struct AVCodecContext *avctx, AVFrame *pic){
|
|||
mpi->usage_count--;
|
||||
}
|
||||
|
||||
if(pic->type!=FF_BUFFER_TYPE_USER){
|
||||
if (pic->type != FF_BUFFER_TYPE_USER) {
|
||||
avcodec_default_release_buffer(avctx, pic);
|
||||
return;
|
||||
}
|
||||
|
||||
for(i=0; i<4; i++){
|
||||
pic->data[i]= NULL;
|
||||
}
|
||||
//printf("R%X %X\n", pic->linesize[0], pic->data[0]);
|
||||
for (int i = 0; i < 4; i++)
|
||||
pic->data[i] = NULL;
|
||||
}
|
||||
|
||||
// copypaste from demux_real.c - it should match to get it working!
|
||||
//FIXME put into some header
|
||||
typedef struct dp_hdr_s {
|
||||
uint32_t chunks; // number of chunks
|
||||
uint32_t timestamp; // timestamp from packet header
|
||||
uint32_t len; // length of actual data
|
||||
uint32_t chunktab; // offset to chunk offset array
|
||||
} dp_hdr_t;
|
||||
|
||||
static av_unused void swap_palette(void *pal)
|
||||
{
|
||||
int i;
|
||||
|
@ -713,33 +666,32 @@ static av_unused void swap_palette(void *pal)
|
|||
p[i] = le2me_32(p[i]);
|
||||
}
|
||||
|
||||
// decode a frame
|
||||
static struct mp_image *decode(struct sh_video *sh, struct demux_packet *packet,
|
||||
void *data, int len, int flags,
|
||||
double *reordered_pts)
|
||||
{
|
||||
int got_picture=0;
|
||||
int got_picture = 0;
|
||||
int ret;
|
||||
vd_ffmpeg_ctx *ctx = sh->context;
|
||||
AVFrame *pic= ctx->pic;
|
||||
AVFrame *pic = ctx->pic;
|
||||
AVCodecContext *avctx = ctx->avctx;
|
||||
struct lavc_param *lavc_param = &sh->opts->lavc_param;
|
||||
mp_image_t *mpi=NULL;
|
||||
int dr1= ctx->do_dr1;
|
||||
mp_image_t *mpi = NULL;
|
||||
int dr1 = ctx->do_dr1;
|
||||
AVPacket pkt;
|
||||
|
||||
if(len<=0) return NULL; // skipped frame
|
||||
if (len <= 0)
|
||||
return NULL; // skipped frame
|
||||
|
||||
//ffmpeg interlace (mpeg2) bug have been fixed. no need of -noslices
|
||||
if (!dr1)
|
||||
avctx->draw_horiz_band=NULL;
|
||||
if(ctx->vo_initialized && !(flags&3) && !dr1){
|
||||
mpi=mpcodecs_get_image(sh, MP_IMGTYPE_EXPORT, MP_IMGFLAG_PRESERVE |
|
||||
(ctx->do_slices?MP_IMGFLAG_DRAW_CALLBACK:0),
|
||||
sh->disp_w, sh->disp_h);
|
||||
if(mpi && mpi->flags&MP_IMGFLAG_DRAW_CALLBACK){
|
||||
avctx->draw_horiz_band = NULL;
|
||||
if (ctx->vo_initialized && !(flags & 3) && !dr1) {
|
||||
mpi = mpcodecs_get_image(sh, MP_IMGTYPE_EXPORT, MP_IMGFLAG_PRESERVE |
|
||||
(ctx->do_slices ? MP_IMGFLAG_DRAW_CALLBACK : 0),
|
||||
sh->disp_w, sh->disp_h);
|
||||
if (mpi && mpi->flags & MP_IMGFLAG_DRAW_CALLBACK) {
|
||||
// vd core likes slices!
|
||||
avctx->draw_horiz_band=draw_slice;
|
||||
avctx->draw_horiz_band = draw_slice;
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -750,8 +702,6 @@ static struct mp_image *decode(struct sh_video *sh, struct demux_packet *packet,
|
|||
else
|
||||
avctx->skip_frame = ctx->skip_frame;
|
||||
|
||||
mp_msg(MSGT_DECVIDEO, MSGL_DBG3, "vd_ffmpeg data: %04x, %04x, %04x, %04x\n",
|
||||
((int *)data)[0], ((int *)data)[1], ((int *)data)[2], ((int *)data)[3]);
|
||||
av_init_packet(&pkt);
|
||||
pkt.data = data;
|
||||
pkt.size = len;
|
||||
|
@ -769,33 +719,33 @@ static struct mp_image *decode(struct sh_video *sh, struct demux_packet *packet,
|
|||
ret = avcodec_decode_video2(avctx, pic, &got_picture, &pkt);
|
||||
*reordered_pts = (union pts){.i = pic->reordered_opaque}.d;
|
||||
|
||||
dr1= ctx->do_dr1;
|
||||
if(ret<0) mp_msg(MSGT_DECVIDEO, MSGL_WARN, "Error while decoding frame!\n");
|
||||
//printf("repeat: %d\n", pic->repeat_pict);
|
||||
//-- vstats generation
|
||||
while(lavc_param->vstats){ // always one time loop
|
||||
static FILE *fvstats=NULL;
|
||||
dr1 = ctx->do_dr1;
|
||||
if (ret < 0)
|
||||
mp_msg(MSGT_DECVIDEO, MSGL_WARN, "Error while decoding frame!\n");
|
||||
//-- vstats generation
|
||||
while (lavc_param->vstats) { // always one time loop
|
||||
static FILE *fvstats = NULL;
|
||||
char filename[20];
|
||||
static long long int all_len=0;
|
||||
static int frame_number=0;
|
||||
static double all_frametime=0.0;
|
||||
AVFrame *pic= avctx->coded_frame;
|
||||
double quality=0.0;
|
||||
static long long int all_len = 0;
|
||||
static int frame_number = 0;
|
||||
static double all_frametime = 0.0;
|
||||
AVFrame *pic = avctx->coded_frame;
|
||||
double quality = 0.0;
|
||||
|
||||
if (!pic)
|
||||
break;
|
||||
|
||||
if(!fvstats) {
|
||||
if (!fvstats) {
|
||||
time_t today2;
|
||||
struct tm *today;
|
||||
today2 = time(NULL);
|
||||
today = localtime(&today2);
|
||||
sprintf(filename, "vstats_%02d%02d%02d.log", today->tm_hour,
|
||||
today->tm_min, today->tm_sec);
|
||||
today->tm_min, today->tm_sec);
|
||||
fvstats = fopen(filename, "w");
|
||||
if(!fvstats) {
|
||||
if (!fvstats) {
|
||||
perror("fopen");
|
||||
lavc_param->vstats=0; // disable block
|
||||
lavc_param->vstats = 0; // disable block
|
||||
break;
|
||||
/*exit(1);*/
|
||||
}
|
||||
|
@ -804,25 +754,25 @@ static struct mp_image *decode(struct sh_video *sh, struct demux_packet *packet,
|
|||
// average MB quantizer
|
||||
{
|
||||
int x, y;
|
||||
int w = ((avctx->width << avctx->lowres)+15) >> 4;
|
||||
int h = ((avctx->height << avctx->lowres)+15) >> 4;
|
||||
int w = ((avctx->width << avctx->lowres) + 15) >> 4;
|
||||
int h = ((avctx->height << avctx->lowres) + 15) >> 4;
|
||||
int8_t *q = pic->qscale_table;
|
||||
for(y = 0; y < h; y++) {
|
||||
for(x = 0; x < w; x++)
|
||||
quality += (double)*(q+x);
|
||||
for (y = 0; y < h; y++) {
|
||||
for (x = 0; x < w; x++)
|
||||
quality += (double)*(q + x);
|
||||
q += pic->qstride;
|
||||
}
|
||||
quality /= w * h;
|
||||
}
|
||||
|
||||
all_len+=len;
|
||||
all_frametime+=sh->frametime;
|
||||
all_len += len;
|
||||
all_frametime += sh->frametime;
|
||||
fprintf(fvstats, "frame= %5d q= %2.2f f_size= %6d s_size= %8.0fkB ",
|
||||
++frame_number, quality, len, (double)all_len/1024);
|
||||
++frame_number, quality, len, (double)all_len / 1024);
|
||||
fprintf(fvstats, "time= %0.3f br= %7.1fkbits/s avg_br= %7.1fkbits/s ",
|
||||
all_frametime, (double)(len*8)/sh->frametime/1000.0,
|
||||
(double)(all_len*8)/all_frametime/1000.0);
|
||||
switch(pic->pict_type){
|
||||
all_frametime, (double)(len * 8) / sh->frametime / 1000.0,
|
||||
(double)(all_len * 8) / all_frametime / 1000.0);
|
||||
switch (pic->pict_type) {
|
||||
case FF_I_TYPE:
|
||||
fprintf(fvstats, "type= I\n");
|
||||
break;
|
||||
|
@ -840,48 +790,50 @@ static struct mp_image *decode(struct sh_video *sh, struct demux_packet *packet,
|
|||
break;
|
||||
}
|
||||
|
||||
ctx->qp_stat[(int)(quality+0.5)]++;
|
||||
ctx->qp_stat[(int)(quality + 0.5)]++;
|
||||
ctx->qp_sum += quality;
|
||||
ctx->inv_qp_sum += 1.0/(double)quality;
|
||||
ctx->inv_qp_sum += 1.0 / (double)quality;
|
||||
|
||||
break;
|
||||
}
|
||||
//--
|
||||
//--
|
||||
|
||||
if(!got_picture) return NULL; // skipped image
|
||||
if (!got_picture)
|
||||
return NULL; // skipped image
|
||||
|
||||
if(init_vo(sh, avctx->pix_fmt) < 0) return NULL;
|
||||
if (init_vo(sh, avctx->pix_fmt) < 0)
|
||||
return NULL;
|
||||
|
||||
if(dr1 && pic->opaque){
|
||||
mpi= (mp_image_t *)pic->opaque;
|
||||
}
|
||||
if (dr1 && pic->opaque)
|
||||
mpi = (mp_image_t *)pic->opaque;
|
||||
|
||||
if(!mpi)
|
||||
mpi=mpcodecs_get_image(sh, MP_IMGTYPE_EXPORT, MP_IMGFLAG_PRESERVE,
|
||||
avctx->width, avctx->height);
|
||||
if(!mpi){ // temporary!
|
||||
mp_tmsg(MSGT_DECVIDEO, MSGL_WARN, "[VD_FFMPEG] Couldn't allocate image for codec.\n");
|
||||
if (!mpi)
|
||||
mpi = mpcodecs_get_image(sh, MP_IMGTYPE_EXPORT, MP_IMGFLAG_PRESERVE,
|
||||
avctx->width, avctx->height);
|
||||
if (!mpi) { // temporary error?
|
||||
mp_tmsg(MSGT_DECVIDEO, MSGL_WARN,
|
||||
"[VD_FFMPEG] Couldn't allocate image for codec.\n");
|
||||
return NULL;
|
||||
}
|
||||
|
||||
if(!dr1){
|
||||
mpi->planes[0]=pic->data[0];
|
||||
mpi->planes[1]=pic->data[1];
|
||||
mpi->planes[2]=pic->data[2];
|
||||
mpi->planes[3]=pic->data[3];
|
||||
mpi->stride[0]=pic->linesize[0];
|
||||
mpi->stride[1]=pic->linesize[1];
|
||||
mpi->stride[2]=pic->linesize[2];
|
||||
mpi->stride[3]=pic->linesize[3];
|
||||
if (!dr1) {
|
||||
mpi->planes[0] = pic->data[0];
|
||||
mpi->planes[1] = pic->data[1];
|
||||
mpi->planes[2] = pic->data[2];
|
||||
mpi->planes[3] = pic->data[3];
|
||||
mpi->stride[0] = pic->linesize[0];
|
||||
mpi->stride[1] = pic->linesize[1];
|
||||
mpi->stride[2] = pic->linesize[2];
|
||||
mpi->stride[3] = pic->linesize[3];
|
||||
}
|
||||
|
||||
if (!mpi->planes[0])
|
||||
return NULL;
|
||||
|
||||
if(ctx->best_csp == IMGFMT_422P && mpi->chroma_y_shift==1){
|
||||
if (ctx->best_csp == IMGFMT_422P && mpi->chroma_y_shift == 1) {
|
||||
// we have 422p but user wants 420p
|
||||
mpi->stride[1]*=2;
|
||||
mpi->stride[2]*=2;
|
||||
mpi->stride[1] *= 2;
|
||||
mpi->stride[2] *= 2;
|
||||
}
|
||||
|
||||
#if HAVE_BIGENDIAN
|
||||
|
@ -889,36 +841,37 @@ static struct mp_image *decode(struct sh_video *sh, struct demux_packet *packet,
|
|||
if (mpi->bpp == 8)
|
||||
swap_palette(mpi->planes[1]);
|
||||
#endif
|
||||
/* to comfirm with newer lavc style */
|
||||
mpi->qscale =pic->qscale_table;
|
||||
mpi->qstride=pic->qstride;
|
||||
mpi->pict_type=pic->pict_type;
|
||||
mpi->qscale_type= pic->qscale_type;
|
||||
|
||||
mpi->qscale = pic->qscale_table;
|
||||
mpi->qstride = pic->qstride;
|
||||
mpi->pict_type = pic->pict_type;
|
||||
mpi->qscale_type = pic->qscale_type;
|
||||
mpi->fields = MP_IMGFIELD_ORDERED;
|
||||
if(pic->interlaced_frame) mpi->fields |= MP_IMGFIELD_INTERLACED;
|
||||
if(pic->top_field_first ) mpi->fields |= MP_IMGFIELD_TOP_FIRST;
|
||||
if(pic->repeat_pict == 1) mpi->fields |= MP_IMGFIELD_REPEAT_FIRST;
|
||||
if (pic->interlaced_frame)
|
||||
mpi->fields |= MP_IMGFIELD_INTERLACED;
|
||||
if (pic->top_field_first)
|
||||
mpi->fields |= MP_IMGFIELD_TOP_FIRST;
|
||||
if (pic->repeat_pict == 1)
|
||||
mpi->fields |= MP_IMGFIELD_REPEAT_FIRST;
|
||||
|
||||
return mpi;
|
||||
}
|
||||
|
||||
static enum PixelFormat get_format(struct AVCodecContext *avctx,
|
||||
const enum PixelFormat *fmt){
|
||||
enum PixelFormat selected_format;
|
||||
int imgfmt;
|
||||
const enum PixelFormat *fmt)
|
||||
{
|
||||
sh_video_t *sh = avctx->opaque;
|
||||
int i;
|
||||
|
||||
for(i=0;fmt[i]!=PIX_FMT_NONE;i++){
|
||||
imgfmt = pixfmt2imgfmt(fmt[i]);
|
||||
if(!IMGFMT_IS_HWACCEL(imgfmt)) continue;
|
||||
for (i = 0; fmt[i] != PIX_FMT_NONE; i++) {
|
||||
int imgfmt = pixfmt2imgfmt(fmt[i]);
|
||||
if (!IMGFMT_IS_HWACCEL(imgfmt))
|
||||
continue;
|
||||
mp_msg(MSGT_DECVIDEO, MSGL_V, "[VD_FFMPEG] Trying pixfmt=%d.\n", i);
|
||||
if(init_vo(sh, fmt[i]) >= 0) {
|
||||
if (init_vo(sh, fmt[i]) >= 0)
|
||||
break;
|
||||
}
|
||||
}
|
||||
selected_format = fmt[i];
|
||||
return selected_format;
|
||||
return fmt[i];
|
||||
}
|
||||
|
||||
const struct vd_functions mpcodecs_vd_ffmpeg = {
|
||||
|
|
Loading…
Reference in New Issue