1
0
mirror of https://github.com/mpv-player/mpv synced 2024-12-14 19:05:33 +00:00
mpv/libmpcodecs/vf_spp.c
michael 52f9770899 spp softthresholding in mmx
git-svn-id: svn://svn.mplayerhq.hu/mplayer/trunk@11995 b3059339-0415-0410-9bf9-f77b7e298cf2
2004-02-24 13:28:42 +00:00

621 lines
18 KiB
C

/*
Copyright (C) 2003 Michael Niedermayer <michaelni@gmx.at>
This program is free software; you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation; either version 2 of the License, or
(at your option) any later version.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program; if not, write to the Free Software
Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
*/
/*
* This implementation is based on an algorithm described in
* "Aria Nosratinia Embedded Post-Processing for
* Enhancement of Compressed Images (1999)"
* (http://citeseer.nj.nec.com/nosratinia99embedded.html)
*/
#include <stdio.h>
#include <stdlib.h>
#include <string.h>
#include <inttypes.h>
#include <math.h>
#include "../config.h"
#ifdef USE_LIBAVCODEC
#include "../mp_msg.h"
#include "../cpudetect.h"
#ifdef USE_LIBAVCODEC_SO
#include <ffmpeg/avcodec.h>
#include <ffmpeg/dsputil.h>
#else
#include "../libavcodec/avcodec.h"
#include "../libavcodec/dsputil.h"
#endif
#ifdef HAVE_MALLOC_H
#include <malloc.h>
#endif
#include "img_format.h"
#include "mp_image.h"
#include "vf.h"
#include "../libvo/fastmemcpy.h"
#define XMIN(a,b) ((a) < (b) ? (a) : (b))
//===========================================================================//
static const uint8_t __attribute__((aligned(8))) dither[8][8]={
{ 0, 48, 12, 60, 3, 51, 15, 63, },
{ 32, 16, 44, 28, 35, 19, 47, 31, },
{ 8, 56, 4, 52, 11, 59, 7, 55, },
{ 40, 24, 36, 20, 43, 27, 39, 23, },
{ 2, 50, 14, 62, 1, 49, 13, 61, },
{ 34, 18, 46, 30, 33, 17, 45, 29, },
{ 10, 58, 6, 54, 9, 57, 5, 53, },
{ 42, 26, 38, 22, 41, 25, 37, 21, },
};
static const uint8_t offset[127][2]= {
{0,0},
{0,0}, {4,4},
{0,0}, {2,2}, {6,4}, {4,6},
{0,0}, {5,1}, {2,2}, {7,3}, {4,4}, {1,5}, {6,6}, {3,7},
{0,0}, {4,0}, {1,1}, {5,1}, {3,2}, {7,2}, {2,3}, {6,3},
{0,4}, {4,4}, {1,5}, {5,5}, {3,6}, {7,6}, {2,7}, {6,7},
{0,0}, {0,2}, {0,4}, {0,6}, {1,1}, {1,3}, {1,5}, {1,7},
{2,0}, {2,2}, {2,4}, {2,6}, {3,1}, {3,3}, {3,5}, {3,7},
{4,0}, {4,2}, {4,4}, {4,6}, {5,1}, {5,3}, {5,5}, {5,7},
{6,0}, {6,2}, {6,4}, {6,6}, {7,1}, {7,3}, {7,5}, {7,7},
{0,0}, {4,4}, {0,4}, {4,0}, {2,2}, {6,6}, {2,6}, {6,2},
{0,2}, {4,6}, {0,6}, {4,2}, {2,0}, {6,4}, {2,4}, {6,0},
{1,1}, {5,5}, {1,5}, {5,1}, {3,3}, {7,7}, {3,7}, {7,3},
{1,3}, {5,7}, {1,7}, {5,3}, {3,1}, {7,5}, {3,5}, {7,1},
{0,1}, {4,5}, {0,5}, {4,1}, {2,3}, {6,7}, {2,7}, {6,3},
{0,3}, {4,7}, {0,7}, {4,3}, {2,1}, {6,5}, {2,5}, {6,1},
{1,0}, {5,4}, {1,4}, {5,0}, {3,2}, {7,6}, {3,6}, {7,2},
{1,2}, {5,6}, {1,6}, {5,2}, {3,0}, {7,4}, {3,4}, {7,0},
};
struct vf_priv_s {
int log2_count;
int qp;
int mode;
int mpeg2;
unsigned int outfmt;
int temp_stride;
uint8_t *src;
int16_t *temp;
AVCodecContext *avctx;
DSPContext dsp;
};
#define SHIFT 22
static void hardthresh_c(DCTELEM dst[64], DCTELEM src[64], int qp, uint8_t *permutation){
int i;
int bias= 0; //FIXME
unsigned int threshold1, threshold2;
threshold1= qp*((1<<4) - bias) - 1;
threshold2= (threshold1<<1);
memset(dst, 0, 64*sizeof(DCTELEM));
dst[0]= (src[0] + 4)>>3;
for(i=1; i<64; i++){
int level= src[i];
if(((unsigned)(level+threshold1))>threshold2){
const int j= permutation[i];
dst[j]= (level + 4)>>3;
}
}
}
static void softthresh_c(DCTELEM dst[64], DCTELEM src[64], int qp, uint8_t *permutation){
int i;
int bias= 0; //FIXME
unsigned int threshold1, threshold2;
threshold1= qp*((1<<4) - bias) - 1;
threshold2= (threshold1<<1);
memset(dst, 0, 64*sizeof(DCTELEM));
dst[0]= (src[0] + 4)>>3;
for(i=1; i<64; i++){
int level= src[i];
if(((unsigned)(level+threshold1))>threshold2){
const int j= permutation[i];
if(level>0)
dst[j]= (level - threshold1 + 4)>>3;
else
dst[j]= (level + threshold1 + 4)>>3;
}
}
}
#ifdef HAVE_MMX
static void hardthresh_mmx(DCTELEM dst[64], DCTELEM src[64], int qp, uint8_t *permutation){
int bias= 0; //FIXME
unsigned int threshold1;
threshold1= qp*((1<<4) - bias) - 1;
asm volatile(
#define REQUANT_CORE(dst0, dst1, dst2, dst3, src0, src1, src2, src3) \
"movq " #src0 ", %%mm0 \n\t"\
"movq " #src1 ", %%mm1 \n\t"\
"movq " #src2 ", %%mm2 \n\t"\
"movq " #src3 ", %%mm3 \n\t"\
"psubw %%mm4, %%mm0 \n\t"\
"psubw %%mm4, %%mm1 \n\t"\
"psubw %%mm4, %%mm2 \n\t"\
"psubw %%mm4, %%mm3 \n\t"\
"paddusw %%mm5, %%mm0 \n\t"\
"paddusw %%mm5, %%mm1 \n\t"\
"paddusw %%mm5, %%mm2 \n\t"\
"paddusw %%mm5, %%mm3 \n\t"\
"paddw %%mm6, %%mm0 \n\t"\
"paddw %%mm6, %%mm1 \n\t"\
"paddw %%mm6, %%mm2 \n\t"\
"paddw %%mm6, %%mm3 \n\t"\
"psubusw %%mm6, %%mm0 \n\t"\
"psubusw %%mm6, %%mm1 \n\t"\
"psubusw %%mm6, %%mm2 \n\t"\
"psubusw %%mm6, %%mm3 \n\t"\
"psraw $3, %%mm0 \n\t"\
"psraw $3, %%mm1 \n\t"\
"psraw $3, %%mm2 \n\t"\
"psraw $3, %%mm3 \n\t"\
\
"movq %%mm0, %%mm7 \n\t"\
"punpcklwd %%mm2, %%mm0 \n\t" /*A*/\
"punpckhwd %%mm2, %%mm7 \n\t" /*C*/\
"movq %%mm1, %%mm2 \n\t"\
"punpcklwd %%mm3, %%mm1 \n\t" /*B*/\
"punpckhwd %%mm3, %%mm2 \n\t" /*D*/\
"movq %%mm0, %%mm3 \n\t"\
"punpcklwd %%mm1, %%mm0 \n\t" /*A*/\
"punpckhwd %%mm7, %%mm3 \n\t" /*C*/\
"punpcklwd %%mm1, %%mm7 \n\t" /*B*/\
"punpckhwd %%mm2, %%mm1 \n\t" /*D*/\
\
"movq %%mm0, " #dst0 " \n\t"\
"movq %%mm7, " #dst1 " \n\t"\
"movq %%mm3, " #dst2 " \n\t"\
"movq %%mm1, " #dst3 " \n\t"
"movd %2, %%mm4 \n\t"
"movd %3, %%mm5 \n\t"
"movd %4, %%mm6 \n\t"
"packssdw %%mm4, %%mm4 \n\t"
"packssdw %%mm5, %%mm5 \n\t"
"packssdw %%mm6, %%mm6 \n\t"
"packssdw %%mm4, %%mm4 \n\t"
"packssdw %%mm5, %%mm5 \n\t"
"packssdw %%mm6, %%mm6 \n\t"
REQUANT_CORE( (%1), 8(%1), 16(%1), 24(%1), (%0), 8(%0), 64(%0), 72(%0))
REQUANT_CORE(32(%1), 40(%1), 48(%1), 56(%1),16(%0),24(%0), 48(%0), 56(%0))
REQUANT_CORE(64(%1), 72(%1), 80(%1), 88(%1),32(%0),40(%0), 96(%0),104(%0))
REQUANT_CORE(96(%1),104(%1),112(%1),120(%1),80(%0),88(%0),112(%0),120(%0))
: : "r" (src), "r" (dst), "g" (threshold1+1), "g" (threshold1+5), "g" (threshold1-4) //FIXME maybe more accurate then needed?
);
dst[0]= (src[0] + 4)>>3;
}
static void softthresh_mmx(DCTELEM dst[64], DCTELEM src[64], int qp, uint8_t *permutation){
int bias= 0; //FIXME
unsigned int threshold1;
threshold1= qp*((1<<4) - bias) - 1;
asm volatile(
#undef REQUANT_CORE
#define REQUANT_CORE(dst0, dst1, dst2, dst3, src0, src1, src2, src3) \
"movq " #src0 ", %%mm0 \n\t"\
"movq " #src1 ", %%mm1 \n\t"\
"pxor %%mm6, %%mm6 \n\t"\
"pxor %%mm7, %%mm7 \n\t"\
"pcmpgtw %%mm0, %%mm6 \n\t"\
"pcmpgtw %%mm1, %%mm7 \n\t"\
"pxor %%mm6, %%mm0 \n\t"\
"pxor %%mm7, %%mm1 \n\t"\
"psubusw %%mm4, %%mm0 \n\t"\
"psubusw %%mm4, %%mm1 \n\t"\
"pxor %%mm6, %%mm0 \n\t"\
"pxor %%mm7, %%mm1 \n\t"\
"movq " #src2 ", %%mm2 \n\t"\
"movq " #src3 ", %%mm3 \n\t"\
"pxor %%mm6, %%mm6 \n\t"\
"pxor %%mm7, %%mm7 \n\t"\
"pcmpgtw %%mm2, %%mm6 \n\t"\
"pcmpgtw %%mm3, %%mm7 \n\t"\
"pxor %%mm6, %%mm2 \n\t"\
"pxor %%mm7, %%mm3 \n\t"\
"psubusw %%mm4, %%mm2 \n\t"\
"psubusw %%mm4, %%mm3 \n\t"\
"pxor %%mm6, %%mm2 \n\t"\
"pxor %%mm7, %%mm3 \n\t"\
\
"paddsw %%mm5, %%mm0 \n\t"\
"paddsw %%mm5, %%mm1 \n\t"\
"paddsw %%mm5, %%mm2 \n\t"\
"paddsw %%mm5, %%mm3 \n\t"\
"psraw $3, %%mm0 \n\t"\
"psraw $3, %%mm1 \n\t"\
"psraw $3, %%mm2 \n\t"\
"psraw $3, %%mm3 \n\t"\
\
"movq %%mm0, %%mm7 \n\t"\
"punpcklwd %%mm2, %%mm0 \n\t" /*A*/\
"punpckhwd %%mm2, %%mm7 \n\t" /*C*/\
"movq %%mm1, %%mm2 \n\t"\
"punpcklwd %%mm3, %%mm1 \n\t" /*B*/\
"punpckhwd %%mm3, %%mm2 \n\t" /*D*/\
"movq %%mm0, %%mm3 \n\t"\
"punpcklwd %%mm1, %%mm0 \n\t" /*A*/\
"punpckhwd %%mm7, %%mm3 \n\t" /*C*/\
"punpcklwd %%mm1, %%mm7 \n\t" /*B*/\
"punpckhwd %%mm2, %%mm1 \n\t" /*D*/\
\
"movq %%mm0, " #dst0 " \n\t"\
"movq %%mm7, " #dst1 " \n\t"\
"movq %%mm3, " #dst2 " \n\t"\
"movq %%mm1, " #dst3 " \n\t"
"movd %2, %%mm4 \n\t"
"movd %3, %%mm5 \n\t"
"packssdw %%mm4, %%mm4 \n\t"
"packssdw %%mm5, %%mm5 \n\t"
"packssdw %%mm4, %%mm4 \n\t"
"packssdw %%mm5, %%mm5 \n\t"
REQUANT_CORE( (%1), 8(%1), 16(%1), 24(%1), (%0), 8(%0), 64(%0), 72(%0))
REQUANT_CORE(32(%1), 40(%1), 48(%1), 56(%1),16(%0),24(%0), 48(%0), 56(%0))
REQUANT_CORE(64(%1), 72(%1), 80(%1), 88(%1),32(%0),40(%0), 96(%0),104(%0))
REQUANT_CORE(96(%1),104(%1),112(%1),120(%1),80(%0),88(%0),112(%0),120(%0))
: : "r" (src), "r" (dst), "g" (threshold1), "rm" (4) //FIXME maybe more accurate then needed?
);
dst[0]= (src[0] + 4)>>3;
}
#endif
static inline void add_block(int16_t *dst, int stride, DCTELEM block[64]){
int y;
for(y=0; y<8; y++){
*(uint32_t*)&dst[0 + y*stride]+= *(uint32_t*)&block[0 + y*8];
*(uint32_t*)&dst[2 + y*stride]+= *(uint32_t*)&block[2 + y*8];
*(uint32_t*)&dst[4 + y*stride]+= *(uint32_t*)&block[4 + y*8];
*(uint32_t*)&dst[6 + y*stride]+= *(uint32_t*)&block[6 + y*8];
}
}
static void store_slice_c(uint8_t *dst, int16_t *src, int dst_stride, int src_stride, int width, int height, int log2_scale){
int y, x;
#define STORE(pos) \
temp= ((src[x + y*src_stride + pos]<<log2_scale) + d[pos])>>6;\
if(temp & 0x100) temp= ~(temp>>31);\
dst[x + y*dst_stride + pos]= temp;
for(y=0; y<height; y++){
uint8_t *d= dither[y];
for(x=0; x<width; x+=8){
int temp;
STORE(0);
STORE(1);
STORE(2);
STORE(3);
STORE(4);
STORE(5);
STORE(6);
STORE(7);
}
}
}
#ifdef HAVE_MMX
static void store_slice_mmx(uint8_t *dst, int16_t *src, int dst_stride, int src_stride, int width, int height, int log2_scale){
int y;
for(y=0; y<height; y++){
uint8_t *dst1= dst;
int16_t *src1= src;
asm volatile(
"movq (%3), %%mm3 \n\t"
"movq (%3), %%mm4 \n\t"
"movd %4, %%mm2 \n\t"
"pxor %%mm0, %%mm0 \n\t"
"punpcklbw %%mm0, %%mm3 \n\t"
"punpckhbw %%mm0, %%mm4 \n\t"
"psraw %%mm2, %%mm3 \n\t"
"psraw %%mm2, %%mm4 \n\t"
"movd %5, %%mm2 \n\t"
"1: \n\t"
"movq (%0), %%mm0 \n\t"
"movq 8(%0), %%mm1 \n\t"
"paddw %%mm3, %%mm0 \n\t"
"paddw %%mm4, %%mm1 \n\t"
"psraw %%mm2, %%mm0 \n\t"
"psraw %%mm2, %%mm1 \n\t"
"packuswb %%mm1, %%mm0 \n\t"
"movq %%mm0, (%1) \n\t"
"addl $16, %0 \n\t"
"addl $8, %1 \n\t"
"cmpl %2, %1 \n\t"
" jb 1b \n\t"
: "+r" (src1), "+r"(dst1)
: "r"(dst + width), "r"(dither[y]), "g"(log2_scale), "g"(6-log2_scale)
);
src += src_stride;
dst += dst_stride;
}
// if(width != mmxw)
// store_slice_c(dst + mmxw, src + mmxw, dst_stride, src_stride, width - mmxw, log2_scale);
}
#endif
static void (*store_slice)(uint8_t *dst, int16_t *src, int dst_stride, int src_stride, int width, int height, int log2_scale)= store_slice_c;
static void (*requantize)(DCTELEM dst[64], DCTELEM src[64], int qp, uint8_t *permutation)= hardthresh_c;
static void filter(struct vf_priv_s *p, uint8_t *dst, uint8_t *src, int dst_stride, int src_stride, int width, int height, uint8_t *qp_store, int qp_stride, int is_luma){
int x, y, i;
const int count= 1<<p->log2_count;
const int stride= is_luma ? p->temp_stride : ((width+16+15)&(~15));
uint64_t block_align[32];
DCTELEM *block = (DCTELEM *)block_align;
DCTELEM *block2= (DCTELEM *)(block_align+16);
for(y=0; y<height; y++){
int index= 8 + 8*stride + y*stride;
memcpy(p->src + index, src + y*src_stride, width);
for(x=0; x<8; x++){
p->src[index - x - 1]= p->src[index + x ];
p->src[index + width + x ]= p->src[index + width - x - 1];
}
}
for(y=0; y<8; y++){
memcpy(p->src + ( 7-y)*stride, p->src + ( y+8)*stride, stride);
memcpy(p->src + (height+8+y)*stride, p->src + (height-y+7)*stride, stride);
}
//FIXME (try edge emu)
for(y=0; y<height+8; y+=8){
memset(p->temp + (8+y)*stride, 0, 8*stride*sizeof(int16_t));
for(x=0; x<width+8; x+=8){
const int qps= 3 + is_luma;
int qp;
if(p->qp)
qp= p->qp;
else{
qp= qp_store[ (XMIN(x, width-1)>>qps) + (XMIN(y, height-1)>>qps) * qp_stride];
if(p->mpeg2) qp>>=1;
}
for(i=0; i<count; i++){
const int x1= x + offset[i+count-1][0];
const int y1= y + offset[i+count-1][1];
const int index= x1 + y1*stride;
p->dsp.get_pixels(block, p->src + index, stride);
p->dsp.fdct(block);
requantize(block2, block, qp, p->dsp.idct_permutation);
p->dsp.idct(block2);
add_block(p->temp + index, stride, block2);
}
}
if(y)
store_slice(dst + (y-8)*dst_stride, p->temp + 8 + y*stride, dst_stride, stride, width, XMIN(8, height+8-y), 6-p->log2_count);
}
#if 0
for(y=0; y<height; y++){
for(x=0; x<width; x++){
if((((x>>6) ^ (y>>6)) & 1) == 0)
dst[x + y*dst_stride]= p->src[8 + 8*stride + x + y*stride];
if((x&63) == 0 || (y&63)==0)
dst[x + y*dst_stride] += 128;
}
}
#endif
//FIXME reorder for better caching
}
static int config(struct vf_instance_s* vf,
int width, int height, int d_width, int d_height,
unsigned int flags, unsigned int outfmt){
int h= (height+16+15)&(~15);
vf->priv->temp_stride= (width+16+15)&(~15);
vf->priv->temp= malloc(vf->priv->temp_stride*h*sizeof(int16_t));
vf->priv->src = malloc(vf->priv->temp_stride*h*sizeof(uint8_t));
return vf_next_config(vf,width,height,d_width,d_height,flags,outfmt);
}
static void get_image(struct vf_instance_s* vf, mp_image_t *mpi){
if(mpi->flags&MP_IMGFLAG_PRESERVE) return; // don't change
if(mpi->imgfmt!=vf->priv->outfmt) return; // colorspace differ
// ok, we can do pp in-place (or pp disabled):
vf->dmpi=vf_get_image(vf->next,mpi->imgfmt,
mpi->type, mpi->flags, mpi->w, mpi->h);
mpi->planes[0]=vf->dmpi->planes[0];
mpi->stride[0]=vf->dmpi->stride[0];
mpi->width=vf->dmpi->width;
if(mpi->flags&MP_IMGFLAG_PLANAR){
mpi->planes[1]=vf->dmpi->planes[1];
mpi->planes[2]=vf->dmpi->planes[2];
mpi->stride[1]=vf->dmpi->stride[1];
mpi->stride[2]=vf->dmpi->stride[2];
}
mpi->flags|=MP_IMGFLAG_DIRECT;
}
static int put_image(struct vf_instance_s* vf, mp_image_t *mpi){
mp_image_t *dmpi;
if(!(mpi->flags&MP_IMGFLAG_DIRECT)){
// no DR, so get a new image! hope we'll get DR buffer:
vf->dmpi=vf_get_image(vf->next,vf->priv->outfmt,
MP_IMGTYPE_TEMP, MP_IMGFLAG_ACCEPT_STRIDE|MP_IMGFLAG_PREFER_ALIGNED_STRIDE,
mpi->w,mpi->h);
}
dmpi= vf->dmpi;
vf->priv->mpeg2= mpi->qscale_type;
if(vf->priv->log2_count || !(mpi->flags&MP_IMGFLAG_DIRECT)){
if(mpi->qscale || vf->priv->qp){
filter(vf->priv, dmpi->planes[0], mpi->planes[0], dmpi->stride[0], mpi->stride[0], mpi->w, mpi->h, mpi->qscale, mpi->qstride, 1);
filter(vf->priv, dmpi->planes[1], mpi->planes[1], dmpi->stride[1], mpi->stride[1], mpi->w>>mpi->chroma_x_shift, mpi->h>>mpi->chroma_y_shift, mpi->qscale, mpi->qstride, 0);
filter(vf->priv, dmpi->planes[2], mpi->planes[2], dmpi->stride[2], mpi->stride[2], mpi->w>>mpi->chroma_x_shift, mpi->h>>mpi->chroma_y_shift, mpi->qscale, mpi->qstride, 0);
}else{
memcpy_pic(dmpi->planes[0], mpi->planes[0], mpi->w, mpi->h, dmpi->stride[0], mpi->stride[0]);
memcpy_pic(dmpi->planes[1], mpi->planes[1], mpi->w>>mpi->chroma_x_shift, mpi->h>>mpi->chroma_y_shift, dmpi->stride[1], mpi->stride[1]);
memcpy_pic(dmpi->planes[2], mpi->planes[2], mpi->w>>mpi->chroma_x_shift, mpi->h>>mpi->chroma_y_shift, dmpi->stride[2], mpi->stride[2]);
}
}
vf_clone_mpi_attributes(dmpi, mpi);
#ifdef HAVE_MMX
if(gCpuCaps.hasMMX) asm volatile ("emms\n\t");
#endif
#ifdef HAVE_MMX2
if(gCpuCaps.hasMMX2) asm volatile ("sfence\n\t");
#endif
return vf_next_put_image(vf,dmpi);
}
static void uninit(struct vf_instance_s* vf){
if(!vf->priv) return;
if(vf->priv->temp) free(vf->priv->temp);
vf->priv->temp= NULL;
if(vf->priv->src) free(vf->priv->src);
vf->priv->src= NULL;
if(vf->priv->avctx) free(vf->priv->avctx);
vf->priv->avctx= NULL;
free(vf->priv);
vf->priv=NULL;
}
//===========================================================================//
static int query_format(struct vf_instance_s* vf, unsigned int fmt){
switch(fmt){
case IMGFMT_YVU9:
case IMGFMT_IF09:
case IMGFMT_YV12:
case IMGFMT_I420:
case IMGFMT_IYUV:
case IMGFMT_CLPL:
case IMGFMT_Y800:
case IMGFMT_Y8:
case IMGFMT_NV12:
case IMGFMT_444P:
case IMGFMT_422P:
case IMGFMT_411P:
return vf_next_query_format(vf,fmt);
}
return 0;
}
static unsigned int fmt_list[]={
IMGFMT_YVU9,
IMGFMT_IF09,
IMGFMT_YV12,
IMGFMT_I420,
IMGFMT_IYUV,
IMGFMT_CLPL,
IMGFMT_Y800,
IMGFMT_Y8,
IMGFMT_NV12,
IMGFMT_444P,
IMGFMT_422P,
IMGFMT_411P,
0
};
static int control(struct vf_instance_s* vf, int request, void* data){
switch(request){
case VFCTRL_QUERY_MAX_PP_LEVEL:
return 6;
case VFCTRL_SET_PP_LEVEL:
vf->priv->log2_count= *((unsigned int*)data);
return CONTROL_TRUE;
}
return vf_next_control(vf,request,data);
}
static int open(vf_instance_t *vf, char* args){
vf->config=config;
vf->put_image=put_image;
vf->get_image=get_image;
vf->query_format=query_format;
vf->uninit=uninit;
vf->control= control;
vf->priv=malloc(sizeof(struct vf_priv_s));
memset(vf->priv, 0, sizeof(struct vf_priv_s));
avcodec_init();
vf->priv->avctx= avcodec_alloc_context();
dsputil_init(&vf->priv->dsp, vf->priv->avctx);
vf->priv->log2_count= 3;
if (args) sscanf(args, "%d:%d:%d", &vf->priv->log2_count, &vf->priv->qp, &vf->priv->mode);
switch(vf->priv->mode){
case 0: requantize= hardthresh_c; break;
case 1: requantize= softthresh_c; break;
}
#ifdef HAVE_MMX
if(gCpuCaps.hasMMX){
store_slice= store_slice_mmx;
switch(vf->priv->mode){
case 0: requantize= hardthresh_mmx; break;
case 1: requantize= softthresh_mmx; break;
}
}
#endif
// check csp:
vf->priv->outfmt=vf_match_csp(&vf->next,fmt_list,IMGFMT_YV12);
if(!vf->priv->outfmt)
{
uninit(vf);
return 0; // no csp match :(
}
return 1;
}
vf_info_t vf_info_spp = {
"simple postprocess",
"spp",
"Michael Niedermayer",
"",
open,
NULL
};
#endif //USE_LIBAVCODEC