1
0
mirror of https://github.com/mpv-player/mpv synced 2024-12-16 20:05:07 +00:00
mpv/RTjpegN.c

3801 lines
98 KiB
C
Raw Normal View History

/*
RTjpeg (C) Justin Schoeman 1998 (justin@suntiger.ee.up.ac.za)
With modifications by:
(c) 1998, 1999 by Joerg Walter <trouble@moes.pmnet.uni-oldenburg.de>
and
(c) 1999 by Wim Taymans <wim.taymans@tvd.be>
This program is free software; you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation; either version 2 of the License, or
(at your option) any later version.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program; if not, write to the Free Software
Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
*/
#include <stdio.h>
#include <stdlib.h>
#include <string.h>
#include "config.h"
#ifdef HAVE_MMX
#define MMX
#endif
#include "RTjpegN.h"
#ifdef MMX
#include "mmx.h"
#endif
//#define SHOWBLOCK 1
#define BETTERCOMPRESSION 1
static const unsigned char RTjpeg_ZZ[64]={
0,
8, 1,
2, 9, 16,
24, 17, 10, 3,
4, 11, 18, 25, 32,
40, 33, 26, 19, 12, 5,
6, 13, 20, 27, 34, 41, 48,
56, 49, 42, 35, 28, 21, 14, 7,
15, 22, 29, 36, 43, 50, 57,
58, 51, 44, 37, 30, 23,
31, 38, 45, 52, 59,
60, 53, 46, 39,
47, 54, 61,
62, 55,
63 };
static const __u64 RTjpeg_aan_tab[64]={
4294967296ULL, 5957222912ULL, 5611718144ULL, 5050464768ULL, 4294967296ULL, 3374581504ULL, 2324432128ULL, 1184891264ULL,
5957222912ULL, 8263040512ULL, 7783580160ULL, 7005009920ULL, 5957222912ULL, 4680582144ULL, 3224107520ULL, 1643641088ULL,
5611718144ULL, 7783580160ULL, 7331904512ULL, 6598688768ULL, 5611718144ULL, 4408998912ULL, 3036936960ULL, 1548224000ULL,
5050464768ULL, 7005009920ULL, 6598688768ULL, 5938608128ULL, 5050464768ULL, 3968072960ULL, 2733115392ULL, 1393296000ULL,
4294967296ULL, 5957222912ULL, 5611718144ULL, 5050464768ULL, 4294967296ULL, 3374581504ULL, 2324432128ULL, 1184891264ULL,
3374581504ULL, 4680582144ULL, 4408998912ULL, 3968072960ULL, 3374581504ULL, 2651326208ULL, 1826357504ULL, 931136000ULL,
2324432128ULL, 3224107520ULL, 3036936960ULL, 2733115392ULL, 2324432128ULL, 1826357504ULL, 1258030336ULL, 641204288ULL,
1184891264ULL, 1643641088ULL, 1548224000ULL, 1393296000ULL, 1184891264ULL, 931136000ULL, 641204288ULL, 326894240ULL,
};
#ifndef MMX
static __s32 RTjpeg_ws[64+31];
#endif
__u8 RTjpeg_alldata[2*64+4*64+4*64+4*64+4*64+32];
static __s16 *block; // rh
static __s16 *RTjpeg_block;
static __s32 *RTjpeg_lqt;
static __s32 *RTjpeg_cqt;
static __u32 *RTjpeg_liqt;
static __u32 *RTjpeg_ciqt;
static unsigned char RTjpeg_lb8;
static unsigned char RTjpeg_cb8;
static int RTjpeg_width, RTjpeg_height;
static int RTjpeg_Ywidth, RTjpeg_Cwidth;
static int RTjpeg_Ysize, RTjpeg_Csize;
static __s16 *RTjpeg_old=NULL;
#ifdef MMX
mmx_t RTjpeg_lmask;
mmx_t RTjpeg_cmask;
#else
__u16 RTjpeg_lmask;
__u16 RTjpeg_cmask;
#endif
int RTjpeg_mtest=0;
static const unsigned char RTjpeg_lum_quant_tbl[64] = {
16, 11, 10, 16, 24, 40, 51, 61,
12, 12, 14, 19, 26, 58, 60, 55,
14, 13, 16, 24, 40, 57, 69, 56,
14, 17, 22, 29, 51, 87, 80, 62,
18, 22, 37, 56, 68, 109, 103, 77,
24, 35, 55, 64, 81, 104, 113, 92,
49, 64, 78, 87, 103, 121, 120, 101,
72, 92, 95, 98, 112, 100, 103, 99
};
static const unsigned char RTjpeg_chrom_quant_tbl[64] = {
17, 18, 24, 47, 99, 99, 99, 99,
18, 21, 26, 66, 99, 99, 99, 99,
24, 26, 56, 99, 99, 99, 99, 99,
47, 66, 99, 99, 99, 99, 99, 99,
99, 99, 99, 99, 99, 99, 99, 99,
99, 99, 99, 99, 99, 99, 99, 99,
99, 99, 99, 99, 99, 99, 99, 99,
99, 99, 99, 99, 99, 99, 99, 99
};
#ifdef BETTERCOMPRESSION
/*--------------------------------------------------*/
/* better encoding, but needs a lot more cpu time */
/* seems to be more effective than old method +lzo */
/* with this encoding lzo isn't efficient anymore */
/* there is still more potential for better */
/* encoding but that would need even more cputime */
/* anyway your mileage may vary */
/* */
/* written by Martin BIELY and Roman HOCHLEITNER */
/*--------------------------------------------------*/
/* +++++++++++++++++++++++++++++++++++++++++++++++++++*/
/* Block to Stream (encoding) */
/* */
int RTjpeg_b2s(__s16 *data, __s8 *strm, __u8 bt8)
{
register int ci, co=1;
register __s16 ZZvalue;
register unsigned char bitten;
register unsigned char bitoff;
#ifdef SHOWBLOCK
int ii;
for (ii=0; ii < 64; ii++) {
fprintf(stdout, "%d ", data[RTjpeg_ZZ[ii]]);
}
fprintf(stdout, "\n\n");
#endif
// *strm++ = 0x10;
// *strm = 0x00;
//
// return 2;
// first byte allways written
(__u8)strm[0]=
(__u8)(data[RTjpeg_ZZ[0]]>254) ? 254:((data[RTjpeg_ZZ[0]]<0)?0:data[RTjpeg_ZZ[0]]);
ci=63;
while (data[RTjpeg_ZZ[ci]]==0 && ci>0) ci--;
bitten = ((unsigned char)ci) << 2;
if (ci==0) {
(__u8)strm[1]= bitten;
co = 2;
return (int)co;
}
/* bitoff=0 because the high 6bit contain first non zero position */
bitoff = 0;
co = 1;
for(; ci>0; ci--) {
ZZvalue = data[RTjpeg_ZZ[ci]];
switch(ZZvalue) {
case 0:
break;
case 1:
bitten |= (0x01<<bitoff);
break;
case -1:
bitten |= (0x03<<bitoff);
break;
default:
bitten |= (0x02<<bitoff);
goto HERZWEH;
break;
}
if( bitoff == 0 ) {
(__u8)strm[co]= bitten;
bitten = 0;
bitoff = 8;
co++;
} /* "fall through" */
bitoff-=2;
}
/* ci must be 0 */
if(bitoff != 6) {
(__u8)strm[co]= bitten;
co++;
}
goto BAUCHWEH;
HERZWEH:
/* ci cannot be 0 */
/* correct bitoff to nibble boundaries */
switch(bitoff){
case 4:
case 6:
bitoff = 0;
break;
case 2:
case 0:
(__u8)strm[co]= bitten;
bitoff = 4;
co++;
bitten = 0; // clear half nibble values in bitten
break;
default:
break;
}
for(; ci>0; ci--) {
ZZvalue = data[RTjpeg_ZZ[ci]];
if( (ZZvalue > 7) || (ZZvalue < -7) ) {
bitten |= (0x08<<bitoff);
goto HIRNWEH;
}
bitten |= (ZZvalue&0xf)<<bitoff;
if( bitoff == 0 ) {
(__u8)strm[co]= bitten;
bitten = 0;
bitoff = 8;
co++;
} /* "fall thru" */
bitoff-=4;
}
/* ci must be 0 */
if( bitoff == 0 ) {
(__u8)strm[co]= bitten;
co++;
}
goto BAUCHWEH;
HIRNWEH:
(__u8)strm[co]= bitten;
co++;
/* bitting is over now we bite */
for(; ci>0; ci--) {
ZZvalue = data[RTjpeg_ZZ[ci]];
if(ZZvalue>0)
{
strm[co++]=(__s8)(ZZvalue>127)?127:ZZvalue;
}
else
{
strm[co++]=(__s8)(ZZvalue<-128)?-128:ZZvalue;
}
}
BAUCHWEH:
/* we gotoo much now we are ill */
#ifdef SHOWBLOCK
{
int i;
fprintf(stdout, "\nco = '%d'\n", co);
for (i=0; i < co+2; i++) {
fprintf(stdout, "%d ", strm[i]);
}
fprintf(stdout, "\n\n");
}
#endif
return (int)co;
}
/* +++++++++++++++++++++++++++++++++++++++++++++++++++*/
/* Stream to Block (decoding) */
/* */
int RTjpeg_s2b(__s16 *data, __s8 *strm, __u8 bt8, __u32 *qtbl)
{
int ci;
register int co;
register int i;
register unsigned char bitten;
register unsigned char bitoff;
/* first byte always read */
i=RTjpeg_ZZ[0];
data[i]=((__u8)strm[0])*qtbl[i];
/* we start at the behind */
bitten = ((unsigned char)strm[1]) >> 2;
co = 63;
for(; co > bitten; co--) {
data[RTjpeg_ZZ[co]] = 0;
}
if (co==0) {
ci = 2;
goto AUTOBAHN;
}
/* we have to read the last 2 bits of the second byte */
ci=1;
bitoff = 0;
for(; co>0; co--) {
bitten = ((unsigned char)strm[ci]) >> bitoff;
bitten &= 0x03;
i=RTjpeg_ZZ[co];
switch( bitten ) {
case 0x03:
data[i]= -qtbl[i];
break;
case 0x02:
goto FUSSWEG;
break;
case 0x01:
data[i]= qtbl[i];
break;
case 0x00:
data[i]= 0;
break;
default:
}
if( bitoff == 0 ) {
bitoff = 8;
ci++;
}
bitoff -= 2;
}
/* co is 0 now */
/* data is written properly */
/* if bitoff!=6 then ci is the index, but should be the byte count, so we increment by 1 */
if (bitoff!=6) ci++;
goto AUTOBAHN;
FUSSWEG:
/* correct bitoff to nibble */
switch(bitoff){
case 4:
case 6:
bitoff = 0;
break;
case 2:
case 0:
/* we have to read from the next byte */
ci++;
bitoff = 4;
break;
default:
break;
}
for(; co>0; co--) {
bitten = ((unsigned char)strm[ci]) >> bitoff;
bitten &= 0x0f;
i=RTjpeg_ZZ[co];
if( bitten == 0x08 ) {
goto STRASSE;
}
/* the compiler cannot do sign extension for signed nibbles */
if( bitten & 0x08 ) {
bitten |= 0xf0;
}
/* the unsigned char bitten now is a valid signed char */
data[i]=((signed char)bitten)*qtbl[i];
if( bitoff == 0 ) {
bitoff = 8;
ci++;
}
bitoff -= 4;
}
/* co is 0 */
/* if bitoff!=4 then ci is the index, but should be the byte count, so we increment by 1 */
if (bitoff!=4) ci++;
goto AUTOBAHN;
STRASSE:
ci++;
for(; co>0; co--) {
i=RTjpeg_ZZ[co];
data[i]=strm[ci++]*qtbl[i];
}
/* ci now is the count, because it points to next element => no incrementing */
AUTOBAHN:
#ifdef SHOWBLOCK
fprintf(stdout, "\nci = '%d'\n", ci);
for (i=0; i < 64; i++) {
fprintf(stdout, "%d ", data[RTjpeg_ZZ[i]]);
}
fprintf(stdout, "\n\n");
#endif
return ci;
}
#else
int RTjpeg_b2s(__s16 *data, __s8 *strm, __u8 bt8)
{
register int ci, co=1, tmp;
register __s16 ZZvalue;
#ifdef SHOWBLOCK
int ii;
for (ii=0; ii < 64; ii++) {
fprintf(stdout, "%d ", data[RTjpeg_ZZ[ii]]);
}
fprintf(stdout, "\n\n");
#endif
(__u8)strm[0]=(__u8)(data[RTjpeg_ZZ[0]]>254) ? 254:((data[RTjpeg_ZZ[0]]<0)?0:data[RTjpeg_ZZ[0]]);
for(ci=1; ci<=bt8; ci++)
{
ZZvalue = data[RTjpeg_ZZ[ci]];
if(ZZvalue>0)
{
strm[co++]=(__s8)(ZZvalue>127)?127:ZZvalue;
}
else
{
strm[co++]=(__s8)(ZZvalue<-128)?-128:ZZvalue;
}
}
for(; ci<64; ci++)
{
ZZvalue = data[RTjpeg_ZZ[ci]];
if(ZZvalue>0)
{
strm[co++]=(__s8)(ZZvalue>63)?63:ZZvalue;
}
else if(ZZvalue<0)
{
strm[co++]=(__s8)(ZZvalue<-64)?-64:ZZvalue;
}
else /* compress zeros */
{
tmp=ci;
do
{
ci++;
}
while((ci<64)&&(data[RTjpeg_ZZ[ci]]==0));
strm[co++]=(__s8)(63+(ci-tmp));
ci--;
}
}
return (int)co;
}
int RTjpeg_s2b(__s16 *data, __s8 *strm, __u8 bt8, __u32 *qtbl)
{
int ci=1, co=1, tmp;
register int i;
i=RTjpeg_ZZ[0];
data[i]=((__u8)strm[0])*qtbl[i];
for(co=1; co<=bt8; co++)
{
i=RTjpeg_ZZ[co];
data[i]=strm[ci++]*qtbl[i];
}
for(; co<64; co++)
{
if(strm[ci]>63)
{
tmp=co+strm[ci]-63;
for(; co<tmp; co++)data[RTjpeg_ZZ[co]]=0;
co--;
} else
{
i=RTjpeg_ZZ[co];
data[i]=strm[ci]*qtbl[i];
}
ci++;
}
return (int)ci;
}
#endif
#if defined(MMX)
void RTjpeg_quant_init(void)
{
int i;
__s16 *qtbl;
qtbl=(__s16 *)RTjpeg_lqt;
for(i=0; i<64; i++)qtbl[i]=(__s16)RTjpeg_lqt[i];
qtbl=(__s16 *)RTjpeg_cqt;
for(i=0; i<64; i++)qtbl[i]=(__s16)RTjpeg_cqt[i];
}
static mmx_t RTjpeg_ones=(mmx_t)(long long)0x0001000100010001LL;
static mmx_t RTjpeg_half=(mmx_t)(long long)0x7fff7fff7fff7fffLL;
void RTjpeg_quant(__s16 *block, __s32 *qtbl)
{
int i;
mmx_t *bl, *ql;
ql=(mmx_t *)qtbl;
bl=(mmx_t *)block;
movq_m2r(RTjpeg_ones, mm6);
movq_m2r(RTjpeg_half, mm7);
for(i=16; i; i--)
{
movq_m2r(*(ql++), mm0); /* quant vals (4) */
movq_m2r(*bl, mm2); /* block vals (4) */
movq_r2r(mm0, mm1);
movq_r2r(mm2, mm3);
punpcklwd_r2r(mm6, mm0); /* 1 qb 1 qa */
punpckhwd_r2r(mm6, mm1); /* 1 qd 1 qc */
punpcklwd_r2r(mm7, mm2); /* 32767 bb 32767 ba */
punpckhwd_r2r(mm7, mm3); /* 32767 bd 32767 bc */
pmaddwd_r2r(mm2, mm0); /* 32767+bb*qb 32767+ba*qa */
pmaddwd_r2r(mm3, mm1); /* 32767+bd*qd 32767+bc*qc */
psrad_i2r(16, mm0);
psrad_i2r(16, mm1);
packssdw_r2r(mm1, mm0);
movq_r2m(mm0, *(bl++));
}
}
#else
void RTjpeg_quant_init(void)
{
}
void RTjpeg_quant(__s16 *block, __s32 *qtbl)
{
int i;
for(i=0; i<64; i++)
block[i]=(__s16)((block[i]*qtbl[i]+32767)>>16);
}
#endif
/*
* Perform the forward DCT on one block of samples.
*/
#ifdef MMX
static mmx_t RTjpeg_C4 =(mmx_t)(long long)0x2D412D412D412D41LL;
static mmx_t RTjpeg_C6 =(mmx_t)(long long)0x187E187E187E187ELL;
static mmx_t RTjpeg_C2mC6=(mmx_t)(long long)0x22A322A322A322A3LL;
static mmx_t RTjpeg_C2pC6=(mmx_t)(long long)0x539F539F539F539FLL;
static mmx_t RTjpeg_zero =(mmx_t)(long long)0x0000000000000000LL;
#else
#define FIX_0_382683433 ((__s32) 98) /* FIX(0.382683433) */
#define FIX_0_541196100 ((__s32) 139) /* FIX(0.541196100) */
#define FIX_0_707106781 ((__s32) 181) /* FIX(0.707106781) */
#define FIX_1_306562965 ((__s32) 334) /* FIX(1.306562965) */
#define DESCALE10(x) (__s16)( ((x)+128) >> 8)
#define DESCALE20(x) (__s16)(((x)+32768) >> 16)
#define D_MULTIPLY(var,const) ((__s32) ((var) * (const)))
#endif
void RTjpeg_dct_init(void)
{
int i;
for(i=0; i<64; i++)
{
RTjpeg_lqt[i]=(((__u64)RTjpeg_lqt[i]<<32)/RTjpeg_aan_tab[i]);
RTjpeg_cqt[i]=(((__u64)RTjpeg_cqt[i]<<32)/RTjpeg_aan_tab[i]);
}
}
void RTjpeg_dctY(__u8 *idata, __s16 *odata, int rskip)
{
#ifndef MMX
__s32 tmp0, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7;
__s32 tmp10, tmp11, tmp12, tmp13;
__s32 z1, z2, z3, z4, z5, z11, z13;
__u8 *idataptr;
__s16 *odataptr;
__s32 *wsptr;
int ctr;
idataptr = idata;
wsptr = RTjpeg_ws;
for (ctr = 7; ctr >= 0; ctr--) {
tmp0 = idataptr[0] + idataptr[7];
tmp7 = idataptr[0] - idataptr[7];
tmp1 = idataptr[1] + idataptr[6];
tmp6 = idataptr[1] - idataptr[6];
tmp2 = idataptr[2] + idataptr[5];
tmp5 = idataptr[2] - idataptr[5];
tmp3 = idataptr[3] + idataptr[4];
tmp4 = idataptr[3] - idataptr[4];
tmp10 = (tmp0 + tmp3); /* phase 2 */
tmp13 = tmp0 - tmp3;
tmp11 = (tmp1 + tmp2);
tmp12 = tmp1 - tmp2;
wsptr[0] = (tmp10 + tmp11)<<8; /* phase 3 */
wsptr[4] = (tmp10 - tmp11)<<8;
z1 = D_MULTIPLY(tmp12 + tmp13, FIX_0_707106781); /* c4 */
wsptr[2] = (tmp13<<8) + z1; /* phase 5 */
wsptr[6] = (tmp13<<8) - z1;
tmp10 = tmp4 + tmp5; /* phase 2 */
tmp11 = tmp5 + tmp6;
tmp12 = tmp6 + tmp7;
z5 = D_MULTIPLY(tmp10 - tmp12, FIX_0_382683433); /* c6 */
z2 = D_MULTIPLY(tmp10, FIX_0_541196100) + z5; /* c2-c6 */
z4 = D_MULTIPLY(tmp12, FIX_1_306562965) + z5; /* c2+c6 */
z3 = D_MULTIPLY(tmp11, FIX_0_707106781); /* c4 */
z11 = (tmp7<<8) + z3; /* phase 5 */
z13 = (tmp7<<8) - z3;
wsptr[5] = z13 + z2; /* phase 6 */
wsptr[3] = z13 - z2;
wsptr[1] = z11 + z4;
wsptr[7] = z11 - z4;
idataptr += rskip<<3; /* advance pointer to next row */
wsptr += 8;
}
wsptr = RTjpeg_ws;
odataptr=odata;
for (ctr = 7; ctr >= 0; ctr--) {
tmp0 = wsptr[0] + wsptr[56];
tmp7 = wsptr[0] - wsptr[56];
tmp1 = wsptr[8] + wsptr[48];
tmp6 = wsptr[8] - wsptr[48];
tmp2 = wsptr[16] + wsptr[40];
tmp5 = wsptr[16] - wsptr[40];
tmp3 = wsptr[24] + wsptr[32];
tmp4 = wsptr[24] - wsptr[32];
tmp10 = tmp0 + tmp3; /* phase 2 */
tmp13 = tmp0 - tmp3;
tmp11 = tmp1 + tmp2;
tmp12 = tmp1 - tmp2;
odataptr[0] = DESCALE10(tmp10 + tmp11); /* phase 3 */
odataptr[32] = DESCALE10(tmp10 - tmp11);
z1 = D_MULTIPLY(tmp12 + tmp13, FIX_0_707106781); /* c4 */
odataptr[16] = DESCALE20((tmp13<<8) + z1); /* phase 5 */
odataptr[48] = DESCALE20((tmp13<<8) - z1);
tmp10 = tmp4 + tmp5; /* phase 2 */
tmp11 = tmp5 + tmp6;
tmp12 = tmp6 + tmp7;
z5 = D_MULTIPLY(tmp10 - tmp12, FIX_0_382683433); /* c6 */
z2 = D_MULTIPLY(tmp10, FIX_0_541196100) + z5; /* c2-c6 */
z4 = D_MULTIPLY(tmp12, FIX_1_306562965) + z5; /* c2+c6 */
z3 = D_MULTIPLY(tmp11, FIX_0_707106781); /* c4 */
z11 = (tmp7<<8) + z3; /* phase 5 */
z13 = (tmp7<<8) - z3;
odataptr[40] = DESCALE20(z13 + z2); /* phase 6 */
odataptr[24] = DESCALE20(z13 - z2);
odataptr[8] = DESCALE20(z11 + z4);
odataptr[56] = DESCALE20(z11 - z4);
odataptr++; /* advance pointer to next column */
wsptr++;
}
#else
volatile mmx_t tmp6, tmp7;
register mmx_t *dataptr = (mmx_t *)odata;
mmx_t *idata2 = (mmx_t *)idata;
// first copy the input 8 bit to the destination 16 bits
movq_m2r(RTjpeg_zero, mm2);
movq_m2r(*idata2, mm0);
movq_r2r(mm0, mm1);
punpcklbw_r2r(mm2, mm0);
movq_r2m(mm0, *(dataptr));
punpckhbw_r2r(mm2, mm1);
movq_r2m(mm1, *(dataptr+1));
idata2 += rskip;
movq_m2r(*idata2, mm0);
movq_r2r(mm0, mm1);
punpcklbw_r2r(mm2, mm0);
movq_r2m(mm0, *(dataptr+2));
punpckhbw_r2r(mm2, mm1);
movq_r2m(mm1, *(dataptr+3));
idata2 += rskip;
movq_m2r(*idata2, mm0);
movq_r2r(mm0, mm1);
punpcklbw_r2r(mm2, mm0);
movq_r2m(mm0, *(dataptr+4));
punpckhbw_r2r(mm2, mm1);
movq_r2m(mm1, *(dataptr+5));
idata2 += rskip;
movq_m2r(*idata2, mm0);
movq_r2r(mm0, mm1);
punpcklbw_r2r(mm2, mm0);
movq_r2m(mm0, *(dataptr+6));
punpckhbw_r2r(mm2, mm1);
movq_r2m(mm1, *(dataptr+7));
idata2 += rskip;
movq_m2r(*idata2, mm0);
movq_r2r(mm0, mm1);
punpcklbw_r2r(mm2, mm0);
movq_r2m(mm0, *(dataptr+8));
punpckhbw_r2r(mm2, mm1);
movq_r2m(mm1, *(dataptr+9));
idata2 += rskip;
movq_m2r(*idata2, mm0);
movq_r2r(mm0, mm1);
punpcklbw_r2r(mm2, mm0);
movq_r2m(mm0, *(dataptr+10));
punpckhbw_r2r(mm2, mm1);
movq_r2m(mm1, *(dataptr+11));
idata2 += rskip;
movq_m2r(*idata2, mm0);
movq_r2r(mm0, mm1);
punpcklbw_r2r(mm2, mm0);
movq_r2m(mm0, *(dataptr+12));
punpckhbw_r2r(mm2, mm1);
movq_r2m(mm1, *(dataptr+13));
idata2 += rskip;
movq_m2r(*idata2, mm0);
movq_r2r(mm0, mm1);
punpcklbw_r2r(mm2, mm0);
movq_r2m(mm0, *(dataptr+14));
punpckhbw_r2r(mm2, mm1);
movq_r2m(mm1, *(dataptr+15));
/* Start Transpose to do calculations on rows */
movq_m2r(*(dataptr+9), mm7); // m03:m02|m01:m00 - first line (line 4)and copy into m5
movq_m2r(*(dataptr+13), mm6); // m23:m22|m21:m20 - third line (line 6)and copy into m2
movq_r2r(mm7, mm5);
punpcklwd_m2r(*(dataptr+11), mm7); // m11:m01|m10:m00 - interleave first and second lines
movq_r2r(mm6, mm2);
punpcklwd_m2r(*(dataptr+15), mm6); // m31:m21|m30:m20 - interleave third and fourth lines
movq_r2r(mm7, mm1);
movq_m2r(*(dataptr+11), mm3); // m13:m13|m11:m10 - second line
punpckldq_r2r(mm6, mm7); // m30:m20|m10:m00 - interleave to produce result 1
movq_m2r(*(dataptr+15), mm0); // m13:m13|m11:m10 - fourth line
punpckhdq_r2r(mm6, mm1); // m31:m21|m11:m01 - interleave to produce result 2
movq_r2m(mm7,*(dataptr+9)); // write result 1
punpckhwd_r2r(mm3, mm5); // m13:m03|m12:m02 - interleave first and second lines
movq_r2m(mm1,*(dataptr+11)); // write result 2
punpckhwd_r2r(mm0, mm2); // m33:m23|m32:m22 - interleave third and fourth lines
movq_r2r(mm5, mm1);
punpckldq_r2r(mm2, mm5); // m32:m22|m12:m02 - interleave to produce result 3
movq_m2r(*(dataptr+1), mm0); // m03:m02|m01:m00 - first line, 4x4
punpckhdq_r2r(mm2, mm1); // m33:m23|m13:m03 - interleave to produce result 4
movq_r2m(mm5,*(dataptr+13)); // write result 3
// last 4x4 done
movq_r2m(mm1, *(dataptr+15)); // write result 4, last 4x4
movq_m2r(*(dataptr+5), mm2); // m23:m22|m21:m20 - third line
movq_r2r(mm0, mm6);
punpcklwd_m2r(*(dataptr+3), mm0); // m11:m01|m10:m00 - interleave first and second lines
movq_r2r(mm2, mm7);
punpcklwd_m2r(*(dataptr+7), mm2); // m31:m21|m30:m20 - interleave third and fourth lines
movq_r2r(mm0, mm4);
//
movq_m2r(*(dataptr+8), mm1); // n03:n02|n01:n00 - first line
punpckldq_r2r(mm2, mm0); // m30:m20|m10:m00 - interleave to produce first result
movq_m2r(*(dataptr+12), mm3); // n23:n22|n21:n20 - third line
punpckhdq_r2r(mm2, mm4); // m31:m21|m11:m01 - interleave to produce second result
punpckhwd_m2r(*(dataptr+3), mm6); // m13:m03|m12:m02 - interleave first and second lines
movq_r2r(mm1, mm2); // copy first line
punpckhwd_m2r(*(dataptr+7), mm7); // m33:m23|m32:m22 - interleave third and fourth lines
movq_r2r(mm6, mm5); // copy first intermediate result
movq_r2m(mm0, *(dataptr+8)); // write result 1
punpckhdq_r2r(mm7, mm5); // m33:m23|m13:m03 - produce third result
punpcklwd_m2r(*(dataptr+10), mm1); // n11:n01|n10:n00 - interleave first and second lines
movq_r2r(mm3, mm0); // copy third line
punpckhwd_m2r(*(dataptr+10), mm2); // n13:n03|n12:n02 - interleave first and second lines
movq_r2m(mm4, *(dataptr+10)); // write result 2 out
punpckldq_r2r(mm7, mm6); // m32:m22|m12:m02 - produce fourth result
punpcklwd_m2r(*(dataptr+14), mm3); // n31:n21|n30:n20 - interleave third and fourth lines
movq_r2r(mm1, mm4);
movq_r2m(mm6, *(dataptr+12)); // write result 3 out
punpckldq_r2r(mm3, mm1); // n30:n20|n10:n00 - produce first result
punpckhwd_m2r(*(dataptr+14), mm0); // n33:n23|n32:n22 - interleave third and fourth lines
movq_r2r(mm2, mm6);
movq_r2m(mm5, *(dataptr+14)); // write result 4 out
punpckhdq_r2r(mm3, mm4); // n31:n21|n11:n01- produce second result
movq_r2m(mm1, *(dataptr+1)); // write result 5 out - (first result for other 4 x 4 block)
punpckldq_r2r(mm0, mm2); // n32:n22|n12:n02- produce third result
movq_r2m(mm4, *(dataptr+3)); // write result 6 out
punpckhdq_r2r(mm0, mm6); // n33:n23|n13:n03 - produce fourth result
movq_r2m(mm2, *(dataptr+5)); // write result 7 out
movq_m2r(*dataptr, mm0); // m03:m02|m01:m00 - first line, first 4x4
movq_r2m(mm6, *(dataptr+7)); // write result 8 out
// Do first 4x4 quadrant, which is used in the beginning of the DCT:
movq_m2r(*(dataptr+4), mm7); // m23:m22|m21:m20 - third line
movq_r2r(mm0, mm2);
punpcklwd_m2r(*(dataptr+2), mm0); // m11:m01|m10:m00 - interleave first and second lines
movq_r2r(mm7, mm4);
punpcklwd_m2r(*(dataptr+6), mm7); // m31:m21|m30:m20 - interleave third and fourth lines
movq_r2r(mm0, mm1);
movq_m2r(*(dataptr+2), mm6); // m13:m12|m11:m10 - second line
punpckldq_r2r(mm7, mm0); // m30:m20|m10:m00 - interleave to produce result 1
movq_m2r(*(dataptr+6), mm5); // m33:m32|m31:m30 - fourth line
punpckhdq_r2r(mm7, mm1); // m31:m21|m11:m01 - interleave to produce result 2
movq_r2r(mm0, mm7); // write result 1
punpckhwd_r2r(mm6, mm2); // m13:m03|m12:m02 - interleave first and second lines
psubw_m2r(*(dataptr+14), mm7); // tmp07=x0-x7 /* Stage 1 */
movq_r2r(mm1, mm6); // write result 2
paddw_m2r(*(dataptr+14), mm0); // tmp00=x0+x7 /* Stage 1 */
punpckhwd_r2r(mm5, mm4); // m33:m23|m32:m22 - interleave third and fourth lines
paddw_m2r(*(dataptr+12), mm1); // tmp01=x1+x6 /* Stage 1 */
movq_r2r(mm2, mm3); // copy first intermediate result
psubw_m2r(*(dataptr+12), mm6); // tmp06=x1-x6 /* Stage 1 */
punpckldq_r2r(mm4, mm2); // m32:m22|m12:m02 - interleave to produce result 3
movq_r2m(mm7, tmp7);
movq_r2r(mm2, mm5); // write result 3
movq_r2m(mm6, tmp6);
punpckhdq_r2r(mm4, mm3); // m33:m23|m13:m03 - interleave to produce result 4
paddw_m2r(*(dataptr+10), mm2); // tmp02=x2+5 /* Stage 1 */
movq_r2r(mm3, mm4); // write result 4
/************************************************************************************************
End of Transpose
************************************************************************************************/
paddw_m2r(*(dataptr+8), mm3); // tmp03=x3+x4 /* stage 1*/
movq_r2r(mm0, mm7);
psubw_m2r(*(dataptr+8), mm4); // tmp04=x3-x4 /* stage 1*/
movq_r2r(mm1, mm6);
paddw_r2r(mm3, mm0); // tmp10 = tmp00 + tmp03 /* even 2 */
psubw_r2r(mm3, mm7); // tmp13 = tmp00 - tmp03 /* even 2 */
psubw_r2r(mm2, mm6); // tmp12 = tmp01 - tmp02 /* even 2 */
paddw_r2r(mm2, mm1); // tmp11 = tmp01 + tmp02 /* even 2 */
psubw_m2r(*(dataptr+10), mm5); // tmp05=x2-x5 /* stage 1*/
paddw_r2r(mm7, mm6); // tmp12 + tmp13
/* stage 3 */
movq_m2r(tmp6, mm2);
movq_r2r(mm0, mm3);
psllw_i2r(2, mm6); // m8 * 2^2
paddw_r2r(mm1, mm0);
pmulhw_m2r(RTjpeg_C4, mm6); // z1
psubw_r2r(mm1, mm3);
movq_r2m(mm0, *dataptr);
movq_r2r(mm7, mm0);
/* Odd part */
movq_r2m(mm3, *(dataptr+8));
paddw_r2r(mm5, mm4); // tmp10
movq_m2r(tmp7, mm3);
paddw_r2r(mm6, mm0); // tmp32
paddw_r2r(mm2, mm5); // tmp11
psubw_r2r(mm6, mm7); // tmp33
movq_r2m(mm0, *(dataptr+4));
paddw_r2r(mm3, mm2); // tmp12
/* stage 4 */
movq_r2m(mm7, *(dataptr+12));
movq_r2r(mm4, mm1); // copy of tmp10
psubw_r2r(mm2, mm1); // tmp10 - tmp12
psllw_i2r(2, mm4); // m8 * 2^2
movq_m2r(RTjpeg_C2mC6, mm0);
psllw_i2r(2, mm1);
pmulhw_m2r(RTjpeg_C6, mm1); // z5
psllw_i2r(2, mm2);
pmulhw_r2r(mm0, mm4); // z5
/* stage 5 */
pmulhw_m2r(RTjpeg_C2pC6, mm2);
psllw_i2r(2, mm5);
pmulhw_m2r(RTjpeg_C4, mm5); // z3
movq_r2r(mm3, mm0); // copy tmp7
movq_m2r(*(dataptr+1), mm7);
paddw_r2r(mm1, mm4); // z2
paddw_r2r(mm1, mm2); // z4
paddw_r2r(mm5, mm0); // z11
psubw_r2r(mm5, mm3); // z13
/* stage 6 */
movq_r2r(mm3, mm5); // copy z13
psubw_r2r(mm4, mm3); // y3=z13 - z2
paddw_r2r(mm4, mm5); // y5=z13 + z2
movq_r2r(mm0, mm6); // copy z11
movq_r2m(mm3, *(dataptr+6)); //save y3
psubw_r2r(mm2, mm0); // y7=z11 - z4
movq_r2m(mm5, *(dataptr+10)); //save y5
paddw_r2r(mm2, mm6); // y1=z11 + z4
movq_r2m(mm0, *(dataptr+14)); //save y7
/************************************************
* End of 1st 4 rows
************************************************/
movq_m2r(*(dataptr+3), mm1); // load x1 /* stage 1 */
movq_r2r(mm7, mm0); // copy x0
movq_r2m(mm6, *(dataptr+2)); //save y1
movq_m2r(*(dataptr+5), mm2); // load x2 /* stage 1 */
movq_r2r(mm1, mm6); // copy x1
paddw_m2r(*(dataptr+15), mm0); // tmp00 = x0 + x7
movq_m2r(*(dataptr+7), mm3); // load x3 /* stage 1 */
movq_r2r(mm2, mm5); // copy x2
psubw_m2r(*(dataptr+15), mm7); // tmp07 = x0 - x7
movq_r2r(mm3, mm4); // copy x3
paddw_m2r(*(dataptr+13), mm1); // tmp01 = x1 + x6
movq_r2m(mm7, tmp7); // save tmp07
movq_r2r(mm0, mm7); // copy tmp00
psubw_m2r(*(dataptr+13), mm6); // tmp06 = x1 - x6
/* stage 2, Even Part */
paddw_m2r(*(dataptr+9), mm3); // tmp03 = x3 + x4
movq_r2m(mm6, tmp6); // save tmp07
movq_r2r(mm1, mm6); // copy tmp01
paddw_m2r(*(dataptr+11), mm2); // tmp02 = x2 + x5
paddw_r2r(mm3, mm0); // tmp10 = tmp00 + tmp03
psubw_r2r(mm3, mm7); // tmp13 = tmp00 - tmp03
psubw_m2r(*(dataptr+9), mm4); // tmp04 = x3 - x4
psubw_r2r(mm2, mm6); // tmp12 = tmp01 - tmp02
paddw_r2r(mm2, mm1); // tmp11 = tmp01 + tmp02
psubw_m2r(*(dataptr+11), mm5); // tmp05 = x2 - x5
paddw_r2r(mm7, mm6); // tmp12 + tmp13
/* stage 3, Even and stage 4 & 5 even */
movq_m2r(tmp6, mm2); // load tmp6
movq_r2r(mm0, mm3); // copy tmp10
psllw_i2r(2, mm6); // shift z1
paddw_r2r(mm1, mm0); // y0=tmp10 + tmp11
pmulhw_m2r(RTjpeg_C4, mm6); // z1
psubw_r2r(mm1, mm3); // y4=tmp10 - tmp11
movq_r2m(mm0, *(dataptr+1)); //save y0
movq_r2r(mm7, mm0); // copy tmp13
/* odd part */
movq_r2m(mm3, *(dataptr+9)); //save y4
paddw_r2r(mm5, mm4); // tmp10 = tmp4 + tmp5
movq_m2r(tmp7, mm3); // load tmp7
paddw_r2r(mm6, mm0); // tmp32 = tmp13 + z1
paddw_r2r(mm2, mm5); // tmp11 = tmp5 + tmp6
psubw_r2r(mm6, mm7); // tmp33 = tmp13 - z1
movq_r2m(mm0, *(dataptr+5)); //save y2
paddw_r2r(mm3, mm2); // tmp12 = tmp6 + tmp7
/* stage 4 */
movq_r2m(mm7, *(dataptr+13)); //save y6
movq_r2r(mm4, mm1); // copy tmp10
psubw_r2r(mm2, mm1); // tmp10 - tmp12
psllw_i2r(2, mm4); // shift tmp10
movq_m2r(RTjpeg_C2mC6, mm0); // load C2mC6
psllw_i2r(2, mm1); // shift (tmp10-tmp12)
pmulhw_m2r(RTjpeg_C6, mm1); // z5
psllw_i2r(2, mm5); // prepare for multiply
pmulhw_r2r(mm0, mm4); // multiply by converted real
/* stage 5 */
pmulhw_m2r(RTjpeg_C4, mm5); // z3
psllw_i2r(2, mm2); // prepare for multiply
pmulhw_m2r(RTjpeg_C2pC6, mm2); // multiply
movq_r2r(mm3, mm0); // copy tmp7
movq_m2r(*(dataptr+9), mm7); // m03:m02|m01:m00 - first line (line 4)and copy into mm7
paddw_r2r(mm1, mm4); // z2
paddw_r2r(mm5, mm0); // z11
psubw_r2r(mm5, mm3); // z13
/* stage 6 */
movq_r2r(mm3, mm5); // copy z13
paddw_r2r(mm1, mm2); // z4
movq_r2r(mm0, mm6); // copy z11
psubw_r2r(mm4, mm5); // y3
paddw_r2r(mm2, mm6); // y1
paddw_r2r(mm4, mm3); // y5
movq_r2m(mm5, *(dataptr+7)); //save y3
movq_r2m(mm6, *(dataptr+3)); //save y1
psubw_r2r(mm2, mm0); // y7
/************************************************************************************************
Start of Transpose
************************************************************************************************/
movq_m2r(*(dataptr+13), mm6); // m23:m22|m21:m20 - third line (line 6)and copy into m2
movq_r2r(mm7, mm5); // copy first line
punpcklwd_r2r(mm3, mm7); // m11:m01|m10:m00 - interleave first and second lines
movq_r2r(mm6, mm2); // copy third line
punpcklwd_r2r(mm0, mm6); // m31:m21|m30:m20 - interleave third and fourth lines
movq_r2r(mm7, mm1); // copy first intermediate result
punpckldq_r2r(mm6, mm7); // m30:m20|m10:m00 - interleave to produce result 1
punpckhdq_r2r(mm6, mm1); // m31:m21|m11:m01 - interleave to produce result 2
movq_r2m(mm7, *(dataptr+9)); // write result 1
punpckhwd_r2r(mm3, mm5); // m13:m03|m12:m02 - interleave first and second lines
movq_r2m(mm1, *(dataptr+11)); // write result 2
punpckhwd_r2r(mm0, mm2); // m33:m23|m32:m22 - interleave third and fourth lines
movq_r2r(mm5, mm1); // copy first intermediate result
punpckldq_r2r(mm2, mm5); // m32:m22|m12:m02 - interleave to produce result 3
movq_m2r(*(dataptr+1), mm0); // m03:m02|m01:m00 - first line, 4x4
punpckhdq_r2r(mm2, mm1); // m33:m23|m13:m03 - interleave to produce result 4
movq_r2m(mm5, *(dataptr+13)); // write result 3
/****** last 4x4 done */
movq_r2m(mm1, *(dataptr+15)); // write result 4, last 4x4
movq_m2r(*(dataptr+5), mm2); // m23:m22|m21:m20 - third line
movq_r2r(mm0, mm6); // copy first line
punpcklwd_m2r(*(dataptr+3), mm0); // m11:m01|m10:m00 - interleave first and second lines
movq_r2r(mm2, mm7); // copy third line
punpcklwd_m2r(*(dataptr+7), mm2); // m31:m21|m30:m20 - interleave third and fourth lines
movq_r2r(mm0, mm4); // copy first intermediate result
movq_m2r(*(dataptr+8), mm1); // n03:n02|n01:n00 - first line
punpckldq_r2r(mm2, mm0); // m30:m20|m10:m00 - interleave to produce first result
movq_m2r(*(dataptr+12), mm3); // n23:n22|n21:n20 - third line
punpckhdq_r2r(mm2, mm4); // m31:m21|m11:m01 - interleave to produce second result
punpckhwd_m2r(*(dataptr+3), mm6); // m13:m03|m12:m02 - interleave first and second lines
movq_r2r(mm1, mm2); // copy first line
punpckhwd_m2r(*(dataptr+7), mm7); // m33:m23|m32:m22 - interleave third and fourth lines
movq_r2r(mm6, mm5); // copy first intermediate result
movq_r2m(mm0, *(dataptr+8)); // write result 1
punpckhdq_r2r(mm7, mm5); // m33:m23|m13:m03 - produce third result
punpcklwd_m2r(*(dataptr+10), mm1); // n11:n01|n10:n00 - interleave first and second lines
movq_r2r(mm3, mm0); // copy third line
punpckhwd_m2r(*(dataptr+10), mm2); // n13:n03|n12:n02 - interleave first and second lines
movq_r2m(mm4, *(dataptr+10)); // write result 2 out
punpckldq_r2r(mm7, mm6); // m32:m22|m12:m02 - produce fourth result
punpcklwd_m2r(*(dataptr+14), mm3); // n33:n23|n32:n22 - interleave third and fourth lines
movq_r2r(mm1, mm4); // copy second intermediate result
movq_r2m(mm6, *(dataptr+12)); // write result 3 out
punpckldq_r2r(mm3, mm1); //
punpckhwd_m2r(*(dataptr+14), mm0); // n33:n23|n32:n22 - interleave third and fourth lines
movq_r2r(mm2, mm6); // copy second intermediate result
movq_r2m(mm5, *(dataptr+14)); // write result 4 out
punpckhdq_r2r(mm3, mm4); // n31:n21|n11:n01- produce second result
movq_r2m(mm1, *(dataptr+1)); // write result 5 out - (first result for other 4 x 4 block)
punpckldq_r2r(mm0, mm2); // n32:n22|n12:n02- produce third result
movq_r2m(mm4, *(dataptr+3)); // write result 6 out
punpckhdq_r2r(mm0, mm6); // n33:n23|n13:n03 - produce fourth result
movq_r2m(mm2, *(dataptr+5)); // write result 7 out
movq_m2r(*dataptr, mm0); // m03:m02|m01:m00 - first line, first 4x4
movq_r2m(mm6, *(dataptr+7)); // write result 8 out
// Do first 4x4 quadrant, which is used in the beginning of the DCT:
movq_m2r(*(dataptr+4), mm7); // m23:m22|m21:m20 - third line
movq_r2r(mm0, mm2); // copy first line
punpcklwd_m2r(*(dataptr+2), mm0); // m11:m01|m10:m00 - interleave first and second lines
movq_r2r(mm7, mm4); // copy third line
punpcklwd_m2r(*(dataptr+6), mm7); // m31:m21|m30:m20 - interleave third and fourth lines
movq_r2r(mm0, mm1); // copy first intermediate result
movq_m2r(*(dataptr+2), mm6); // m13:m12|m11:m10 - second line
punpckldq_r2r(mm7, mm0); // m30:m20|m10:m00 - interleave to produce result 1
movq_m2r(*(dataptr+6), mm5); // m33:m32|m31:m30 - fourth line
punpckhdq_r2r(mm7, mm1); // m31:m21|m11:m01 - interleave to produce result 2
movq_r2r(mm0, mm7); // write result 1
punpckhwd_r2r(mm6, mm2); // m13:m03|m12:m02 - interleave first and second lines
psubw_m2r(*(dataptr+14), mm7); // tmp07=x0-x7 /* Stage 1 */
movq_r2r(mm1, mm6); // write result 2
paddw_m2r(*(dataptr+14), mm0); // tmp00=x0+x7 /* Stage 1 */
punpckhwd_r2r(mm5, mm4); // m33:m23|m32:m22 - interleave third and fourth lines
paddw_m2r(*(dataptr+12), mm1); // tmp01=x1+x6 /* Stage 1 */
movq_r2r(mm2, mm3); // copy first intermediate result
psubw_m2r(*(dataptr+12), mm6); // tmp06=x1-x6 /* Stage 1 */
punpckldq_r2r(mm4, mm2); // m32:m22|m12:m02 - interleave to produce result 3
movq_r2m(mm7, tmp7); // save tmp07
movq_r2r(mm2, mm5); // write result 3
movq_r2m(mm6, tmp6); // save tmp06
punpckhdq_r2r(mm4, mm3); // m33:m23|m13:m03 - interleave to produce result 4
paddw_m2r(*(dataptr+10), mm2); // tmp02=x2+x5 /* stage 1 */
movq_r2r(mm3, mm4); // write result 4
/************************************************************************************************
End of Transpose 2
************************************************************************************************/
paddw_m2r(*(dataptr+8), mm3); // tmp03=x3+x4 /* stage 1*/
movq_r2r(mm0, mm7);
psubw_m2r(*(dataptr+8), mm4); // tmp04=x3-x4 /* stage 1*/
movq_r2r(mm1, mm6);
paddw_r2r(mm3, mm0); // tmp10 = tmp00 + tmp03 /* even 2 */
psubw_r2r(mm3, mm7); // tmp13 = tmp00 - tmp03 /* even 2 */
psubw_r2r(mm2, mm6); // tmp12 = tmp01 - tmp02 /* even 2 */
paddw_r2r(mm2, mm1); // tmp11 = tmp01 + tmp02 /* even 2 */
psubw_m2r(*(dataptr+10), mm5); // tmp05=x2-x5 /* stage 1*/
paddw_r2r(mm7, mm6); // tmp12 + tmp13
/* stage 3 */
movq_m2r(tmp6, mm2);
movq_r2r(mm0, mm3);
psllw_i2r(2, mm6); // m8 * 2^2
paddw_r2r(mm1, mm0);
pmulhw_m2r(RTjpeg_C4, mm6); // z1
psubw_r2r(mm1, mm3);
movq_r2m(mm0, *dataptr);
movq_r2r(mm7, mm0);
/* Odd part */
movq_r2m(mm3, *(dataptr+8));
paddw_r2r(mm5, mm4); // tmp10
movq_m2r(tmp7, mm3);
paddw_r2r(mm6, mm0); // tmp32
paddw_r2r(mm2, mm5); // tmp11
psubw_r2r(mm6, mm7); // tmp33
movq_r2m(mm0, *(dataptr+4));
paddw_r2r(mm3, mm2); // tmp12
/* stage 4 */
movq_r2m(mm7, *(dataptr+12));
movq_r2r(mm4, mm1); // copy of tmp10
psubw_r2r(mm2, mm1); // tmp10 - tmp12
psllw_i2r(2, mm4); // m8 * 2^2
movq_m2r(RTjpeg_C2mC6, mm0);
psllw_i2r(2, mm1);
pmulhw_m2r(RTjpeg_C6, mm1); // z5
psllw_i2r(2, mm2);
pmulhw_r2r(mm0, mm4); // z5
/* stage 5 */
pmulhw_m2r(RTjpeg_C2pC6, mm2);
psllw_i2r(2, mm5);
pmulhw_m2r(RTjpeg_C4, mm5); // z3
movq_r2r(mm3, mm0); // copy tmp7
movq_m2r(*(dataptr+1), mm7);
paddw_r2r(mm1, mm4); // z2
paddw_r2r(mm1, mm2); // z4
paddw_r2r(mm5, mm0); // z11
psubw_r2r(mm5, mm3); // z13
/* stage 6 */
movq_r2r(mm3, mm5); // copy z13
psubw_r2r(mm4, mm3); // y3=z13 - z2
paddw_r2r(mm4, mm5); // y5=z13 + z2
movq_r2r(mm0, mm6); // copy z11
movq_r2m(mm3, *(dataptr+6)); //save y3
psubw_r2r(mm2, mm0); // y7=z11 - z4
movq_r2m(mm5, *(dataptr+10)); //save y5
paddw_r2r(mm2, mm6); // y1=z11 + z4
movq_r2m(mm0, *(dataptr+14)); //save y7
/************************************************
* End of 1st 4 rows
************************************************/
movq_m2r(*(dataptr+3), mm1); // load x1 /* stage 1 */
movq_r2r(mm7, mm0); // copy x0
movq_r2m(mm6, *(dataptr+2)); //save y1
movq_m2r(*(dataptr+5), mm2); // load x2 /* stage 1 */
movq_r2r(mm1, mm6); // copy x1
paddw_m2r(*(dataptr+15), mm0); // tmp00 = x0 + x7
movq_m2r(*(dataptr+7), mm3); // load x3 /* stage 1 */
movq_r2r(mm2, mm5); // copy x2
psubw_m2r(*(dataptr+15), mm7); // tmp07 = x0 - x7
movq_r2r(mm3, mm4); // copy x3
paddw_m2r(*(dataptr+13), mm1); // tmp01 = x1 + x6
movq_r2m(mm7, tmp7); // save tmp07
movq_r2r(mm0, mm7); // copy tmp00
psubw_m2r(*(dataptr+13), mm6); // tmp06 = x1 - x6
/* stage 2, Even Part */
paddw_m2r(*(dataptr+9), mm3); // tmp03 = x3 + x4
movq_r2m(mm6, tmp6); // save tmp07
movq_r2r(mm1, mm6); // copy tmp01
paddw_m2r(*(dataptr+11), mm2); // tmp02 = x2 + x5
paddw_r2r(mm3, mm0); // tmp10 = tmp00 + tmp03
psubw_r2r(mm3, mm7); // tmp13 = tmp00 - tmp03
psubw_m2r(*(dataptr+9), mm4); // tmp04 = x3 - x4
psubw_r2r(mm2, mm6); // tmp12 = tmp01 - tmp02
paddw_r2r(mm2, mm1); // tmp11 = tmp01 + tmp02
psubw_m2r(*(dataptr+11), mm5); // tmp05 = x2 - x5
paddw_r2r(mm7, mm6); // tmp12 + tmp13
/* stage 3, Even and stage 4 & 5 even */
movq_m2r(tmp6, mm2); // load tmp6
movq_r2r(mm0, mm3); // copy tmp10
psllw_i2r(2, mm6); // shift z1
paddw_r2r(mm1, mm0); // y0=tmp10 + tmp11
pmulhw_m2r(RTjpeg_C4, mm6); // z1
psubw_r2r(mm1, mm3); // y4=tmp10 - tmp11
movq_r2m(mm0, *(dataptr+1)); //save y0
movq_r2r(mm7, mm0); // copy tmp13
/* odd part */
movq_r2m(mm3, *(dataptr+9)); //save y4
paddw_r2r(mm5, mm4); // tmp10 = tmp4 + tmp5
movq_m2r(tmp7, mm3); // load tmp7
paddw_r2r(mm6, mm0); // tmp32 = tmp13 + z1
paddw_r2r(mm2, mm5); // tmp11 = tmp5 + tmp6
psubw_r2r(mm6, mm7); // tmp33 = tmp13 - z1
movq_r2m(mm0, *(dataptr+5)); //save y2
paddw_r2r(mm3, mm2); // tmp12 = tmp6 + tmp7
/* stage 4 */
movq_r2m(mm7, *(dataptr+13)); //save y6
movq_r2r(mm4, mm1); // copy tmp10
psubw_r2r(mm2, mm1); // tmp10 - tmp12
psllw_i2r(2, mm4); // shift tmp10
movq_m2r(RTjpeg_C2mC6, mm0); // load C2mC6
psllw_i2r(2, mm1); // shift (tmp10-tmp12)
pmulhw_m2r(RTjpeg_C6, mm1); // z5
psllw_i2r(2, mm5); // prepare for multiply
pmulhw_r2r(mm0, mm4); // multiply by converted real
/* stage 5 */
pmulhw_m2r(RTjpeg_C4, mm5); // z3
psllw_i2r(2, mm2); // prepare for multiply
pmulhw_m2r(RTjpeg_C2pC6, mm2); // multiply
movq_r2r(mm3, mm0); // copy tmp7
movq_m2r(*(dataptr+9), mm7); // m03:m02|m01:m00 - first line (line 4)and copy into mm7
paddw_r2r(mm1, mm4); // z2
paddw_r2r(mm5, mm0); // z11
psubw_r2r(mm5, mm3); // z13
/* stage 6 */
movq_r2r(mm3, mm5); // copy z13
paddw_r2r(mm1, mm2); // z4
movq_r2r(mm0, mm6); // copy z11
psubw_r2r(mm4, mm5); // y3
paddw_r2r(mm2, mm6); // y1
paddw_r2r(mm4, mm3); // y5
movq_r2m(mm5, *(dataptr+7)); //save y3
psubw_r2r(mm2, mm0); // y<>=z11 - z4
movq_r2m(mm3, *(dataptr+11)); //save y5
movq_r2m(mm6, *(dataptr+3)); //save y1
movq_r2m(mm0, *(dataptr+15)); //save y7
#endif
}
#define FIX_1_082392200 ((__s32) 277) /* FIX(1.082392200) */
#define FIX_1_414213562 ((__s32) 362) /* FIX(1.414213562) */
#define FIX_1_847759065 ((__s32) 473) /* FIX(1.847759065) */
#define FIX_2_613125930 ((__s32) 669) /* FIX(2.613125930) */
#define DESCALE(x) (__s16)( ((x)+4) >> 3)
/* clip yuv to 16..235 (should be 16..240 for cr/cb but ... */
#define RL(x) ((x)>235) ? 235 : (((x)<16) ? 16 : (x))
#define MULTIPLY(var,const) (((__s32) ((var) * (const)) + 128)>>8)
void RTjpeg_idct_init(void)
{
int i;
for(i=0; i<64; i++)
{
RTjpeg_liqt[i]=((__u64)RTjpeg_liqt[i]*RTjpeg_aan_tab[i])>>32;
RTjpeg_ciqt[i]=((__u64)RTjpeg_ciqt[i]*RTjpeg_aan_tab[i])>>32;
}
}
void RTjpeg_idct(__u8 *odata, __s16 *data, int rskip)
{
#ifdef MMX
static mmx_t fix_141 = (mmx_t)(long long)0x5a825a825a825a82LL;
static mmx_t fix_184n261 = (mmx_t)(long long)0xcf04cf04cf04cf04LL;
static mmx_t fix_184 = (mmx_t)(long long)0x7641764176417641LL;
static mmx_t fix_n184 = (mmx_t)(long long)0x896f896f896f896fLL;
static mmx_t fix_108n184 = (mmx_t)(long long)0xcf04cf04cf04cf04LL;
mmx_t workspace[64];
mmx_t *wsptr = workspace;
register mmx_t *dataptr = (mmx_t *)odata;
mmx_t *idata = (mmx_t *)data;
rskip = rskip>>3;
/*
* Perform inverse DCT on one block of coefficients.
*/
/* Odd part */
movq_m2r(*(idata+10), mm1); // load idata[DCTSIZE*5]
movq_m2r(*(idata+6), mm0); // load idata[DCTSIZE*3]
movq_m2r(*(idata+2), mm3); // load idata[DCTSIZE*1]
movq_r2r(mm1, mm2); // copy tmp6 /* phase 6 */
movq_m2r(*(idata+14), mm4); // load idata[DCTSIZE*7]
paddw_r2r(mm0, mm1); // z13 = tmp6 + tmp5;
psubw_r2r(mm0, mm2); // z10 = tmp6 - tmp5
psllw_i2r(2, mm2); // shift z10
movq_r2r(mm2, mm0); // copy z10
pmulhw_m2r(fix_184n261, mm2); // MULTIPLY( z12, FIX_1_847759065); /* 2*c2 */
movq_r2r(mm3, mm5); // copy tmp4
pmulhw_m2r(fix_n184, mm0); // MULTIPLY(z10, -FIX_1_847759065); /* 2*c2 */
paddw_r2r(mm4, mm3); // z11 = tmp4 + tmp7;
movq_r2r(mm3, mm6); // copy z11 /* phase 5 */
psubw_r2r(mm4, mm5); // z12 = tmp4 - tmp7;
psubw_r2r(mm1, mm6); // z11-z13
psllw_i2r(2, mm5); // shift z12
movq_m2r(*(idata+12), mm4); // load idata[DCTSIZE*6], even part
movq_r2r(mm5, mm7); // copy z12
pmulhw_m2r(fix_108n184, mm5); // MULT(z12, (FIX_1_08-FIX_1_84)) //- z5; /* 2*(c2-c6) */ even part
paddw_r2r(mm1, mm3); // tmp7 = z11 + z13;
//ok
/* Even part */
pmulhw_m2r(fix_184, mm7); // MULTIPLY(z10,(FIX_1_847759065 - FIX_2_613125930)) //+ z5; /* -2*(c2+c6) */
psllw_i2r(2, mm6);
movq_m2r(*(idata+4), mm1); // load idata[DCTSIZE*2]
paddw_r2r(mm5, mm0); // tmp10
paddw_r2r(mm7, mm2); // tmp12
pmulhw_m2r(fix_141, mm6); // tmp11 = MULTIPLY(z11 - z13, FIX_1_414213562); /* 2*c4 */
psubw_r2r(mm3, mm2); // tmp6 = tmp12 - tmp7
movq_r2r(mm1, mm5); // copy tmp1
paddw_r2r(mm4, mm1); // tmp13= tmp1 + tmp3; /* phases 5-3 */
psubw_r2r(mm4, mm5); // tmp1-tmp3
psubw_r2r(mm2, mm6); // tmp5 = tmp11 - tmp6;
movq_r2m(mm1, *(wsptr)); // save tmp13 in workspace
psllw_i2r(2, mm5); // shift tmp1-tmp3
movq_m2r(*(idata), mm7); // load idata[DCTSIZE*0]
pmulhw_m2r(fix_141, mm5); // MULTIPLY(tmp1 - tmp3, FIX_1_414213562)
paddw_r2r(mm6, mm0); // tmp4 = tmp10 + tmp5;
movq_m2r(*(idata+8), mm4); // load idata[DCTSIZE*4]
psubw_r2r(mm1, mm5); // tmp12 = MULTIPLY(tmp1 - tmp3, FIX_1_414213562) - tmp13; /* 2*c4 */
movq_r2m(mm0, *(wsptr+4)); // save tmp4 in workspace
movq_r2r(mm7, mm1); // copy tmp0 /* phase 3 */
movq_r2m(mm5, *(wsptr+2)); // save tmp12 in workspace
psubw_r2r(mm4, mm1); // tmp11 = tmp0 - tmp2;
paddw_r2r(mm4, mm7); // tmp10 = tmp0 + tmp2;
movq_r2r(mm1, mm5); // copy tmp11
paddw_m2r(*(wsptr+2), mm1); // tmp1 = tmp11 + tmp12;
movq_r2r(mm7, mm4); // copy tmp10 /* phase 2 */
paddw_m2r(*(wsptr), mm7); // tmp0 = tmp10 + tmp13;
psubw_m2r(*(wsptr), mm4); // tmp3 = tmp10 - tmp13;
movq_r2r(mm7, mm0); // copy tmp0
psubw_m2r(*(wsptr+2), mm5); // tmp2 = tmp11 - tmp12;
paddw_r2r(mm3, mm7); // wsptr[DCTSIZE*0] = (int) (tmp0 + tmp7);
psubw_r2r(mm3, mm0); // wsptr[DCTSIZE*7] = (int) (tmp0 - tmp7);
movq_r2m(mm7, *(wsptr)); // wsptr[DCTSIZE*0]
movq_r2r(mm1, mm3); // copy tmp1
movq_r2m(mm0, *(wsptr+14)); // wsptr[DCTSIZE*7]
paddw_r2r(mm2, mm1); // wsptr[DCTSIZE*1] = (int) (tmp1 + tmp6);
psubw_r2r(mm2, mm3); // wsptr[DCTSIZE*6] = (int) (tmp1 - tmp6);
movq_r2m(mm1, *(wsptr+2)); // wsptr[DCTSIZE*1]
movq_r2r(mm4, mm1); // copy tmp3
movq_r2m(mm3, *(wsptr+12)); // wsptr[DCTSIZE*6]
paddw_m2r(*(wsptr+4), mm4); // wsptr[DCTSIZE*4] = (int) (tmp3 + tmp4);
psubw_m2r(*(wsptr+4), mm1); // wsptr[DCTSIZE*3] = (int) (tmp3 - tmp4);
movq_r2m(mm4, *(wsptr+8));
movq_r2r(mm5, mm7); // copy tmp2
paddw_r2r(mm6, mm5); // wsptr[DCTSIZE*2] = (int) (tmp2 + tmp5)
movq_r2m(mm1, *(wsptr+6));
psubw_r2r(mm6, mm7); // wsptr[DCTSIZE*5] = (int) (tmp2 - tmp5);
movq_r2m(mm5, *(wsptr+4));
movq_r2m(mm7, *(wsptr+10));
//ok
/*****************************************************************/
idata++;
wsptr++;
/*****************************************************************/
movq_m2r(*(idata+10), mm1); // load idata[DCTSIZE*5]
movq_m2r(*(idata+6), mm0); // load idata[DCTSIZE*3]
movq_m2r(*(idata+2), mm3); // load idata[DCTSIZE*1]
movq_r2r(mm1, mm2); // copy tmp6 /* phase 6 */
movq_m2r(*(idata+14), mm4); // load idata[DCTSIZE*7]
paddw_r2r(mm0, mm1); // z13 = tmp6 + tmp5;
psubw_r2r(mm0, mm2); // z10 = tmp6 - tmp5
psllw_i2r(2, mm2); // shift z10
movq_r2r(mm2, mm0); // copy z10
pmulhw_m2r(fix_184n261, mm2); // MULTIPLY( z12, FIX_1_847759065); /* 2*c2 */
movq_r2r(mm3, mm5); // copy tmp4
pmulhw_m2r(fix_n184, mm0); // MULTIPLY(z10, -FIX_1_847759065); /* 2*c2 */
paddw_r2r(mm4, mm3); // z11 = tmp4 + tmp7;
movq_r2r(mm3, mm6); // copy z11 /* phase 5 */
psubw_r2r(mm4, mm5); // z12 = tmp4 - tmp7;
psubw_r2r(mm1, mm6); // z11-z13
psllw_i2r(2, mm5); // shift z12
movq_m2r(*(idata+12), mm4); // load idata[DCTSIZE*6], even part
movq_r2r(mm5, mm7); // copy z12
pmulhw_m2r(fix_108n184, mm5); // MULT(z12, (FIX_1_08-FIX_1_84)) //- z5; /* 2*(c2-c6) */ even part
paddw_r2r(mm1, mm3); // tmp7 = z11 + z13;
//ok
/* Even part */
pmulhw_m2r(fix_184, mm7); // MULTIPLY(z10,(FIX_1_847759065 - FIX_2_613125930)) //+ z5; /* -2*(c2+c6) */
psllw_i2r(2, mm6);
movq_m2r(*(idata+4), mm1); // load idata[DCTSIZE*2]
paddw_r2r(mm5, mm0); // tmp10
paddw_r2r(mm7, mm2); // tmp12
pmulhw_m2r(fix_141, mm6); // tmp11 = MULTIPLY(z11 - z13, FIX_1_414213562); /* 2*c4 */
psubw_r2r(mm3, mm2); // tmp6 = tmp12 - tmp7
movq_r2r(mm1, mm5); // copy tmp1
paddw_r2r(mm4, mm1); // tmp13= tmp1 + tmp3; /* phases 5-3 */
psubw_r2r(mm4, mm5); // tmp1-tmp3
psubw_r2r(mm2, mm6); // tmp5 = tmp11 - tmp6;
movq_r2m(mm1, *(wsptr)); // save tmp13 in workspace
psllw_i2r(2, mm5); // shift tmp1-tmp3
movq_m2r(*(idata), mm7); // load idata[DCTSIZE*0]
paddw_r2r(mm6, mm0); // tmp4 = tmp10 + tmp5;
pmulhw_m2r(fix_141, mm5); // MULTIPLY(tmp1 - tmp3, FIX_1_414213562)
movq_m2r(*(idata+8), mm4); // load idata[DCTSIZE*4]
psubw_r2r(mm1, mm5); // tmp12 = MULTIPLY(tmp1 - tmp3, FIX_1_414213562) - tmp13; /* 2*c4 */
movq_r2m(mm0, *(wsptr+4)); // save tmp4 in workspace
movq_r2r(mm7, mm1); // copy tmp0 /* phase 3 */
movq_r2m(mm5, *(wsptr+2)); // save tmp12 in workspace
psubw_r2r(mm4, mm1); // tmp11 = tmp0 - tmp2;
paddw_r2r(mm4, mm7); // tmp10 = tmp0 + tmp2;
movq_r2r(mm1, mm5); // copy tmp11
paddw_m2r(*(wsptr+2), mm1); // tmp1 = tmp11 + tmp12;
movq_r2r(mm7, mm4); // copy tmp10 /* phase 2 */
paddw_m2r(*(wsptr), mm7); // tmp0 = tmp10 + tmp13;
psubw_m2r(*(wsptr), mm4); // tmp3 = tmp10 - tmp13;
movq_r2r(mm7, mm0); // copy tmp0
psubw_m2r(*(wsptr+2), mm5); // tmp2 = tmp11 - tmp12;
paddw_r2r(mm3, mm7); // wsptr[DCTSIZE*0] = (int) (tmp0 + tmp7);
psubw_r2r(mm3, mm0); // wsptr[DCTSIZE*7] = (int) (tmp0 - tmp7);
movq_r2m(mm7, *(wsptr)); // wsptr[DCTSIZE*0]
movq_r2r(mm1, mm3); // copy tmp1
movq_r2m(mm0, *(wsptr+14)); // wsptr[DCTSIZE*7]
paddw_r2r(mm2, mm1); // wsptr[DCTSIZE*1] = (int) (tmp1 + tmp6);
psubw_r2r(mm2, mm3); // wsptr[DCTSIZE*6] = (int) (tmp1 - tmp6);
movq_r2m(mm1, *(wsptr+2)); // wsptr[DCTSIZE*1]
movq_r2r(mm4, mm1); // copy tmp3
movq_r2m(mm3, *(wsptr+12)); // wsptr[DCTSIZE*6]
paddw_m2r(*(wsptr+4), mm4); // wsptr[DCTSIZE*4] = (int) (tmp3 + tmp4);
psubw_m2r(*(wsptr+4), mm1); // wsptr[DCTSIZE*3] = (int) (tmp3 - tmp4);
movq_r2m(mm4, *(wsptr+8));
movq_r2r(mm5, mm7); // copy tmp2
paddw_r2r(mm6, mm5); // wsptr[DCTSIZE*2] = (int) (tmp2 + tmp5)
movq_r2m(mm1, *(wsptr+6));
psubw_r2r(mm6, mm7); // wsptr[DCTSIZE*5] = (int) (tmp2 - tmp5);
movq_r2m(mm5, *(wsptr+4));
movq_r2m(mm7, *(wsptr+10));
/*****************************************************************/
/* Pass 2: process rows from work array, store into output array. */
/* Note that we must descale the results by a factor of 8 == 2**3, */
/* and also undo the PASS1_BITS scaling. */
/*****************************************************************/
/* Even part */
wsptr--;
// tmp10 = ((DCTELEM) wsptr[0] + (DCTELEM) wsptr[4]);
// tmp13 = ((DCTELEM) wsptr[2] + (DCTELEM) wsptr[6]);
// tmp11 = ((DCTELEM) wsptr[0] - (DCTELEM) wsptr[4]);
// tmp14 = ((DCTELEM) wsptr[2] - (DCTELEM) wsptr[6]);
movq_m2r(*(wsptr), mm0); // wsptr[0,0],[0,1],[0,2],[0,3]
movq_m2r(*(wsptr+1), mm1); // wsptr[0,4],[0,5],[0,6],[0,7]
movq_r2r(mm0, mm2);
movq_m2r(*(wsptr+2), mm3); // wsptr[1,0],[1,1],[1,2],[1,3]
paddw_r2r(mm1, mm0); // wsptr[0,tmp10],[xxx],[0,tmp13],[xxx]
movq_m2r(*(wsptr+3), mm4); // wsptr[1,4],[1,5],[1,6],[1,7]
psubw_r2r(mm1, mm2); // wsptr[0,tmp11],[xxx],[0,tmp14],[xxx]
movq_r2r(mm0, mm6);
movq_r2r(mm3, mm5);
paddw_r2r(mm4, mm3); // wsptr[1,tmp10],[xxx],[1,tmp13],[xxx]
movq_r2r(mm2, mm1);
psubw_r2r(mm4, mm5); // wsptr[1,tmp11],[xxx],[1,tmp14],[xxx]
punpcklwd_r2r(mm3, mm0); // wsptr[0,tmp10],[1,tmp10],[xxx],[xxx]
movq_m2r(*(wsptr+7), mm7); // wsptr[3,4],[3,5],[3,6],[3,7]
punpckhwd_r2r(mm3, mm6); // wsptr[0,tmp13],[1,tmp13],[xxx],[xxx]
movq_m2r(*(wsptr+4), mm3); // wsptr[2,0],[2,1],[2,2],[2,3]
punpckldq_r2r(mm6, mm0); // wsptr[0,tmp10],[1,tmp10],[0,tmp13],[1,tmp13]
punpcklwd_r2r(mm5, mm1); // wsptr[0,tmp11],[1,tmp11],[xxx],[xxx]
movq_r2r(mm3, mm4);
movq_m2r(*(wsptr+6), mm6); // wsptr[3,0],[3,1],[3,2],[3,3]
punpckhwd_r2r(mm5, mm2); // wsptr[0,tmp14],[1,tmp14],[xxx],[xxx]
movq_m2r(*(wsptr+5), mm5); // wsptr[2,4],[2,5],[2,6],[2,7]
punpckldq_r2r(mm2, mm1); // wsptr[0,tmp11],[1,tmp11],[0,tmp14],[1,tmp14]
paddw_r2r(mm5, mm3); // wsptr[2,tmp10],[xxx],[2,tmp13],[xxx]
movq_r2r(mm6, mm2);
psubw_r2r(mm5, mm4); // wsptr[2,tmp11],[xxx],[2,tmp14],[xxx]
paddw_r2r(mm7, mm6); // wsptr[3,tmp10],[xxx],[3,tmp13],[xxx]
movq_r2r(mm3, mm5);
punpcklwd_r2r(mm6, mm3); // wsptr[2,tmp10],[3,tmp10],[xxx],[xxx]
psubw_r2r(mm7, mm2); // wsptr[3,tmp11],[xxx],[3,tmp14],[xxx]
punpckhwd_r2r(mm6, mm5); // wsptr[2,tmp13],[3,tmp13],[xxx],[xxx]
movq_r2r(mm4, mm7);
punpckldq_r2r(mm5, mm3); // wsptr[2,tmp10],[3,tmp10],[2,tmp13],[3,tmp13]
punpcklwd_r2r(mm2, mm4); // wsptr[2,tmp11],[3,tmp11],[xxx],[xxx]
punpckhwd_r2r(mm2, mm7); // wsptr[2,tmp14],[3,tmp14],[xxx],[xxx]
punpckldq_r2r(mm7, mm4); // wsptr[2,tmp11],[3,tmp11],[2,tmp14],[3,tmp14]
movq_r2r(mm1, mm6);
//ok
// mm0 = ;wsptr[0,tmp10],[1,tmp10],[0,tmp13],[1,tmp13]
// mm1 = ;wsptr[0,tmp11],[1,tmp11],[0,tmp14],[1,tmp14]
movq_r2r(mm0, mm2);
punpckhdq_r2r(mm4, mm6); // wsptr[0,tmp14],[1,tmp14],[2,tmp14],[3,tmp14]
punpckldq_r2r(mm4, mm1); // wsptr[0,tmp11],[1,tmp11],[2,tmp11],[3,tmp11]
psllw_i2r(2, mm6);
pmulhw_m2r(fix_141, mm6);
punpckldq_r2r(mm3, mm0); // wsptr[0,tmp10],[1,tmp10],[2,tmp10],[3,tmp10]
punpckhdq_r2r(mm3, mm2); // wsptr[0,tmp13],[1,tmp13],[2,tmp13],[3,tmp13]
movq_r2r(mm0, mm7);
// tmp0 = tmp10 + tmp13;
// tmp3 = tmp10 - tmp13;
paddw_r2r(mm2, mm0); // [0,tmp0],[1,tmp0],[2,tmp0],[3,tmp0]
psubw_r2r(mm2, mm7); // [0,tmp3],[1,tmp3],[2,tmp3],[3,tmp3]
// tmp12 = MULTIPLY(tmp14, FIX_1_414213562) - tmp13;
psubw_r2r(mm2, mm6); // wsptr[0,tmp12],[1,tmp12],[2,tmp12],[3,tmp12]
// tmp1 = tmp11 + tmp12;
// tmp2 = tmp11 - tmp12;
movq_r2r(mm1, mm5);
//OK
/* Odd part */
// z13 = (DCTELEM) wsptr[5] + (DCTELEM) wsptr[3];
// z10 = (DCTELEM) wsptr[5] - (DCTELEM) wsptr[3];
// z11 = (DCTELEM) wsptr[1] + (DCTELEM) wsptr[7];
// z12 = (DCTELEM) wsptr[1] - (DCTELEM) wsptr[7];
movq_m2r(*(wsptr), mm3); // wsptr[0,0],[0,1],[0,2],[0,3]
paddw_r2r(mm6, mm1); // [0,tmp1],[1,tmp1],[2,tmp1],[3,tmp1]
movq_m2r(*(wsptr+1), mm4); // wsptr[0,4],[0,5],[0,6],[0,7]
psubw_r2r(mm6, mm5); // [0,tmp2],[1,tmp2],[2,tmp2],[3,tmp2]
movq_r2r(mm3, mm6);
punpckldq_r2r(mm4, mm3); // wsptr[0,0],[0,1],[0,4],[0,5]
punpckhdq_r2r(mm6, mm4); // wsptr[0,6],[0,7],[0,2],[0,3]
movq_r2r(mm3, mm2);
//Save tmp0 and tmp1 in wsptr
movq_r2m(mm0, *(wsptr)); // save tmp0
paddw_r2r(mm4, mm2); // wsptr[xxx],[0,z11],[xxx],[0,z13]
//Continue with z10 --- z13
movq_m2r(*(wsptr+2), mm6); // wsptr[1,0],[1,1],[1,2],[1,3]
psubw_r2r(mm4, mm3); // wsptr[xxx],[0,z12],[xxx],[0,z10]
movq_m2r(*(wsptr+3), mm0); // wsptr[1,4],[1,5],[1,6],[1,7]
movq_r2r(mm6, mm4);
movq_r2m(mm1, *(wsptr+1)); // save tmp1
punpckldq_r2r(mm0, mm6); // wsptr[1,0],[1,1],[1,4],[1,5]
punpckhdq_r2r(mm4, mm0); // wsptr[1,6],[1,7],[1,2],[1,3]
movq_r2r(mm6, mm1);
//Save tmp2 and tmp3 in wsptr
paddw_r2r(mm0, mm6); // wsptr[xxx],[1,z11],[xxx],[1,z13]
movq_r2r(mm2, mm4);
//Continue with z10 --- z13
movq_r2m(mm5, *(wsptr+2)); // save tmp2
punpcklwd_r2r(mm6, mm2); // wsptr[xxx],[xxx],[0,z11],[1,z11]
psubw_r2r(mm0, mm1); // wsptr[xxx],[1,z12],[xxx],[1,z10]
punpckhwd_r2r(mm6, mm4); // wsptr[xxx],[xxx],[0,z13],[1,z13]
movq_r2r(mm3, mm0);
punpcklwd_r2r(mm1, mm3); // wsptr[xxx],[xxx],[0,z12],[1,z12]
movq_r2m(mm7, *(wsptr+3)); // save tmp3
punpckhwd_r2r(mm1, mm0); // wsptr[xxx],[xxx],[0,z10],[1,z10]
movq_m2r(*(wsptr+4), mm6); // wsptr[2,0],[2,1],[2,2],[2,3]
punpckhdq_r2r(mm2, mm0); // wsptr[0,z10],[1,z10],[0,z11],[1,z11]
movq_m2r(*(wsptr+5), mm7); // wsptr[2,4],[2,5],[2,6],[2,7]
punpckhdq_r2r(mm4, mm3); // wsptr[0,z12],[1,z12],[0,z13],[1,z13]
movq_m2r(*(wsptr+6), mm1); // wsptr[3,0],[3,1],[3,2],[3,3]
movq_r2r(mm6, mm4);
punpckldq_r2r(mm7, mm6); // wsptr[2,0],[2,1],[2,4],[2,5]
movq_r2r(mm1, mm5);
punpckhdq_r2r(mm4, mm7); // wsptr[2,6],[2,7],[2,2],[2,3]
movq_r2r(mm6, mm2);
movq_m2r(*(wsptr+7), mm4); // wsptr[3,4],[3,5],[3,6],[3,7]
paddw_r2r(mm7, mm6); // wsptr[xxx],[2,z11],[xxx],[2,z13]
psubw_r2r(mm7, mm2); // wsptr[xxx],[2,z12],[xxx],[2,z10]
punpckldq_r2r(mm4, mm1); // wsptr[3,0],[3,1],[3,4],[3,5]
punpckhdq_r2r(mm5, mm4); // wsptr[3,6],[3,7],[3,2],[3,3]
movq_r2r(mm1, mm7);
paddw_r2r(mm4, mm1); // wsptr[xxx],[3,z11],[xxx],[3,z13]
psubw_r2r(mm4, mm7); // wsptr[xxx],[3,z12],[xxx],[3,z10]
movq_r2r(mm6, mm5);
punpcklwd_r2r(mm1, mm6); // wsptr[xxx],[xxx],[2,z11],[3,z11]
punpckhwd_r2r(mm1, mm5); // wsptr[xxx],[xxx],[2,z13],[3,z13]
movq_r2r(mm2, mm4);
punpcklwd_r2r(mm7, mm2); // wsptr[xxx],[xxx],[2,z12],[3,z12]
punpckhwd_r2r(mm7, mm4); // wsptr[xxx],[xxx],[2,z10],[3,z10]
punpckhdq_r2r(mm6, mm4); /// wsptr[2,z10],[3,z10],[2,z11],[3,z11]
punpckhdq_r2r(mm5, mm2); // wsptr[2,z12],[3,z12],[2,z13],[3,z13]
movq_r2r(mm0, mm5);
punpckldq_r2r(mm4, mm0); // wsptr[0,z10],[1,z10],[2,z10],[3,z10]
punpckhdq_r2r(mm4, mm5); // wsptr[0,z11],[1,z11],[2,z11],[3,z11]
movq_r2r(mm3, mm4);
punpckhdq_r2r(mm2, mm4); // wsptr[0,z13],[1,z13],[2,z13],[3,z13]
movq_r2r(mm5, mm1);
punpckldq_r2r(mm2, mm3); // wsptr[0,z12],[1,z12],[2,z12],[3,z12]
// tmp7 = z11 + z13; /* phase 5 */
// tmp8 = z11 - z13; /* phase 5 */
psubw_r2r(mm4, mm1); // tmp8
paddw_r2r(mm4, mm5); // tmp7
// tmp21 = MULTIPLY(tmp8, FIX_1_414213562); /* 2*c4 */
psllw_i2r(2, mm1);
psllw_i2r(2, mm0);
pmulhw_m2r(fix_141, mm1); // tmp21
// tmp20 = MULTIPLY(z12, (FIX_1_082392200- FIX_1_847759065)) /* 2*(c2-c6) */
// + MULTIPLY(z10, - FIX_1_847759065); /* 2*c2 */
psllw_i2r(2, mm3);
movq_r2r(mm0, mm7);
pmulhw_m2r(fix_n184, mm7);
movq_r2r(mm3, mm6);
movq_m2r(*(wsptr), mm2); // tmp0,final1
pmulhw_m2r(fix_108n184, mm6);
// tmp22 = MULTIPLY(z10,(FIX_1_847759065 - FIX_2_613125930)) /* -2*(c2+c6) */
// + MULTIPLY(z12, FIX_1_847759065); /* 2*c2 */
movq_r2r(mm2, mm4); // final1
pmulhw_m2r(fix_184n261, mm0);
paddw_r2r(mm5, mm2); // tmp0+tmp7,final1
pmulhw_m2r(fix_184, mm3);
psubw_r2r(mm5, mm4); // tmp0-tmp7,final1
// tmp6 = tmp22 - tmp7; /* phase 2 */
psraw_i2r(3, mm2); // outptr[0,0],[1,0],[2,0],[3,0],final1
paddw_r2r(mm6, mm7); // tmp20
psraw_i2r(3, mm4); // outptr[0,7],[1,7],[2,7],[3,7],final1
paddw_r2r(mm0, mm3); // tmp22
// tmp5 = tmp21 - tmp6;
psubw_r2r(mm5, mm3); // tmp6
// tmp4 = tmp20 + tmp5;
movq_m2r(*(wsptr+1), mm0); // tmp1,final2
psubw_r2r(mm3, mm1); // tmp5
movq_r2r(mm0, mm6); // final2
paddw_r2r(mm3, mm0); // tmp1+tmp6,final2
/* Final output stage: scale down by a factor of 8 and range-limit */
// outptr[0] = range_limit[IDESCALE(tmp0 + tmp7, PASS1_BITS+3)
// & RANGE_MASK];
// outptr[7] = range_limit[IDESCALE(tmp0 - tmp7, PASS1_BITS+3)
// & RANGE_MASK]; final1
// outptr[1] = range_limit[IDESCALE(tmp1 + tmp6, PASS1_BITS+3)
// & RANGE_MASK];
// outptr[6] = range_limit[IDESCALE(tmp1 - tmp6, PASS1_BITS+3)
// & RANGE_MASK]; final2
psubw_r2r(mm3, mm6); // tmp1-tmp6,final2
psraw_i2r(3, mm0); // outptr[0,1],[1,1],[2,1],[3,1]
psraw_i2r(3, mm6); // outptr[0,6],[1,6],[2,6],[3,6]
packuswb_r2r(mm4, mm0); // out[0,1],[1,1],[2,1],[3,1],[0,7],[1,7],[2,7],[3,7]
movq_m2r(*(wsptr+2), mm5); // tmp2,final3
packuswb_r2r(mm6, mm2); // out[0,0],[1,0],[2,0],[3,0],[0,6],[1,6],[2,6],[3,6]
// outptr[2] = range_limit[IDESCALE(tmp2 + tmp5, PASS1_BITS+3)
// & RANGE_MASK];
// outptr[5] = range_limit[IDESCALE(tmp2 - tmp5, PASS1_BITS+3)
// & RANGE_MASK]; final3
paddw_r2r(mm1, mm7); // tmp4
movq_r2r(mm5, mm3);
paddw_r2r(mm1, mm5); // tmp2+tmp5
psubw_r2r(mm1, mm3); // tmp2-tmp5
psraw_i2r(3, mm5); // outptr[0,2],[1,2],[2,2],[3,2]
movq_m2r(*(wsptr+3), mm4); // tmp3,final4
psraw_i2r(3, mm3); // outptr[0,5],[1,5],[2,5],[3,5]
// outptr[4] = range_limit[IDESCALE(tmp3 + tmp4, PASS1_BITS+3)
// & RANGE_MASK];
// outptr[3] = range_limit[IDESCALE(tmp3 - tmp4, PASS1_BITS+3)
// & RANGE_MASK]; final4
movq_r2r(mm4, mm6);
paddw_r2r(mm7, mm4); // tmp3+tmp4
psubw_r2r(mm7, mm6); // tmp3-tmp4
psraw_i2r(3, mm4); // outptr[0,4],[1,4],[2,4],[3,4]
// mov ecx, [dataptr]
psraw_i2r(3, mm6); // outptr[0,3],[1,3],[2,3],[3,3]
packuswb_r2r(mm4, mm5); // out[0,2],[1,2],[2,2],[3,2],[0,4],[1,4],[2,4],[3,4]
packuswb_r2r(mm3, mm6); // out[0,3],[1,3],[2,3],[3,3],[0,5],[1,5],[2,5],[3,5]
movq_r2r(mm2, mm4);
movq_r2r(mm5, mm7);
punpcklbw_r2r(mm0, mm2); // out[0,0],[0,1],[1,0],[1,1],[2,0],[2,1],[3,0],[3,1]
punpckhbw_r2r(mm0, mm4); // out[0,6],[0,7],[1,6],[1,7],[2,6],[2,7],[3,6],[3,7]
movq_r2r(mm2, mm1);
punpcklbw_r2r(mm6, mm5); // out[0,2],[0,3],[1,2],[1,3],[2,2],[2,3],[3,2],[3,3]
// add dataptr, 4
punpckhbw_r2r(mm6, mm7); // out[0,4],[0,5],[1,4],[1,5],[2,4],[2,5],[3,4],[3,5]
punpcklwd_r2r(mm5, mm2); // out[0,0],[0,1],[0,2],[0,3],[1,0],[1,1],[1,2],[1,3]
// add ecx, output_col
movq_r2r(mm7, mm6);
punpckhwd_r2r(mm5, mm1); // out[2,0],[2,1],[2,2],[2,3],[3,0],[3,1],[3,2],[3,3]
movq_r2r(mm2, mm0);
punpcklwd_r2r(mm4, mm6); // out[0,4],[0,5],[0,6],[0,7],[1,4],[1,5],[1,6],[1,7]
// mov idata, [dataptr]
punpckldq_r2r(mm6, mm2); // out[0,0],[0,1],[0,2],[0,3],[0,4],[0,5],[0,6],[0,7]
// add dataptr, 4
movq_r2r(mm1, mm3);
// add idata, output_col
punpckhwd_r2r(mm4, mm7); // out[2,4],[2,5],[2,6],[2,7],[3,4],[3,5],[3,6],[3,7]
movq_r2m(mm2, *(dataptr));
punpckhdq_r2r(mm6, mm0); // out[1,0],[1,1],[1,2],[1,3],[1,4],[1,5],[1,6],[1,7]
dataptr += rskip;
movq_r2m(mm0, *(dataptr));
punpckldq_r2r(mm7, mm1); // out[2,0],[2,1],[2,2],[2,3],[2,4],[2,5],[2,6],[2,7]
punpckhdq_r2r(mm7, mm3); // out[3,0],[3,1],[3,2],[3,3],[3,4],[3,5],[3,6],[3,7]
dataptr += rskip;
movq_r2m(mm1, *(dataptr));
dataptr += rskip;
movq_r2m(mm3, *(dataptr));
/*******************************************************************/
wsptr += 8;
/*******************************************************************/
// tmp10 = ((DCTELEM) wsptr[0] + (DCTELEM) wsptr[4]);
// tmp13 = ((DCTELEM) wsptr[2] + (DCTELEM) wsptr[6]);
// tmp11 = ((DCTELEM) wsptr[0] - (DCTELEM) wsptr[4]);
// tmp14 = ((DCTELEM) wsptr[2] - (DCTELEM) wsptr[6]);
movq_m2r(*(wsptr), mm0); // wsptr[0,0],[0,1],[0,2],[0,3]
movq_m2r(*(wsptr+1), mm1); // wsptr[0,4],[0,5],[0,6],[0,7]
movq_r2r(mm0, mm2);
movq_m2r(*(wsptr+2), mm3); // wsptr[1,0],[1,1],[1,2],[1,3]
paddw_r2r(mm1, mm0); // wsptr[0,tmp10],[xxx],[0,tmp13],[xxx]
movq_m2r(*(wsptr+3), mm4); // wsptr[1,4],[1,5],[1,6],[1,7]
psubw_r2r(mm1, mm2); // wsptr[0,tmp11],[xxx],[0,tmp14],[xxx]
movq_r2r(mm0, mm6);
movq_r2r(mm3, mm5);
paddw_r2r(mm4, mm3); // wsptr[1,tmp10],[xxx],[1,tmp13],[xxx]
movq_r2r(mm2, mm1);
psubw_r2r(mm4, mm5); // wsptr[1,tmp11],[xxx],[1,tmp14],[xxx]
punpcklwd_r2r(mm3, mm0); // wsptr[0,tmp10],[1,tmp10],[xxx],[xxx]
movq_m2r(*(wsptr+7), mm7); // wsptr[3,4],[3,5],[3,6],[3,7]
punpckhwd_r2r(mm3, mm6); // wsptr[0,tmp13],[1,tmp13],[xxx],[xxx]
movq_m2r(*(wsptr+4), mm3); // wsptr[2,0],[2,1],[2,2],[2,3]
punpckldq_r2r(mm6, mm0); // wsptr[0,tmp10],[1,tmp10],[0,tmp13],[1,tmp13]
punpcklwd_r2r(mm5, mm1); // wsptr[0,tmp11],[1,tmp11],[xxx],[xxx]
movq_r2r(mm3, mm4);
movq_m2r(*(wsptr+6), mm6); // wsptr[3,0],[3,1],[3,2],[3,3]
punpckhwd_r2r(mm5, mm2); // wsptr[0,tmp14],[1,tmp14],[xxx],[xxx]
movq_m2r(*(wsptr+5), mm5); // wsptr[2,4],[2,5],[2,6],[2,7]
punpckldq_r2r(mm2, mm1); // wsptr[0,tmp11],[1,tmp11],[0,tmp14],[1,tmp14]
paddw_r2r(mm5, mm3); // wsptr[2,tmp10],[xxx],[2,tmp13],[xxx]
movq_r2r(mm6, mm2);
psubw_r2r(mm5, mm4); // wsptr[2,tmp11],[xxx],[2,tmp14],[xxx]
paddw_r2r(mm7, mm6); // wsptr[3,tmp10],[xxx],[3,tmp13],[xxx]
movq_r2r(mm3, mm5);
punpcklwd_r2r(mm6, mm3); // wsptr[2,tmp10],[3,tmp10],[xxx],[xxx]
psubw_r2r(mm7, mm2); // wsptr[3,tmp11],[xxx],[3,tmp14],[xxx]
punpckhwd_r2r(mm6, mm5); // wsptr[2,tmp13],[3,tmp13],[xxx],[xxx]
movq_r2r(mm4, mm7);
punpckldq_r2r(mm5, mm3); // wsptr[2,tmp10],[3,tmp10],[2,tmp13],[3,tmp13]
punpcklwd_r2r(mm2, mm4); // wsptr[2,tmp11],[3,tmp11],[xxx],[xxx]
punpckhwd_r2r(mm2, mm7); // wsptr[2,tmp14],[3,tmp14],[xxx],[xxx]
punpckldq_r2r(mm7, mm4); // wsptr[2,tmp11],[3,tmp11],[2,tmp14],[3,tmp14]
movq_r2r(mm1, mm6);
//OK
// mm0 = ;wsptr[0,tmp10],[1,tmp10],[0,tmp13],[1,tmp13]
// mm1 = ;wsptr[0,tmp11],[1,tmp11],[0,tmp14],[1,tmp14]
movq_r2r(mm0, mm2);
punpckhdq_r2r(mm4, mm6); // wsptr[0,tmp14],[1,tmp14],[2,tmp14],[3,tmp14]
punpckldq_r2r(mm4, mm1); // wsptr[0,tmp11],[1,tmp11],[2,tmp11],[3,tmp11]
psllw_i2r(2, mm6);
pmulhw_m2r(fix_141, mm6);
punpckldq_r2r(mm3, mm0); // wsptr[0,tmp10],[1,tmp10],[2,tmp10],[3,tmp10]
punpckhdq_r2r(mm3, mm2); // wsptr[0,tmp13],[1,tmp13],[2,tmp13],[3,tmp13]
movq_r2r(mm0, mm7);
// tmp0 = tmp10 + tmp13;
// tmp3 = tmp10 - tmp13;
paddw_r2r(mm2, mm0); // [0,tmp0],[1,tmp0],[2,tmp0],[3,tmp0]
psubw_r2r(mm2, mm7); // [0,tmp3],[1,tmp3],[2,tmp3],[3,tmp3]
// tmp12 = MULTIPLY(tmp14, FIX_1_414213562) - tmp13;
psubw_r2r(mm2, mm6); // wsptr[0,tmp12],[1,tmp12],[2,tmp12],[3,tmp12]
// tmp1 = tmp11 + tmp12;
// tmp2 = tmp11 - tmp12;
movq_r2r(mm1, mm5);
//OK
/* Odd part */
// z13 = (DCTELEM) wsptr[5] + (DCTELEM) wsptr[3];
// z10 = (DCTELEM) wsptr[5] - (DCTELEM) wsptr[3];
// z11 = (DCTELEM) wsptr[1] + (DCTELEM) wsptr[7];
// z12 = (DCTELEM) wsptr[1] - (DCTELEM) wsptr[7];
movq_m2r(*(wsptr), mm3); // wsptr[0,0],[0,1],[0,2],[0,3]
paddw_r2r(mm6, mm1); // [0,tmp1],[1,tmp1],[2,tmp1],[3,tmp1]
movq_m2r(*(wsptr+1), mm4); // wsptr[0,4],[0,5],[0,6],[0,7]
psubw_r2r(mm6, mm5); // [0,tmp2],[1,tmp2],[2,tmp2],[3,tmp2]
movq_r2r(mm3, mm6);
punpckldq_r2r(mm4, mm3); // wsptr[0,0],[0,1],[0,4],[0,5]
punpckhdq_r2r(mm6, mm4); // wsptr[0,6],[0,7],[0,2],[0,3]
movq_r2r(mm3, mm2);
//Save tmp0 and tmp1 in wsptr
movq_r2m(mm0, *(wsptr)); // save tmp0
paddw_r2r(mm4, mm2); // wsptr[xxx],[0,z11],[xxx],[0,z13]
//Continue with z10 --- z13
movq_m2r(*(wsptr+2), mm6); // wsptr[1,0],[1,1],[1,2],[1,3]
psubw_r2r(mm4, mm3); // wsptr[xxx],[0,z12],[xxx],[0,z10]
movq_m2r(*(wsptr+3), mm0); // wsptr[1,4],[1,5],[1,6],[1,7]
movq_r2r(mm6, mm4);
movq_r2m(mm1, *(wsptr+1)); // save tmp1
punpckldq_r2r(mm0, mm6); // wsptr[1,0],[1,1],[1,4],[1,5]
punpckhdq_r2r(mm4, mm0); // wsptr[1,6],[1,7],[1,2],[1,3]
movq_r2r(mm6, mm1);
//Save tmp2 and tmp3 in wsptr
paddw_r2r(mm0, mm6); // wsptr[xxx],[1,z11],[xxx],[1,z13]
movq_r2r(mm2, mm4);
//Continue with z10 --- z13
movq_r2m(mm5, *(wsptr+2)); // save tmp2
punpcklwd_r2r(mm6, mm2); // wsptr[xxx],[xxx],[0,z11],[1,z11]
psubw_r2r(mm0, mm1); // wsptr[xxx],[1,z12],[xxx],[1,z10]
punpckhwd_r2r(mm6, mm4); // wsptr[xxx],[xxx],[0,z13],[1,z13]
movq_r2r(mm3, mm0);
punpcklwd_r2r(mm1, mm3); // wsptr[xxx],[xxx],[0,z12],[1,z12]
movq_r2m(mm7, *(wsptr+3)); // save tmp3
punpckhwd_r2r(mm1, mm0); // wsptr[xxx],[xxx],[0,z10],[1,z10]
movq_m2r(*(wsptr+4), mm6); // wsptr[2,0],[2,1],[2,2],[2,3]
punpckhdq_r2r(mm2, mm0); // wsptr[0,z10],[1,z10],[0,z11],[1,z11]
movq_m2r(*(wsptr+5), mm7); // wsptr[2,4],[2,5],[2,6],[2,7]
punpckhdq_r2r(mm4, mm3); // wsptr[0,z12],[1,z12],[0,z13],[1,z13]
movq_m2r(*(wsptr+6), mm1); // wsptr[3,0],[3,1],[3,2],[3,3]
movq_r2r(mm6, mm4);
punpckldq_r2r(mm7, mm6); // wsptr[2,0],[2,1],[2,4],[2,5]
movq_r2r(mm1, mm5);
punpckhdq_r2r(mm4, mm7); // wsptr[2,6],[2,7],[2,2],[2,3]
movq_r2r(mm6, mm2);
movq_m2r(*(wsptr+7), mm4); // wsptr[3,4],[3,5],[3,6],[3,7]
paddw_r2r(mm7, mm6); // wsptr[xxx],[2,z11],[xxx],[2,z13]
psubw_r2r(mm7, mm2); // wsptr[xxx],[2,z12],[xxx],[2,z10]
punpckldq_r2r(mm4, mm1); // wsptr[3,0],[3,1],[3,4],[3,5]
punpckhdq_r2r(mm5, mm4); // wsptr[3,6],[3,7],[3,2],[3,3]
movq_r2r(mm1, mm7);
paddw_r2r(mm4, mm1); // wsptr[xxx],[3,z11],[xxx],[3,z13]
psubw_r2r(mm4, mm7); // wsptr[xxx],[3,z12],[xxx],[3,z10]
movq_r2r(mm6, mm5);
punpcklwd_r2r(mm1, mm6); // wsptr[xxx],[xxx],[2,z11],[3,z11]
punpckhwd_r2r(mm1, mm5); // wsptr[xxx],[xxx],[2,z13],[3,z13]
movq_r2r(mm2, mm4);
punpcklwd_r2r(mm7, mm2); // wsptr[xxx],[xxx],[2,z12],[3,z12]
punpckhwd_r2r(mm7, mm4); // wsptr[xxx],[xxx],[2,z10],[3,z10]
punpckhdq_r2r(mm6, mm4); // wsptr[2,z10],[3,z10],[2,z11],[3,z11]
punpckhdq_r2r(mm5, mm2); // wsptr[2,z12],[3,z12],[2,z13],[3,z13]
movq_r2r(mm0, mm5);
punpckldq_r2r(mm4, mm0); // wsptr[0,z10],[1,z10],[2,z10],[3,z10]
punpckhdq_r2r(mm4, mm5); // wsptr[0,z11],[1,z11],[2,z11],[3,z11]
movq_r2r(mm3, mm4);
punpckhdq_r2r(mm2, mm4); // wsptr[0,z13],[1,z13],[2,z13],[3,z13]
movq_r2r(mm5, mm1);
punpckldq_r2r(mm2, mm3); // wsptr[0,z12],[1,z12],[2,z12],[3,z12]
// tmp7 = z11 + z13; /* phase 5 */
// tmp8 = z11 - z13; /* phase 5 */
psubw_r2r(mm4, mm1); // tmp8
paddw_r2r(mm4, mm5); // tmp7
// tmp21 = MULTIPLY(tmp8, FIX_1_414213562); /* 2*c4 */
psllw_i2r(2, mm1);
psllw_i2r(2, mm0);
pmulhw_m2r(fix_141, mm1); // tmp21
// tmp20 = MULTIPLY(z12, (FIX_1_082392200- FIX_1_847759065)) /* 2*(c2-c6) */
// + MULTIPLY(z10, - FIX_1_847759065); /* 2*c2 */
psllw_i2r(2, mm3);
movq_r2r(mm0, mm7);
pmulhw_m2r(fix_n184, mm7);
movq_r2r(mm3, mm6);
movq_m2r(*(wsptr), mm2); // tmp0,final1
pmulhw_m2r(fix_108n184, mm6);
// tmp22 = MULTIPLY(z10,(FIX_1_847759065 - FIX_2_613125930)) /* -2*(c2+c6) */
// + MULTIPLY(z12, FIX_1_847759065); /* 2*c2 */
movq_r2r(mm2, mm4); // final1
pmulhw_m2r(fix_184n261, mm0);
paddw_r2r(mm5, mm2); // tmp0+tmp7,final1
pmulhw_m2r(fix_184, mm3);
psubw_r2r(mm5, mm4); // tmp0-tmp7,final1
// tmp6 = tmp22 - tmp7; /* phase 2 */
psraw_i2r(3, mm2); // outptr[0,0],[1,0],[2,0],[3,0],final1
paddw_r2r(mm6, mm7); // tmp20
psraw_i2r(3, mm4); // outptr[0,7],[1,7],[2,7],[3,7],final1
paddw_r2r(mm0, mm3); // tmp22
// tmp5 = tmp21 - tmp6;
psubw_r2r(mm5, mm3); // tmp6
// tmp4 = tmp20 + tmp5;
movq_m2r(*(wsptr+1), mm0); // tmp1,final2
psubw_r2r(mm3, mm1); // tmp5
movq_r2r(mm0, mm6); // final2
paddw_r2r(mm3, mm0); // tmp1+tmp6,final2
/* Final output stage: scale down by a factor of 8 and range-limit */
// outptr[0] = range_limit[IDESCALE(tmp0 + tmp7, PASS1_BITS+3)
// & RANGE_MASK];
// outptr[7] = range_limit[IDESCALE(tmp0 - tmp7, PASS1_BITS+3)
// & RANGE_MASK]; final1
// outptr[1] = range_limit[IDESCALE(tmp1 + tmp6, PASS1_BITS+3)
// & RANGE_MASK];
// outptr[6] = range_limit[IDESCALE(tmp1 - tmp6, PASS1_BITS+3)
// & RANGE_MASK]; final2
psubw_r2r(mm3, mm6); // tmp1-tmp6,final2
psraw_i2r(3, mm0); // outptr[0,1],[1,1],[2,1],[3,1]
psraw_i2r(3, mm6); // outptr[0,6],[1,6],[2,6],[3,6]
packuswb_r2r(mm4, mm0); // out[0,1],[1,1],[2,1],[3,1],[0,7],[1,7],[2,7],[3,7]
movq_m2r(*(wsptr+2), mm5); // tmp2,final3
packuswb_r2r(mm6, mm2); // out[0,0],[1,0],[2,0],[3,0],[0,6],[1,6],[2,6],[3,6]
// outptr[2] = range_limit[IDESCALE(tmp2 + tmp5, PASS1_BITS+3)
// & RANGE_MASK];
// outptr[5] = range_limit[IDESCALE(tmp2 - tmp5, PASS1_BITS+3)
// & RANGE_MASK]; final3
paddw_r2r(mm1, mm7); // tmp4
movq_r2r(mm5, mm3);
paddw_r2r(mm1, mm5); // tmp2+tmp5
psubw_r2r(mm1, mm3); // tmp2-tmp5
psraw_i2r(3, mm5); // outptr[0,2],[1,2],[2,2],[3,2]
movq_m2r(*(wsptr+3), mm4); // tmp3,final4
psraw_i2r(3, mm3); // outptr[0,5],[1,5],[2,5],[3,5]
// outptr[4] = range_limit[IDESCALE(tmp3 + tmp4, PASS1_BITS+3)
// & RANGE_MASK];
// outptr[3] = range_limit[IDESCALE(tmp3 - tmp4, PASS1_BITS+3)
// & RANGE_MASK]; final4
movq_r2r(mm4, mm6);
paddw_r2r(mm7, mm4); // tmp3+tmp4
psubw_r2r(mm7, mm6); // tmp3-tmp4
psraw_i2r(3, mm4); // outptr[0,4],[1,4],[2,4],[3,4]
psraw_i2r(3, mm6); // outptr[0,3],[1,3],[2,3],[3,3]
/*
movq_r2m(mm4, *dummy);
fprintf(stderr, "3-4 %016llx\n", dummy);
movq_r2m(mm4, *dummy);
fprintf(stderr, "3+4 %016llx\n", dummy);
*/
packuswb_r2r(mm4, mm5); // out[0,2],[1,2],[2,2],[3,2],[0,4],[1,4],[2,4],[3,4]
packuswb_r2r(mm3, mm6); // out[0,3],[1,3],[2,3],[3,3],[0,5],[1,5],[2,5],[3,5]
movq_r2r(mm2, mm4);
movq_r2r(mm5, mm7);
punpcklbw_r2r(mm0, mm2); // out[0,0],[0,1],[1,0],[1,1],[2,0],[2,1],[3,0],[3,1]
punpckhbw_r2r(mm0, mm4); // out[0,6],[0,7],[1,6],[1,7],[2,6],[2,7],[3,6],[3,7]
movq_r2r(mm2, mm1);
punpcklbw_r2r(mm6, mm5); // out[0,2],[0,3],[1,2],[1,3],[2,2],[2,3],[3,2],[3,3]
punpckhbw_r2r(mm6, mm7); // out[0,4],[0,5],[1,4],[1,5],[2,4],[2,5],[3,4],[3,5]
punpcklwd_r2r(mm5, mm2); // out[0,0],[0,1],[0,2],[0,3],[1,0],[1,1],[1,2],[1,3]
movq_r2r(mm7, mm6);
punpckhwd_r2r(mm5, mm1); // out[2,0],[2,1],[2,2],[2,3],[3,0],[3,1],[3,2],[3,3]
movq_r2r(mm2, mm0);
punpcklwd_r2r(mm4, mm6); // out[0,4],[0,5],[0,6],[0,7],[1,4],[1,5],[1,6],[1,7]
punpckldq_r2r(mm6, mm2); // out[0,0],[0,1],[0,2],[0,3],[0,4],[0,5],[0,6],[0,7]
movq_r2r(mm1, mm3);
punpckhwd_r2r(mm4, mm7); // out[2,4],[2,5],[2,6],[2,7],[3,4],[3,5],[3,6],[3,7]
dataptr += rskip;
movq_r2m(mm2, *(dataptr));
punpckhdq_r2r(mm6, mm0); // out[1,0],[1,1],[1,2],[1,3],[1,4],[1,5],[1,6],[1,7]
dataptr += rskip;
movq_r2m(mm0, *(dataptr));
punpckldq_r2r(mm7, mm1); // out[2,0],[2,1],[2,2],[2,3],[2,4],[2,5],[2,6],[2,7]
punpckhdq_r2r(mm7, mm3); // out[3,0],[3,1],[3,2],[3,3],[3,4],[3,5],[3,6],[3,7]
dataptr += rskip;
movq_r2m(mm1, *(dataptr));
dataptr += rskip;
movq_r2m(mm3, *(dataptr));
#else
__s32 tmp0, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7;
__s32 tmp10, tmp11, tmp12, tmp13;
__s32 z5, z10, z11, z12, z13;
__s16 *inptr;
__s32 *wsptr;
__u8 *outptr;
int ctr;
__s32 dcval;
__s32 workspace[64];
inptr = data;
wsptr = workspace;
for (ctr = 8; ctr > 0; ctr--) {
if ((inptr[8] | inptr[16] | inptr[24] |
inptr[32] | inptr[40] | inptr[48] | inptr[56]) == 0) {
dcval = inptr[0];
wsptr[0] = dcval;
wsptr[8] = dcval;
wsptr[16] = dcval;
wsptr[24] = dcval;
wsptr[32] = dcval;
wsptr[40] = dcval;
wsptr[48] = dcval;
wsptr[56] = dcval;
inptr++;
wsptr++;
continue;
}
tmp0 = inptr[0];
tmp1 = inptr[16];
tmp2 = inptr[32];
tmp3 = inptr[48];
tmp10 = tmp0 + tmp2;
tmp11 = tmp0 - tmp2;
tmp13 = tmp1 + tmp3;
tmp12 = MULTIPLY(tmp1 - tmp3, FIX_1_414213562) - tmp13;
tmp0 = tmp10 + tmp13;
tmp3 = tmp10 - tmp13;
tmp1 = tmp11 + tmp12;
tmp2 = tmp11 - tmp12;
tmp4 = inptr[8];
tmp5 = inptr[24];
tmp6 = inptr[40];
tmp7 = inptr[56];
z13 = tmp6 + tmp5;
z10 = tmp6 - tmp5;
z11 = tmp4 + tmp7;
z12 = tmp4 - tmp7;
tmp7 = z11 + z13;
tmp11 = MULTIPLY(z11 - z13, FIX_1_414213562);
z5 = MULTIPLY(z10 + z12, FIX_1_847759065);
tmp10 = MULTIPLY(z12, FIX_1_082392200) - z5;
tmp12 = MULTIPLY(z10, - FIX_2_613125930) + z5;
tmp6 = tmp12 - tmp7;
tmp5 = tmp11 - tmp6;
tmp4 = tmp10 + tmp5;
wsptr[0] = (__s32) (tmp0 + tmp7);
wsptr[56] = (__s32) (tmp0 - tmp7);
wsptr[8] = (__s32) (tmp1 + tmp6);
wsptr[48] = (__s32) (tmp1 - tmp6);
wsptr[16] = (__s32) (tmp2 + tmp5);
wsptr[40] = (__s32) (tmp2 - tmp5);
wsptr[32] = (__s32) (tmp3 + tmp4);
wsptr[24] = (__s32) (tmp3 - tmp4);
inptr++;
wsptr++;
}
wsptr = workspace;
for (ctr = 0; ctr < 8; ctr++) {
outptr = &(odata[ctr*rskip]);
tmp10 = wsptr[0] + wsptr[4];
tmp11 = wsptr[0] - wsptr[4];
tmp13 = wsptr[2] + wsptr[6];
tmp12 = MULTIPLY(wsptr[2] - wsptr[6], FIX_1_414213562) - tmp13;
tmp0 = tmp10 + tmp13;
tmp3 = tmp10 - tmp13;
tmp1 = tmp11 + tmp12;
tmp2 = tmp11 - tmp12;
z13 = wsptr[5] + wsptr[3];
z10 = wsptr[5] - wsptr[3];
z11 = wsptr[1] + wsptr[7];
z12 = wsptr[1] - wsptr[7];
tmp7 = z11 + z13;
tmp11 = MULTIPLY(z11 - z13, FIX_1_414213562);
z5 = MULTIPLY(z10 + z12, FIX_1_847759065);
tmp10 = MULTIPLY(z12, FIX_1_082392200) - z5;
tmp12 = MULTIPLY(z10, - FIX_2_613125930) + z5;
tmp6 = tmp12 - tmp7;
tmp5 = tmp11 - tmp6;
tmp4 = tmp10 + tmp5;
outptr[0] = RL(DESCALE(tmp0 + tmp7));
outptr[7] = RL(DESCALE(tmp0 - tmp7));
outptr[1] = RL(DESCALE(tmp1 + tmp6));
outptr[6] = RL(DESCALE(tmp1 - tmp6));
outptr[2] = RL(DESCALE(tmp2 + tmp5));
outptr[5] = RL(DESCALE(tmp2 - tmp5));
outptr[4] = RL(DESCALE(tmp3 + tmp4));
outptr[3] = RL(DESCALE(tmp3 - tmp4));
wsptr += 8;
}
#endif
}
/*
Main Routines
This file contains most of the initialisation and control functions
(C) Justin Schoeman 1998
*/
/*
Private function
Initialise all the cache-aliged data blocks
*/
void RTjpeg_init_data(void)
{
unsigned long dptr;
dptr=(unsigned long)&(RTjpeg_alldata[0]);
dptr+=32;
dptr=dptr>>5;
dptr=dptr<<5; /* cache align data */
RTjpeg_block=(__s16 *)dptr;
dptr+=sizeof(__s16)*64;
RTjpeg_lqt=(__s32 *)dptr;
dptr+=sizeof(__s32)*64;
RTjpeg_cqt=(__s32 *)dptr;
dptr+=sizeof(__s32)*64;
RTjpeg_liqt=(__u32 *)dptr;
dptr+=sizeof(__u32)*64;
RTjpeg_ciqt=(__u32 *)dptr;
}
/*
External Function
Re-set quality factor
Input: buf -> pointer to 128 ints for quant values store to pass back to
init_decompress.
Q -> quality factor (192=best, 32=worst)
*/
void RTjpeg_init_Q(__u8 Q)
{
int i;
__u64 qual;
qual=(__u64)Q<<(32-7); /* 32 bit FP, 255=2, 0=0 */
for(i=0; i<64; i++)
{
RTjpeg_lqt[i]=(__s32)((qual/((__u64)RTjpeg_lum_quant_tbl[i]<<16))>>3);
if(RTjpeg_lqt[i]==0)RTjpeg_lqt[i]=1;
RTjpeg_cqt[i]=(__s32)((qual/((__u64)RTjpeg_chrom_quant_tbl[i]<<16))>>3);
if(RTjpeg_cqt[i]==0)RTjpeg_cqt[i]=1;
RTjpeg_liqt[i]=(1<<16)/(RTjpeg_lqt[i]<<3);
RTjpeg_ciqt[i]=(1<<16)/(RTjpeg_cqt[i]<<3);
RTjpeg_lqt[i]=((1<<16)/RTjpeg_liqt[i])>>3;
RTjpeg_cqt[i]=((1<<16)/RTjpeg_ciqt[i])>>3;
}
RTjpeg_lb8=0;
while(RTjpeg_liqt[RTjpeg_ZZ[++RTjpeg_lb8]]<=8);
RTjpeg_lb8--;
RTjpeg_cb8=0;
while(RTjpeg_ciqt[RTjpeg_ZZ[++RTjpeg_cb8]]<=8);
RTjpeg_cb8--;
RTjpeg_dct_init();
RTjpeg_idct_init();
RTjpeg_quant_init();
}
/*
External Function
Initialise compression.
Input: buf -> pointer to 128 ints for quant values store to pass back to
init_decompress.
width -> width of image
height -> height of image
Q -> quality factor (192=best, 32=worst)
*/
void RTjpeg_init_compress(__u32 *buf, int width, int height, __u8 Q)
{
int i;
__u64 qual;
RTjpeg_init_data();
RTjpeg_width=width;
RTjpeg_height=height;
RTjpeg_Ywidth = RTjpeg_width>>3;
RTjpeg_Ysize=width * height;
RTjpeg_Cwidth = RTjpeg_width>>4;
RTjpeg_Csize= (width>>1) * height;
qual=(__u64)Q<<(32-7); /* 32 bit FP, 255=2, 0=0 */
for(i=0; i<64; i++)
{
RTjpeg_lqt[i]=(__s32)((qual/((__u64)RTjpeg_lum_quant_tbl[i]<<16))>>3);
if(RTjpeg_lqt[i]==0)RTjpeg_lqt[i]=1;
RTjpeg_cqt[i]=(__s32)((qual/((__u64)RTjpeg_chrom_quant_tbl[i]<<16))>>3);
if(RTjpeg_cqt[i]==0)RTjpeg_cqt[i]=1;
RTjpeg_liqt[i]=(1<<16)/(RTjpeg_lqt[i]<<3);
RTjpeg_ciqt[i]=(1<<16)/(RTjpeg_cqt[i]<<3);
RTjpeg_lqt[i]=((1<<16)/RTjpeg_liqt[i])>>3;
RTjpeg_cqt[i]=((1<<16)/RTjpeg_ciqt[i])>>3;
}
RTjpeg_lb8=0;
while(RTjpeg_liqt[RTjpeg_ZZ[++RTjpeg_lb8]]<=8);
RTjpeg_lb8--;
RTjpeg_cb8=0;
while(RTjpeg_ciqt[RTjpeg_ZZ[++RTjpeg_cb8]]<=8);
RTjpeg_cb8--;
RTjpeg_dct_init();
RTjpeg_quant_init();
for(i=0; i<64; i++)
buf[i]=RTjpeg_liqt[i];
for(i=0; i<64; i++)
buf[64+i]=RTjpeg_ciqt[i];
}
void RTjpeg_init_decompress(__u32 *buf, int width, int height)
{
int i;
RTjpeg_init_data();
RTjpeg_width=width;
RTjpeg_height=height;
RTjpeg_Ywidth = RTjpeg_width>>3;
RTjpeg_Ysize=width * height;
RTjpeg_Cwidth = RTjpeg_width>>4;
RTjpeg_Csize= (width>>1) * height;
for(i=0; i<64; i++)
{
RTjpeg_liqt[i]=buf[i];
RTjpeg_ciqt[i]=buf[i+64];
}
RTjpeg_lb8=0;
while(RTjpeg_liqt[RTjpeg_ZZ[++RTjpeg_lb8]]<=8);
RTjpeg_lb8--;
RTjpeg_cb8=0;
while(RTjpeg_ciqt[RTjpeg_ZZ[++RTjpeg_cb8]]<=8);
RTjpeg_cb8--;
RTjpeg_idct_init();
// RTjpeg_color_init();
}
int RTjpeg_compressYUV420(__s8 *sp, unsigned char *bp)
{
__s8 * sb;
register __s8 * bp1 = bp + (RTjpeg_width<<3);
register __s8 * bp2 = bp + RTjpeg_Ysize;
register __s8 * bp3 = bp2 + (RTjpeg_Csize>>1);
register int i, j, k;
#ifdef MMX
emms();
#endif
sb=sp;
/* Y */
for(i=RTjpeg_height>>1; i; i-=8)
{
for(j=0, k=0; j<RTjpeg_width; j+=16, k+=8)
{
RTjpeg_dctY(bp+j, RTjpeg_block, RTjpeg_Ywidth);
RTjpeg_quant(RTjpeg_block, RTjpeg_lqt);
sp+=RTjpeg_b2s(RTjpeg_block, sp, RTjpeg_lb8);
RTjpeg_dctY(bp+j+8, RTjpeg_block, RTjpeg_Ywidth);
RTjpeg_quant(RTjpeg_block, RTjpeg_lqt);
sp+=RTjpeg_b2s(RTjpeg_block, sp, RTjpeg_lb8);
RTjpeg_dctY(bp1+j, RTjpeg_block, RTjpeg_Ywidth);
RTjpeg_quant(RTjpeg_block, RTjpeg_lqt);
sp+=RTjpeg_b2s(RTjpeg_block, sp, RTjpeg_lb8);
RTjpeg_dctY(bp1+j+8, RTjpeg_block, RTjpeg_Ywidth);
RTjpeg_quant(RTjpeg_block, RTjpeg_lqt);
sp+=RTjpeg_b2s(RTjpeg_block, sp, RTjpeg_lb8);
RTjpeg_dctY(bp2+k, RTjpeg_block, RTjpeg_Cwidth);
RTjpeg_quant(RTjpeg_block, RTjpeg_cqt);
sp+=RTjpeg_b2s(RTjpeg_block, sp, RTjpeg_cb8);
RTjpeg_dctY(bp3+k, RTjpeg_block, RTjpeg_Cwidth);
RTjpeg_quant(RTjpeg_block, RTjpeg_cqt);
sp+=RTjpeg_b2s(RTjpeg_block, sp, RTjpeg_cb8);
}
bp+=RTjpeg_width<<4;
bp1+=RTjpeg_width<<4;
bp2+=RTjpeg_width<<2;
bp3+=RTjpeg_width<<2;
}
#ifdef MMX
emms();
#endif
return (sp-sb);
}
int RTjpeg_compressYUV422(__s8 *sp, unsigned char *bp)
{
__s8 * sb;
register __s8 * bp2 = bp + RTjpeg_Ysize;
register __s8 * bp3 = bp2 + RTjpeg_Csize;
register int i, j, k;
#ifdef MMX
emms();
#endif
sb=sp;
/* Y */
for(i=RTjpeg_height; i; i-=8)
{
for(j=0, k=0; j<RTjpeg_width; j+=16, k+=8)
{
RTjpeg_dctY(bp+j, RTjpeg_block, RTjpeg_Ywidth);
RTjpeg_quant(RTjpeg_block, RTjpeg_lqt);
sp+=RTjpeg_b2s(RTjpeg_block, sp, RTjpeg_lb8);
RTjpeg_dctY(bp+j+8, RTjpeg_block, RTjpeg_Ywidth);
RTjpeg_quant(RTjpeg_block, RTjpeg_lqt);
sp+=RTjpeg_b2s(RTjpeg_block, sp, RTjpeg_lb8);
RTjpeg_dctY(bp2+k, RTjpeg_block, RTjpeg_Cwidth);
RTjpeg_quant(RTjpeg_block, RTjpeg_cqt);
sp+=RTjpeg_b2s(RTjpeg_block, sp, RTjpeg_cb8);
RTjpeg_dctY(bp3+k, RTjpeg_block, RTjpeg_Cwidth);
RTjpeg_quant(RTjpeg_block, RTjpeg_cqt);
sp+=RTjpeg_b2s(RTjpeg_block, sp, RTjpeg_cb8);
}
bp+=RTjpeg_width<<3;
bp2+=RTjpeg_width<<2;
bp3+=RTjpeg_width<<2;
}
#ifdef MMX
emms();
#endif
return (sp-sb);
}
int RTjpeg_compress8(__s8 *sp, unsigned char *bp)
{
__s8 * sb;
int i, j;
#ifdef MMX
emms();
#endif
sb=sp;
/* Y */
for(i=0; i<RTjpeg_height; i+=8)
{
for(j=0; j<RTjpeg_width; j+=8)
{
RTjpeg_dctY(bp+j, RTjpeg_block, RTjpeg_width);
RTjpeg_quant(RTjpeg_block, RTjpeg_lqt);
sp+=RTjpeg_b2s(RTjpeg_block, sp, RTjpeg_lb8);
}
bp+=RTjpeg_width;
}
#ifdef MMX
emms();
#endif
return (sp-sb);
}
void RTjpeg_decompressYUV422(__s8 *sp, __u8 *bp)
{
register __s8 * bp2 = bp + RTjpeg_Ysize;
register __s8 * bp3 = bp2 + (RTjpeg_Csize);
int i, j,k;
#ifdef MMX
emms();
#endif
/* Y */
for(i=RTjpeg_height; i; i-=8)
{
for(k=0, j=0; j<RTjpeg_width; j+=16, k+=8) {
if(*sp==-1)sp++;
else
{
sp+=RTjpeg_s2b(RTjpeg_block, sp, RTjpeg_lb8, RTjpeg_liqt);
RTjpeg_idct(bp+j, RTjpeg_block, RTjpeg_width);
}
if(*sp==-1)sp++;
else
{
sp+=RTjpeg_s2b(RTjpeg_block, sp, RTjpeg_lb8, RTjpeg_liqt);
RTjpeg_idct(bp+j+8, RTjpeg_block, RTjpeg_width);
}
if(*sp==-1)sp++;
else
{
sp+=RTjpeg_s2b(RTjpeg_block, sp, RTjpeg_cb8, RTjpeg_ciqt);
RTjpeg_idct(bp2+k, RTjpeg_block, RTjpeg_width>>1);
}
if(*sp==-1)sp++;
else
{
sp+=RTjpeg_s2b(RTjpeg_block, sp, RTjpeg_cb8, RTjpeg_ciqt);
RTjpeg_idct(bp3+k, RTjpeg_block, RTjpeg_width>>1);
}
}
bp+=RTjpeg_width<<3;
bp2+=RTjpeg_width<<2;
bp3+=RTjpeg_width<<2;
}
#ifdef MMX
emms();
#endif
}
void RTjpeg_decompressYUV420(__s8 *sp, __u8 *bp)
{
register __s8 * bp1 = bp + (RTjpeg_width<<3);
register __s8 * bp2 = bp + RTjpeg_Ysize;
register __s8 * bp3 = bp2 + (RTjpeg_Csize>>1);
int i, j,k;
#ifdef MMX
emms();
#endif
/* Y */
for(i=RTjpeg_height>>1; i; i-=8)
{
for(k=0, j=0; j<RTjpeg_width; j+=16, k+=8) {
if(*sp==-1)sp++;
else
{
sp+=RTjpeg_s2b(RTjpeg_block, sp, RTjpeg_lb8, RTjpeg_liqt);
RTjpeg_idct(bp+j, RTjpeg_block, RTjpeg_width);
}
if(*sp==-1)sp++;
else
{
sp+=RTjpeg_s2b(RTjpeg_block, sp, RTjpeg_lb8, RTjpeg_liqt);
RTjpeg_idct(bp+j+8, RTjpeg_block, RTjpeg_width);
}
if(*sp==-1)sp++;
else
{
sp+=RTjpeg_s2b(RTjpeg_block, sp, RTjpeg_lb8, RTjpeg_liqt);
RTjpeg_idct(bp1+j, RTjpeg_block, RTjpeg_width);
}
if(*sp==-1)sp++;
else
{
sp+=RTjpeg_s2b(RTjpeg_block, sp, RTjpeg_lb8, RTjpeg_liqt);
RTjpeg_idct(bp1+j+8, RTjpeg_block, RTjpeg_width);
}
if(*sp==-1)sp++;
else
{
sp+=RTjpeg_s2b(RTjpeg_block, sp, RTjpeg_cb8, RTjpeg_ciqt);
RTjpeg_idct(bp2+k, RTjpeg_block, RTjpeg_width>>1);
}
if(*sp==-1)sp++;
else
{
sp+=RTjpeg_s2b(RTjpeg_block, sp, RTjpeg_cb8, RTjpeg_ciqt);
RTjpeg_idct(bp3+k, RTjpeg_block, RTjpeg_width>>1);
}
}
bp+=RTjpeg_width<<4;
bp1+=RTjpeg_width<<4;
bp2+=RTjpeg_width<<2;
bp3+=RTjpeg_width<<2;
}
#ifdef MMX
emms();
#endif
}
void RTjpeg_decompress8(__s8 *sp, __u8 *bp)
{
int i, j;
#ifdef MMX
emms();
#endif
/* Y */
for(i=0; i<RTjpeg_height; i+=8)
{
for(j=0; j<RTjpeg_width; j+=8)
if(*sp==-1)sp++;
else
{
sp+=RTjpeg_s2b(RTjpeg_block, sp, RTjpeg_lb8, RTjpeg_liqt);
RTjpeg_idct(bp+j, RTjpeg_block, RTjpeg_width);
}
bp+=RTjpeg_width<<3;
}
}
/*
External Function
Initialise additional data structures for motion compensation
*/
void RTjpeg_init_mcompress(void)
{
unsigned long tmp;
if(!RTjpeg_old)
{
RTjpeg_old=malloc((4*RTjpeg_width*RTjpeg_height)+32);
tmp=(unsigned long)RTjpeg_old;
tmp+=32;
tmp=tmp>>5;
RTjpeg_old=(__s16 *)(tmp<<5);
}
if (!RTjpeg_old)
{
fprintf(stderr, "RTjpeg: Could not allocate memory\n");
exit(-1);
}
bzero(RTjpeg_old, ((4*RTjpeg_width*RTjpeg_height)));
}
#ifdef MMX
int RTjpeg_bcomp(__s16 *old, mmx_t *mask)
{
int i;
mmx_t *mold=(mmx_t *)old;
mmx_t *mblock=(mmx_t *)RTjpeg_block;
volatile mmx_t result;
static mmx_t neg=(mmx_t)(unsigned long long)0xffffffffffffffffULL;
movq_m2r(*mask, mm7);
movq_m2r(neg, mm6);
pxor_r2r(mm5, mm5);
for(i=0; i<8; i++)
{
movq_m2r(*(mblock++), mm0);
movq_m2r(*(mblock++), mm2);
movq_m2r(*(mold++), mm1);
movq_m2r(*(mold++), mm3);
psubsw_r2r(mm1, mm0);
psubsw_r2r(mm3, mm2);
movq_r2r(mm0, mm1);
movq_r2r(mm2, mm3);
pcmpgtw_r2r(mm7, mm0);
pcmpgtw_r2r(mm7, mm2);
pxor_r2r(mm6, mm1);
pxor_r2r(mm6, mm3);
pcmpgtw_r2r(mm7, mm1);
pcmpgtw_r2r(mm7, mm3);
por_r2r(mm0, mm5);
por_r2r(mm2, mm5);
por_r2r(mm1, mm5);
por_r2r(mm3, mm5);
}
movq_r2m(mm5, result);
if(result.q)
{
// if(!RTjpeg_mtest)
// for(i=0; i<16; i++)((__u64 *)old)[i]=((__u64 *)RTjpeg_block)[i];
return 0;
}
// printf(".");
return 1;
}
#else
int RTjpeg_bcomp(__s16 *old, __u16 *mask)
{
int i;
for(i=0; i<64; i++)
if(abs(old[i]-RTjpeg_block[i])>*mask)
{
if(!RTjpeg_mtest)
for(i=0; i<16; i++)((__u64 *)old)[i]=((__u64 *)RTjpeg_block)[i];
return 0;
}
return 1;
}
#endif
void RTjpeg_set_test(int i)
{
RTjpeg_mtest=i;
}
int RTjpeg_mcompressYUV420(__s8 *sp, unsigned char *bp, __u16 lmask, __u16 cmask)
{
__s8 * sb;
//rh __s16 *block;
register __s8 * bp1 = bp + (RTjpeg_width<<3);
register __s8 * bp2 = bp + RTjpeg_Ysize;
register __s8 * bp3 = bp2 + (RTjpeg_Csize>>1);
register int i, j, k;
#ifdef MMX
emms();
RTjpeg_lmask=(mmx_t)(((__u64)lmask<<48)|((__u64)lmask<<32)|((__u64)lmask<<16)|lmask);
RTjpeg_cmask=(mmx_t)(((__u64)cmask<<48)|((__u64)cmask<<32)|((__u64)cmask<<16)|cmask);
#else
RTjpeg_lmask=lmask;
RTjpeg_cmask=cmask;
#endif
sb=sp;
block=RTjpeg_old;
/* Y */
for(i=RTjpeg_height>>1; i; i-=8)
{
for(j=0, k=0; j<RTjpeg_width; j+=16, k+=8)
{
RTjpeg_dctY(bp+j, RTjpeg_block, RTjpeg_Ywidth);
RTjpeg_quant(RTjpeg_block, RTjpeg_lqt);
if(RTjpeg_bcomp(block, &RTjpeg_lmask))
{
*((__u8 *)sp++)=255;
}
else sp+=RTjpeg_b2s(RTjpeg_block, sp, RTjpeg_lb8);
block+=64;
RTjpeg_dctY(bp+j+8, RTjpeg_block, RTjpeg_Ywidth);
RTjpeg_quant(RTjpeg_block, RTjpeg_lqt);
if(RTjpeg_bcomp(block, &RTjpeg_lmask))
{
*((__u8 *)sp++)=255;
}
else sp+=RTjpeg_b2s(RTjpeg_block, sp, RTjpeg_lb8);
block+=64;
RTjpeg_dctY(bp1+j, RTjpeg_block, RTjpeg_Ywidth);
RTjpeg_quant(RTjpeg_block, RTjpeg_lqt);
if(RTjpeg_bcomp(block, &RTjpeg_lmask))
{
*((__u8 *)sp++)=255;
}
else sp+=RTjpeg_b2s(RTjpeg_block, sp, RTjpeg_lb8);
block+=64;
RTjpeg_dctY(bp1+j+8, RTjpeg_block, RTjpeg_Ywidth);
RTjpeg_quant(RTjpeg_block, RTjpeg_lqt);
if(RTjpeg_bcomp(block, &RTjpeg_lmask))
{
*((__u8 *)sp++)=255;
}
else sp+=RTjpeg_b2s(RTjpeg_block, sp, RTjpeg_lb8);
block+=64;
RTjpeg_dctY(bp2+k, RTjpeg_block, RTjpeg_Cwidth);
RTjpeg_quant(RTjpeg_block, RTjpeg_cqt);
if(RTjpeg_bcomp(block, &RTjpeg_cmask))
{
*((__u8 *)sp++)=255;
}
else sp+=RTjpeg_b2s(RTjpeg_block, sp, RTjpeg_cb8);
block+=64;
RTjpeg_dctY(bp3+k, RTjpeg_block, RTjpeg_Cwidth);
RTjpeg_quant(RTjpeg_block, RTjpeg_cqt);
if(RTjpeg_bcomp(block, &RTjpeg_cmask))
{
*((__u8 *)sp++)=255;
}
else sp+=RTjpeg_b2s(RTjpeg_block, sp, RTjpeg_cb8);
block+=64;
}
bp+=RTjpeg_width<<4;
bp1+=RTjpeg_width<<4;
bp2+=RTjpeg_width<<2;
bp3+=RTjpeg_width<<2;
}
#ifdef MMX
emms();
#endif
return (sp-sb);
}
int RTjpeg_mcompressYUV422(__s8 *sp, unsigned char *bp, __u16 lmask, __u16 cmask)
{
__s8 * sb;
__s16 *block;
register __s8 * bp2;
register __s8 * bp3;
register int i, j, k;
#ifdef MMX
emms();
RTjpeg_lmask=(mmx_t)(((__u64)lmask<<48)|((__u64)lmask<<32)|((__u64)lmask<<16)|lmask);
RTjpeg_cmask=(mmx_t)(((__u64)cmask<<48)|((__u64)cmask<<32)|((__u64)cmask<<16)|cmask);
#else
RTjpeg_lmask=lmask;
RTjpeg_cmask=cmask;
#endif
bp = bp - RTjpeg_width*0;
bp2 = bp + RTjpeg_Ysize-RTjpeg_width*0;
bp3 = bp2 + RTjpeg_Csize;
sb=sp;
block=RTjpeg_old;
/* Y */
for(i=RTjpeg_height; i; i-=8)
{
for(j=0, k=0; j<RTjpeg_width; j+=16, k+=8)
{
RTjpeg_dctY(bp+j, RTjpeg_block, RTjpeg_Ywidth);
RTjpeg_quant(RTjpeg_block, RTjpeg_lqt);
if(RTjpeg_bcomp(block, &RTjpeg_lmask))
{
*((__u8 *)sp++)=255;
}
else sp+=RTjpeg_b2s(RTjpeg_block, sp, RTjpeg_lb8);
block+=64;
RTjpeg_dctY(bp+j+8, RTjpeg_block, RTjpeg_Ywidth);
RTjpeg_quant(RTjpeg_block, RTjpeg_lqt);
if(RTjpeg_bcomp(block, &RTjpeg_lmask))
{
*((__u8 *)sp++)=255;
}
else sp+=RTjpeg_b2s(RTjpeg_block, sp, RTjpeg_lb8);
block+=64;
RTjpeg_dctY(bp2+k, RTjpeg_block, RTjpeg_Cwidth);
RTjpeg_quant(RTjpeg_block, RTjpeg_cqt);
if(RTjpeg_bcomp(block, &RTjpeg_cmask))
{
*((__u8 *)sp++)=255;
}
else sp+=RTjpeg_b2s(RTjpeg_block, sp, RTjpeg_cb8);
block+=64;
RTjpeg_dctY(bp3+k, RTjpeg_block, RTjpeg_Cwidth);
RTjpeg_quant(RTjpeg_block, RTjpeg_cqt);
if(RTjpeg_bcomp(block, &RTjpeg_cmask))
{
*((__u8 *)sp++)=255;
}
else sp+=RTjpeg_b2s(RTjpeg_block, sp, RTjpeg_cb8);
block+=64;
}
bp+=RTjpeg_width<<3;
bp2+=RTjpeg_width<<2;
bp3+=RTjpeg_width<<2;
}
printf ("%d\n", block - RTjpeg_old);
#ifdef MMX
emms();
#endif
return (sp-sb);
}
int RTjpeg_mcompress8(__s8 *sp, unsigned char *bp, __u16 lmask)
{
__s8 * sb;
__s16 *block;
int i, j;
#ifdef MMX
emms();
RTjpeg_lmask=(mmx_t)(((__u64)lmask<<48)|((__u64)lmask<<32)|((__u64)lmask<<16)|lmask);
#else
RTjpeg_lmask=lmask;
#endif
sb=sp;
block=RTjpeg_old;
/* Y */
for(i=0; i<RTjpeg_height; i+=8)
{
for(j=0; j<RTjpeg_width; j+=8)
{
RTjpeg_dctY(bp+j, RTjpeg_block, RTjpeg_width);
RTjpeg_quant(RTjpeg_block, RTjpeg_lqt);
if(RTjpeg_bcomp(block, &RTjpeg_lmask))
{
*((__u8 *)sp++)=255;
// printf("* %d ", sp[-1]);
} else sp+=RTjpeg_b2s(RTjpeg_block, sp, RTjpeg_lb8);
block+=64;
}
bp+=RTjpeg_width<<3;
}
#ifdef MMX
emms();
#endif
return (sp-sb);
}
void RTjpeg_color_init(void)
{
}
#define KcrR 76284
#define KcrG 53281
#define KcbG 25625
#define KcbB 132252
#define Ky 76284
void RTjpeg_yuv422rgb(__u8 *buf, __u8 *rgb, int stride)
{
int tmp;
int i, j;
__s32 y, crR, crG, cbG, cbB;
__u8 *bufcr, *bufcb, *bufy, *bufoute;
int yskip;
yskip=RTjpeg_width;
bufcb=&buf[RTjpeg_width*RTjpeg_height];
bufcr=&buf[RTjpeg_width*RTjpeg_height+(RTjpeg_width*RTjpeg_height)/2];
bufy=&buf[0];
bufoute=rgb;
for(i=0; i<(RTjpeg_height); i++)
{
for(j=0; j<RTjpeg_width; j+=2)
{
crR=(*bufcr-128)*KcrR;
crG=(*(bufcr++)-128)*KcrG;
cbG=(*bufcb-128)*KcbG;
cbB=(*(bufcb++)-128)*KcbB;
y=(bufy[j]-16)*Ky;
tmp=(y+crR)>>16;
*(bufoute++)=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(y-crG-cbG)>>16;
*(bufoute++)=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(y+cbB)>>16;
*(bufoute++)=(tmp>255)?255:((tmp<0)?0:tmp);
y=(bufy[j+1]-16)*Ky;
tmp=(y+crR)>>16;
*(bufoute++)=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(y-crG-cbG)>>16;
*(bufoute++)=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(y+cbB)>>16;
*(bufoute++)=(tmp>255)?255:((tmp<0)?0:tmp);
}
bufy+=yskip;
}
}
void RTjpeg_yuv420rgb(__u8 *buf, __u8 *rgb, int stride)
{
int tmp;
int i, j;
__s32 y, crR, crG, cbG, cbB;
__u8 *bufcr, *bufcb, *bufy, *bufoute, *bufouto;
int oskip, yskip;
if(stride==0)
oskip=RTjpeg_width*3;
else
oskip=2*stride-RTjpeg_width*3;
yskip=RTjpeg_width;
bufcb=&buf[RTjpeg_width*RTjpeg_height];
bufcr=&buf[RTjpeg_width*RTjpeg_height+(RTjpeg_width*RTjpeg_height)/4];
bufy=&buf[0];
bufoute=rgb;
bufouto=rgb+RTjpeg_width*3;
for(i=0; i<(RTjpeg_height>>1); i++)
{
for(j=0; j<RTjpeg_width; j+=2)
{
crR=(*bufcr-128)*KcrR;
crG=(*(bufcr++)-128)*KcrG;
cbG=(*bufcb-128)*KcbG;
cbB=(*(bufcb++)-128)*KcbB;
y=(bufy[j]-16)*Ky;
tmp=(y+crR)>>16;
*(bufoute++)=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(y-crG-cbG)>>16;
*(bufoute++)=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(y+cbB)>>16;
*(bufoute++)=(tmp>255)?255:((tmp<0)?0:tmp);
y=(bufy[j+1]-16)*Ky;
tmp=(y+crR)>>16;
*(bufoute++)=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(y-crG-cbG)>>16;
*(bufoute++)=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(y+cbB)>>16;
*(bufoute++)=(tmp>255)?255:((tmp<0)?0:tmp);
y=(bufy[j+yskip]-16)*Ky;
tmp=(y+crR)>>16;
*(bufouto++)=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(y-crG-cbG)>>16;
*(bufouto++)=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(y+cbB)>>16;
*(bufouto++)=(tmp>255)?255:((tmp<0)?0:tmp);
y=(bufy[j+1+yskip]-16)*Ky;
tmp=(y+crR)>>16;
*(bufouto++)=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(y-crG-cbG)>>16;
*(bufouto++)=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(y+cbB)>>16;
*(bufouto++)=(tmp>255)?255:((tmp<0)?0:tmp);
}
bufoute+=oskip;
bufouto+=oskip;
bufy+=yskip<<1;
}
}
void RTjpeg_yuvrgb32(__u8 *buf, __u8 *rgb, int stride)
{
int tmp;
int i, j;
__s32 y, crR, crG, cbG, cbB;
__u8 *bufcr, *bufcb, *bufy, *bufoute, *bufouto;
int oskip, yskip;
if(stride==0)
oskip=RTjpeg_width*4;
else
oskip = 2*stride-RTjpeg_width*4;
yskip=RTjpeg_width;
bufcb=&buf[RTjpeg_width*RTjpeg_height];
bufcr=&buf[RTjpeg_width*RTjpeg_height+(RTjpeg_width*RTjpeg_height)/2];
bufy=&buf[0];
bufoute=rgb;
bufouto=rgb+RTjpeg_width*4;
for(i=0; i<(RTjpeg_height>>1); i++)
{
for(j=0; j<RTjpeg_width; j+=2)
{
crR=(*bufcr-128)*KcrR;
crG=(*(bufcr++)-128)*KcrG;
cbG=(*bufcb-128)*KcbG;
cbB=(*(bufcb++)-128)*KcbB;
y=(bufy[j]-16)*Ky;
tmp=(y+cbB)>>16;
*(bufoute++)=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(y-crG-cbG)>>16;
*(bufoute++)=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(y+crR)>>16;
*(bufoute++)=(tmp>255)?255:((tmp<0)?0:tmp);
bufoute++;
y=(bufy[j+1]-16)*Ky;
tmp=(y+cbB)>>16;
*(bufoute++)=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(y-crG-cbG)>>16;
*(bufoute++)=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(y+crR)>>16;
*(bufoute++)=(tmp>255)?255:((tmp<0)?0:tmp);
bufoute++;
y=(bufy[j+yskip]-16)*Ky;
tmp=(y+cbB)>>16;
*(bufouto++)=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(y-crG-cbG)>>16;
*(bufouto++)=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(y+crR)>>16;
*(bufouto++)=(tmp>255)?255:((tmp<0)?0:tmp);
bufouto++;
y=(bufy[j+1+yskip]-16)*Ky;
tmp=(y+cbB)>>16;
*(bufouto++)=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(y-crG-cbG)>>16;
*(bufouto++)=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(y+crR)>>16;
*(bufouto++)=(tmp>255)?255:((tmp<0)?0:tmp);
bufouto++;
}
bufoute+=oskip;
bufouto+=oskip;
bufy+=yskip<<1;
}
}
void RTjpeg_yuvrgb24(__u8 *buf, __u8 *rgb, int stride)
{
int tmp;
int i, j;
__s32 y, crR, crG, cbG, cbB;
__u8 *bufcr, *bufcb, *bufy, *bufoute, *bufouto;
int oskip, yskip;
if(stride==0)
oskip=RTjpeg_width*3;
else
oskip=2*stride - RTjpeg_width*3;
yskip=RTjpeg_width;
bufcb=&buf[RTjpeg_width*RTjpeg_height];
bufcr=&buf[RTjpeg_width*RTjpeg_height+(RTjpeg_width*RTjpeg_height)/4];
bufy=&buf[0];
bufoute=rgb;
bufouto=rgb+RTjpeg_width*3;
for(i=0; i<(RTjpeg_height>>1); i++)
{
for(j=0; j<RTjpeg_width; j+=2)
{
crR=(*bufcr-128)*KcrR;
crG=(*(bufcr++)-128)*KcrG;
cbG=(*bufcb-128)*KcbG;
cbB=(*(bufcb++)-128)*KcbB;
y=(bufy[j]-16)*Ky;
tmp=(y+cbB)>>16;
*(bufoute++)=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(y-crG-cbG)>>16;
*(bufoute++)=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(y+crR)>>16;
*(bufoute++)=(tmp>255)?255:((tmp<0)?0:tmp);
y=(bufy[j+1]-16)*Ky;
tmp=(y+cbB)>>16;
*(bufoute++)=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(y-crG-cbG)>>16;
*(bufoute++)=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(y+crR)>>16;
*(bufoute++)=(tmp>255)?255:((tmp<0)?0:tmp);
y=(bufy[j+yskip]-16)*Ky;
tmp=(y+cbB)>>16;
*(bufouto++)=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(y-crG-cbG)>>16;
*(bufouto++)=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(y+crR)>>16;
*(bufouto++)=(tmp>255)?255:((tmp<0)?0:tmp);
y=(bufy[j+1+yskip]-16)*Ky;
tmp=(y+cbB)>>16;
*(bufouto++)=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(y-crG-cbG)>>16;
*(bufouto++)=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(y+crR)>>16;
*(bufouto++)=(tmp>255)?255:((tmp<0)?0:tmp);
}
bufoute+=oskip;
bufouto+=oskip;
bufy+=yskip<<1;
}
}
void RTjpeg_yuvrgb16(__u8 *buf, __u8 *rgb, int stride)
{
int tmp;
int i, j;
__s32 y, crR, crG, cbG, cbB;
__u8 *bufcr, *bufcb, *bufy, *bufoute, *bufouto;
int oskip, yskip;
unsigned char r, g, b;
if(stride==0)
oskip=RTjpeg_width*2;
else
oskip=2*stride-RTjpeg_width*2;
yskip=RTjpeg_width;
bufcb=&buf[RTjpeg_width*RTjpeg_height];
bufcr=&buf[RTjpeg_width*RTjpeg_height+(RTjpeg_width*RTjpeg_height)/4];
bufy=&buf[0];
bufoute=rgb;
bufouto=rgb+RTjpeg_width*2;
for(i=0; i<(RTjpeg_height>>1); i++)
{
for(j=0; j<RTjpeg_width; j+=2)
{
crR=(*bufcr-128)*KcrR;
crG=(*(bufcr++)-128)*KcrG;
cbG=(*bufcb-128)*KcbG;
cbB=(*(bufcb++)-128)*KcbB;
y=(bufy[j]-16)*Ky;
tmp=(y+cbB)>>16;
b=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(y-crG-cbG)>>16;
g=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(y+crR)>>16;
r=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(int)((int)b >> 3);
tmp|=(int)(((int)g >> 2) << 5);
tmp|=(int)(((int)r >> 3) << 11);
*(bufoute++)=tmp&0xff;
*(bufoute++)=tmp>>8;
y=(bufy[j+1]-16)*Ky;
tmp=(y+cbB)>>16;
b=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(y-crG-cbG)>>16;
g=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(y+crR)>>16;
r=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(int)((int)b >> 3);
tmp|=(int)(((int)g >> 2) << 5);
tmp|=(int)(((int)r >> 3) << 11);
*(bufoute++)=tmp&0xff;
*(bufoute++)=tmp>>8;
y=(bufy[j+yskip]-16)*Ky;
tmp=(y+cbB)>>16;
b=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(y-crG-cbG)>>16;
g=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(y+crR)>>16;
r=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(int)((int)b >> 3);
tmp|=(int)(((int)g >> 2) << 5);
tmp|=(int)(((int)r >> 3) << 11);
*(bufouto++)=tmp&0xff;
*(bufouto++)=tmp>>8;
y=(bufy[j+1+yskip]-16)*Ky;
tmp=(y+cbB)>>16;
b=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(y-crG-cbG)>>16;
g=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(y+crR)>>16;
r=(tmp>255)?255:((tmp<0)?0:tmp);
tmp=(int)((int)b >> 3);
tmp|=(int)(((int)g >> 2) << 5);
tmp|=(int)(((int)r >> 3) << 11);
*(bufouto++)=tmp&0xff;
*(bufouto++)=tmp>>8;
}
bufoute+=oskip;
bufouto+=oskip;
bufy+=yskip<<1;
}
}
/* fix stride */
void RTjpeg_yuvrgb8(__u8 *buf, __u8 *rgb, int stride)
{
bcopy(buf, rgb, RTjpeg_width*RTjpeg_height);
}