mirror of
https://git.ffmpeg.org/ffmpeg.git
synced 2024-12-25 16:52:31 +00:00
bb198e198a
interlaced mpeg2 encoding P & B frames rate distored interlaced mb decission alternate scantable support 4mv encoding fixes (thats also why the regression tests change) passing height to most dsp functions interlaced mpeg4 encoding (no direct mode MBs yet) various related cleanups disabled old motion estimaton algorithms (log, full, ...) they will either be fixed or removed Originally committed as revision 2638 to svn://svn.ffmpeg.org/ffmpeg/trunk
2046 lines
76 KiB
C
2046 lines
76 KiB
C
/*
|
|
* MMX optimized DSP utils
|
|
* Copyright (c) 2000, 2001 Fabrice Bellard.
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2 of the License, or (at your option) any later version.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with this library; if not, write to the Free Software
|
|
* Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
|
|
*
|
|
* MMX optimization by Nick Kurshev <nickols_k@mail.ru>
|
|
*/
|
|
|
|
#include "../dsputil.h"
|
|
#include "../simple_idct.h"
|
|
|
|
extern const uint8_t ff_h263_loop_filter_strength[32];
|
|
|
|
int mm_flags; /* multimedia extension flags */
|
|
|
|
/* pixel operations */
|
|
static const uint64_t mm_bone __attribute__ ((aligned(8))) = 0x0101010101010101ULL;
|
|
static const uint64_t mm_wone __attribute__ ((aligned(8))) = 0x0001000100010001ULL;
|
|
static const uint64_t mm_wtwo __attribute__ ((aligned(8))) = 0x0002000200020002ULL;
|
|
|
|
static const uint64_t ff_pw_20 __attribute__ ((aligned(8))) = 0x0014001400140014ULL;
|
|
static const uint64_t ff_pw_3 __attribute__ ((aligned(8))) = 0x0003000300030003ULL;
|
|
static const uint64_t ff_pw_16 __attribute__ ((aligned(8))) = 0x0010001000100010ULL;
|
|
static const uint64_t ff_pw_15 __attribute__ ((aligned(8))) = 0x000F000F000F000FULL;
|
|
|
|
static const uint64_t ff_pb_FC __attribute__ ((aligned(8))) = 0xFCFCFCFCFCFCFCFCULL;
|
|
|
|
#define JUMPALIGN() __asm __volatile (".balign 8"::)
|
|
#define MOVQ_ZERO(regd) __asm __volatile ("pxor %%" #regd ", %%" #regd ::)
|
|
|
|
#define MOVQ_WONE(regd) \
|
|
__asm __volatile ( \
|
|
"pcmpeqd %%" #regd ", %%" #regd " \n\t" \
|
|
"psrlw $15, %%" #regd ::)
|
|
|
|
#define MOVQ_BFE(regd) \
|
|
__asm __volatile ( \
|
|
"pcmpeqd %%" #regd ", %%" #regd " \n\t"\
|
|
"paddb %%" #regd ", %%" #regd " \n\t" ::)
|
|
|
|
#ifndef PIC
|
|
#define MOVQ_BONE(regd) __asm __volatile ("movq %0, %%" #regd " \n\t" ::"m"(mm_bone))
|
|
#define MOVQ_WTWO(regd) __asm __volatile ("movq %0, %%" #regd " \n\t" ::"m"(mm_wtwo))
|
|
#else
|
|
// for shared library it's better to use this way for accessing constants
|
|
// pcmpeqd -> -1
|
|
#define MOVQ_BONE(regd) \
|
|
__asm __volatile ( \
|
|
"pcmpeqd %%" #regd ", %%" #regd " \n\t" \
|
|
"psrlw $15, %%" #regd " \n\t" \
|
|
"packuswb %%" #regd ", %%" #regd " \n\t" ::)
|
|
|
|
#define MOVQ_WTWO(regd) \
|
|
__asm __volatile ( \
|
|
"pcmpeqd %%" #regd ", %%" #regd " \n\t" \
|
|
"psrlw $15, %%" #regd " \n\t" \
|
|
"psllw $1, %%" #regd " \n\t"::)
|
|
|
|
#endif
|
|
|
|
// using regr as temporary and for the output result
|
|
// first argument is unmodifed and second is trashed
|
|
// regfe is supposed to contain 0xfefefefefefefefe
|
|
#define PAVGB_MMX_NO_RND(rega, regb, regr, regfe) \
|
|
"movq " #rega ", " #regr " \n\t"\
|
|
"pand " #regb ", " #regr " \n\t"\
|
|
"pxor " #rega ", " #regb " \n\t"\
|
|
"pand " #regfe "," #regb " \n\t"\
|
|
"psrlq $1, " #regb " \n\t"\
|
|
"paddb " #regb ", " #regr " \n\t"
|
|
|
|
#define PAVGB_MMX(rega, regb, regr, regfe) \
|
|
"movq " #rega ", " #regr " \n\t"\
|
|
"por " #regb ", " #regr " \n\t"\
|
|
"pxor " #rega ", " #regb " \n\t"\
|
|
"pand " #regfe "," #regb " \n\t"\
|
|
"psrlq $1, " #regb " \n\t"\
|
|
"psubb " #regb ", " #regr " \n\t"
|
|
|
|
// mm6 is supposed to contain 0xfefefefefefefefe
|
|
#define PAVGBP_MMX_NO_RND(rega, regb, regr, regc, regd, regp) \
|
|
"movq " #rega ", " #regr " \n\t"\
|
|
"movq " #regc ", " #regp " \n\t"\
|
|
"pand " #regb ", " #regr " \n\t"\
|
|
"pand " #regd ", " #regp " \n\t"\
|
|
"pxor " #rega ", " #regb " \n\t"\
|
|
"pxor " #regc ", " #regd " \n\t"\
|
|
"pand %%mm6, " #regb " \n\t"\
|
|
"pand %%mm6, " #regd " \n\t"\
|
|
"psrlq $1, " #regb " \n\t"\
|
|
"psrlq $1, " #regd " \n\t"\
|
|
"paddb " #regb ", " #regr " \n\t"\
|
|
"paddb " #regd ", " #regp " \n\t"
|
|
|
|
#define PAVGBP_MMX(rega, regb, regr, regc, regd, regp) \
|
|
"movq " #rega ", " #regr " \n\t"\
|
|
"movq " #regc ", " #regp " \n\t"\
|
|
"por " #regb ", " #regr " \n\t"\
|
|
"por " #regd ", " #regp " \n\t"\
|
|
"pxor " #rega ", " #regb " \n\t"\
|
|
"pxor " #regc ", " #regd " \n\t"\
|
|
"pand %%mm6, " #regb " \n\t"\
|
|
"pand %%mm6, " #regd " \n\t"\
|
|
"psrlq $1, " #regd " \n\t"\
|
|
"psrlq $1, " #regb " \n\t"\
|
|
"psubb " #regb ", " #regr " \n\t"\
|
|
"psubb " #regd ", " #regp " \n\t"
|
|
|
|
/***********************************/
|
|
/* MMX no rounding */
|
|
#define DEF(x, y) x ## _no_rnd_ ## y ##_mmx
|
|
#define SET_RND MOVQ_WONE
|
|
#define PAVGBP(a, b, c, d, e, f) PAVGBP_MMX_NO_RND(a, b, c, d, e, f)
|
|
#define PAVGB(a, b, c, e) PAVGB_MMX_NO_RND(a, b, c, e)
|
|
|
|
#include "dsputil_mmx_rnd.h"
|
|
|
|
#undef DEF
|
|
#undef SET_RND
|
|
#undef PAVGBP
|
|
#undef PAVGB
|
|
/***********************************/
|
|
/* MMX rounding */
|
|
|
|
#define DEF(x, y) x ## _ ## y ##_mmx
|
|
#define SET_RND MOVQ_WTWO
|
|
#define PAVGBP(a, b, c, d, e, f) PAVGBP_MMX(a, b, c, d, e, f)
|
|
#define PAVGB(a, b, c, e) PAVGB_MMX(a, b, c, e)
|
|
|
|
#include "dsputil_mmx_rnd.h"
|
|
|
|
#undef DEF
|
|
#undef SET_RND
|
|
#undef PAVGBP
|
|
#undef PAVGB
|
|
|
|
/***********************************/
|
|
/* 3Dnow specific */
|
|
|
|
#define DEF(x) x ## _3dnow
|
|
/* for Athlons PAVGUSB is prefered */
|
|
#define PAVGB "pavgusb"
|
|
|
|
#include "dsputil_mmx_avg.h"
|
|
|
|
#undef DEF
|
|
#undef PAVGB
|
|
|
|
/***********************************/
|
|
/* MMX2 specific */
|
|
|
|
#define DEF(x) x ## _mmx2
|
|
|
|
/* Introduced only in MMX2 set */
|
|
#define PAVGB "pavgb"
|
|
|
|
#include "dsputil_mmx_avg.h"
|
|
|
|
#undef DEF
|
|
#undef PAVGB
|
|
|
|
/***********************************/
|
|
/* standard MMX */
|
|
|
|
#ifdef CONFIG_ENCODERS
|
|
static void get_pixels_mmx(DCTELEM *block, const uint8_t *pixels, int line_size)
|
|
{
|
|
asm volatile(
|
|
"movl $-128, %%eax \n\t"
|
|
"pxor %%mm7, %%mm7 \n\t"
|
|
".balign 16 \n\t"
|
|
"1: \n\t"
|
|
"movq (%0), %%mm0 \n\t"
|
|
"movq (%0, %2), %%mm2 \n\t"
|
|
"movq %%mm0, %%mm1 \n\t"
|
|
"movq %%mm2, %%mm3 \n\t"
|
|
"punpcklbw %%mm7, %%mm0 \n\t"
|
|
"punpckhbw %%mm7, %%mm1 \n\t"
|
|
"punpcklbw %%mm7, %%mm2 \n\t"
|
|
"punpckhbw %%mm7, %%mm3 \n\t"
|
|
"movq %%mm0, (%1, %%eax)\n\t"
|
|
"movq %%mm1, 8(%1, %%eax)\n\t"
|
|
"movq %%mm2, 16(%1, %%eax)\n\t"
|
|
"movq %%mm3, 24(%1, %%eax)\n\t"
|
|
"addl %3, %0 \n\t"
|
|
"addl $32, %%eax \n\t"
|
|
"js 1b \n\t"
|
|
: "+r" (pixels)
|
|
: "r" (block+64), "r" (line_size), "r" (line_size*2)
|
|
: "%eax"
|
|
);
|
|
}
|
|
|
|
static inline void diff_pixels_mmx(DCTELEM *block, const uint8_t *s1, const uint8_t *s2, int stride)
|
|
{
|
|
asm volatile(
|
|
"pxor %%mm7, %%mm7 \n\t"
|
|
"movl $-128, %%eax \n\t"
|
|
".balign 16 \n\t"
|
|
"1: \n\t"
|
|
"movq (%0), %%mm0 \n\t"
|
|
"movq (%1), %%mm2 \n\t"
|
|
"movq %%mm0, %%mm1 \n\t"
|
|
"movq %%mm2, %%mm3 \n\t"
|
|
"punpcklbw %%mm7, %%mm0 \n\t"
|
|
"punpckhbw %%mm7, %%mm1 \n\t"
|
|
"punpcklbw %%mm7, %%mm2 \n\t"
|
|
"punpckhbw %%mm7, %%mm3 \n\t"
|
|
"psubw %%mm2, %%mm0 \n\t"
|
|
"psubw %%mm3, %%mm1 \n\t"
|
|
"movq %%mm0, (%2, %%eax)\n\t"
|
|
"movq %%mm1, 8(%2, %%eax)\n\t"
|
|
"addl %3, %0 \n\t"
|
|
"addl %3, %1 \n\t"
|
|
"addl $16, %%eax \n\t"
|
|
"jnz 1b \n\t"
|
|
: "+r" (s1), "+r" (s2)
|
|
: "r" (block+64), "r" (stride)
|
|
: "%eax"
|
|
);
|
|
}
|
|
#endif //CONFIG_ENCODERS
|
|
|
|
void put_pixels_clamped_mmx(const DCTELEM *block, uint8_t *pixels, int line_size)
|
|
{
|
|
const DCTELEM *p;
|
|
uint8_t *pix;
|
|
|
|
/* read the pixels */
|
|
p = block;
|
|
pix = pixels;
|
|
/* unrolled loop */
|
|
__asm __volatile(
|
|
"movq %3, %%mm0\n\t"
|
|
"movq 8%3, %%mm1\n\t"
|
|
"movq 16%3, %%mm2\n\t"
|
|
"movq 24%3, %%mm3\n\t"
|
|
"movq 32%3, %%mm4\n\t"
|
|
"movq 40%3, %%mm5\n\t"
|
|
"movq 48%3, %%mm6\n\t"
|
|
"movq 56%3, %%mm7\n\t"
|
|
"packuswb %%mm1, %%mm0\n\t"
|
|
"packuswb %%mm3, %%mm2\n\t"
|
|
"packuswb %%mm5, %%mm4\n\t"
|
|
"packuswb %%mm7, %%mm6\n\t"
|
|
"movq %%mm0, (%0)\n\t"
|
|
"movq %%mm2, (%0, %1)\n\t"
|
|
"movq %%mm4, (%0, %1, 2)\n\t"
|
|
"movq %%mm6, (%0, %2)\n\t"
|
|
::"r" (pix), "r" (line_size), "r" (line_size*3), "m"(*p)
|
|
:"memory");
|
|
pix += line_size*4;
|
|
p += 32;
|
|
|
|
// if here would be an exact copy of the code above
|
|
// compiler would generate some very strange code
|
|
// thus using "r"
|
|
__asm __volatile(
|
|
"movq (%3), %%mm0\n\t"
|
|
"movq 8(%3), %%mm1\n\t"
|
|
"movq 16(%3), %%mm2\n\t"
|
|
"movq 24(%3), %%mm3\n\t"
|
|
"movq 32(%3), %%mm4\n\t"
|
|
"movq 40(%3), %%mm5\n\t"
|
|
"movq 48(%3), %%mm6\n\t"
|
|
"movq 56(%3), %%mm7\n\t"
|
|
"packuswb %%mm1, %%mm0\n\t"
|
|
"packuswb %%mm3, %%mm2\n\t"
|
|
"packuswb %%mm5, %%mm4\n\t"
|
|
"packuswb %%mm7, %%mm6\n\t"
|
|
"movq %%mm0, (%0)\n\t"
|
|
"movq %%mm2, (%0, %1)\n\t"
|
|
"movq %%mm4, (%0, %1, 2)\n\t"
|
|
"movq %%mm6, (%0, %2)\n\t"
|
|
::"r" (pix), "r" (line_size), "r" (line_size*3), "r"(p)
|
|
:"memory");
|
|
}
|
|
|
|
void add_pixels_clamped_mmx(const DCTELEM *block, uint8_t *pixels, int line_size)
|
|
{
|
|
const DCTELEM *p;
|
|
uint8_t *pix;
|
|
int i;
|
|
|
|
/* read the pixels */
|
|
p = block;
|
|
pix = pixels;
|
|
MOVQ_ZERO(mm7);
|
|
i = 4;
|
|
do {
|
|
__asm __volatile(
|
|
"movq (%2), %%mm0\n\t"
|
|
"movq 8(%2), %%mm1\n\t"
|
|
"movq 16(%2), %%mm2\n\t"
|
|
"movq 24(%2), %%mm3\n\t"
|
|
"movq %0, %%mm4\n\t"
|
|
"movq %1, %%mm6\n\t"
|
|
"movq %%mm4, %%mm5\n\t"
|
|
"punpcklbw %%mm7, %%mm4\n\t"
|
|
"punpckhbw %%mm7, %%mm5\n\t"
|
|
"paddsw %%mm4, %%mm0\n\t"
|
|
"paddsw %%mm5, %%mm1\n\t"
|
|
"movq %%mm6, %%mm5\n\t"
|
|
"punpcklbw %%mm7, %%mm6\n\t"
|
|
"punpckhbw %%mm7, %%mm5\n\t"
|
|
"paddsw %%mm6, %%mm2\n\t"
|
|
"paddsw %%mm5, %%mm3\n\t"
|
|
"packuswb %%mm1, %%mm0\n\t"
|
|
"packuswb %%mm3, %%mm2\n\t"
|
|
"movq %%mm0, %0\n\t"
|
|
"movq %%mm2, %1\n\t"
|
|
:"+m"(*pix), "+m"(*(pix+line_size))
|
|
:"r"(p)
|
|
:"memory");
|
|
pix += line_size*2;
|
|
p += 16;
|
|
} while (--i);
|
|
}
|
|
|
|
static void put_pixels8_mmx(uint8_t *block, const uint8_t *pixels, int line_size, int h)
|
|
{
|
|
__asm __volatile(
|
|
"lea (%3, %3), %%eax \n\t"
|
|
".balign 8 \n\t"
|
|
"1: \n\t"
|
|
"movq (%1), %%mm0 \n\t"
|
|
"movq (%1, %3), %%mm1 \n\t"
|
|
"movq %%mm0, (%2) \n\t"
|
|
"movq %%mm1, (%2, %3) \n\t"
|
|
"addl %%eax, %1 \n\t"
|
|
"addl %%eax, %2 \n\t"
|
|
"movq (%1), %%mm0 \n\t"
|
|
"movq (%1, %3), %%mm1 \n\t"
|
|
"movq %%mm0, (%2) \n\t"
|
|
"movq %%mm1, (%2, %3) \n\t"
|
|
"addl %%eax, %1 \n\t"
|
|
"addl %%eax, %2 \n\t"
|
|
"subl $4, %0 \n\t"
|
|
"jnz 1b \n\t"
|
|
: "+g"(h), "+r" (pixels), "+r" (block)
|
|
: "r"(line_size)
|
|
: "%eax", "memory"
|
|
);
|
|
}
|
|
|
|
static void put_pixels16_mmx(uint8_t *block, const uint8_t *pixels, int line_size, int h)
|
|
{
|
|
__asm __volatile(
|
|
"lea (%3, %3), %%eax \n\t"
|
|
".balign 8 \n\t"
|
|
"1: \n\t"
|
|
"movq (%1), %%mm0 \n\t"
|
|
"movq 8(%1), %%mm4 \n\t"
|
|
"movq (%1, %3), %%mm1 \n\t"
|
|
"movq 8(%1, %3), %%mm5 \n\t"
|
|
"movq %%mm0, (%2) \n\t"
|
|
"movq %%mm4, 8(%2) \n\t"
|
|
"movq %%mm1, (%2, %3) \n\t"
|
|
"movq %%mm5, 8(%2, %3) \n\t"
|
|
"addl %%eax, %1 \n\t"
|
|
"addl %%eax, %2 \n\t"
|
|
"movq (%1), %%mm0 \n\t"
|
|
"movq 8(%1), %%mm4 \n\t"
|
|
"movq (%1, %3), %%mm1 \n\t"
|
|
"movq 8(%1, %3), %%mm5 \n\t"
|
|
"movq %%mm0, (%2) \n\t"
|
|
"movq %%mm4, 8(%2) \n\t"
|
|
"movq %%mm1, (%2, %3) \n\t"
|
|
"movq %%mm5, 8(%2, %3) \n\t"
|
|
"addl %%eax, %1 \n\t"
|
|
"addl %%eax, %2 \n\t"
|
|
"subl $4, %0 \n\t"
|
|
"jnz 1b \n\t"
|
|
: "+g"(h), "+r" (pixels), "+r" (block)
|
|
: "r"(line_size)
|
|
: "%eax", "memory"
|
|
);
|
|
}
|
|
|
|
static void clear_blocks_mmx(DCTELEM *blocks)
|
|
{
|
|
__asm __volatile(
|
|
"pxor %%mm7, %%mm7 \n\t"
|
|
"movl $-128*6, %%eax \n\t"
|
|
"1: \n\t"
|
|
"movq %%mm7, (%0, %%eax) \n\t"
|
|
"movq %%mm7, 8(%0, %%eax) \n\t"
|
|
"movq %%mm7, 16(%0, %%eax) \n\t"
|
|
"movq %%mm7, 24(%0, %%eax) \n\t"
|
|
"addl $32, %%eax \n\t"
|
|
" js 1b \n\t"
|
|
: : "r" (((int)blocks)+128*6)
|
|
: "%eax"
|
|
);
|
|
}
|
|
|
|
#ifdef CONFIG_ENCODERS
|
|
static int pix_sum16_mmx(uint8_t * pix, int line_size){
|
|
const int h=16;
|
|
int sum;
|
|
int index= -line_size*h;
|
|
|
|
__asm __volatile(
|
|
"pxor %%mm7, %%mm7 \n\t"
|
|
"pxor %%mm6, %%mm6 \n\t"
|
|
"1: \n\t"
|
|
"movq (%2, %1), %%mm0 \n\t"
|
|
"movq (%2, %1), %%mm1 \n\t"
|
|
"movq 8(%2, %1), %%mm2 \n\t"
|
|
"movq 8(%2, %1), %%mm3 \n\t"
|
|
"punpcklbw %%mm7, %%mm0 \n\t"
|
|
"punpckhbw %%mm7, %%mm1 \n\t"
|
|
"punpcklbw %%mm7, %%mm2 \n\t"
|
|
"punpckhbw %%mm7, %%mm3 \n\t"
|
|
"paddw %%mm0, %%mm1 \n\t"
|
|
"paddw %%mm2, %%mm3 \n\t"
|
|
"paddw %%mm1, %%mm3 \n\t"
|
|
"paddw %%mm3, %%mm6 \n\t"
|
|
"addl %3, %1 \n\t"
|
|
" js 1b \n\t"
|
|
"movq %%mm6, %%mm5 \n\t"
|
|
"psrlq $32, %%mm6 \n\t"
|
|
"paddw %%mm5, %%mm6 \n\t"
|
|
"movq %%mm6, %%mm5 \n\t"
|
|
"psrlq $16, %%mm6 \n\t"
|
|
"paddw %%mm5, %%mm6 \n\t"
|
|
"movd %%mm6, %0 \n\t"
|
|
"andl $0xFFFF, %0 \n\t"
|
|
: "=&r" (sum), "+r" (index)
|
|
: "r" (pix - index), "r" (line_size)
|
|
);
|
|
|
|
return sum;
|
|
}
|
|
#endif //CONFIG_ENCODERS
|
|
|
|
static void add_bytes_mmx(uint8_t *dst, uint8_t *src, int w){
|
|
int i=0;
|
|
asm volatile(
|
|
"1: \n\t"
|
|
"movq (%1, %0), %%mm0 \n\t"
|
|
"movq (%2, %0), %%mm1 \n\t"
|
|
"paddb %%mm0, %%mm1 \n\t"
|
|
"movq %%mm1, (%2, %0) \n\t"
|
|
"movq 8(%1, %0), %%mm0 \n\t"
|
|
"movq 8(%2, %0), %%mm1 \n\t"
|
|
"paddb %%mm0, %%mm1 \n\t"
|
|
"movq %%mm1, 8(%2, %0) \n\t"
|
|
"addl $16, %0 \n\t"
|
|
"cmpl %3, %0 \n\t"
|
|
" jb 1b \n\t"
|
|
: "+r" (i)
|
|
: "r"(src), "r"(dst), "r"(w-15)
|
|
);
|
|
for(; i<w; i++)
|
|
dst[i+0] += src[i+0];
|
|
}
|
|
|
|
#define H263_LOOP_FILTER \
|
|
"pxor %%mm7, %%mm7 \n\t"\
|
|
"movq %0, %%mm0 \n\t"\
|
|
"movq %0, %%mm1 \n\t"\
|
|
"movq %3, %%mm2 \n\t"\
|
|
"movq %3, %%mm3 \n\t"\
|
|
"punpcklbw %%mm7, %%mm0 \n\t"\
|
|
"punpckhbw %%mm7, %%mm1 \n\t"\
|
|
"punpcklbw %%mm7, %%mm2 \n\t"\
|
|
"punpckhbw %%mm7, %%mm3 \n\t"\
|
|
"psubw %%mm2, %%mm0 \n\t"\
|
|
"psubw %%mm3, %%mm1 \n\t"\
|
|
"movq %1, %%mm2 \n\t"\
|
|
"movq %1, %%mm3 \n\t"\
|
|
"movq %2, %%mm4 \n\t"\
|
|
"movq %2, %%mm5 \n\t"\
|
|
"punpcklbw %%mm7, %%mm2 \n\t"\
|
|
"punpckhbw %%mm7, %%mm3 \n\t"\
|
|
"punpcklbw %%mm7, %%mm4 \n\t"\
|
|
"punpckhbw %%mm7, %%mm5 \n\t"\
|
|
"psubw %%mm2, %%mm4 \n\t"\
|
|
"psubw %%mm3, %%mm5 \n\t"\
|
|
"psllw $2, %%mm4 \n\t"\
|
|
"psllw $2, %%mm5 \n\t"\
|
|
"paddw %%mm0, %%mm4 \n\t"\
|
|
"paddw %%mm1, %%mm5 \n\t"\
|
|
"pxor %%mm6, %%mm6 \n\t"\
|
|
"pcmpgtw %%mm4, %%mm6 \n\t"\
|
|
"pcmpgtw %%mm5, %%mm7 \n\t"\
|
|
"pxor %%mm6, %%mm4 \n\t"\
|
|
"pxor %%mm7, %%mm5 \n\t"\
|
|
"psubw %%mm6, %%mm4 \n\t"\
|
|
"psubw %%mm7, %%mm5 \n\t"\
|
|
"psrlw $3, %%mm4 \n\t"\
|
|
"psrlw $3, %%mm5 \n\t"\
|
|
"packuswb %%mm5, %%mm4 \n\t"\
|
|
"packsswb %%mm7, %%mm6 \n\t"\
|
|
"pxor %%mm7, %%mm7 \n\t"\
|
|
"movd %4, %%mm2 \n\t"\
|
|
"punpcklbw %%mm2, %%mm2 \n\t"\
|
|
"punpcklbw %%mm2, %%mm2 \n\t"\
|
|
"punpcklbw %%mm2, %%mm2 \n\t"\
|
|
"psubusb %%mm4, %%mm2 \n\t"\
|
|
"movq %%mm2, %%mm3 \n\t"\
|
|
"psubusb %%mm4, %%mm3 \n\t"\
|
|
"psubb %%mm3, %%mm2 \n\t"\
|
|
"movq %1, %%mm3 \n\t"\
|
|
"movq %2, %%mm4 \n\t"\
|
|
"pxor %%mm6, %%mm3 \n\t"\
|
|
"pxor %%mm6, %%mm4 \n\t"\
|
|
"paddusb %%mm2, %%mm3 \n\t"\
|
|
"psubusb %%mm2, %%mm4 \n\t"\
|
|
"pxor %%mm6, %%mm3 \n\t"\
|
|
"pxor %%mm6, %%mm4 \n\t"\
|
|
"paddusb %%mm2, %%mm2 \n\t"\
|
|
"packsswb %%mm1, %%mm0 \n\t"\
|
|
"pcmpgtb %%mm0, %%mm7 \n\t"\
|
|
"pxor %%mm7, %%mm0 \n\t"\
|
|
"psubb %%mm7, %%mm0 \n\t"\
|
|
"movq %%mm0, %%mm1 \n\t"\
|
|
"psubusb %%mm2, %%mm0 \n\t"\
|
|
"psubb %%mm0, %%mm1 \n\t"\
|
|
"pand %5, %%mm1 \n\t"\
|
|
"psrlw $2, %%mm1 \n\t"\
|
|
"pxor %%mm7, %%mm1 \n\t"\
|
|
"psubb %%mm7, %%mm1 \n\t"\
|
|
"movq %0, %%mm5 \n\t"\
|
|
"movq %3, %%mm6 \n\t"\
|
|
"psubb %%mm1, %%mm5 \n\t"\
|
|
"paddb %%mm1, %%mm6 \n\t"
|
|
|
|
static void h263_v_loop_filter_mmx(uint8_t *src, int stride, int qscale){
|
|
const int strength= ff_h263_loop_filter_strength[qscale];
|
|
|
|
asm volatile(
|
|
|
|
H263_LOOP_FILTER
|
|
|
|
"movq %%mm3, %1 \n\t"
|
|
"movq %%mm4, %2 \n\t"
|
|
"movq %%mm5, %0 \n\t"
|
|
"movq %%mm6, %3 \n\t"
|
|
: "+m" (*(uint64_t*)(src - 2*stride)),
|
|
"+m" (*(uint64_t*)(src - 1*stride)),
|
|
"+m" (*(uint64_t*)(src + 0*stride)),
|
|
"+m" (*(uint64_t*)(src + 1*stride))
|
|
: "g" (2*strength), "m"(ff_pb_FC)
|
|
);
|
|
}
|
|
|
|
static inline void transpose4x4(uint8_t *dst, uint8_t *src, int dst_stride, int src_stride){
|
|
asm volatile( //FIXME could save 1 instruction if done as 8x4 ...
|
|
"movd %4, %%mm0 \n\t"
|
|
"movd %5, %%mm1 \n\t"
|
|
"movd %6, %%mm2 \n\t"
|
|
"movd %7, %%mm3 \n\t"
|
|
"punpcklbw %%mm1, %%mm0 \n\t"
|
|
"punpcklbw %%mm3, %%mm2 \n\t"
|
|
"movq %%mm0, %%mm1 \n\t"
|
|
"punpcklwd %%mm2, %%mm0 \n\t"
|
|
"punpckhwd %%mm2, %%mm1 \n\t"
|
|
"movd %%mm0, %0 \n\t"
|
|
"punpckhdq %%mm0, %%mm0 \n\t"
|
|
"movd %%mm0, %1 \n\t"
|
|
"movd %%mm1, %2 \n\t"
|
|
"punpckhdq %%mm1, %%mm1 \n\t"
|
|
"movd %%mm1, %3 \n\t"
|
|
|
|
: "=m" (*(uint32_t*)(dst + 0*dst_stride)),
|
|
"=m" (*(uint32_t*)(dst + 1*dst_stride)),
|
|
"=m" (*(uint32_t*)(dst + 2*dst_stride)),
|
|
"=m" (*(uint32_t*)(dst + 3*dst_stride))
|
|
: "m" (*(uint32_t*)(src + 0*src_stride)),
|
|
"m" (*(uint32_t*)(src + 1*src_stride)),
|
|
"m" (*(uint32_t*)(src + 2*src_stride)),
|
|
"m" (*(uint32_t*)(src + 3*src_stride))
|
|
);
|
|
}
|
|
|
|
static void h263_h_loop_filter_mmx(uint8_t *src, int stride, int qscale){
|
|
const int strength= ff_h263_loop_filter_strength[qscale];
|
|
uint64_t temp[4] __attribute__ ((aligned(8)));
|
|
uint8_t *btemp= (uint8_t*)temp;
|
|
|
|
src -= 2;
|
|
|
|
transpose4x4(btemp , src , 8, stride);
|
|
transpose4x4(btemp+4, src + 4*stride, 8, stride);
|
|
asm volatile(
|
|
H263_LOOP_FILTER // 5 3 4 6
|
|
|
|
: "+m" (temp[0]),
|
|
"+m" (temp[1]),
|
|
"+m" (temp[2]),
|
|
"+m" (temp[3])
|
|
: "g" (2*strength), "m"(ff_pb_FC)
|
|
);
|
|
|
|
asm volatile(
|
|
"movq %%mm5, %%mm1 \n\t"
|
|
"movq %%mm4, %%mm0 \n\t"
|
|
"punpcklbw %%mm3, %%mm5 \n\t"
|
|
"punpcklbw %%mm6, %%mm4 \n\t"
|
|
"punpckhbw %%mm3, %%mm1 \n\t"
|
|
"punpckhbw %%mm6, %%mm0 \n\t"
|
|
"movq %%mm5, %%mm3 \n\t"
|
|
"movq %%mm1, %%mm6 \n\t"
|
|
"punpcklwd %%mm4, %%mm5 \n\t"
|
|
"punpcklwd %%mm0, %%mm1 \n\t"
|
|
"punpckhwd %%mm4, %%mm3 \n\t"
|
|
"punpckhwd %%mm0, %%mm6 \n\t"
|
|
"movd %%mm5, %0 \n\t"
|
|
"punpckhdq %%mm5, %%mm5 \n\t"
|
|
"movd %%mm5, %1 \n\t"
|
|
"movd %%mm3, %2 \n\t"
|
|
"punpckhdq %%mm3, %%mm3 \n\t"
|
|
"movd %%mm3, %3 \n\t"
|
|
"movd %%mm1, %4 \n\t"
|
|
"punpckhdq %%mm1, %%mm1 \n\t"
|
|
"movd %%mm1, %5 \n\t"
|
|
"movd %%mm6, %6 \n\t"
|
|
"punpckhdq %%mm6, %%mm6 \n\t"
|
|
"movd %%mm6, %7 \n\t"
|
|
: "=m" (*(uint32_t*)(src + 0*stride)),
|
|
"=m" (*(uint32_t*)(src + 1*stride)),
|
|
"=m" (*(uint32_t*)(src + 2*stride)),
|
|
"=m" (*(uint32_t*)(src + 3*stride)),
|
|
"=m" (*(uint32_t*)(src + 4*stride)),
|
|
"=m" (*(uint32_t*)(src + 5*stride)),
|
|
"=m" (*(uint32_t*)(src + 6*stride)),
|
|
"=m" (*(uint32_t*)(src + 7*stride))
|
|
);
|
|
}
|
|
|
|
#ifdef CONFIG_ENCODERS
|
|
static int pix_norm1_mmx(uint8_t *pix, int line_size) {
|
|
int tmp;
|
|
asm volatile (
|
|
"movl $16,%%ecx\n"
|
|
"pxor %%mm0,%%mm0\n"
|
|
"pxor %%mm7,%%mm7\n"
|
|
"1:\n"
|
|
"movq (%0),%%mm2\n" /* mm2 = pix[0-7] */
|
|
"movq 8(%0),%%mm3\n" /* mm3 = pix[8-15] */
|
|
|
|
"movq %%mm2,%%mm1\n" /* mm1 = mm2 = pix[0-7] */
|
|
|
|
"punpckhbw %%mm0,%%mm1\n" /* mm1 = [pix4-7] */
|
|
"punpcklbw %%mm0,%%mm2\n" /* mm2 = [pix0-3] */
|
|
|
|
"movq %%mm3,%%mm4\n" /* mm4 = mm3 = pix[8-15] */
|
|
"punpckhbw %%mm0,%%mm3\n" /* mm3 = [pix12-15] */
|
|
"punpcklbw %%mm0,%%mm4\n" /* mm4 = [pix8-11] */
|
|
|
|
"pmaddwd %%mm1,%%mm1\n" /* mm1 = (pix0^2+pix1^2,pix2^2+pix3^2) */
|
|
"pmaddwd %%mm2,%%mm2\n" /* mm2 = (pix4^2+pix5^2,pix6^2+pix7^2) */
|
|
|
|
"pmaddwd %%mm3,%%mm3\n"
|
|
"pmaddwd %%mm4,%%mm4\n"
|
|
|
|
"paddd %%mm1,%%mm2\n" /* mm2 = (pix0^2+pix1^2+pix4^2+pix5^2,
|
|
pix2^2+pix3^2+pix6^2+pix7^2) */
|
|
"paddd %%mm3,%%mm4\n"
|
|
"paddd %%mm2,%%mm7\n"
|
|
|
|
"addl %2, %0\n"
|
|
"paddd %%mm4,%%mm7\n"
|
|
"dec %%ecx\n"
|
|
"jnz 1b\n"
|
|
|
|
"movq %%mm7,%%mm1\n"
|
|
"psrlq $32, %%mm7\n" /* shift hi dword to lo */
|
|
"paddd %%mm7,%%mm1\n"
|
|
"movd %%mm1,%1\n"
|
|
: "+r" (pix), "=r"(tmp) : "r" (line_size) : "%ecx" );
|
|
return tmp;
|
|
}
|
|
|
|
static int sse16_mmx(void *v, uint8_t * pix1, uint8_t * pix2, int line_size, int h) {
|
|
int tmp;
|
|
asm volatile (
|
|
"movl %4,%%ecx\n"
|
|
"pxor %%mm0,%%mm0\n" /* mm0 = 0 */
|
|
"pxor %%mm7,%%mm7\n" /* mm7 holds the sum */
|
|
"1:\n"
|
|
"movq (%0),%%mm1\n" /* mm1 = pix1[0-7] */
|
|
"movq (%1),%%mm2\n" /* mm2 = pix2[0-7] */
|
|
"movq 8(%0),%%mm3\n" /* mm3 = pix1[8-15] */
|
|
"movq 8(%1),%%mm4\n" /* mm4 = pix2[8-15] */
|
|
|
|
/* todo: mm1-mm2, mm3-mm4 */
|
|
/* algo: substract mm1 from mm2 with saturation and vice versa */
|
|
/* OR the results to get absolute difference */
|
|
"movq %%mm1,%%mm5\n"
|
|
"movq %%mm3,%%mm6\n"
|
|
"psubusb %%mm2,%%mm1\n"
|
|
"psubusb %%mm4,%%mm3\n"
|
|
"psubusb %%mm5,%%mm2\n"
|
|
"psubusb %%mm6,%%mm4\n"
|
|
|
|
"por %%mm1,%%mm2\n"
|
|
"por %%mm3,%%mm4\n"
|
|
|
|
/* now convert to 16-bit vectors so we can square them */
|
|
"movq %%mm2,%%mm1\n"
|
|
"movq %%mm4,%%mm3\n"
|
|
|
|
"punpckhbw %%mm0,%%mm2\n"
|
|
"punpckhbw %%mm0,%%mm4\n"
|
|
"punpcklbw %%mm0,%%mm1\n" /* mm1 now spread over (mm1,mm2) */
|
|
"punpcklbw %%mm0,%%mm3\n" /* mm4 now spread over (mm3,mm4) */
|
|
|
|
"pmaddwd %%mm2,%%mm2\n"
|
|
"pmaddwd %%mm4,%%mm4\n"
|
|
"pmaddwd %%mm1,%%mm1\n"
|
|
"pmaddwd %%mm3,%%mm3\n"
|
|
|
|
"addl %3,%0\n"
|
|
"addl %3,%1\n"
|
|
|
|
"paddd %%mm2,%%mm1\n"
|
|
"paddd %%mm4,%%mm3\n"
|
|
"paddd %%mm1,%%mm7\n"
|
|
"paddd %%mm3,%%mm7\n"
|
|
|
|
"decl %%ecx\n"
|
|
"jnz 1b\n"
|
|
|
|
"movq %%mm7,%%mm1\n"
|
|
"psrlq $32, %%mm7\n" /* shift hi dword to lo */
|
|
"paddd %%mm7,%%mm1\n"
|
|
"movd %%mm1,%2\n"
|
|
: "+r" (pix1), "+r" (pix2), "=r"(tmp)
|
|
: "r" (line_size) , "m" (h)
|
|
: "%ecx");
|
|
return tmp;
|
|
}
|
|
|
|
static void diff_bytes_mmx(uint8_t *dst, uint8_t *src1, uint8_t *src2, int w){
|
|
int i=0;
|
|
asm volatile(
|
|
"1: \n\t"
|
|
"movq (%2, %0), %%mm0 \n\t"
|
|
"movq (%1, %0), %%mm1 \n\t"
|
|
"psubb %%mm0, %%mm1 \n\t"
|
|
"movq %%mm1, (%3, %0) \n\t"
|
|
"movq 8(%2, %0), %%mm0 \n\t"
|
|
"movq 8(%1, %0), %%mm1 \n\t"
|
|
"psubb %%mm0, %%mm1 \n\t"
|
|
"movq %%mm1, 8(%3, %0) \n\t"
|
|
"addl $16, %0 \n\t"
|
|
"cmpl %4, %0 \n\t"
|
|
" jb 1b \n\t"
|
|
: "+r" (i)
|
|
: "r"(src1), "r"(src2), "r"(dst), "r"(w-15)
|
|
);
|
|
for(; i<w; i++)
|
|
dst[i+0] = src1[i+0]-src2[i+0];
|
|
}
|
|
|
|
static void sub_hfyu_median_prediction_mmx2(uint8_t *dst, uint8_t *src1, uint8_t *src2, int w, int *left, int *left_top){
|
|
int i=0;
|
|
uint8_t l, lt;
|
|
|
|
asm volatile(
|
|
"1: \n\t"
|
|
"movq -1(%1, %0), %%mm0 \n\t" // LT
|
|
"movq (%1, %0), %%mm1 \n\t" // T
|
|
"movq -1(%2, %0), %%mm2 \n\t" // L
|
|
"movq (%2, %0), %%mm3 \n\t" // X
|
|
"movq %%mm2, %%mm4 \n\t" // L
|
|
"psubb %%mm0, %%mm2 \n\t"
|
|
"paddb %%mm1, %%mm2 \n\t" // L + T - LT
|
|
"movq %%mm4, %%mm5 \n\t" // L
|
|
"pmaxub %%mm1, %%mm4 \n\t" // max(T, L)
|
|
"pminub %%mm5, %%mm1 \n\t" // min(T, L)
|
|
"pminub %%mm2, %%mm4 \n\t"
|
|
"pmaxub %%mm1, %%mm4 \n\t"
|
|
"psubb %%mm4, %%mm3 \n\t" // dst - pred
|
|
"movq %%mm3, (%3, %0) \n\t"
|
|
"addl $8, %0 \n\t"
|
|
"cmpl %4, %0 \n\t"
|
|
" jb 1b \n\t"
|
|
: "+r" (i)
|
|
: "r"(src1), "r"(src2), "r"(dst), "r"(w)
|
|
);
|
|
|
|
l= *left;
|
|
lt= *left_top;
|
|
|
|
dst[0]= src2[0] - mid_pred(l, src1[0], (l + src1[0] - lt)&0xFF);
|
|
|
|
*left_top= src1[w-1];
|
|
*left = src2[w-1];
|
|
}
|
|
|
|
#define LBUTTERFLY2(a1,b1,a2,b2)\
|
|
"paddw " #b1 ", " #a1 " \n\t"\
|
|
"paddw " #b2 ", " #a2 " \n\t"\
|
|
"paddw " #b1 ", " #b1 " \n\t"\
|
|
"paddw " #b2 ", " #b2 " \n\t"\
|
|
"psubw " #a1 ", " #b1 " \n\t"\
|
|
"psubw " #a2 ", " #b2 " \n\t"
|
|
|
|
#define HADAMARD48\
|
|
LBUTTERFLY2(%%mm0, %%mm1, %%mm2, %%mm3)\
|
|
LBUTTERFLY2(%%mm4, %%mm5, %%mm6, %%mm7)\
|
|
LBUTTERFLY2(%%mm0, %%mm2, %%mm1, %%mm3)\
|
|
LBUTTERFLY2(%%mm4, %%mm6, %%mm5, %%mm7)\
|
|
LBUTTERFLY2(%%mm0, %%mm4, %%mm1, %%mm5)\
|
|
LBUTTERFLY2(%%mm2, %%mm6, %%mm3, %%mm7)\
|
|
|
|
#define MMABS(a,z)\
|
|
"pxor " #z ", " #z " \n\t"\
|
|
"pcmpgtw " #a ", " #z " \n\t"\
|
|
"pxor " #z ", " #a " \n\t"\
|
|
"psubw " #z ", " #a " \n\t"
|
|
|
|
#define MMABS_SUM(a,z, sum)\
|
|
"pxor " #z ", " #z " \n\t"\
|
|
"pcmpgtw " #a ", " #z " \n\t"\
|
|
"pxor " #z ", " #a " \n\t"\
|
|
"psubw " #z ", " #a " \n\t"\
|
|
"paddusw " #a ", " #sum " \n\t"
|
|
|
|
#define MMABS_MMX2(a,z)\
|
|
"pxor " #z ", " #z " \n\t"\
|
|
"psubw " #a ", " #z " \n\t"\
|
|
"pmaxsw " #z ", " #a " \n\t"
|
|
|
|
#define MMABS_SUM_MMX2(a,z, sum)\
|
|
"pxor " #z ", " #z " \n\t"\
|
|
"psubw " #a ", " #z " \n\t"\
|
|
"pmaxsw " #z ", " #a " \n\t"\
|
|
"paddusw " #a ", " #sum " \n\t"
|
|
|
|
#define SBUTTERFLY(a,b,t,n)\
|
|
"movq " #a ", " #t " \n\t" /* abcd */\
|
|
"punpckl" #n " " #b ", " #a " \n\t" /* aebf */\
|
|
"punpckh" #n " " #b ", " #t " \n\t" /* cgdh */\
|
|
|
|
#define TRANSPOSE4(a,b,c,d,t)\
|
|
SBUTTERFLY(a,b,t,wd) /* a=aebf t=cgdh */\
|
|
SBUTTERFLY(c,d,b,wd) /* c=imjn b=kolp */\
|
|
SBUTTERFLY(a,c,d,dq) /* a=aeim d=bfjn */\
|
|
SBUTTERFLY(t,b,c,dq) /* t=cgko c=dhlp */
|
|
|
|
#define LOAD4(o, a, b, c, d)\
|
|
"movq "#o"(%1), " #a " \n\t"\
|
|
"movq "#o"+16(%1), " #b " \n\t"\
|
|
"movq "#o"+32(%1), " #c " \n\t"\
|
|
"movq "#o"+48(%1), " #d " \n\t"
|
|
|
|
#define STORE4(o, a, b, c, d)\
|
|
"movq "#a", "#o"(%1) \n\t"\
|
|
"movq "#b", "#o"+16(%1) \n\t"\
|
|
"movq "#c", "#o"+32(%1) \n\t"\
|
|
"movq "#d", "#o"+48(%1) \n\t"\
|
|
|
|
static int hadamard8_diff_mmx(void *s, uint8_t *src1, uint8_t *src2, int stride, int h){
|
|
uint64_t temp[16] __align8;
|
|
int sum=0;
|
|
|
|
assert(h==8);
|
|
|
|
diff_pixels_mmx((DCTELEM*)temp, src1, src2, stride);
|
|
|
|
asm volatile(
|
|
LOAD4(0 , %%mm0, %%mm1, %%mm2, %%mm3)
|
|
LOAD4(64, %%mm4, %%mm5, %%mm6, %%mm7)
|
|
|
|
HADAMARD48
|
|
|
|
"movq %%mm7, 112(%1) \n\t"
|
|
|
|
TRANSPOSE4(%%mm0, %%mm1, %%mm2, %%mm3, %%mm7)
|
|
STORE4(0 , %%mm0, %%mm3, %%mm7, %%mm2)
|
|
|
|
"movq 112(%1), %%mm7 \n\t"
|
|
TRANSPOSE4(%%mm4, %%mm5, %%mm6, %%mm7, %%mm0)
|
|
STORE4(64, %%mm4, %%mm7, %%mm0, %%mm6)
|
|
|
|
LOAD4(8 , %%mm0, %%mm1, %%mm2, %%mm3)
|
|
LOAD4(72, %%mm4, %%mm5, %%mm6, %%mm7)
|
|
|
|
HADAMARD48
|
|
|
|
"movq %%mm7, 120(%1) \n\t"
|
|
|
|
TRANSPOSE4(%%mm0, %%mm1, %%mm2, %%mm3, %%mm7)
|
|
STORE4(8 , %%mm0, %%mm3, %%mm7, %%mm2)
|
|
|
|
"movq 120(%1), %%mm7 \n\t"
|
|
TRANSPOSE4(%%mm4, %%mm5, %%mm6, %%mm7, %%mm0)
|
|
"movq %%mm7, %%mm5 \n\t"//FIXME remove
|
|
"movq %%mm6, %%mm7 \n\t"
|
|
"movq %%mm0, %%mm6 \n\t"
|
|
// STORE4(72, %%mm4, %%mm7, %%mm0, %%mm6) //FIXME remove
|
|
|
|
LOAD4(64, %%mm0, %%mm1, %%mm2, %%mm3)
|
|
// LOAD4(72, %%mm4, %%mm5, %%mm6, %%mm7)
|
|
|
|
HADAMARD48
|
|
"movq %%mm7, 64(%1) \n\t"
|
|
MMABS(%%mm0, %%mm7)
|
|
MMABS_SUM(%%mm1, %%mm7, %%mm0)
|
|
MMABS_SUM(%%mm2, %%mm7, %%mm0)
|
|
MMABS_SUM(%%mm3, %%mm7, %%mm0)
|
|
MMABS_SUM(%%mm4, %%mm7, %%mm0)
|
|
MMABS_SUM(%%mm5, %%mm7, %%mm0)
|
|
MMABS_SUM(%%mm6, %%mm7, %%mm0)
|
|
"movq 64(%1), %%mm1 \n\t"
|
|
MMABS_SUM(%%mm1, %%mm7, %%mm0)
|
|
"movq %%mm0, 64(%1) \n\t"
|
|
|
|
LOAD4(0 , %%mm0, %%mm1, %%mm2, %%mm3)
|
|
LOAD4(8 , %%mm4, %%mm5, %%mm6, %%mm7)
|
|
|
|
HADAMARD48
|
|
"movq %%mm7, (%1) \n\t"
|
|
MMABS(%%mm0, %%mm7)
|
|
MMABS_SUM(%%mm1, %%mm7, %%mm0)
|
|
MMABS_SUM(%%mm2, %%mm7, %%mm0)
|
|
MMABS_SUM(%%mm3, %%mm7, %%mm0)
|
|
MMABS_SUM(%%mm4, %%mm7, %%mm0)
|
|
MMABS_SUM(%%mm5, %%mm7, %%mm0)
|
|
MMABS_SUM(%%mm6, %%mm7, %%mm0)
|
|
"movq (%1), %%mm1 \n\t"
|
|
MMABS_SUM(%%mm1, %%mm7, %%mm0)
|
|
"movq 64(%1), %%mm1 \n\t"
|
|
MMABS_SUM(%%mm1, %%mm7, %%mm0)
|
|
|
|
"movq %%mm0, %%mm1 \n\t"
|
|
"psrlq $32, %%mm0 \n\t"
|
|
"paddusw %%mm1, %%mm0 \n\t"
|
|
"movq %%mm0, %%mm1 \n\t"
|
|
"psrlq $16, %%mm0 \n\t"
|
|
"paddusw %%mm1, %%mm0 \n\t"
|
|
"movd %%mm0, %0 \n\t"
|
|
|
|
: "=r" (sum)
|
|
: "r"(temp)
|
|
);
|
|
return sum&0xFFFF;
|
|
}
|
|
|
|
static int hadamard8_diff_mmx2(void *s, uint8_t *src1, uint8_t *src2, int stride, int h){
|
|
uint64_t temp[16] __align8;
|
|
int sum=0;
|
|
|
|
assert(h==8);
|
|
|
|
diff_pixels_mmx((DCTELEM*)temp, src1, src2, stride);
|
|
|
|
asm volatile(
|
|
LOAD4(0 , %%mm0, %%mm1, %%mm2, %%mm3)
|
|
LOAD4(64, %%mm4, %%mm5, %%mm6, %%mm7)
|
|
|
|
HADAMARD48
|
|
|
|
"movq %%mm7, 112(%1) \n\t"
|
|
|
|
TRANSPOSE4(%%mm0, %%mm1, %%mm2, %%mm3, %%mm7)
|
|
STORE4(0 , %%mm0, %%mm3, %%mm7, %%mm2)
|
|
|
|
"movq 112(%1), %%mm7 \n\t"
|
|
TRANSPOSE4(%%mm4, %%mm5, %%mm6, %%mm7, %%mm0)
|
|
STORE4(64, %%mm4, %%mm7, %%mm0, %%mm6)
|
|
|
|
LOAD4(8 , %%mm0, %%mm1, %%mm2, %%mm3)
|
|
LOAD4(72, %%mm4, %%mm5, %%mm6, %%mm7)
|
|
|
|
HADAMARD48
|
|
|
|
"movq %%mm7, 120(%1) \n\t"
|
|
|
|
TRANSPOSE4(%%mm0, %%mm1, %%mm2, %%mm3, %%mm7)
|
|
STORE4(8 , %%mm0, %%mm3, %%mm7, %%mm2)
|
|
|
|
"movq 120(%1), %%mm7 \n\t"
|
|
TRANSPOSE4(%%mm4, %%mm5, %%mm6, %%mm7, %%mm0)
|
|
"movq %%mm7, %%mm5 \n\t"//FIXME remove
|
|
"movq %%mm6, %%mm7 \n\t"
|
|
"movq %%mm0, %%mm6 \n\t"
|
|
// STORE4(72, %%mm4, %%mm7, %%mm0, %%mm6) //FIXME remove
|
|
|
|
LOAD4(64, %%mm0, %%mm1, %%mm2, %%mm3)
|
|
// LOAD4(72, %%mm4, %%mm5, %%mm6, %%mm7)
|
|
|
|
HADAMARD48
|
|
"movq %%mm7, 64(%1) \n\t"
|
|
MMABS_MMX2(%%mm0, %%mm7)
|
|
MMABS_SUM_MMX2(%%mm1, %%mm7, %%mm0)
|
|
MMABS_SUM_MMX2(%%mm2, %%mm7, %%mm0)
|
|
MMABS_SUM_MMX2(%%mm3, %%mm7, %%mm0)
|
|
MMABS_SUM_MMX2(%%mm4, %%mm7, %%mm0)
|
|
MMABS_SUM_MMX2(%%mm5, %%mm7, %%mm0)
|
|
MMABS_SUM_MMX2(%%mm6, %%mm7, %%mm0)
|
|
"movq 64(%1), %%mm1 \n\t"
|
|
MMABS_SUM_MMX2(%%mm1, %%mm7, %%mm0)
|
|
"movq %%mm0, 64(%1) \n\t"
|
|
|
|
LOAD4(0 , %%mm0, %%mm1, %%mm2, %%mm3)
|
|
LOAD4(8 , %%mm4, %%mm5, %%mm6, %%mm7)
|
|
|
|
HADAMARD48
|
|
"movq %%mm7, (%1) \n\t"
|
|
MMABS_MMX2(%%mm0, %%mm7)
|
|
MMABS_SUM_MMX2(%%mm1, %%mm7, %%mm0)
|
|
MMABS_SUM_MMX2(%%mm2, %%mm7, %%mm0)
|
|
MMABS_SUM_MMX2(%%mm3, %%mm7, %%mm0)
|
|
MMABS_SUM_MMX2(%%mm4, %%mm7, %%mm0)
|
|
MMABS_SUM_MMX2(%%mm5, %%mm7, %%mm0)
|
|
MMABS_SUM_MMX2(%%mm6, %%mm7, %%mm0)
|
|
"movq (%1), %%mm1 \n\t"
|
|
MMABS_SUM_MMX2(%%mm1, %%mm7, %%mm0)
|
|
"movq 64(%1), %%mm1 \n\t"
|
|
MMABS_SUM_MMX2(%%mm1, %%mm7, %%mm0)
|
|
|
|
"movq %%mm0, %%mm1 \n\t"
|
|
"psrlq $32, %%mm0 \n\t"
|
|
"paddusw %%mm1, %%mm0 \n\t"
|
|
"movq %%mm0, %%mm1 \n\t"
|
|
"psrlq $16, %%mm0 \n\t"
|
|
"paddusw %%mm1, %%mm0 \n\t"
|
|
"movd %%mm0, %0 \n\t"
|
|
|
|
: "=r" (sum)
|
|
: "r"(temp)
|
|
);
|
|
return sum&0xFFFF;
|
|
}
|
|
|
|
|
|
WARPER8_16_SQ(hadamard8_diff_mmx, hadamard8_diff16_mmx)
|
|
WARPER8_16_SQ(hadamard8_diff_mmx2, hadamard8_diff16_mmx2)
|
|
#endif //CONFIG_ENCODERS
|
|
|
|
#define put_no_rnd_pixels8_mmx(a,b,c,d) put_pixels8_mmx(a,b,c,d)
|
|
#define put_no_rnd_pixels16_mmx(a,b,c,d) put_pixels16_mmx(a,b,c,d)
|
|
|
|
#define QPEL_V_LOW(m3,m4,m5,m6, pw_20, pw_3, rnd, in0, in1, in2, in7, out, OP)\
|
|
"paddw " #m4 ", " #m3 " \n\t" /* x1 */\
|
|
"movq "MANGLE(ff_pw_20)", %%mm4 \n\t" /* 20 */\
|
|
"pmullw " #m3 ", %%mm4 \n\t" /* 20x1 */\
|
|
"movq "#in7", " #m3 " \n\t" /* d */\
|
|
"movq "#in0", %%mm5 \n\t" /* D */\
|
|
"paddw " #m3 ", %%mm5 \n\t" /* x4 */\
|
|
"psubw %%mm5, %%mm4 \n\t" /* 20x1 - x4 */\
|
|
"movq "#in1", %%mm5 \n\t" /* C */\
|
|
"movq "#in2", %%mm6 \n\t" /* B */\
|
|
"paddw " #m6 ", %%mm5 \n\t" /* x3 */\
|
|
"paddw " #m5 ", %%mm6 \n\t" /* x2 */\
|
|
"paddw %%mm6, %%mm6 \n\t" /* 2x2 */\
|
|
"psubw %%mm6, %%mm5 \n\t" /* -2x2 + x3 */\
|
|
"pmullw "MANGLE(ff_pw_3)", %%mm5 \n\t" /* -6x2 + 3x3 */\
|
|
"paddw " #rnd ", %%mm4 \n\t" /* x2 */\
|
|
"paddw %%mm4, %%mm5 \n\t" /* 20x1 - 6x2 + 3x3 - x4 */\
|
|
"psraw $5, %%mm5 \n\t"\
|
|
"packuswb %%mm5, %%mm5 \n\t"\
|
|
OP(%%mm5, out, %%mm7, d)
|
|
|
|
#define QPEL_BASE(OPNAME, ROUNDER, RND, OP_MMX2, OP_3DNOW)\
|
|
static void OPNAME ## mpeg4_qpel16_h_lowpass_mmx2(uint8_t *dst, uint8_t *src, int dstStride, int srcStride, int h){\
|
|
uint64_t temp;\
|
|
\
|
|
asm volatile(\
|
|
"pxor %%mm7, %%mm7 \n\t"\
|
|
"1: \n\t"\
|
|
"movq (%0), %%mm0 \n\t" /* ABCDEFGH */\
|
|
"movq %%mm0, %%mm1 \n\t" /* ABCDEFGH */\
|
|
"movq %%mm0, %%mm2 \n\t" /* ABCDEFGH */\
|
|
"punpcklbw %%mm7, %%mm0 \n\t" /* 0A0B0C0D */\
|
|
"punpckhbw %%mm7, %%mm1 \n\t" /* 0E0F0G0H */\
|
|
"pshufw $0x90, %%mm0, %%mm5 \n\t" /* 0A0A0B0C */\
|
|
"pshufw $0x41, %%mm0, %%mm6 \n\t" /* 0B0A0A0B */\
|
|
"movq %%mm2, %%mm3 \n\t" /* ABCDEFGH */\
|
|
"movq %%mm2, %%mm4 \n\t" /* ABCDEFGH */\
|
|
"psllq $8, %%mm2 \n\t" /* 0ABCDEFG */\
|
|
"psllq $16, %%mm3 \n\t" /* 00ABCDEF */\
|
|
"psllq $24, %%mm4 \n\t" /* 000ABCDE */\
|
|
"punpckhbw %%mm7, %%mm2 \n\t" /* 0D0E0F0G */\
|
|
"punpckhbw %%mm7, %%mm3 \n\t" /* 0C0D0E0F */\
|
|
"punpckhbw %%mm7, %%mm4 \n\t" /* 0B0C0D0E */\
|
|
"paddw %%mm3, %%mm5 \n\t" /* b */\
|
|
"paddw %%mm2, %%mm6 \n\t" /* c */\
|
|
"paddw %%mm5, %%mm5 \n\t" /* 2b */\
|
|
"psubw %%mm5, %%mm6 \n\t" /* c - 2b */\
|
|
"pshufw $0x06, %%mm0, %%mm5 \n\t" /* 0C0B0A0A */\
|
|
"pmullw "MANGLE(ff_pw_3)", %%mm6 \n\t" /* 3c - 6b */\
|
|
"paddw %%mm4, %%mm0 \n\t" /* a */\
|
|
"paddw %%mm1, %%mm5 \n\t" /* d */\
|
|
"pmullw "MANGLE(ff_pw_20)", %%mm0 \n\t" /* 20a */\
|
|
"psubw %%mm5, %%mm0 \n\t" /* 20a - d */\
|
|
"paddw %6, %%mm6 \n\t"\
|
|
"paddw %%mm6, %%mm0 \n\t" /* 20a - 6b + 3c - d */\
|
|
"psraw $5, %%mm0 \n\t"\
|
|
"movq %%mm0, %5 \n\t"\
|
|
/* mm1=EFGH, mm2=DEFG, mm3=CDEF, mm4=BCDE, mm7=0 */\
|
|
\
|
|
"movq 5(%0), %%mm0 \n\t" /* FGHIJKLM */\
|
|
"movq %%mm0, %%mm5 \n\t" /* FGHIJKLM */\
|
|
"movq %%mm0, %%mm6 \n\t" /* FGHIJKLM */\
|
|
"psrlq $8, %%mm0 \n\t" /* GHIJKLM0 */\
|
|
"psrlq $16, %%mm5 \n\t" /* HIJKLM00 */\
|
|
"punpcklbw %%mm7, %%mm0 \n\t" /* 0G0H0I0J */\
|
|
"punpcklbw %%mm7, %%mm5 \n\t" /* 0H0I0J0K */\
|
|
"paddw %%mm0, %%mm2 \n\t" /* b */\
|
|
"paddw %%mm5, %%mm3 \n\t" /* c */\
|
|
"paddw %%mm2, %%mm2 \n\t" /* 2b */\
|
|
"psubw %%mm2, %%mm3 \n\t" /* c - 2b */\
|
|
"movq %%mm6, %%mm2 \n\t" /* FGHIJKLM */\
|
|
"psrlq $24, %%mm6 \n\t" /* IJKLM000 */\
|
|
"punpcklbw %%mm7, %%mm2 \n\t" /* 0F0G0H0I */\
|
|
"punpcklbw %%mm7, %%mm6 \n\t" /* 0I0J0K0L */\
|
|
"pmullw "MANGLE(ff_pw_3)", %%mm3 \n\t" /* 3c - 6b */\
|
|
"paddw %%mm2, %%mm1 \n\t" /* a */\
|
|
"paddw %%mm6, %%mm4 \n\t" /* d */\
|
|
"pmullw "MANGLE(ff_pw_20)", %%mm1 \n\t" /* 20a */\
|
|
"psubw %%mm4, %%mm3 \n\t" /* - 6b +3c - d */\
|
|
"paddw %6, %%mm1 \n\t"\
|
|
"paddw %%mm1, %%mm3 \n\t" /* 20a - 6b +3c - d */\
|
|
"psraw $5, %%mm3 \n\t"\
|
|
"movq %5, %%mm1 \n\t"\
|
|
"packuswb %%mm3, %%mm1 \n\t"\
|
|
OP_MMX2(%%mm1, (%1),%%mm4, q)\
|
|
/* mm0= GHIJ, mm2=FGHI, mm5=HIJK, mm6=IJKL, mm7=0 */\
|
|
\
|
|
"movq 9(%0), %%mm1 \n\t" /* JKLMNOPQ */\
|
|
"movq %%mm1, %%mm4 \n\t" /* JKLMNOPQ */\
|
|
"movq %%mm1, %%mm3 \n\t" /* JKLMNOPQ */\
|
|
"psrlq $8, %%mm1 \n\t" /* KLMNOPQ0 */\
|
|
"psrlq $16, %%mm4 \n\t" /* LMNOPQ00 */\
|
|
"punpcklbw %%mm7, %%mm1 \n\t" /* 0K0L0M0N */\
|
|
"punpcklbw %%mm7, %%mm4 \n\t" /* 0L0M0N0O */\
|
|
"paddw %%mm1, %%mm5 \n\t" /* b */\
|
|
"paddw %%mm4, %%mm0 \n\t" /* c */\
|
|
"paddw %%mm5, %%mm5 \n\t" /* 2b */\
|
|
"psubw %%mm5, %%mm0 \n\t" /* c - 2b */\
|
|
"movq %%mm3, %%mm5 \n\t" /* JKLMNOPQ */\
|
|
"psrlq $24, %%mm3 \n\t" /* MNOPQ000 */\
|
|
"pmullw "MANGLE(ff_pw_3)", %%mm0 \n\t" /* 3c - 6b */\
|
|
"punpcklbw %%mm7, %%mm3 \n\t" /* 0M0N0O0P */\
|
|
"paddw %%mm3, %%mm2 \n\t" /* d */\
|
|
"psubw %%mm2, %%mm0 \n\t" /* -6b + 3c - d */\
|
|
"movq %%mm5, %%mm2 \n\t" /* JKLMNOPQ */\
|
|
"punpcklbw %%mm7, %%mm2 \n\t" /* 0J0K0L0M */\
|
|
"punpckhbw %%mm7, %%mm5 \n\t" /* 0N0O0P0Q */\
|
|
"paddw %%mm2, %%mm6 \n\t" /* a */\
|
|
"pmullw "MANGLE(ff_pw_20)", %%mm6 \n\t" /* 20a */\
|
|
"paddw %6, %%mm0 \n\t"\
|
|
"paddw %%mm6, %%mm0 \n\t" /* 20a - 6b + 3c - d */\
|
|
"psraw $5, %%mm0 \n\t"\
|
|
/* mm1=KLMN, mm2=JKLM, mm3=MNOP, mm4=LMNO, mm5=NOPQ mm7=0 */\
|
|
\
|
|
"paddw %%mm5, %%mm3 \n\t" /* a */\
|
|
"pshufw $0xF9, %%mm5, %%mm6 \n\t" /* 0O0P0Q0Q */\
|
|
"paddw %%mm4, %%mm6 \n\t" /* b */\
|
|
"pshufw $0xBE, %%mm5, %%mm4 \n\t" /* 0P0Q0Q0P */\
|
|
"pshufw $0x6F, %%mm5, %%mm5 \n\t" /* 0Q0Q0P0O */\
|
|
"paddw %%mm1, %%mm4 \n\t" /* c */\
|
|
"paddw %%mm2, %%mm5 \n\t" /* d */\
|
|
"paddw %%mm6, %%mm6 \n\t" /* 2b */\
|
|
"psubw %%mm6, %%mm4 \n\t" /* c - 2b */\
|
|
"pmullw "MANGLE(ff_pw_20)", %%mm3 \n\t" /* 20a */\
|
|
"pmullw "MANGLE(ff_pw_3)", %%mm4 \n\t" /* 3c - 6b */\
|
|
"psubw %%mm5, %%mm3 \n\t" /* -6b + 3c - d */\
|
|
"paddw %6, %%mm4 \n\t"\
|
|
"paddw %%mm3, %%mm4 \n\t" /* 20a - 6b + 3c - d */\
|
|
"psraw $5, %%mm4 \n\t"\
|
|
"packuswb %%mm4, %%mm0 \n\t"\
|
|
OP_MMX2(%%mm0, 8(%1), %%mm4, q)\
|
|
\
|
|
"addl %3, %0 \n\t"\
|
|
"addl %4, %1 \n\t"\
|
|
"decl %2 \n\t"\
|
|
" jnz 1b \n\t"\
|
|
: "+a"(src), "+c"(dst), "+m"(h)\
|
|
: "d"(srcStride), "S"(dstStride), /*"m"(ff_pw_20), "m"(ff_pw_3),*/ "m"(temp), "m"(ROUNDER)\
|
|
: "memory"\
|
|
);\
|
|
}\
|
|
\
|
|
static void OPNAME ## mpeg4_qpel16_h_lowpass_3dnow(uint8_t *dst, uint8_t *src, int dstStride, int srcStride, int h){\
|
|
int i;\
|
|
int16_t temp[16];\
|
|
/* quick HACK, XXX FIXME MUST be optimized */\
|
|
for(i=0; i<h; i++)\
|
|
{\
|
|
temp[ 0]= (src[ 0]+src[ 1])*20 - (src[ 0]+src[ 2])*6 + (src[ 1]+src[ 3])*3 - (src[ 2]+src[ 4]);\
|
|
temp[ 1]= (src[ 1]+src[ 2])*20 - (src[ 0]+src[ 3])*6 + (src[ 0]+src[ 4])*3 - (src[ 1]+src[ 5]);\
|
|
temp[ 2]= (src[ 2]+src[ 3])*20 - (src[ 1]+src[ 4])*6 + (src[ 0]+src[ 5])*3 - (src[ 0]+src[ 6]);\
|
|
temp[ 3]= (src[ 3]+src[ 4])*20 - (src[ 2]+src[ 5])*6 + (src[ 1]+src[ 6])*3 - (src[ 0]+src[ 7]);\
|
|
temp[ 4]= (src[ 4]+src[ 5])*20 - (src[ 3]+src[ 6])*6 + (src[ 2]+src[ 7])*3 - (src[ 1]+src[ 8]);\
|
|
temp[ 5]= (src[ 5]+src[ 6])*20 - (src[ 4]+src[ 7])*6 + (src[ 3]+src[ 8])*3 - (src[ 2]+src[ 9]);\
|
|
temp[ 6]= (src[ 6]+src[ 7])*20 - (src[ 5]+src[ 8])*6 + (src[ 4]+src[ 9])*3 - (src[ 3]+src[10]);\
|
|
temp[ 7]= (src[ 7]+src[ 8])*20 - (src[ 6]+src[ 9])*6 + (src[ 5]+src[10])*3 - (src[ 4]+src[11]);\
|
|
temp[ 8]= (src[ 8]+src[ 9])*20 - (src[ 7]+src[10])*6 + (src[ 6]+src[11])*3 - (src[ 5]+src[12]);\
|
|
temp[ 9]= (src[ 9]+src[10])*20 - (src[ 8]+src[11])*6 + (src[ 7]+src[12])*3 - (src[ 6]+src[13]);\
|
|
temp[10]= (src[10]+src[11])*20 - (src[ 9]+src[12])*6 + (src[ 8]+src[13])*3 - (src[ 7]+src[14]);\
|
|
temp[11]= (src[11]+src[12])*20 - (src[10]+src[13])*6 + (src[ 9]+src[14])*3 - (src[ 8]+src[15]);\
|
|
temp[12]= (src[12]+src[13])*20 - (src[11]+src[14])*6 + (src[10]+src[15])*3 - (src[ 9]+src[16]);\
|
|
temp[13]= (src[13]+src[14])*20 - (src[12]+src[15])*6 + (src[11]+src[16])*3 - (src[10]+src[16]);\
|
|
temp[14]= (src[14]+src[15])*20 - (src[13]+src[16])*6 + (src[12]+src[16])*3 - (src[11]+src[15]);\
|
|
temp[15]= (src[15]+src[16])*20 - (src[14]+src[16])*6 + (src[13]+src[15])*3 - (src[12]+src[14]);\
|
|
asm volatile(\
|
|
"movq (%0), %%mm0 \n\t"\
|
|
"movq 8(%0), %%mm1 \n\t"\
|
|
"paddw %2, %%mm0 \n\t"\
|
|
"paddw %2, %%mm1 \n\t"\
|
|
"psraw $5, %%mm0 \n\t"\
|
|
"psraw $5, %%mm1 \n\t"\
|
|
"packuswb %%mm1, %%mm0 \n\t"\
|
|
OP_3DNOW(%%mm0, (%1), %%mm1, q)\
|
|
"movq 16(%0), %%mm0 \n\t"\
|
|
"movq 24(%0), %%mm1 \n\t"\
|
|
"paddw %2, %%mm0 \n\t"\
|
|
"paddw %2, %%mm1 \n\t"\
|
|
"psraw $5, %%mm0 \n\t"\
|
|
"psraw $5, %%mm1 \n\t"\
|
|
"packuswb %%mm1, %%mm0 \n\t"\
|
|
OP_3DNOW(%%mm0, 8(%1), %%mm1, q)\
|
|
:: "r"(temp), "r"(dst), "m"(ROUNDER)\
|
|
: "memory"\
|
|
);\
|
|
dst+=dstStride;\
|
|
src+=srcStride;\
|
|
}\
|
|
}\
|
|
\
|
|
static void OPNAME ## mpeg4_qpel8_h_lowpass_mmx2(uint8_t *dst, uint8_t *src, int dstStride, int srcStride, int h){\
|
|
uint64_t temp;\
|
|
\
|
|
asm volatile(\
|
|
"pxor %%mm7, %%mm7 \n\t"\
|
|
"1: \n\t"\
|
|
"movq (%0), %%mm0 \n\t" /* ABCDEFGH */\
|
|
"movq %%mm0, %%mm1 \n\t" /* ABCDEFGH */\
|
|
"movq %%mm0, %%mm2 \n\t" /* ABCDEFGH */\
|
|
"punpcklbw %%mm7, %%mm0 \n\t" /* 0A0B0C0D */\
|
|
"punpckhbw %%mm7, %%mm1 \n\t" /* 0E0F0G0H */\
|
|
"pshufw $0x90, %%mm0, %%mm5 \n\t" /* 0A0A0B0C */\
|
|
"pshufw $0x41, %%mm0, %%mm6 \n\t" /* 0B0A0A0B */\
|
|
"movq %%mm2, %%mm3 \n\t" /* ABCDEFGH */\
|
|
"movq %%mm2, %%mm4 \n\t" /* ABCDEFGH */\
|
|
"psllq $8, %%mm2 \n\t" /* 0ABCDEFG */\
|
|
"psllq $16, %%mm3 \n\t" /* 00ABCDEF */\
|
|
"psllq $24, %%mm4 \n\t" /* 000ABCDE */\
|
|
"punpckhbw %%mm7, %%mm2 \n\t" /* 0D0E0F0G */\
|
|
"punpckhbw %%mm7, %%mm3 \n\t" /* 0C0D0E0F */\
|
|
"punpckhbw %%mm7, %%mm4 \n\t" /* 0B0C0D0E */\
|
|
"paddw %%mm3, %%mm5 \n\t" /* b */\
|
|
"paddw %%mm2, %%mm6 \n\t" /* c */\
|
|
"paddw %%mm5, %%mm5 \n\t" /* 2b */\
|
|
"psubw %%mm5, %%mm6 \n\t" /* c - 2b */\
|
|
"pshufw $0x06, %%mm0, %%mm5 \n\t" /* 0C0B0A0A */\
|
|
"pmullw "MANGLE(ff_pw_3)", %%mm6 \n\t" /* 3c - 6b */\
|
|
"paddw %%mm4, %%mm0 \n\t" /* a */\
|
|
"paddw %%mm1, %%mm5 \n\t" /* d */\
|
|
"pmullw "MANGLE(ff_pw_20)", %%mm0 \n\t" /* 20a */\
|
|
"psubw %%mm5, %%mm0 \n\t" /* 20a - d */\
|
|
"paddw %6, %%mm6 \n\t"\
|
|
"paddw %%mm6, %%mm0 \n\t" /* 20a - 6b + 3c - d */\
|
|
"psraw $5, %%mm0 \n\t"\
|
|
/* mm1=EFGH, mm2=DEFG, mm3=CDEF, mm4=BCDE, mm7=0 */\
|
|
\
|
|
"movd 5(%0), %%mm5 \n\t" /* FGHI */\
|
|
"punpcklbw %%mm7, %%mm5 \n\t" /* 0F0G0H0I */\
|
|
"pshufw $0xF9, %%mm5, %%mm6 \n\t" /* 0G0H0I0I */\
|
|
"paddw %%mm5, %%mm1 \n\t" /* a */\
|
|
"paddw %%mm6, %%mm2 \n\t" /* b */\
|
|
"pshufw $0xBE, %%mm5, %%mm6 \n\t" /* 0H0I0I0H */\
|
|
"pshufw $0x6F, %%mm5, %%mm5 \n\t" /* 0I0I0H0G */\
|
|
"paddw %%mm6, %%mm3 \n\t" /* c */\
|
|
"paddw %%mm5, %%mm4 \n\t" /* d */\
|
|
"paddw %%mm2, %%mm2 \n\t" /* 2b */\
|
|
"psubw %%mm2, %%mm3 \n\t" /* c - 2b */\
|
|
"pmullw "MANGLE(ff_pw_20)", %%mm1 \n\t" /* 20a */\
|
|
"pmullw "MANGLE(ff_pw_3)", %%mm3 \n\t" /* 3c - 6b */\
|
|
"psubw %%mm4, %%mm3 \n\t" /* -6b + 3c - d */\
|
|
"paddw %6, %%mm1 \n\t"\
|
|
"paddw %%mm1, %%mm3 \n\t" /* 20a - 6b + 3c - d */\
|
|
"psraw $5, %%mm3 \n\t"\
|
|
"packuswb %%mm3, %%mm0 \n\t"\
|
|
OP_MMX2(%%mm0, (%1), %%mm4, q)\
|
|
\
|
|
"addl %3, %0 \n\t"\
|
|
"addl %4, %1 \n\t"\
|
|
"decl %2 \n\t"\
|
|
" jnz 1b \n\t"\
|
|
: "+a"(src), "+c"(dst), "+m"(h)\
|
|
: "S"(srcStride), "D"(dstStride), /*"m"(ff_pw_20), "m"(ff_pw_3),*/ "m"(temp), "m"(ROUNDER)\
|
|
: "memory"\
|
|
);\
|
|
}\
|
|
\
|
|
static void OPNAME ## mpeg4_qpel8_h_lowpass_3dnow(uint8_t *dst, uint8_t *src, int dstStride, int srcStride, int h){\
|
|
int i;\
|
|
int16_t temp[8];\
|
|
/* quick HACK, XXX FIXME MUST be optimized */\
|
|
for(i=0; i<h; i++)\
|
|
{\
|
|
temp[ 0]= (src[ 0]+src[ 1])*20 - (src[ 0]+src[ 2])*6 + (src[ 1]+src[ 3])*3 - (src[ 2]+src[ 4]);\
|
|
temp[ 1]= (src[ 1]+src[ 2])*20 - (src[ 0]+src[ 3])*6 + (src[ 0]+src[ 4])*3 - (src[ 1]+src[ 5]);\
|
|
temp[ 2]= (src[ 2]+src[ 3])*20 - (src[ 1]+src[ 4])*6 + (src[ 0]+src[ 5])*3 - (src[ 0]+src[ 6]);\
|
|
temp[ 3]= (src[ 3]+src[ 4])*20 - (src[ 2]+src[ 5])*6 + (src[ 1]+src[ 6])*3 - (src[ 0]+src[ 7]);\
|
|
temp[ 4]= (src[ 4]+src[ 5])*20 - (src[ 3]+src[ 6])*6 + (src[ 2]+src[ 7])*3 - (src[ 1]+src[ 8]);\
|
|
temp[ 5]= (src[ 5]+src[ 6])*20 - (src[ 4]+src[ 7])*6 + (src[ 3]+src[ 8])*3 - (src[ 2]+src[ 8]);\
|
|
temp[ 6]= (src[ 6]+src[ 7])*20 - (src[ 5]+src[ 8])*6 + (src[ 4]+src[ 8])*3 - (src[ 3]+src[ 7]);\
|
|
temp[ 7]= (src[ 7]+src[ 8])*20 - (src[ 6]+src[ 8])*6 + (src[ 5]+src[ 7])*3 - (src[ 4]+src[ 6]);\
|
|
asm volatile(\
|
|
"movq (%0), %%mm0 \n\t"\
|
|
"movq 8(%0), %%mm1 \n\t"\
|
|
"paddw %2, %%mm0 \n\t"\
|
|
"paddw %2, %%mm1 \n\t"\
|
|
"psraw $5, %%mm0 \n\t"\
|
|
"psraw $5, %%mm1 \n\t"\
|
|
"packuswb %%mm1, %%mm0 \n\t"\
|
|
OP_3DNOW(%%mm0, (%1), %%mm1, q)\
|
|
:: "r"(temp), "r"(dst), "m"(ROUNDER)\
|
|
:"memory"\
|
|
);\
|
|
dst+=dstStride;\
|
|
src+=srcStride;\
|
|
}\
|
|
}
|
|
|
|
#define QPEL_OP(OPNAME, ROUNDER, RND, OP, MMX)\
|
|
\
|
|
static void OPNAME ## mpeg4_qpel16_v_lowpass_ ## MMX(uint8_t *dst, uint8_t *src, int dstStride, int srcStride){\
|
|
uint64_t temp[17*4];\
|
|
uint64_t *temp_ptr= temp;\
|
|
int count= 17;\
|
|
\
|
|
/*FIXME unroll */\
|
|
asm volatile(\
|
|
"pxor %%mm7, %%mm7 \n\t"\
|
|
"1: \n\t"\
|
|
"movq (%0), %%mm0 \n\t"\
|
|
"movq (%0), %%mm1 \n\t"\
|
|
"movq 8(%0), %%mm2 \n\t"\
|
|
"movq 8(%0), %%mm3 \n\t"\
|
|
"punpcklbw %%mm7, %%mm0 \n\t"\
|
|
"punpckhbw %%mm7, %%mm1 \n\t"\
|
|
"punpcklbw %%mm7, %%mm2 \n\t"\
|
|
"punpckhbw %%mm7, %%mm3 \n\t"\
|
|
"movq %%mm0, (%1) \n\t"\
|
|
"movq %%mm1, 17*8(%1) \n\t"\
|
|
"movq %%mm2, 2*17*8(%1) \n\t"\
|
|
"movq %%mm3, 3*17*8(%1) \n\t"\
|
|
"addl $8, %1 \n\t"\
|
|
"addl %3, %0 \n\t"\
|
|
"decl %2 \n\t"\
|
|
" jnz 1b \n\t"\
|
|
: "+r" (src), "+r" (temp_ptr), "+r"(count)\
|
|
: "r" (srcStride)\
|
|
: "memory"\
|
|
);\
|
|
\
|
|
temp_ptr= temp;\
|
|
count=4;\
|
|
\
|
|
/*FIXME reorder for speed */\
|
|
asm volatile(\
|
|
/*"pxor %%mm7, %%mm7 \n\t"*/\
|
|
"1: \n\t"\
|
|
"movq (%0), %%mm0 \n\t"\
|
|
"movq 8(%0), %%mm1 \n\t"\
|
|
"movq 16(%0), %%mm2 \n\t"\
|
|
"movq 24(%0), %%mm3 \n\t"\
|
|
QPEL_V_LOW(%%mm0, %%mm1, %%mm2, %%mm3, %5, %6, %5, 16(%0), 8(%0), (%0), 32(%0), (%1), OP)\
|
|
QPEL_V_LOW(%%mm1, %%mm2, %%mm3, %%mm0, %5, %6, %5, 8(%0), (%0), (%0), 40(%0), (%1, %3), OP)\
|
|
"addl %4, %1 \n\t"\
|
|
QPEL_V_LOW(%%mm2, %%mm3, %%mm0, %%mm1, %5, %6, %5, (%0), (%0), 8(%0), 48(%0), (%1), OP)\
|
|
\
|
|
QPEL_V_LOW(%%mm3, %%mm0, %%mm1, %%mm2, %5, %6, %5, (%0), 8(%0), 16(%0), 56(%0), (%1, %3), OP)\
|
|
"addl %4, %1 \n\t"\
|
|
QPEL_V_LOW(%%mm0, %%mm1, %%mm2, %%mm3, %5, %6, %5, 8(%0), 16(%0), 24(%0), 64(%0), (%1), OP)\
|
|
QPEL_V_LOW(%%mm1, %%mm2, %%mm3, %%mm0, %5, %6, %5, 16(%0), 24(%0), 32(%0), 72(%0), (%1, %3), OP)\
|
|
"addl %4, %1 \n\t"\
|
|
QPEL_V_LOW(%%mm2, %%mm3, %%mm0, %%mm1, %5, %6, %5, 24(%0), 32(%0), 40(%0), 80(%0), (%1), OP)\
|
|
QPEL_V_LOW(%%mm3, %%mm0, %%mm1, %%mm2, %5, %6, %5, 32(%0), 40(%0), 48(%0), 88(%0), (%1, %3), OP)\
|
|
"addl %4, %1 \n\t"\
|
|
QPEL_V_LOW(%%mm0, %%mm1, %%mm2, %%mm3, %5, %6, %5, 40(%0), 48(%0), 56(%0), 96(%0), (%1), OP)\
|
|
QPEL_V_LOW(%%mm1, %%mm2, %%mm3, %%mm0, %5, %6, %5, 48(%0), 56(%0), 64(%0),104(%0), (%1, %3), OP)\
|
|
"addl %4, %1 \n\t"\
|
|
QPEL_V_LOW(%%mm2, %%mm3, %%mm0, %%mm1, %5, %6, %5, 56(%0), 64(%0), 72(%0),112(%0), (%1), OP)\
|
|
QPEL_V_LOW(%%mm3, %%mm0, %%mm1, %%mm2, %5, %6, %5, 64(%0), 72(%0), 80(%0),120(%0), (%1, %3), OP)\
|
|
"addl %4, %1 \n\t"\
|
|
QPEL_V_LOW(%%mm0, %%mm1, %%mm2, %%mm3, %5, %6, %5, 72(%0), 80(%0), 88(%0),128(%0), (%1), OP)\
|
|
\
|
|
QPEL_V_LOW(%%mm1, %%mm2, %%mm3, %%mm0, %5, %6, %5, 80(%0), 88(%0), 96(%0),128(%0), (%1, %3), OP)\
|
|
"addl %4, %1 \n\t" \
|
|
QPEL_V_LOW(%%mm2, %%mm3, %%mm0, %%mm1, %5, %6, %5, 88(%0), 96(%0),104(%0),120(%0), (%1), OP)\
|
|
QPEL_V_LOW(%%mm3, %%mm0, %%mm1, %%mm2, %5, %6, %5, 96(%0),104(%0),112(%0),112(%0), (%1, %3), OP)\
|
|
\
|
|
"addl $136, %0 \n\t"\
|
|
"addl %6, %1 \n\t"\
|
|
"decl %2 \n\t"\
|
|
" jnz 1b \n\t"\
|
|
\
|
|
: "+r"(temp_ptr), "+r"(dst), "+g"(count)\
|
|
: "r"(dstStride), "r"(2*dstStride), /*"m"(ff_pw_20), "m"(ff_pw_3),*/ "m"(ROUNDER), "g"(4-14*dstStride)\
|
|
:"memory"\
|
|
);\
|
|
}\
|
|
\
|
|
static void OPNAME ## mpeg4_qpel8_v_lowpass_ ## MMX(uint8_t *dst, uint8_t *src, int dstStride, int srcStride){\
|
|
uint64_t temp[9*4];\
|
|
uint64_t *temp_ptr= temp;\
|
|
int count= 9;\
|
|
\
|
|
/*FIXME unroll */\
|
|
asm volatile(\
|
|
"pxor %%mm7, %%mm7 \n\t"\
|
|
"1: \n\t"\
|
|
"movq (%0), %%mm0 \n\t"\
|
|
"movq (%0), %%mm1 \n\t"\
|
|
"punpcklbw %%mm7, %%mm0 \n\t"\
|
|
"punpckhbw %%mm7, %%mm1 \n\t"\
|
|
"movq %%mm0, (%1) \n\t"\
|
|
"movq %%mm1, 9*8(%1) \n\t"\
|
|
"addl $8, %1 \n\t"\
|
|
"addl %3, %0 \n\t"\
|
|
"decl %2 \n\t"\
|
|
" jnz 1b \n\t"\
|
|
: "+r" (src), "+r" (temp_ptr), "+r"(count)\
|
|
: "r" (srcStride)\
|
|
: "memory"\
|
|
);\
|
|
\
|
|
temp_ptr= temp;\
|
|
count=2;\
|
|
\
|
|
/*FIXME reorder for speed */\
|
|
asm volatile(\
|
|
/*"pxor %%mm7, %%mm7 \n\t"*/\
|
|
"1: \n\t"\
|
|
"movq (%0), %%mm0 \n\t"\
|
|
"movq 8(%0), %%mm1 \n\t"\
|
|
"movq 16(%0), %%mm2 \n\t"\
|
|
"movq 24(%0), %%mm3 \n\t"\
|
|
QPEL_V_LOW(%%mm0, %%mm1, %%mm2, %%mm3, %5, %6, %5, 16(%0), 8(%0), (%0), 32(%0), (%1), OP)\
|
|
QPEL_V_LOW(%%mm1, %%mm2, %%mm3, %%mm0, %5, %6, %5, 8(%0), (%0), (%0), 40(%0), (%1, %3), OP)\
|
|
"addl %4, %1 \n\t"\
|
|
QPEL_V_LOW(%%mm2, %%mm3, %%mm0, %%mm1, %5, %6, %5, (%0), (%0), 8(%0), 48(%0), (%1), OP)\
|
|
\
|
|
QPEL_V_LOW(%%mm3, %%mm0, %%mm1, %%mm2, %5, %6, %5, (%0), 8(%0), 16(%0), 56(%0), (%1, %3), OP)\
|
|
"addl %4, %1 \n\t"\
|
|
QPEL_V_LOW(%%mm0, %%mm1, %%mm2, %%mm3, %5, %6, %5, 8(%0), 16(%0), 24(%0), 64(%0), (%1), OP)\
|
|
\
|
|
QPEL_V_LOW(%%mm1, %%mm2, %%mm3, %%mm0, %5, %6, %5, 16(%0), 24(%0), 32(%0), 64(%0), (%1, %3), OP)\
|
|
"addl %4, %1 \n\t"\
|
|
QPEL_V_LOW(%%mm2, %%mm3, %%mm0, %%mm1, %5, %6, %5, 24(%0), 32(%0), 40(%0), 56(%0), (%1), OP)\
|
|
QPEL_V_LOW(%%mm3, %%mm0, %%mm1, %%mm2, %5, %6, %5, 32(%0), 40(%0), 48(%0), 48(%0), (%1, %3), OP)\
|
|
\
|
|
"addl $72, %0 \n\t"\
|
|
"addl %6, %1 \n\t"\
|
|
"decl %2 \n\t"\
|
|
" jnz 1b \n\t"\
|
|
\
|
|
: "+r"(temp_ptr), "+r"(dst), "+g"(count)\
|
|
: "r"(dstStride), "r"(2*dstStride), /*"m"(ff_pw_20), "m"(ff_pw_3),*/ "m"(ROUNDER), "g"(4-6*dstStride)\
|
|
: "memory"\
|
|
);\
|
|
}\
|
|
\
|
|
static void OPNAME ## qpel8_mc00_ ## MMX (uint8_t *dst, uint8_t *src, int stride){\
|
|
OPNAME ## pixels8_mmx(dst, src, stride, 8);\
|
|
}\
|
|
\
|
|
static void OPNAME ## qpel8_mc10_ ## MMX(uint8_t *dst, uint8_t *src, int stride){\
|
|
uint64_t temp[8];\
|
|
uint8_t * const half= (uint8_t*)temp;\
|
|
put ## RND ## mpeg4_qpel8_h_lowpass_ ## MMX(half, src, 8, stride, 8);\
|
|
OPNAME ## pixels8_l2_mmx(dst, src, half, stride, stride, 8);\
|
|
}\
|
|
\
|
|
static void OPNAME ## qpel8_mc20_ ## MMX(uint8_t *dst, uint8_t *src, int stride){\
|
|
OPNAME ## mpeg4_qpel8_h_lowpass_ ## MMX(dst, src, stride, stride, 8);\
|
|
}\
|
|
\
|
|
static void OPNAME ## qpel8_mc30_ ## MMX(uint8_t *dst, uint8_t *src, int stride){\
|
|
uint64_t temp[8];\
|
|
uint8_t * const half= (uint8_t*)temp;\
|
|
put ## RND ## mpeg4_qpel8_h_lowpass_ ## MMX(half, src, 8, stride, 8);\
|
|
OPNAME ## pixels8_l2_mmx(dst, src+1, half, stride, stride, 8);\
|
|
}\
|
|
\
|
|
static void OPNAME ## qpel8_mc01_ ## MMX(uint8_t *dst, uint8_t *src, int stride){\
|
|
uint64_t temp[8];\
|
|
uint8_t * const half= (uint8_t*)temp;\
|
|
put ## RND ## mpeg4_qpel8_v_lowpass_ ## MMX(half, src, 8, stride);\
|
|
OPNAME ## pixels8_l2_mmx(dst, src, half, stride, stride, 8);\
|
|
}\
|
|
\
|
|
static void OPNAME ## qpel8_mc02_ ## MMX(uint8_t *dst, uint8_t *src, int stride){\
|
|
OPNAME ## mpeg4_qpel8_v_lowpass_ ## MMX(dst, src, stride, stride);\
|
|
}\
|
|
\
|
|
static void OPNAME ## qpel8_mc03_ ## MMX(uint8_t *dst, uint8_t *src, int stride){\
|
|
uint64_t temp[8];\
|
|
uint8_t * const half= (uint8_t*)temp;\
|
|
put ## RND ## mpeg4_qpel8_v_lowpass_ ## MMX(half, src, 8, stride);\
|
|
OPNAME ## pixels8_l2_mmx(dst, src+stride, half, stride, stride, 8);\
|
|
}\
|
|
static void OPNAME ## qpel8_mc11_ ## MMX(uint8_t *dst, uint8_t *src, int stride){\
|
|
uint64_t half[8 + 9];\
|
|
uint8_t * const halfH= ((uint8_t*)half) + 64;\
|
|
uint8_t * const halfHV= ((uint8_t*)half);\
|
|
put ## RND ## mpeg4_qpel8_h_lowpass_ ## MMX(halfH, src, 8, stride, 9);\
|
|
put ## RND ## pixels8_l2_mmx(halfH, src, halfH, 8, stride, 9);\
|
|
put ## RND ## mpeg4_qpel8_v_lowpass_ ## MMX(halfHV, halfH, 8, 8);\
|
|
OPNAME ## pixels8_l2_mmx(dst, halfH, halfHV, stride, 8, 8);\
|
|
}\
|
|
static void OPNAME ## qpel8_mc31_ ## MMX(uint8_t *dst, uint8_t *src, int stride){\
|
|
uint64_t half[8 + 9];\
|
|
uint8_t * const halfH= ((uint8_t*)half) + 64;\
|
|
uint8_t * const halfHV= ((uint8_t*)half);\
|
|
put ## RND ## mpeg4_qpel8_h_lowpass_ ## MMX(halfH, src, 8, stride, 9);\
|
|
put ## RND ## pixels8_l2_mmx(halfH, src+1, halfH, 8, stride, 9);\
|
|
put ## RND ## mpeg4_qpel8_v_lowpass_ ## MMX(halfHV, halfH, 8, 8);\
|
|
OPNAME ## pixels8_l2_mmx(dst, halfH, halfHV, stride, 8, 8);\
|
|
}\
|
|
static void OPNAME ## qpel8_mc13_ ## MMX(uint8_t *dst, uint8_t *src, int stride){\
|
|
uint64_t half[8 + 9];\
|
|
uint8_t * const halfH= ((uint8_t*)half) + 64;\
|
|
uint8_t * const halfHV= ((uint8_t*)half);\
|
|
put ## RND ## mpeg4_qpel8_h_lowpass_ ## MMX(halfH, src, 8, stride, 9);\
|
|
put ## RND ## pixels8_l2_mmx(halfH, src, halfH, 8, stride, 9);\
|
|
put ## RND ## mpeg4_qpel8_v_lowpass_ ## MMX(halfHV, halfH, 8, 8);\
|
|
OPNAME ## pixels8_l2_mmx(dst, halfH+8, halfHV, stride, 8, 8);\
|
|
}\
|
|
static void OPNAME ## qpel8_mc33_ ## MMX(uint8_t *dst, uint8_t *src, int stride){\
|
|
uint64_t half[8 + 9];\
|
|
uint8_t * const halfH= ((uint8_t*)half) + 64;\
|
|
uint8_t * const halfHV= ((uint8_t*)half);\
|
|
put ## RND ## mpeg4_qpel8_h_lowpass_ ## MMX(halfH, src, 8, stride, 9);\
|
|
put ## RND ## pixels8_l2_mmx(halfH, src+1, halfH, 8, stride, 9);\
|
|
put ## RND ## mpeg4_qpel8_v_lowpass_ ## MMX(halfHV, halfH, 8, 8);\
|
|
OPNAME ## pixels8_l2_mmx(dst, halfH+8, halfHV, stride, 8, 8);\
|
|
}\
|
|
static void OPNAME ## qpel8_mc21_ ## MMX(uint8_t *dst, uint8_t *src, int stride){\
|
|
uint64_t half[8 + 9];\
|
|
uint8_t * const halfH= ((uint8_t*)half) + 64;\
|
|
uint8_t * const halfHV= ((uint8_t*)half);\
|
|
put ## RND ## mpeg4_qpel8_h_lowpass_ ## MMX(halfH, src, 8, stride, 9);\
|
|
put ## RND ## mpeg4_qpel8_v_lowpass_ ## MMX(halfHV, halfH, 8, 8);\
|
|
OPNAME ## pixels8_l2_mmx(dst, halfH, halfHV, stride, 8, 8);\
|
|
}\
|
|
static void OPNAME ## qpel8_mc23_ ## MMX(uint8_t *dst, uint8_t *src, int stride){\
|
|
uint64_t half[8 + 9];\
|
|
uint8_t * const halfH= ((uint8_t*)half) + 64;\
|
|
uint8_t * const halfHV= ((uint8_t*)half);\
|
|
put ## RND ## mpeg4_qpel8_h_lowpass_ ## MMX(halfH, src, 8, stride, 9);\
|
|
put ## RND ## mpeg4_qpel8_v_lowpass_ ## MMX(halfHV, halfH, 8, 8);\
|
|
OPNAME ## pixels8_l2_mmx(dst, halfH+8, halfHV, stride, 8, 8);\
|
|
}\
|
|
static void OPNAME ## qpel8_mc12_ ## MMX(uint8_t *dst, uint8_t *src, int stride){\
|
|
uint64_t half[8 + 9];\
|
|
uint8_t * const halfH= ((uint8_t*)half);\
|
|
put ## RND ## mpeg4_qpel8_h_lowpass_ ## MMX(halfH, src, 8, stride, 9);\
|
|
put ## RND ## pixels8_l2_mmx(halfH, src, halfH, 8, stride, 9);\
|
|
OPNAME ## mpeg4_qpel8_v_lowpass_ ## MMX(dst, halfH, stride, 8);\
|
|
}\
|
|
static void OPNAME ## qpel8_mc32_ ## MMX(uint8_t *dst, uint8_t *src, int stride){\
|
|
uint64_t half[8 + 9];\
|
|
uint8_t * const halfH= ((uint8_t*)half);\
|
|
put ## RND ## mpeg4_qpel8_h_lowpass_ ## MMX(halfH, src, 8, stride, 9);\
|
|
put ## RND ## pixels8_l2_mmx(halfH, src+1, halfH, 8, stride, 9);\
|
|
OPNAME ## mpeg4_qpel8_v_lowpass_ ## MMX(dst, halfH, stride, 8);\
|
|
}\
|
|
static void OPNAME ## qpel8_mc22_ ## MMX(uint8_t *dst, uint8_t *src, int stride){\
|
|
uint64_t half[9];\
|
|
uint8_t * const halfH= ((uint8_t*)half);\
|
|
put ## RND ## mpeg4_qpel8_h_lowpass_ ## MMX(halfH, src, 8, stride, 9);\
|
|
OPNAME ## mpeg4_qpel8_v_lowpass_ ## MMX(dst, halfH, stride, 8);\
|
|
}\
|
|
static void OPNAME ## qpel16_mc00_ ## MMX (uint8_t *dst, uint8_t *src, int stride){\
|
|
OPNAME ## pixels16_mmx(dst, src, stride, 16);\
|
|
}\
|
|
\
|
|
static void OPNAME ## qpel16_mc10_ ## MMX(uint8_t *dst, uint8_t *src, int stride){\
|
|
uint64_t temp[32];\
|
|
uint8_t * const half= (uint8_t*)temp;\
|
|
put ## RND ## mpeg4_qpel16_h_lowpass_ ## MMX(half, src, 16, stride, 16);\
|
|
OPNAME ## pixels16_l2_mmx(dst, src, half, stride, stride, 16);\
|
|
}\
|
|
\
|
|
static void OPNAME ## qpel16_mc20_ ## MMX(uint8_t *dst, uint8_t *src, int stride){\
|
|
OPNAME ## mpeg4_qpel16_h_lowpass_ ## MMX(dst, src, stride, stride, 16);\
|
|
}\
|
|
\
|
|
static void OPNAME ## qpel16_mc30_ ## MMX(uint8_t *dst, uint8_t *src, int stride){\
|
|
uint64_t temp[32];\
|
|
uint8_t * const half= (uint8_t*)temp;\
|
|
put ## RND ## mpeg4_qpel16_h_lowpass_ ## MMX(half, src, 16, stride, 16);\
|
|
OPNAME ## pixels16_l2_mmx(dst, src+1, half, stride, stride, 16);\
|
|
}\
|
|
\
|
|
static void OPNAME ## qpel16_mc01_ ## MMX(uint8_t *dst, uint8_t *src, int stride){\
|
|
uint64_t temp[32];\
|
|
uint8_t * const half= (uint8_t*)temp;\
|
|
put ## RND ## mpeg4_qpel16_v_lowpass_ ## MMX(half, src, 16, stride);\
|
|
OPNAME ## pixels16_l2_mmx(dst, src, half, stride, stride, 16);\
|
|
}\
|
|
\
|
|
static void OPNAME ## qpel16_mc02_ ## MMX(uint8_t *dst, uint8_t *src, int stride){\
|
|
OPNAME ## mpeg4_qpel16_v_lowpass_ ## MMX(dst, src, stride, stride);\
|
|
}\
|
|
\
|
|
static void OPNAME ## qpel16_mc03_ ## MMX(uint8_t *dst, uint8_t *src, int stride){\
|
|
uint64_t temp[32];\
|
|
uint8_t * const half= (uint8_t*)temp;\
|
|
put ## RND ## mpeg4_qpel16_v_lowpass_ ## MMX(half, src, 16, stride);\
|
|
OPNAME ## pixels16_l2_mmx(dst, src+stride, half, stride, stride, 16);\
|
|
}\
|
|
static void OPNAME ## qpel16_mc11_ ## MMX(uint8_t *dst, uint8_t *src, int stride){\
|
|
uint64_t half[16*2 + 17*2];\
|
|
uint8_t * const halfH= ((uint8_t*)half) + 256;\
|
|
uint8_t * const halfHV= ((uint8_t*)half);\
|
|
put ## RND ## mpeg4_qpel16_h_lowpass_ ## MMX(halfH, src, 16, stride, 17);\
|
|
put ## RND ## pixels16_l2_mmx(halfH, src, halfH, 16, stride, 17);\
|
|
put ## RND ## mpeg4_qpel16_v_lowpass_ ## MMX(halfHV, halfH, 16, 16);\
|
|
OPNAME ## pixels16_l2_mmx(dst, halfH, halfHV, stride, 16, 16);\
|
|
}\
|
|
static void OPNAME ## qpel16_mc31_ ## MMX(uint8_t *dst, uint8_t *src, int stride){\
|
|
uint64_t half[16*2 + 17*2];\
|
|
uint8_t * const halfH= ((uint8_t*)half) + 256;\
|
|
uint8_t * const halfHV= ((uint8_t*)half);\
|
|
put ## RND ## mpeg4_qpel16_h_lowpass_ ## MMX(halfH, src, 16, stride, 17);\
|
|
put ## RND ## pixels16_l2_mmx(halfH, src+1, halfH, 16, stride, 17);\
|
|
put ## RND ## mpeg4_qpel16_v_lowpass_ ## MMX(halfHV, halfH, 16, 16);\
|
|
OPNAME ## pixels16_l2_mmx(dst, halfH, halfHV, stride, 16, 16);\
|
|
}\
|
|
static void OPNAME ## qpel16_mc13_ ## MMX(uint8_t *dst, uint8_t *src, int stride){\
|
|
uint64_t half[16*2 + 17*2];\
|
|
uint8_t * const halfH= ((uint8_t*)half) + 256;\
|
|
uint8_t * const halfHV= ((uint8_t*)half);\
|
|
put ## RND ## mpeg4_qpel16_h_lowpass_ ## MMX(halfH, src, 16, stride, 17);\
|
|
put ## RND ## pixels16_l2_mmx(halfH, src, halfH, 16, stride, 17);\
|
|
put ## RND ## mpeg4_qpel16_v_lowpass_ ## MMX(halfHV, halfH, 16, 16);\
|
|
OPNAME ## pixels16_l2_mmx(dst, halfH+16, halfHV, stride, 16, 16);\
|
|
}\
|
|
static void OPNAME ## qpel16_mc33_ ## MMX(uint8_t *dst, uint8_t *src, int stride){\
|
|
uint64_t half[16*2 + 17*2];\
|
|
uint8_t * const halfH= ((uint8_t*)half) + 256;\
|
|
uint8_t * const halfHV= ((uint8_t*)half);\
|
|
put ## RND ## mpeg4_qpel16_h_lowpass_ ## MMX(halfH, src, 16, stride, 17);\
|
|
put ## RND ## pixels16_l2_mmx(halfH, src+1, halfH, 16, stride, 17);\
|
|
put ## RND ## mpeg4_qpel16_v_lowpass_ ## MMX(halfHV, halfH, 16, 16);\
|
|
OPNAME ## pixels16_l2_mmx(dst, halfH+16, halfHV, stride, 16, 16);\
|
|
}\
|
|
static void OPNAME ## qpel16_mc21_ ## MMX(uint8_t *dst, uint8_t *src, int stride){\
|
|
uint64_t half[16*2 + 17*2];\
|
|
uint8_t * const halfH= ((uint8_t*)half) + 256;\
|
|
uint8_t * const halfHV= ((uint8_t*)half);\
|
|
put ## RND ## mpeg4_qpel16_h_lowpass_ ## MMX(halfH, src, 16, stride, 17);\
|
|
put ## RND ## mpeg4_qpel16_v_lowpass_ ## MMX(halfHV, halfH, 16, 16);\
|
|
OPNAME ## pixels16_l2_mmx(dst, halfH, halfHV, stride, 16, 16);\
|
|
}\
|
|
static void OPNAME ## qpel16_mc23_ ## MMX(uint8_t *dst, uint8_t *src, int stride){\
|
|
uint64_t half[16*2 + 17*2];\
|
|
uint8_t * const halfH= ((uint8_t*)half) + 256;\
|
|
uint8_t * const halfHV= ((uint8_t*)half);\
|
|
put ## RND ## mpeg4_qpel16_h_lowpass_ ## MMX(halfH, src, 16, stride, 17);\
|
|
put ## RND ## mpeg4_qpel16_v_lowpass_ ## MMX(halfHV, halfH, 16, 16);\
|
|
OPNAME ## pixels16_l2_mmx(dst, halfH+16, halfHV, stride, 16, 16);\
|
|
}\
|
|
static void OPNAME ## qpel16_mc12_ ## MMX(uint8_t *dst, uint8_t *src, int stride){\
|
|
uint64_t half[17*2];\
|
|
uint8_t * const halfH= ((uint8_t*)half);\
|
|
put ## RND ## mpeg4_qpel16_h_lowpass_ ## MMX(halfH, src, 16, stride, 17);\
|
|
put ## RND ## pixels16_l2_mmx(halfH, src, halfH, 16, stride, 17);\
|
|
OPNAME ## mpeg4_qpel16_v_lowpass_ ## MMX(dst, halfH, stride, 16);\
|
|
}\
|
|
static void OPNAME ## qpel16_mc32_ ## MMX(uint8_t *dst, uint8_t *src, int stride){\
|
|
uint64_t half[17*2];\
|
|
uint8_t * const halfH= ((uint8_t*)half);\
|
|
put ## RND ## mpeg4_qpel16_h_lowpass_ ## MMX(halfH, src, 16, stride, 17);\
|
|
put ## RND ## pixels16_l2_mmx(halfH, src+1, halfH, 16, stride, 17);\
|
|
OPNAME ## mpeg4_qpel16_v_lowpass_ ## MMX(dst, halfH, stride, 16);\
|
|
}\
|
|
static void OPNAME ## qpel16_mc22_ ## MMX(uint8_t *dst, uint8_t *src, int stride){\
|
|
uint64_t half[17*2];\
|
|
uint8_t * const halfH= ((uint8_t*)half);\
|
|
put ## RND ## mpeg4_qpel16_h_lowpass_ ## MMX(halfH, src, 16, stride, 17);\
|
|
OPNAME ## mpeg4_qpel16_v_lowpass_ ## MMX(dst, halfH, stride, 16);\
|
|
}
|
|
|
|
|
|
#define PUT_OP(a,b,temp, size) "mov" #size " " #a ", " #b " \n\t"
|
|
#define AVG_3DNOW_OP(a,b,temp, size) \
|
|
"mov" #size " " #b ", " #temp " \n\t"\
|
|
"pavgusb " #temp ", " #a " \n\t"\
|
|
"mov" #size " " #a ", " #b " \n\t"
|
|
#define AVG_MMX2_OP(a,b,temp, size) \
|
|
"mov" #size " " #b ", " #temp " \n\t"\
|
|
"pavgb " #temp ", " #a " \n\t"\
|
|
"mov" #size " " #a ", " #b " \n\t"
|
|
|
|
QPEL_BASE(put_ , ff_pw_16, _ , PUT_OP, PUT_OP)
|
|
QPEL_BASE(avg_ , ff_pw_16, _ , AVG_MMX2_OP, AVG_3DNOW_OP)
|
|
QPEL_BASE(put_no_rnd_, ff_pw_15, _no_rnd_, PUT_OP, PUT_OP)
|
|
QPEL_OP(put_ , ff_pw_16, _ , PUT_OP, 3dnow)
|
|
QPEL_OP(avg_ , ff_pw_16, _ , AVG_3DNOW_OP, 3dnow)
|
|
QPEL_OP(put_no_rnd_, ff_pw_15, _no_rnd_, PUT_OP, 3dnow)
|
|
QPEL_OP(put_ , ff_pw_16, _ , PUT_OP, mmx2)
|
|
QPEL_OP(avg_ , ff_pw_16, _ , AVG_MMX2_OP, mmx2)
|
|
QPEL_OP(put_no_rnd_, ff_pw_15, _no_rnd_, PUT_OP, mmx2)
|
|
|
|
#if 0
|
|
static void just_return() { return; }
|
|
#endif
|
|
|
|
#define SET_QPEL_FUNC(postfix1, postfix2) \
|
|
c->put_ ## postfix1 = put_ ## postfix2;\
|
|
c->put_no_rnd_ ## postfix1 = put_no_rnd_ ## postfix2;\
|
|
c->avg_ ## postfix1 = avg_ ## postfix2;
|
|
|
|
/* external functions, from idct_mmx.c */
|
|
void ff_mmx_idct(DCTELEM *block);
|
|
void ff_mmxext_idct(DCTELEM *block);
|
|
|
|
/* XXX: those functions should be suppressed ASAP when all IDCTs are
|
|
converted */
|
|
static void ff_libmpeg2mmx_idct_put(uint8_t *dest, int line_size, DCTELEM *block)
|
|
{
|
|
ff_mmx_idct (block);
|
|
put_pixels_clamped_mmx(block, dest, line_size);
|
|
}
|
|
static void ff_libmpeg2mmx_idct_add(uint8_t *dest, int line_size, DCTELEM *block)
|
|
{
|
|
ff_mmx_idct (block);
|
|
add_pixels_clamped_mmx(block, dest, line_size);
|
|
}
|
|
static void ff_libmpeg2mmx2_idct_put(uint8_t *dest, int line_size, DCTELEM *block)
|
|
{
|
|
ff_mmxext_idct (block);
|
|
put_pixels_clamped_mmx(block, dest, line_size);
|
|
}
|
|
static void ff_libmpeg2mmx2_idct_add(uint8_t *dest, int line_size, DCTELEM *block)
|
|
{
|
|
ff_mmxext_idct (block);
|
|
add_pixels_clamped_mmx(block, dest, line_size);
|
|
}
|
|
|
|
void dsputil_init_mmx(DSPContext* c, AVCodecContext *avctx)
|
|
{
|
|
mm_flags = mm_support();
|
|
|
|
if (avctx->dsp_mask) {
|
|
if (avctx->dsp_mask & FF_MM_FORCE)
|
|
mm_flags |= (avctx->dsp_mask & 0xffff);
|
|
else
|
|
mm_flags &= ~(avctx->dsp_mask & 0xffff);
|
|
}
|
|
|
|
#if 0
|
|
fprintf(stderr, "libavcodec: CPU flags:");
|
|
if (mm_flags & MM_MMX)
|
|
fprintf(stderr, " mmx");
|
|
if (mm_flags & MM_MMXEXT)
|
|
fprintf(stderr, " mmxext");
|
|
if (mm_flags & MM_3DNOW)
|
|
fprintf(stderr, " 3dnow");
|
|
if (mm_flags & MM_SSE)
|
|
fprintf(stderr, " sse");
|
|
if (mm_flags & MM_SSE2)
|
|
fprintf(stderr, " sse2");
|
|
fprintf(stderr, "\n");
|
|
#endif
|
|
|
|
if (mm_flags & MM_MMX) {
|
|
const int dct_algo = avctx->dct_algo;
|
|
const int idct_algo= avctx->idct_algo;
|
|
|
|
#ifdef CONFIG_ENCODERS
|
|
if(dct_algo==FF_DCT_AUTO || dct_algo==FF_DCT_MMX){
|
|
if(mm_flags & MM_MMXEXT){
|
|
c->fdct = ff_fdct_mmx2;
|
|
}else{
|
|
c->fdct = ff_fdct_mmx;
|
|
}
|
|
}
|
|
#endif //CONFIG_ENCODERS
|
|
|
|
if(idct_algo==FF_IDCT_AUTO || idct_algo==FF_IDCT_SIMPLEMMX){
|
|
c->idct_put= ff_simple_idct_put_mmx;
|
|
c->idct_add= ff_simple_idct_add_mmx;
|
|
c->idct = ff_simple_idct_mmx;
|
|
c->idct_permutation_type= FF_SIMPLE_IDCT_PERM;
|
|
}else if(idct_algo==FF_IDCT_LIBMPEG2MMX){
|
|
if(mm_flags & MM_MMXEXT){
|
|
c->idct_put= ff_libmpeg2mmx2_idct_put;
|
|
c->idct_add= ff_libmpeg2mmx2_idct_add;
|
|
c->idct = ff_mmxext_idct;
|
|
}else{
|
|
c->idct_put= ff_libmpeg2mmx_idct_put;
|
|
c->idct_add= ff_libmpeg2mmx_idct_add;
|
|
c->idct = ff_mmx_idct;
|
|
}
|
|
c->idct_permutation_type= FF_LIBMPEG2_IDCT_PERM;
|
|
}
|
|
|
|
#ifdef CONFIG_ENCODERS
|
|
c->get_pixels = get_pixels_mmx;
|
|
c->diff_pixels = diff_pixels_mmx;
|
|
#endif //CONFIG_ENCODERS
|
|
c->put_pixels_clamped = put_pixels_clamped_mmx;
|
|
c->add_pixels_clamped = add_pixels_clamped_mmx;
|
|
c->clear_blocks = clear_blocks_mmx;
|
|
#ifdef CONFIG_ENCODERS
|
|
c->pix_sum = pix_sum16_mmx;
|
|
#endif //CONFIG_ENCODERS
|
|
|
|
c->put_pixels_tab[0][0] = put_pixels16_mmx;
|
|
c->put_pixels_tab[0][1] = put_pixels16_x2_mmx;
|
|
c->put_pixels_tab[0][2] = put_pixels16_y2_mmx;
|
|
c->put_pixels_tab[0][3] = put_pixels16_xy2_mmx;
|
|
|
|
c->put_no_rnd_pixels_tab[0][0] = put_pixels16_mmx;
|
|
c->put_no_rnd_pixels_tab[0][1] = put_no_rnd_pixels16_x2_mmx;
|
|
c->put_no_rnd_pixels_tab[0][2] = put_no_rnd_pixels16_y2_mmx;
|
|
c->put_no_rnd_pixels_tab[0][3] = put_no_rnd_pixels16_xy2_mmx;
|
|
|
|
c->avg_pixels_tab[0][0] = avg_pixels16_mmx;
|
|
c->avg_pixels_tab[0][1] = avg_pixels16_x2_mmx;
|
|
c->avg_pixels_tab[0][2] = avg_pixels16_y2_mmx;
|
|
c->avg_pixels_tab[0][3] = avg_pixels16_xy2_mmx;
|
|
|
|
c->avg_no_rnd_pixels_tab[0][0] = avg_no_rnd_pixels16_mmx;
|
|
c->avg_no_rnd_pixels_tab[0][1] = avg_no_rnd_pixels16_x2_mmx;
|
|
c->avg_no_rnd_pixels_tab[0][2] = avg_no_rnd_pixels16_y2_mmx;
|
|
c->avg_no_rnd_pixels_tab[0][3] = avg_no_rnd_pixels16_xy2_mmx;
|
|
|
|
c->put_pixels_tab[1][0] = put_pixels8_mmx;
|
|
c->put_pixels_tab[1][1] = put_pixels8_x2_mmx;
|
|
c->put_pixels_tab[1][2] = put_pixels8_y2_mmx;
|
|
c->put_pixels_tab[1][3] = put_pixels8_xy2_mmx;
|
|
|
|
c->put_no_rnd_pixels_tab[1][0] = put_pixels8_mmx;
|
|
c->put_no_rnd_pixels_tab[1][1] = put_no_rnd_pixels8_x2_mmx;
|
|
c->put_no_rnd_pixels_tab[1][2] = put_no_rnd_pixels8_y2_mmx;
|
|
c->put_no_rnd_pixels_tab[1][3] = put_no_rnd_pixels8_xy2_mmx;
|
|
|
|
c->avg_pixels_tab[1][0] = avg_pixels8_mmx;
|
|
c->avg_pixels_tab[1][1] = avg_pixels8_x2_mmx;
|
|
c->avg_pixels_tab[1][2] = avg_pixels8_y2_mmx;
|
|
c->avg_pixels_tab[1][3] = avg_pixels8_xy2_mmx;
|
|
|
|
c->avg_no_rnd_pixels_tab[1][0] = avg_no_rnd_pixels8_mmx;
|
|
c->avg_no_rnd_pixels_tab[1][1] = avg_no_rnd_pixels8_x2_mmx;
|
|
c->avg_no_rnd_pixels_tab[1][2] = avg_no_rnd_pixels8_y2_mmx;
|
|
c->avg_no_rnd_pixels_tab[1][3] = avg_no_rnd_pixels8_xy2_mmx;
|
|
|
|
c->add_bytes= add_bytes_mmx;
|
|
#ifdef CONFIG_ENCODERS
|
|
c->diff_bytes= diff_bytes_mmx;
|
|
|
|
c->hadamard8_diff[0]= hadamard8_diff16_mmx;
|
|
c->hadamard8_diff[1]= hadamard8_diff_mmx;
|
|
|
|
c->pix_norm1 = pix_norm1_mmx;
|
|
c->sse[0] = sse16_mmx;
|
|
#endif //CONFIG_ENCODERS
|
|
|
|
c->h263_v_loop_filter= h263_v_loop_filter_mmx;
|
|
c->h263_h_loop_filter= h263_h_loop_filter_mmx;
|
|
|
|
if (mm_flags & MM_MMXEXT) {
|
|
c->put_pixels_tab[0][1] = put_pixels16_x2_mmx2;
|
|
c->put_pixels_tab[0][2] = put_pixels16_y2_mmx2;
|
|
|
|
c->avg_pixels_tab[0][0] = avg_pixels16_mmx2;
|
|
c->avg_pixels_tab[0][1] = avg_pixels16_x2_mmx2;
|
|
c->avg_pixels_tab[0][2] = avg_pixels16_y2_mmx2;
|
|
|
|
c->put_pixels_tab[1][1] = put_pixels8_x2_mmx2;
|
|
c->put_pixels_tab[1][2] = put_pixels8_y2_mmx2;
|
|
|
|
c->avg_pixels_tab[1][0] = avg_pixels8_mmx2;
|
|
c->avg_pixels_tab[1][1] = avg_pixels8_x2_mmx2;
|
|
c->avg_pixels_tab[1][2] = avg_pixels8_y2_mmx2;
|
|
|
|
#ifdef CONFIG_ENCODERS
|
|
c->hadamard8_diff[0]= hadamard8_diff16_mmx2;
|
|
c->hadamard8_diff[1]= hadamard8_diff_mmx2;
|
|
#endif //CONFIG_ENCODERS
|
|
|
|
if(!(avctx->flags & CODEC_FLAG_BITEXACT)){
|
|
c->put_no_rnd_pixels_tab[0][1] = put_no_rnd_pixels16_x2_mmx2;
|
|
c->put_no_rnd_pixels_tab[0][2] = put_no_rnd_pixels16_y2_mmx2;
|
|
c->put_no_rnd_pixels_tab[1][1] = put_no_rnd_pixels8_x2_mmx2;
|
|
c->put_no_rnd_pixels_tab[1][2] = put_no_rnd_pixels8_y2_mmx2;
|
|
c->avg_pixels_tab[0][3] = avg_pixels16_xy2_mmx2;
|
|
c->avg_pixels_tab[1][3] = avg_pixels8_xy2_mmx2;
|
|
}
|
|
|
|
#if 1
|
|
SET_QPEL_FUNC(qpel_pixels_tab[0][ 0], qpel16_mc00_mmx2)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[0][ 1], qpel16_mc10_mmx2)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[0][ 2], qpel16_mc20_mmx2)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[0][ 3], qpel16_mc30_mmx2)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[0][ 4], qpel16_mc01_mmx2)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[0][ 5], qpel16_mc11_mmx2)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[0][ 6], qpel16_mc21_mmx2)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[0][ 7], qpel16_mc31_mmx2)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[0][ 8], qpel16_mc02_mmx2)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[0][ 9], qpel16_mc12_mmx2)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[0][10], qpel16_mc22_mmx2)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[0][11], qpel16_mc32_mmx2)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[0][12], qpel16_mc03_mmx2)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[0][13], qpel16_mc13_mmx2)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[0][14], qpel16_mc23_mmx2)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[0][15], qpel16_mc33_mmx2)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[1][ 0], qpel8_mc00_mmx2)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[1][ 1], qpel8_mc10_mmx2)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[1][ 2], qpel8_mc20_mmx2)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[1][ 3], qpel8_mc30_mmx2)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[1][ 4], qpel8_mc01_mmx2)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[1][ 5], qpel8_mc11_mmx2)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[1][ 6], qpel8_mc21_mmx2)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[1][ 7], qpel8_mc31_mmx2)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[1][ 8], qpel8_mc02_mmx2)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[1][ 9], qpel8_mc12_mmx2)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[1][10], qpel8_mc22_mmx2)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[1][11], qpel8_mc32_mmx2)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[1][12], qpel8_mc03_mmx2)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[1][13], qpel8_mc13_mmx2)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[1][14], qpel8_mc23_mmx2)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[1][15], qpel8_mc33_mmx2)
|
|
#endif
|
|
|
|
#ifdef CONFIG_ENCODERS
|
|
c->sub_hfyu_median_prediction= sub_hfyu_median_prediction_mmx2;
|
|
#endif //CONFIG_ENCODERS
|
|
} else if (mm_flags & MM_3DNOW) {
|
|
c->put_pixels_tab[0][1] = put_pixels16_x2_3dnow;
|
|
c->put_pixels_tab[0][2] = put_pixels16_y2_3dnow;
|
|
|
|
c->avg_pixels_tab[0][0] = avg_pixels16_3dnow;
|
|
c->avg_pixels_tab[0][1] = avg_pixels16_x2_3dnow;
|
|
c->avg_pixels_tab[0][2] = avg_pixels16_y2_3dnow;
|
|
|
|
c->put_pixels_tab[1][1] = put_pixels8_x2_3dnow;
|
|
c->put_pixels_tab[1][2] = put_pixels8_y2_3dnow;
|
|
|
|
c->avg_pixels_tab[1][0] = avg_pixels8_3dnow;
|
|
c->avg_pixels_tab[1][1] = avg_pixels8_x2_3dnow;
|
|
c->avg_pixels_tab[1][2] = avg_pixels8_y2_3dnow;
|
|
|
|
if(!(avctx->flags & CODEC_FLAG_BITEXACT)){
|
|
c->put_no_rnd_pixels_tab[0][1] = put_no_rnd_pixels16_x2_3dnow;
|
|
c->put_no_rnd_pixels_tab[0][2] = put_no_rnd_pixels16_y2_3dnow;
|
|
c->put_no_rnd_pixels_tab[1][1] = put_no_rnd_pixels8_x2_3dnow;
|
|
c->put_no_rnd_pixels_tab[1][2] = put_no_rnd_pixels8_y2_3dnow;
|
|
c->avg_pixels_tab[0][3] = avg_pixels16_xy2_3dnow;
|
|
c->avg_pixels_tab[1][3] = avg_pixels8_xy2_3dnow;
|
|
}
|
|
|
|
SET_QPEL_FUNC(qpel_pixels_tab[0][ 0], qpel16_mc00_3dnow)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[0][ 1], qpel16_mc10_3dnow)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[0][ 2], qpel16_mc20_3dnow)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[0][ 3], qpel16_mc30_3dnow)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[0][ 4], qpel16_mc01_3dnow)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[0][ 5], qpel16_mc11_3dnow)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[0][ 6], qpel16_mc21_3dnow)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[0][ 7], qpel16_mc31_3dnow)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[0][ 8], qpel16_mc02_3dnow)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[0][ 9], qpel16_mc12_3dnow)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[0][10], qpel16_mc22_3dnow)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[0][11], qpel16_mc32_3dnow)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[0][12], qpel16_mc03_3dnow)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[0][13], qpel16_mc13_3dnow)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[0][14], qpel16_mc23_3dnow)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[0][15], qpel16_mc33_3dnow)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[1][ 0], qpel8_mc00_3dnow)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[1][ 1], qpel8_mc10_3dnow)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[1][ 2], qpel8_mc20_3dnow)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[1][ 3], qpel8_mc30_3dnow)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[1][ 4], qpel8_mc01_3dnow)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[1][ 5], qpel8_mc11_3dnow)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[1][ 6], qpel8_mc21_3dnow)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[1][ 7], qpel8_mc31_3dnow)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[1][ 8], qpel8_mc02_3dnow)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[1][ 9], qpel8_mc12_3dnow)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[1][10], qpel8_mc22_3dnow)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[1][11], qpel8_mc32_3dnow)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[1][12], qpel8_mc03_3dnow)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[1][13], qpel8_mc13_3dnow)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[1][14], qpel8_mc23_3dnow)
|
|
SET_QPEL_FUNC(qpel_pixels_tab[1][15], qpel8_mc33_3dnow)
|
|
}
|
|
}
|
|
|
|
#ifdef CONFIG_ENCODERS
|
|
dsputil_init_pix_mmx(c, avctx);
|
|
#endif //CONFIG_ENCODERS
|
|
#if 0
|
|
// for speed testing
|
|
get_pixels = just_return;
|
|
put_pixels_clamped = just_return;
|
|
add_pixels_clamped = just_return;
|
|
|
|
pix_abs16x16 = just_return;
|
|
pix_abs16x16_x2 = just_return;
|
|
pix_abs16x16_y2 = just_return;
|
|
pix_abs16x16_xy2 = just_return;
|
|
|
|
put_pixels_tab[0] = just_return;
|
|
put_pixels_tab[1] = just_return;
|
|
put_pixels_tab[2] = just_return;
|
|
put_pixels_tab[3] = just_return;
|
|
|
|
put_no_rnd_pixels_tab[0] = just_return;
|
|
put_no_rnd_pixels_tab[1] = just_return;
|
|
put_no_rnd_pixels_tab[2] = just_return;
|
|
put_no_rnd_pixels_tab[3] = just_return;
|
|
|
|
avg_pixels_tab[0] = just_return;
|
|
avg_pixels_tab[1] = just_return;
|
|
avg_pixels_tab[2] = just_return;
|
|
avg_pixels_tab[3] = just_return;
|
|
|
|
avg_no_rnd_pixels_tab[0] = just_return;
|
|
avg_no_rnd_pixels_tab[1] = just_return;
|
|
avg_no_rnd_pixels_tab[2] = just_return;
|
|
avg_no_rnd_pixels_tab[3] = just_return;
|
|
|
|
//av_fdct = just_return;
|
|
//ff_idct = just_return;
|
|
#endif
|
|
}
|