mirror of
https://git.ffmpeg.org/ffmpeg.git
synced 2024-12-24 16:22:37 +00:00
bb198e198a
interlaced mpeg2 encoding P & B frames rate distored interlaced mb decission alternate scantable support 4mv encoding fixes (thats also why the regression tests change) passing height to most dsp functions interlaced mpeg4 encoding (no direct mode MBs yet) various related cleanups disabled old motion estimaton algorithms (log, full, ...) they will either be fixed or removed Originally committed as revision 2638 to svn://svn.ffmpeg.org/ffmpeg/trunk
344 lines
9.5 KiB
C
344 lines
9.5 KiB
C
/*
|
|
* Alpha optimized DSP utils
|
|
* Copyright (c) 2002 Falk Hueffner <falk@debian.org>
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2 of the License, or (at your option) any later version.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with this library; if not, write to the Free Software
|
|
* Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
|
|
*/
|
|
|
|
#include "asm.h"
|
|
#include "../dsputil.h"
|
|
|
|
void get_pixels_mvi(DCTELEM *restrict block,
|
|
const uint8_t *restrict pixels, int line_size)
|
|
{
|
|
int h = 8;
|
|
|
|
do {
|
|
uint64_t p;
|
|
|
|
p = ldq(pixels);
|
|
stq(unpkbw(p), block);
|
|
stq(unpkbw(p >> 32), block + 4);
|
|
|
|
pixels += line_size;
|
|
block += 8;
|
|
} while (--h);
|
|
}
|
|
|
|
void diff_pixels_mvi(DCTELEM *block, const uint8_t *s1, const uint8_t *s2,
|
|
int stride) {
|
|
int h = 8;
|
|
uint64_t mask = 0x4040;
|
|
|
|
mask |= mask << 16;
|
|
mask |= mask << 32;
|
|
do {
|
|
uint64_t x, y, c, d, a;
|
|
uint64_t signs;
|
|
|
|
x = ldq(s1);
|
|
y = ldq(s2);
|
|
c = cmpbge(x, y);
|
|
d = x - y;
|
|
a = zap(mask, c); /* We use 0x4040404040404040 here... */
|
|
d += 4 * a; /* ...so we can use s4addq here. */
|
|
signs = zap(-1, c);
|
|
|
|
stq(unpkbw(d) | (unpkbw(signs) << 8), block);
|
|
stq(unpkbw(d >> 32) | (unpkbw(signs >> 32) << 8), block + 4);
|
|
|
|
s1 += stride;
|
|
s2 += stride;
|
|
block += 8;
|
|
} while (--h);
|
|
}
|
|
|
|
static inline uint64_t avg2(uint64_t a, uint64_t b)
|
|
{
|
|
return (a | b) - (((a ^ b) & BYTE_VEC(0xfe)) >> 1);
|
|
}
|
|
|
|
static inline uint64_t avg4(uint64_t l1, uint64_t l2, uint64_t l3, uint64_t l4)
|
|
{
|
|
uint64_t r1 = ((l1 & ~BYTE_VEC(0x03)) >> 2)
|
|
+ ((l2 & ~BYTE_VEC(0x03)) >> 2)
|
|
+ ((l3 & ~BYTE_VEC(0x03)) >> 2)
|
|
+ ((l4 & ~BYTE_VEC(0x03)) >> 2);
|
|
uint64_t r2 = (( (l1 & BYTE_VEC(0x03))
|
|
+ (l2 & BYTE_VEC(0x03))
|
|
+ (l3 & BYTE_VEC(0x03))
|
|
+ (l4 & BYTE_VEC(0x03))
|
|
+ BYTE_VEC(0x02)) >> 2) & BYTE_VEC(0x03);
|
|
return r1 + r2;
|
|
}
|
|
|
|
int pix_abs8x8_mvi(void *v, uint8_t *pix1, uint8_t *pix2, int line_size, int h)
|
|
{
|
|
int result = 0;
|
|
|
|
if ((size_t) pix2 & 0x7) {
|
|
/* works only when pix2 is actually unaligned */
|
|
do { /* do 8 pixel a time */
|
|
uint64_t p1, p2;
|
|
|
|
p1 = ldq(pix1);
|
|
p2 = uldq(pix2);
|
|
result += perr(p1, p2);
|
|
|
|
pix1 += line_size;
|
|
pix2 += line_size;
|
|
} while (--h);
|
|
} else {
|
|
do {
|
|
uint64_t p1, p2;
|
|
|
|
p1 = ldq(pix1);
|
|
p2 = ldq(pix2);
|
|
result += perr(p1, p2);
|
|
|
|
pix1 += line_size;
|
|
pix2 += line_size;
|
|
} while (--h);
|
|
}
|
|
|
|
return result;
|
|
}
|
|
|
|
#if 0 /* now done in assembly */
|
|
int pix_abs16x16_mvi(uint8_t *pix1, uint8_t *pix2, int line_size)
|
|
{
|
|
int result = 0;
|
|
int h = 16;
|
|
|
|
if ((size_t) pix2 & 0x7) {
|
|
/* works only when pix2 is actually unaligned */
|
|
do { /* do 16 pixel a time */
|
|
uint64_t p1_l, p1_r, p2_l, p2_r;
|
|
uint64_t t;
|
|
|
|
p1_l = ldq(pix1);
|
|
p1_r = ldq(pix1 + 8);
|
|
t = ldq_u(pix2 + 8);
|
|
p2_l = extql(ldq_u(pix2), pix2) | extqh(t, pix2);
|
|
p2_r = extql(t, pix2) | extqh(ldq_u(pix2 + 16), pix2);
|
|
pix1 += line_size;
|
|
pix2 += line_size;
|
|
|
|
result += perr(p1_l, p2_l)
|
|
+ perr(p1_r, p2_r);
|
|
} while (--h);
|
|
} else {
|
|
do {
|
|
uint64_t p1_l, p1_r, p2_l, p2_r;
|
|
|
|
p1_l = ldq(pix1);
|
|
p1_r = ldq(pix1 + 8);
|
|
p2_l = ldq(pix2);
|
|
p2_r = ldq(pix2 + 8);
|
|
pix1 += line_size;
|
|
pix2 += line_size;
|
|
|
|
result += perr(p1_l, p2_l)
|
|
+ perr(p1_r, p2_r);
|
|
} while (--h);
|
|
}
|
|
|
|
return result;
|
|
}
|
|
#endif
|
|
|
|
int pix_abs16x16_x2_mvi(void *v, uint8_t *pix1, uint8_t *pix2, int line_size, int h)
|
|
{
|
|
int result = 0;
|
|
uint64_t disalign = (size_t) pix2 & 0x7;
|
|
|
|
switch (disalign) {
|
|
case 0:
|
|
do {
|
|
uint64_t p1_l, p1_r, p2_l, p2_r;
|
|
uint64_t l, r;
|
|
|
|
p1_l = ldq(pix1);
|
|
p1_r = ldq(pix1 + 8);
|
|
l = ldq(pix2);
|
|
r = ldq(pix2 + 8);
|
|
p2_l = avg2(l, (l >> 8) | ((uint64_t) r << 56));
|
|
p2_r = avg2(r, (r >> 8) | ((uint64_t) pix2[16] << 56));
|
|
pix1 += line_size;
|
|
pix2 += line_size;
|
|
|
|
result += perr(p1_l, p2_l)
|
|
+ perr(p1_r, p2_r);
|
|
} while (--h);
|
|
break;
|
|
case 7:
|
|
/* |.......l|lllllllr|rrrrrrr*|
|
|
This case is special because disalign1 would be 8, which
|
|
gets treated as 0 by extqh. At least it is a bit faster
|
|
that way :) */
|
|
do {
|
|
uint64_t p1_l, p1_r, p2_l, p2_r;
|
|
uint64_t l, m, r;
|
|
|
|
p1_l = ldq(pix1);
|
|
p1_r = ldq(pix1 + 8);
|
|
l = ldq_u(pix2);
|
|
m = ldq_u(pix2 + 8);
|
|
r = ldq_u(pix2 + 16);
|
|
p2_l = avg2(extql(l, disalign) | extqh(m, disalign), m);
|
|
p2_r = avg2(extql(m, disalign) | extqh(r, disalign), r);
|
|
pix1 += line_size;
|
|
pix2 += line_size;
|
|
|
|
result += perr(p1_l, p2_l)
|
|
+ perr(p1_r, p2_r);
|
|
} while (--h);
|
|
break;
|
|
default:
|
|
do {
|
|
uint64_t disalign1 = disalign + 1;
|
|
uint64_t p1_l, p1_r, p2_l, p2_r;
|
|
uint64_t l, m, r;
|
|
|
|
p1_l = ldq(pix1);
|
|
p1_r = ldq(pix1 + 8);
|
|
l = ldq_u(pix2);
|
|
m = ldq_u(pix2 + 8);
|
|
r = ldq_u(pix2 + 16);
|
|
p2_l = avg2(extql(l, disalign) | extqh(m, disalign),
|
|
extql(l, disalign1) | extqh(m, disalign1));
|
|
p2_r = avg2(extql(m, disalign) | extqh(r, disalign),
|
|
extql(m, disalign1) | extqh(r, disalign1));
|
|
pix1 += line_size;
|
|
pix2 += line_size;
|
|
|
|
result += perr(p1_l, p2_l)
|
|
+ perr(p1_r, p2_r);
|
|
} while (--h);
|
|
break;
|
|
}
|
|
return result;
|
|
}
|
|
|
|
int pix_abs16x16_y2_mvi(void *v, uint8_t *pix1, uint8_t *pix2, int line_size, int h)
|
|
{
|
|
int result = 0;
|
|
|
|
if ((size_t) pix2 & 0x7) {
|
|
uint64_t t, p2_l, p2_r;
|
|
t = ldq_u(pix2 + 8);
|
|
p2_l = extql(ldq_u(pix2), pix2) | extqh(t, pix2);
|
|
p2_r = extql(t, pix2) | extqh(ldq_u(pix2 + 16), pix2);
|
|
|
|
do {
|
|
uint64_t p1_l, p1_r, np2_l, np2_r;
|
|
uint64_t t;
|
|
|
|
p1_l = ldq(pix1);
|
|
p1_r = ldq(pix1 + 8);
|
|
pix2 += line_size;
|
|
t = ldq_u(pix2 + 8);
|
|
np2_l = extql(ldq_u(pix2), pix2) | extqh(t, pix2);
|
|
np2_r = extql(t, pix2) | extqh(ldq_u(pix2 + 16), pix2);
|
|
|
|
result += perr(p1_l, avg2(p2_l, np2_l))
|
|
+ perr(p1_r, avg2(p2_r, np2_r));
|
|
|
|
pix1 += line_size;
|
|
p2_l = np2_l;
|
|
p2_r = np2_r;
|
|
|
|
} while (--h);
|
|
} else {
|
|
uint64_t p2_l, p2_r;
|
|
p2_l = ldq(pix2);
|
|
p2_r = ldq(pix2 + 8);
|
|
do {
|
|
uint64_t p1_l, p1_r, np2_l, np2_r;
|
|
|
|
p1_l = ldq(pix1);
|
|
p1_r = ldq(pix1 + 8);
|
|
pix2 += line_size;
|
|
np2_l = ldq(pix2);
|
|
np2_r = ldq(pix2 + 8);
|
|
|
|
result += perr(p1_l, avg2(p2_l, np2_l))
|
|
+ perr(p1_r, avg2(p2_r, np2_r));
|
|
|
|
pix1 += line_size;
|
|
p2_l = np2_l;
|
|
p2_r = np2_r;
|
|
} while (--h);
|
|
}
|
|
return result;
|
|
}
|
|
|
|
int pix_abs16x16_xy2_mvi(void *v, uint8_t *pix1, uint8_t *pix2, int line_size, int h)
|
|
{
|
|
int result = 0;
|
|
|
|
uint64_t p1_l, p1_r;
|
|
uint64_t p2_l, p2_r, p2_x;
|
|
|
|
p1_l = ldq(pix1);
|
|
p1_r = ldq(pix1 + 8);
|
|
|
|
if ((size_t) pix2 & 0x7) { /* could be optimized a lot */
|
|
p2_l = uldq(pix2);
|
|
p2_r = uldq(pix2 + 8);
|
|
p2_x = (uint64_t) pix2[16] << 56;
|
|
} else {
|
|
p2_l = ldq(pix2);
|
|
p2_r = ldq(pix2 + 8);
|
|
p2_x = ldq(pix2 + 16) << 56;
|
|
}
|
|
|
|
do {
|
|
uint64_t np1_l, np1_r;
|
|
uint64_t np2_l, np2_r, np2_x;
|
|
|
|
pix1 += line_size;
|
|
pix2 += line_size;
|
|
|
|
np1_l = ldq(pix1);
|
|
np1_r = ldq(pix1 + 8);
|
|
|
|
if ((size_t) pix2 & 0x7) { /* could be optimized a lot */
|
|
np2_l = uldq(pix2);
|
|
np2_r = uldq(pix2 + 8);
|
|
np2_x = (uint64_t) pix2[16] << 56;
|
|
} else {
|
|
np2_l = ldq(pix2);
|
|
np2_r = ldq(pix2 + 8);
|
|
np2_x = ldq(pix2 + 16) << 56;
|
|
}
|
|
|
|
result += perr(p1_l,
|
|
avg4( p2_l, ( p2_l >> 8) | ((uint64_t) p2_r << 56),
|
|
np2_l, (np2_l >> 8) | ((uint64_t) np2_r << 56)))
|
|
+ perr(p1_r,
|
|
avg4( p2_r, ( p2_r >> 8) | ((uint64_t) p2_x),
|
|
np2_r, (np2_r >> 8) | ((uint64_t) np2_x)));
|
|
|
|
p1_l = np1_l;
|
|
p1_r = np1_r;
|
|
p2_l = np2_l;
|
|
p2_r = np2_r;
|
|
p2_x = np2_x;
|
|
} while (--h);
|
|
|
|
return result;
|
|
}
|