2011-09-28 20:53:03 +00:00
|
|
|
/*
|
|
|
|
* Copyright (c) 2011 Mans Rullgard <mans@mansr.com>
|
|
|
|
*
|
|
|
|
* This file is part of Libav.
|
|
|
|
*
|
|
|
|
* Libav is free software; you can redistribute it and/or
|
|
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
|
|
* License as published by the Free Software Foundation; either
|
|
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
|
|
*
|
|
|
|
* Libav is distributed in the hope that it will be useful,
|
|
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
|
|
* Lesser General Public License for more details.
|
|
|
|
*
|
|
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
|
|
* License along with Libav; if not, write to the Free Software
|
|
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
|
|
*/
|
|
|
|
|
|
|
|
#ifndef AVCODEC_ARM_DCA_H
|
|
|
|
#define AVCODEC_ARM_DCA_H
|
|
|
|
|
|
|
|
#include <stdint.h>
|
|
|
|
#include "config.h"
|
2011-10-23 16:39:49 +00:00
|
|
|
#include "libavutil/intmath.h"
|
|
|
|
|
2011-12-15 00:29:49 +00:00
|
|
|
#if HAVE_ARMV6 && HAVE_INLINE_ASM && AV_GCC_VERSION_AT_LEAST(4,4)
|
2011-10-23 16:39:49 +00:00
|
|
|
|
|
|
|
#define decode_blockcodes decode_blockcodes
|
|
|
|
static inline int decode_blockcodes(int code1, int code2, int levels,
|
|
|
|
int *values)
|
|
|
|
{
|
|
|
|
int v0, v1, v2, v3, v4, v5;
|
|
|
|
|
|
|
|
__asm__ ("smmul %8, %14, %18 \n"
|
|
|
|
"smmul %11, %15, %18 \n"
|
|
|
|
"smlabb %14, %8, %17, %14 \n"
|
|
|
|
"smlabb %15, %11, %17, %15 \n"
|
|
|
|
"smmul %9, %8, %18 \n"
|
|
|
|
"smmul %12, %11, %18 \n"
|
|
|
|
"sub %14, %14, %16, lsr #1 \n"
|
|
|
|
"sub %15, %15, %16, lsr #1 \n"
|
|
|
|
"smlabb %8, %9, %17, %8 \n"
|
|
|
|
"smlabb %11, %12, %17, %11 \n"
|
|
|
|
"smmul %10, %9, %18 \n"
|
|
|
|
"smmul %13, %12, %18 \n"
|
|
|
|
"str %14, %0 \n"
|
|
|
|
"str %15, %4 \n"
|
|
|
|
"sub %8, %8, %16, lsr #1 \n"
|
|
|
|
"sub %11, %11, %16, lsr #1 \n"
|
|
|
|
"smlabb %9, %10, %17, %9 \n"
|
|
|
|
"smlabb %12, %13, %17, %12 \n"
|
|
|
|
"smmul %14, %10, %18 \n"
|
|
|
|
"smmul %15, %13, %18 \n"
|
|
|
|
"str %8, %1 \n"
|
|
|
|
"str %11, %5 \n"
|
|
|
|
"sub %9, %9, %16, lsr #1 \n"
|
|
|
|
"sub %12, %12, %16, lsr #1 \n"
|
|
|
|
"smlabb %10, %14, %17, %10 \n"
|
|
|
|
"smlabb %13, %15, %17, %13 \n"
|
|
|
|
"str %9, %2 \n"
|
|
|
|
"str %12, %6 \n"
|
|
|
|
"sub %10, %10, %16, lsr #1 \n"
|
|
|
|
"sub %13, %13, %16, lsr #1 \n"
|
|
|
|
"str %10, %3 \n"
|
|
|
|
"str %13, %7 \n"
|
|
|
|
: "=m"(values[0]), "=m"(values[1]),
|
|
|
|
"=m"(values[2]), "=m"(values[3]),
|
|
|
|
"=m"(values[4]), "=m"(values[5]),
|
|
|
|
"=m"(values[6]), "=m"(values[7]),
|
|
|
|
"=&r"(v0), "=&r"(v1), "=&r"(v2),
|
|
|
|
"=&r"(v3), "=&r"(v4), "=&r"(v5),
|
|
|
|
"+&r"(code1), "+&r"(code2)
|
|
|
|
: "r"(levels - 1), "r"(-levels), "r"(ff_inverse[levels]));
|
|
|
|
|
|
|
|
return code1 | code2;
|
|
|
|
}
|
|
|
|
|
|
|
|
#endif
|
2011-09-28 20:53:03 +00:00
|
|
|
|
2011-10-02 22:38:32 +00:00
|
|
|
#if HAVE_NEON && HAVE_INLINE_ASM && HAVE_ASM_MOD_Y
|
2011-09-28 20:53:03 +00:00
|
|
|
|
|
|
|
#define int8x8_fmul_int32 int8x8_fmul_int32
|
|
|
|
static inline void int8x8_fmul_int32(float *dst, const int8_t *src, int scale)
|
|
|
|
{
|
|
|
|
__asm__ ("vcvt.f32.s32 %2, %2, #4 \n"
|
|
|
|
"vld1.8 {d0}, [%1,:64] \n"
|
|
|
|
"vmovl.s8 q0, d0 \n"
|
|
|
|
"vmovl.s16 q1, d1 \n"
|
|
|
|
"vmovl.s16 q0, d0 \n"
|
|
|
|
"vcvt.f32.s32 q0, q0 \n"
|
|
|
|
"vcvt.f32.s32 q1, q1 \n"
|
|
|
|
"vmul.f32 q0, q0, %y2 \n"
|
|
|
|
"vmul.f32 q1, q1, %y2 \n"
|
|
|
|
"vst1.32 {q0-q1}, [%m0,:128] \n"
|
|
|
|
: "=Um"(*(float (*)[8])dst)
|
|
|
|
: "r"(src), "x"(scale)
|
|
|
|
: "d0", "d1", "d2", "d3");
|
|
|
|
}
|
|
|
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#endif /* AVCODEC_ARM_DCA_H */
|