mirror of
https://git.ffmpeg.org/ffmpeg.git
synced 2024-12-30 19:32:13 +00:00
157d6f0d5b
This mainly consists of replacing all the pointer arithmatic 'addiu' instructions with PTR_ADDIU which will handle the differences in pointer sizes when compiled on 64 bit mips systems. The header asmdefs.h contains the PTR_ macros which expend to the correct mips instructions to manipulate registers containing pointers. Signed-off-by: James Cowgill <james410@cowgill.org.uk> Reviewed-by: Nedeljko Babic <Nedeljko.Babic@imgtec.com> Signed-off-by: Michael Niedermayer <michaelni@gmx.at>
443 lines
18 KiB
C
443 lines
18 KiB
C
/*
|
|
* Copyright (c) 2012
|
|
* MIPS Technologies, Inc., California.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
* 3. Neither the name of the MIPS Technologies, Inc., nor the names of its
|
|
* contributors may be used to endorse or promote products derived from
|
|
* this software without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE MIPS TECHNOLOGIES, INC. ``AS IS'' AND
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE MIPS TECHNOLOGIES, INC. BE LIABLE
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
* SUCH DAMAGE.
|
|
*
|
|
* Authors: Darko Laus (darko@mips.com)
|
|
* Djordje Pesut (djordje@mips.com)
|
|
* Mirjana Vulin (mvulin@mips.com)
|
|
*
|
|
* This file is part of FFmpeg.
|
|
*
|
|
* FFmpeg is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with FFmpeg; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
/**
|
|
* @file
|
|
* Reference: libavcodec/aacdec.c
|
|
*/
|
|
|
|
#include "libavcodec/aac.h"
|
|
#include "aacdec_mips.h"
|
|
#include "libavcodec/aactab.h"
|
|
#include "libavcodec/sinewin.h"
|
|
#include "libavutil/mips/asmdefs.h"
|
|
|
|
#if HAVE_INLINE_ASM
|
|
static av_always_inline void float_copy(float *dst, const float *src, int count)
|
|
{
|
|
// Copy 'count' floats from src to dst
|
|
const float *loop_end = src + count;
|
|
int temp[8];
|
|
|
|
// count must be a multiple of 8
|
|
av_assert2(count % 8 == 0);
|
|
|
|
// loop unrolled 8 times
|
|
__asm__ volatile (
|
|
".set push \n\t"
|
|
".set noreorder \n\t"
|
|
"1: \n\t"
|
|
"lw %[temp0], 0(%[src]) \n\t"
|
|
"lw %[temp1], 4(%[src]) \n\t"
|
|
"lw %[temp2], 8(%[src]) \n\t"
|
|
"lw %[temp3], 12(%[src]) \n\t"
|
|
"lw %[temp4], 16(%[src]) \n\t"
|
|
"lw %[temp5], 20(%[src]) \n\t"
|
|
"lw %[temp6], 24(%[src]) \n\t"
|
|
"lw %[temp7], 28(%[src]) \n\t"
|
|
PTR_ADDIU "%[src], %[src], 32 \n\t"
|
|
"sw %[temp0], 0(%[dst]) \n\t"
|
|
"sw %[temp1], 4(%[dst]) \n\t"
|
|
"sw %[temp2], 8(%[dst]) \n\t"
|
|
"sw %[temp3], 12(%[dst]) \n\t"
|
|
"sw %[temp4], 16(%[dst]) \n\t"
|
|
"sw %[temp5], 20(%[dst]) \n\t"
|
|
"sw %[temp6], 24(%[dst]) \n\t"
|
|
"sw %[temp7], 28(%[dst]) \n\t"
|
|
"bne %[src], %[loop_end], 1b \n\t"
|
|
PTR_ADDIU "%[dst], %[dst], 32 \n\t"
|
|
".set pop \n\t"
|
|
|
|
: [temp0]"=&r"(temp[0]), [temp1]"=&r"(temp[1]),
|
|
[temp2]"=&r"(temp[2]), [temp3]"=&r"(temp[3]),
|
|
[temp4]"=&r"(temp[4]), [temp5]"=&r"(temp[5]),
|
|
[temp6]"=&r"(temp[6]), [temp7]"=&r"(temp[7]),
|
|
[src]"+r"(src), [dst]"+r"(dst)
|
|
: [loop_end]"r"(loop_end)
|
|
: "memory"
|
|
);
|
|
}
|
|
|
|
static av_always_inline int lcg_random(unsigned previous_val)
|
|
{
|
|
union { unsigned u; int s; } v = { previous_val * 1664525u + 1013904223 };
|
|
return v.s;
|
|
}
|
|
|
|
static void imdct_and_windowing_mips(AACContext *ac, SingleChannelElement *sce)
|
|
{
|
|
IndividualChannelStream *ics = &sce->ics;
|
|
float *in = sce->coeffs;
|
|
float *out = sce->ret;
|
|
float *saved = sce->saved;
|
|
const float *swindow = ics->use_kb_window[0] ? ff_aac_kbd_short_128 : ff_sine_128;
|
|
const float *lwindow_prev = ics->use_kb_window[1] ? ff_aac_kbd_long_1024 : ff_sine_1024;
|
|
const float *swindow_prev = ics->use_kb_window[1] ? ff_aac_kbd_short_128 : ff_sine_128;
|
|
float *buf = ac->buf_mdct;
|
|
int i;
|
|
|
|
if (ics->window_sequence[0] == EIGHT_SHORT_SEQUENCE) {
|
|
for (i = 0; i < 1024; i += 128)
|
|
ac->mdct_small.imdct_half(&ac->mdct_small, buf + i, in + i);
|
|
} else
|
|
ac->mdct.imdct_half(&ac->mdct, buf, in);
|
|
|
|
/* window overlapping
|
|
* NOTE: To simplify the overlapping code, all 'meaningless' short to long
|
|
* and long to short transitions are considered to be short to short
|
|
* transitions. This leaves just two cases (long to long and short to short)
|
|
* with a little special sauce for EIGHT_SHORT_SEQUENCE.
|
|
*/
|
|
if ((ics->window_sequence[1] == ONLY_LONG_SEQUENCE || ics->window_sequence[1] == LONG_STOP_SEQUENCE) &&
|
|
(ics->window_sequence[0] == ONLY_LONG_SEQUENCE || ics->window_sequence[0] == LONG_START_SEQUENCE)) {
|
|
ac->fdsp->vector_fmul_window( out, saved, buf, lwindow_prev, 512);
|
|
} else {
|
|
float_copy(out, saved, 448);
|
|
|
|
if (ics->window_sequence[0] == EIGHT_SHORT_SEQUENCE) {
|
|
{
|
|
float wi;
|
|
float wj;
|
|
int i;
|
|
float temp0, temp1, temp2, temp3;
|
|
float *dst0 = out + 448 + 0*128;
|
|
float *dst1 = dst0 + 64 + 63;
|
|
float *dst2 = saved + 63;
|
|
float *win0 = (float*)swindow;
|
|
float *win1 = win0 + 64 + 63;
|
|
float *win0_prev = (float*)swindow_prev;
|
|
float *win1_prev = win0_prev + 64 + 63;
|
|
float *src0_prev = saved + 448;
|
|
float *src1_prev = buf + 0*128 + 63;
|
|
float *src0 = buf + 0*128 + 64;
|
|
float *src1 = buf + 1*128 + 63;
|
|
|
|
for(i = 0; i < 64; i++)
|
|
{
|
|
temp0 = src0_prev[0];
|
|
temp1 = src1_prev[0];
|
|
wi = *win0_prev;
|
|
wj = *win1_prev;
|
|
temp2 = src0[0];
|
|
temp3 = src1[0];
|
|
dst0[0] = temp0 * wj - temp1 * wi;
|
|
dst1[0] = temp0 * wi + temp1 * wj;
|
|
|
|
wi = *win0;
|
|
wj = *win1;
|
|
|
|
temp0 = src0[128];
|
|
temp1 = src1[128];
|
|
dst0[128] = temp2 * wj - temp3 * wi;
|
|
dst1[128] = temp2 * wi + temp3 * wj;
|
|
|
|
temp2 = src0[256];
|
|
temp3 = src1[256];
|
|
dst0[256] = temp0 * wj - temp1 * wi;
|
|
dst1[256] = temp0 * wi + temp1 * wj;
|
|
dst0[384] = temp2 * wj - temp3 * wi;
|
|
dst1[384] = temp2 * wi + temp3 * wj;
|
|
|
|
temp0 = src0[384];
|
|
temp1 = src1[384];
|
|
dst0[512] = temp0 * wj - temp1 * wi;
|
|
dst2[0] = temp0 * wi + temp1 * wj;
|
|
|
|
src0++;
|
|
src1--;
|
|
src0_prev++;
|
|
src1_prev--;
|
|
win0++;
|
|
win1--;
|
|
win0_prev++;
|
|
win1_prev--;
|
|
dst0++;
|
|
dst1--;
|
|
dst2--;
|
|
}
|
|
}
|
|
} else {
|
|
ac->fdsp->vector_fmul_window(out + 448, saved + 448, buf, swindow_prev, 64);
|
|
float_copy(out + 576, buf + 64, 448);
|
|
}
|
|
}
|
|
|
|
// buffer update
|
|
if (ics->window_sequence[0] == EIGHT_SHORT_SEQUENCE) {
|
|
ac->fdsp->vector_fmul_window(saved + 64, buf + 4*128 + 64, buf + 5*128, swindow, 64);
|
|
ac->fdsp->vector_fmul_window(saved + 192, buf + 5*128 + 64, buf + 6*128, swindow, 64);
|
|
ac->fdsp->vector_fmul_window(saved + 320, buf + 6*128 + 64, buf + 7*128, swindow, 64);
|
|
float_copy(saved + 448, buf + 7*128 + 64, 64);
|
|
} else if (ics->window_sequence[0] == LONG_START_SEQUENCE) {
|
|
float_copy(saved, buf + 512, 448);
|
|
float_copy(saved + 448, buf + 7*128 + 64, 64);
|
|
} else { // LONG_STOP or ONLY_LONG
|
|
float_copy(saved, buf + 512, 512);
|
|
}
|
|
}
|
|
|
|
static void apply_ltp_mips(AACContext *ac, SingleChannelElement *sce)
|
|
{
|
|
const LongTermPrediction *ltp = &sce->ics.ltp;
|
|
const uint16_t *offsets = sce->ics.swb_offset;
|
|
int i, sfb;
|
|
int j, k;
|
|
|
|
if (sce->ics.window_sequence[0] != EIGHT_SHORT_SEQUENCE) {
|
|
float *predTime = sce->ret;
|
|
float *predFreq = ac->buf_mdct;
|
|
float *p_predTime;
|
|
int16_t num_samples = 2048;
|
|
|
|
if (ltp->lag < 1024)
|
|
num_samples = ltp->lag + 1024;
|
|
j = (2048 - num_samples) >> 2;
|
|
k = (2048 - num_samples) & 3;
|
|
p_predTime = &predTime[num_samples];
|
|
|
|
for (i = 0; i < num_samples; i++)
|
|
predTime[i] = sce->ltp_state[i + 2048 - ltp->lag] * ltp->coef;
|
|
for (i = 0; i < j; i++) {
|
|
|
|
/* loop unrolled 4 times */
|
|
__asm__ volatile (
|
|
"sw $0, 0(%[p_predTime]) \n\t"
|
|
"sw $0, 4(%[p_predTime]) \n\t"
|
|
"sw $0, 8(%[p_predTime]) \n\t"
|
|
"sw $0, 12(%[p_predTime]) \n\t"
|
|
PTR_ADDIU "%[p_predTime], %[p_predTime], 16 \n\t"
|
|
|
|
: [p_predTime]"+r"(p_predTime)
|
|
:
|
|
: "memory"
|
|
);
|
|
}
|
|
for (i = 0; i < k; i++) {
|
|
|
|
__asm__ volatile (
|
|
"sw $0, 0(%[p_predTime]) \n\t"
|
|
PTR_ADDIU "%[p_predTime], %[p_predTime], 4 \n\t"
|
|
|
|
: [p_predTime]"+r"(p_predTime)
|
|
:
|
|
: "memory"
|
|
);
|
|
}
|
|
|
|
ac->windowing_and_mdct_ltp(ac, predFreq, predTime, &sce->ics);
|
|
|
|
if (sce->tns.present)
|
|
ac->apply_tns(predFreq, &sce->tns, &sce->ics, 0);
|
|
|
|
for (sfb = 0; sfb < FFMIN(sce->ics.max_sfb, MAX_LTP_LONG_SFB); sfb++)
|
|
if (ltp->used[sfb])
|
|
for (i = offsets[sfb]; i < offsets[sfb + 1]; i++)
|
|
sce->coeffs[i] += predFreq[i];
|
|
}
|
|
}
|
|
|
|
#if HAVE_MIPSFPU
|
|
static av_always_inline void fmul_and_reverse(float *dst, const float *src0, const float *src1, int count)
|
|
{
|
|
/* Multiply 'count' floats in src0 by src1 and store the results in dst in reverse */
|
|
/* This should be equivalent to a normal fmul, followed by reversing dst */
|
|
|
|
// count must be a multiple of 4
|
|
av_assert2(count % 4 == 0);
|
|
|
|
// move src0 and src1 to the last element of their arrays
|
|
src0 += count - 1;
|
|
src1 += count - 1;
|
|
|
|
for (; count > 0; count -= 4){
|
|
float temp[12];
|
|
|
|
/* loop unrolled 4 times */
|
|
__asm__ volatile (
|
|
"lwc1 %[temp0], 0(%[ptr2]) \n\t"
|
|
"lwc1 %[temp1], -4(%[ptr2]) \n\t"
|
|
"lwc1 %[temp2], -8(%[ptr2]) \n\t"
|
|
"lwc1 %[temp3], -12(%[ptr2]) \n\t"
|
|
"lwc1 %[temp4], 0(%[ptr3]) \n\t"
|
|
"lwc1 %[temp5], -4(%[ptr3]) \n\t"
|
|
"lwc1 %[temp6], -8(%[ptr3]) \n\t"
|
|
"lwc1 %[temp7], -12(%[ptr3]) \n\t"
|
|
"mul.s %[temp8], %[temp0], %[temp4] \n\t"
|
|
"mul.s %[temp9], %[temp1], %[temp5] \n\t"
|
|
"mul.s %[temp10], %[temp2], %[temp6] \n\t"
|
|
"mul.s %[temp11], %[temp3], %[temp7] \n\t"
|
|
"swc1 %[temp8], 0(%[ptr1]) \n\t"
|
|
"swc1 %[temp9], 4(%[ptr1]) \n\t"
|
|
"swc1 %[temp10], 8(%[ptr1]) \n\t"
|
|
"swc1 %[temp11], 12(%[ptr1]) \n\t"
|
|
PTR_ADDIU "%[ptr1], %[ptr1], 16 \n\t"
|
|
PTR_ADDIU "%[ptr2], %[ptr2], -16 \n\t"
|
|
PTR_ADDIU "%[ptr3], %[ptr3], -16 \n\t"
|
|
|
|
: [temp0]"=&f"(temp[0]), [temp1]"=&f"(temp[1]),
|
|
[temp2]"=&f"(temp[2]), [temp3]"=&f"(temp[3]),
|
|
[temp4]"=&f"(temp[4]), [temp5]"=&f"(temp[5]),
|
|
[temp6]"=&f"(temp[6]), [temp7]"=&f"(temp[7]),
|
|
[temp8]"=&f"(temp[8]), [temp9]"=&f"(temp[9]),
|
|
[temp10]"=&f"(temp[10]), [temp11]"=&f"(temp[11]),
|
|
[ptr1]"+r"(dst), [ptr2]"+r"(src0), [ptr3]"+r"(src1)
|
|
:
|
|
: "memory"
|
|
);
|
|
}
|
|
}
|
|
|
|
static void update_ltp_mips(AACContext *ac, SingleChannelElement *sce)
|
|
{
|
|
IndividualChannelStream *ics = &sce->ics;
|
|
float *saved = sce->saved;
|
|
float *saved_ltp = sce->coeffs;
|
|
const float *lwindow = ics->use_kb_window[0] ? ff_aac_kbd_long_1024 : ff_sine_1024;
|
|
const float *swindow = ics->use_kb_window[0] ? ff_aac_kbd_short_128 : ff_sine_128;
|
|
float temp0, temp1, temp2, temp3, temp4, temp5, temp6, temp7;
|
|
|
|
if (ics->window_sequence[0] == EIGHT_SHORT_SEQUENCE) {
|
|
float *p_saved_ltp = saved_ltp + 576;
|
|
float *loop_end1 = p_saved_ltp + 448;
|
|
|
|
float_copy(saved_ltp, saved, 512);
|
|
|
|
/* loop unrolled 8 times */
|
|
__asm__ volatile (
|
|
"1: \n\t"
|
|
"sw $0, 0(%[p_saved_ltp]) \n\t"
|
|
"sw $0, 4(%[p_saved_ltp]) \n\t"
|
|
"sw $0, 8(%[p_saved_ltp]) \n\t"
|
|
"sw $0, 12(%[p_saved_ltp]) \n\t"
|
|
"sw $0, 16(%[p_saved_ltp]) \n\t"
|
|
"sw $0, 20(%[p_saved_ltp]) \n\t"
|
|
"sw $0, 24(%[p_saved_ltp]) \n\t"
|
|
"sw $0, 28(%[p_saved_ltp]) \n\t"
|
|
PTR_ADDIU "%[p_saved_ltp],%[p_saved_ltp], 32 \n\t"
|
|
"bne %[p_saved_ltp], %[loop_end1], 1b \n\t"
|
|
|
|
: [p_saved_ltp]"+r"(p_saved_ltp)
|
|
: [loop_end1]"r"(loop_end1)
|
|
: "memory"
|
|
);
|
|
|
|
ac->fdsp->vector_fmul_reverse(saved_ltp + 448, ac->buf_mdct + 960, &swindow[64], 64);
|
|
fmul_and_reverse(saved_ltp + 512, ac->buf_mdct + 960, swindow, 64);
|
|
} else if (ics->window_sequence[0] == LONG_START_SEQUENCE) {
|
|
float *buff0 = saved;
|
|
float *buff1 = saved_ltp;
|
|
float *loop_end = saved + 448;
|
|
|
|
/* loop unrolled 8 times */
|
|
__asm__ volatile (
|
|
".set push \n\t"
|
|
".set noreorder \n\t"
|
|
"1: \n\t"
|
|
"lw %[temp0], 0(%[src]) \n\t"
|
|
"lw %[temp1], 4(%[src]) \n\t"
|
|
"lw %[temp2], 8(%[src]) \n\t"
|
|
"lw %[temp3], 12(%[src]) \n\t"
|
|
"lw %[temp4], 16(%[src]) \n\t"
|
|
"lw %[temp5], 20(%[src]) \n\t"
|
|
"lw %[temp6], 24(%[src]) \n\t"
|
|
"lw %[temp7], 28(%[src]) \n\t"
|
|
PTR_ADDIU "%[src], %[src], 32 \n\t"
|
|
"sw %[temp0], 0(%[dst]) \n\t"
|
|
"sw %[temp1], 4(%[dst]) \n\t"
|
|
"sw %[temp2], 8(%[dst]) \n\t"
|
|
"sw %[temp3], 12(%[dst]) \n\t"
|
|
"sw %[temp4], 16(%[dst]) \n\t"
|
|
"sw %[temp5], 20(%[dst]) \n\t"
|
|
"sw %[temp6], 24(%[dst]) \n\t"
|
|
"sw %[temp7], 28(%[dst]) \n\t"
|
|
"sw $0, 2304(%[dst]) \n\t"
|
|
"sw $0, 2308(%[dst]) \n\t"
|
|
"sw $0, 2312(%[dst]) \n\t"
|
|
"sw $0, 2316(%[dst]) \n\t"
|
|
"sw $0, 2320(%[dst]) \n\t"
|
|
"sw $0, 2324(%[dst]) \n\t"
|
|
"sw $0, 2328(%[dst]) \n\t"
|
|
"sw $0, 2332(%[dst]) \n\t"
|
|
"bne %[src], %[loop_end], 1b \n\t"
|
|
PTR_ADDIU "%[dst], %[dst], 32 \n\t"
|
|
".set pop \n\t"
|
|
|
|
: [temp0]"=&r"(temp0), [temp1]"=&r"(temp1),
|
|
[temp2]"=&r"(temp2), [temp3]"=&r"(temp3),
|
|
[temp4]"=&r"(temp4), [temp5]"=&r"(temp5),
|
|
[temp6]"=&r"(temp6), [temp7]"=&r"(temp7),
|
|
[src]"+r"(buff0), [dst]"+r"(buff1)
|
|
: [loop_end]"r"(loop_end)
|
|
: "memory"
|
|
);
|
|
ac->fdsp->vector_fmul_reverse(saved_ltp + 448, ac->buf_mdct + 960, &swindow[64], 64);
|
|
fmul_and_reverse(saved_ltp + 512, ac->buf_mdct + 960, swindow, 64);
|
|
} else { // LONG_STOP or ONLY_LONG
|
|
ac->fdsp->vector_fmul_reverse(saved_ltp, ac->buf_mdct + 512, &lwindow[512], 512);
|
|
fmul_and_reverse(saved_ltp + 512, ac->buf_mdct + 512, lwindow, 512);
|
|
}
|
|
|
|
float_copy(sce->ltp_state, sce->ltp_state + 1024, 1024);
|
|
float_copy(sce->ltp_state + 1024, sce->ret, 1024);
|
|
float_copy(sce->ltp_state + 2048, saved_ltp, 1024);
|
|
}
|
|
#endif /* HAVE_MIPSFPU */
|
|
#endif /* HAVE_INLINE_ASM */
|
|
|
|
void ff_aacdec_init_mips(AACContext *c)
|
|
{
|
|
#if HAVE_INLINE_ASM
|
|
c->imdct_and_windowing = imdct_and_windowing_mips;
|
|
c->apply_ltp = apply_ltp_mips;
|
|
#if HAVE_MIPSFPU
|
|
c->update_ltp = update_ltp_mips;
|
|
#endif /* HAVE_MIPSFPU */
|
|
#endif /* HAVE_INLINE_ASM */
|
|
}
|