850adb40d7
git-svn-id: svn://svn.rockbox.org/rockbox/trunk@18413 a1c6a512-1295-4272-9138-f99709370657
230 lines
6 KiB
C
230 lines
6 KiB
C
/********************************************************************
|
|
* *
|
|
* THIS FILE IS PART OF THE OggVorbis 'TREMOR' CODEC SOURCE CODE. *
|
|
* *
|
|
* USE, DISTRIBUTION AND REPRODUCTION OF THIS LIBRARY SOURCE IS *
|
|
* GOVERNED BY A BSD-STYLE SOURCE LICENSE INCLUDED WITH THIS SOURCE *
|
|
* IN 'COPYING'. PLEASE READ THESE TERMS BEFORE DISTRIBUTING. *
|
|
* *
|
|
* THE OggVorbis 'TREMOR' SOURCE CODE IS (C) COPYRIGHT 1994-2002 *
|
|
* BY THE Xiph.Org FOUNDATION http://www.xiph.org/ *
|
|
* *
|
|
********************************************************************
|
|
|
|
function: arm7 and later wide math functions
|
|
|
|
********************************************************************/
|
|
#ifdef CPU_ARM
|
|
|
|
#if !defined(_V_WIDE_MATH) && !defined(_LOW_ACCURACY_)
|
|
#define _V_WIDE_MATH
|
|
|
|
static inline int32_t MULT32(int32_t x, int32_t y) {
|
|
int lo,hi;
|
|
asm volatile("smull\t%0, %1, %2, %3"
|
|
: "=&r"(lo),"=&r"(hi)
|
|
: "%r"(x),"r"(y)
|
|
: "cc");
|
|
return(hi);
|
|
}
|
|
|
|
static inline int32_t MULT31(int32_t x, int32_t y) {
|
|
return MULT32(x,y)<<1;
|
|
}
|
|
|
|
static inline int32_t MULT31_SHIFT15(int32_t x, int32_t y) {
|
|
int lo,hi;
|
|
asm volatile("smull %0, %1, %2, %3\n\t"
|
|
"movs %0, %0, lsr #15\n\t"
|
|
"adc %1, %0, %1, lsl #17\n\t"
|
|
: "=&r"(lo),"=&r"(hi)
|
|
: "%r"(x),"r"(y)
|
|
: "cc");
|
|
return(hi);
|
|
}
|
|
|
|
#define MB() asm volatile ("" : : : "memory")
|
|
|
|
#define XPROD32(a, b, t, v, x, y) \
|
|
{ \
|
|
long l; \
|
|
asm( "smull %0, %1, %4, %6\n\t" \
|
|
"smlal %0, %1, %5, %7\n\t" \
|
|
"rsb %3, %4, #0\n\t" \
|
|
"smull %0, %2, %5, %6\n\t" \
|
|
"smlal %0, %2, %3, %7" \
|
|
: "=&r" (l), "=&r" (x), "=&r" (y), "=r" ((a)) \
|
|
: "3" ((a)), "r" ((b)), "r" ((t)), "r" ((v)) \
|
|
: "cc" ); \
|
|
}
|
|
|
|
static inline void XPROD31(int32_t a, int32_t b,
|
|
int32_t t, int32_t v,
|
|
int32_t *x, int32_t *y)
|
|
{
|
|
int x1, y1, l;
|
|
asm( "smull %0, %1, %4, %6\n\t"
|
|
"smlal %0, %1, %5, %7\n\t"
|
|
"rsb %3, %4, #0\n\t"
|
|
"smull %0, %2, %5, %6\n\t"
|
|
"smlal %0, %2, %3, %7"
|
|
: "=&r" (l), "=&r" (x1), "=&r" (y1), "=r" (a)
|
|
: "3" (a), "r" (b), "r" (t), "r" (v)
|
|
: "cc" );
|
|
*x = x1 << 1;
|
|
MB();
|
|
*y = y1 << 1;
|
|
}
|
|
|
|
static inline void XNPROD31(int32_t a, int32_t b,
|
|
int32_t t, int32_t v,
|
|
int32_t *x, int32_t *y)
|
|
{
|
|
int x1, y1, l;
|
|
asm( "rsb %2, %4, #0\n\t"
|
|
"smull %0, %1, %3, %5\n\t"
|
|
"smlal %0, %1, %2, %6\n\t"
|
|
"smull %0, %2, %4, %5\n\t"
|
|
"smlal %0, %2, %3, %6"
|
|
: "=&r" (l), "=&r" (x1), "=&r" (y1)
|
|
: "r" (a), "r" (b), "r" (t), "r" (v)
|
|
: "cc" );
|
|
*x = x1 << 1;
|
|
MB();
|
|
*y = y1 << 1;
|
|
}
|
|
|
|
#ifndef _V_VECT_OPS
|
|
#define _V_VECT_OPS
|
|
|
|
/* asm versions of vector operations for block.c, window.c */
|
|
static inline
|
|
void vect_add(int32_t *x, int32_t *y, int n)
|
|
{
|
|
while (n>=4) {
|
|
asm volatile ("ldmia %[x], {r0, r1, r2, r3};"
|
|
"ldmia %[y]!, {r4, r5, r6, r7};"
|
|
"add r0, r0, r4;"
|
|
"add r1, r1, r5;"
|
|
"add r2, r2, r6;"
|
|
"add r3, r3, r7;"
|
|
"stmia %[x]!, {r0, r1, r2, r3};"
|
|
: [x] "+r" (x), [y] "+r" (y)
|
|
: : "r0", "r1", "r2", "r3",
|
|
"r4", "r5", "r6", "r7",
|
|
"memory");
|
|
n -= 4;
|
|
}
|
|
/* add final elements */
|
|
while (n>0) {
|
|
*x++ += *y++;
|
|
n--;
|
|
}
|
|
}
|
|
|
|
static inline
|
|
void vect_copy(int32_t *x, int32_t *y, int n)
|
|
{
|
|
while (n>=4) {
|
|
asm volatile ("ldmia %[y]!, {r0, r1, r2, r3};"
|
|
"stmia %[x]!, {r0, r1, r2, r3};"
|
|
: [x] "+r" (x), [y] "+r" (y)
|
|
: : "r0", "r1", "r2", "r3",
|
|
"memory");
|
|
n -= 4;
|
|
}
|
|
/* copy final elements */
|
|
while (n>0) {
|
|
*x++ = *y++;
|
|
n--;
|
|
}
|
|
}
|
|
|
|
static inline
|
|
void vect_mult_fw(int32_t *data, int32_t *window, int n)
|
|
{
|
|
while (n>=4) {
|
|
asm volatile ("ldmia %[d], {r0, r1, r2, r3};"
|
|
"ldmia %[w]!, {r4, r5, r6, r7};"
|
|
"smull r8, r9, r0, r4;"
|
|
"mov r0, r9, lsl #1;"
|
|
"smull r8, r9, r1, r5;"
|
|
"mov r1, r9, lsl #1;"
|
|
"smull r8, r9, r2, r6;"
|
|
"mov r2, r9, lsl #1;"
|
|
"smull r8, r9, r3, r7;"
|
|
"mov r3, r9, lsl #1;"
|
|
"stmia %[d]!, {r0, r1, r2, r3};"
|
|
: [d] "+r" (data), [w] "+r" (window)
|
|
: : "r0", "r1", "r2", "r3",
|
|
"r4", "r5", "r6", "r7", "r8", "r9",
|
|
"memory", "cc");
|
|
n -= 4;
|
|
}
|
|
while(n>0) {
|
|
*data = MULT31(*data, *window);
|
|
data++;
|
|
window++;
|
|
n--;
|
|
}
|
|
}
|
|
|
|
static inline
|
|
void vect_mult_bw(int32_t *data, int32_t *window, int n)
|
|
{
|
|
while (n>=4) {
|
|
asm volatile ("ldmia %[d], {r0, r1, r2, r3};"
|
|
"ldmda %[w]!, {r4, r5, r6, r7};"
|
|
"smull r8, r9, r0, r7;"
|
|
"mov r0, r9, lsl #1;"
|
|
"smull r8, r9, r1, r6;"
|
|
"mov r1, r9, lsl #1;"
|
|
"smull r8, r9, r2, r5;"
|
|
"mov r2, r9, lsl #1;"
|
|
"smull r8, r9, r3, r4;"
|
|
"mov r3, r9, lsl #1;"
|
|
"stmia %[d]!, {r0, r1, r2, r3};"
|
|
: [d] "+r" (data), [w] "+r" (window)
|
|
: : "r0", "r1", "r2", "r3",
|
|
"r4", "r5", "r6", "r7", "r8", "r9",
|
|
"memory", "cc");
|
|
n -= 4;
|
|
}
|
|
while(n>0) {
|
|
*data = MULT31(*data, *window);
|
|
data++;
|
|
window--;
|
|
n--;
|
|
}
|
|
}
|
|
|
|
#endif
|
|
|
|
#endif
|
|
|
|
#ifndef _V_CLIP_MATH
|
|
#define _V_CLIP_MATH
|
|
|
|
static inline int32_t CLIP_TO_15(int32_t x) {
|
|
int tmp;
|
|
asm volatile("subs %1, %0, #32768\n\t"
|
|
"movpl %0, #0x7f00\n\t"
|
|
"orrpl %0, %0, #0xff\n"
|
|
"adds %1, %0, #32768\n\t"
|
|
"movmi %0, #0x8000"
|
|
: "+r"(x),"=r"(tmp)
|
|
:
|
|
: "cc");
|
|
return(x);
|
|
}
|
|
|
|
#endif
|
|
|
|
#ifndef _V_LSP_MATH_ASM
|
|
#define _V_LSP_MATH_ASM
|
|
|
|
|
|
|
|
#endif
|
|
#endif
|
|
|