mame/src/emu/eigccppc.h
Aaron Giles 58d832602b Having decided to look at the MAME source again, I have a fairly minor patch:
* Fix build of ldplayer on OS X.  Since the CUSTOM sound module no longer exists, I 
arbitrarily changed it to WAVE, as ar gets upset if it has no input files.  I also 
removed the -all_load flag for ldplayer from the main makefile as it upsets the linker 
on OS X.

* Fix build for PPC64 Linux.  (This slightly messes up static branch prediction hints 
on OS X and AIX, but OS X for PPC64 is dead, and no- one builds MAME for AIX, and 
it will still build, anyway.)

* Paramaterise the arguments to check for NULL in the ATTR_NONNULL macro rather than 
just checking the first argument.  This requires compiler support for C99 variadic 
macros (MSVC2005 and GCC4 have this AFAIK).

Vas
2009-03-19 06:15:14 +00:00

474 lines
13 KiB
C

/***************************************************************************
eigccppc.h
PowerPC (32 and 64-bit) inline implementations for GCC compilers. This
code is automatically included if appropriate by eminline.h.
Copyright Nicola Salmoria and the MAME Team.
Visit http://mamedev.org for licensing and usage restrictions.
***************************************************************************/
#ifndef __EIGCCPPC__
#define __EIGCCPPC__
/***************************************************************************
INLINE MATH FUNCTIONS
***************************************************************************/
/*-------------------------------------------------
mul_32x32 - perform a signed 32 bit x 32 bit
multiply and return the full 64 bit result
-------------------------------------------------*/
/* GCC can do a good job of this. */
/*-------------------------------------------------
mulu_32x32 - perform an unsigned 32 bit x
32 bit multiply and return the full 64 bit
result
-------------------------------------------------*/
/* GCC can do a good job of this */
/*-------------------------------------------------
mul_32x32_hi - perform a signed 32 bit x 32 bit
multiply and return the upper 32 bits of the
result
-------------------------------------------------*/
#define mul_32x32_hi _mul_32x32_hi
INLINE INT32 ATTR_CONST ATTR_FORCE_INLINE
_mul_32x32_hi(INT32 val1, INT32 val2)
{
register INT32 result;
__asm__ (
" mulhw %[result], %[val1], %[val2] \n"
: [result] "=r" (result)
: [val1] "%r" (val1)
, [val2] "r" (val2)
);
return result;
}
/*-------------------------------------------------
mulu_32x32_hi - perform an unsigned 32 bit x
32 bit multiply and return the upper 32 bits
of the result
-------------------------------------------------*/
#define mulu_32x32_hi _mulu_32x32_hi
INLINE UINT32 ATTR_CONST ATTR_FORCE_INLINE
_mulu_32x32_hi(UINT32 val1, UINT32 val2)
{
register UINT32 result;
__asm__ (
" mulhwu %[result], %[val1], %[val2] \n"
: [result] "=r" (result)
: [val1] "%r" (val1)
, [val2] "r" (val2)
);
return result;
}
/*-------------------------------------------------
mul_32x32_shift - perform a signed 32 bit x
32 bit multiply and shift the result by the
given number of bits before truncating the
result to 32 bits
-------------------------------------------------*/
#if !defined(__ppc64__) && !defined(__PPC64__) && !defined(_ARCH_PPC64)
#define mul_32x32_shift _mul_32x32_shift
INLINE INT32 ATTR_CONST ATTR_FORCE_INLINE
_mul_32x32_shift(INT32 val1, INT32 val2, UINT8 shift)
{
register INT32 result;
/* Valid for (0 <= shift <= 32) */
__asm__ (
" mullw %[result], %[val1], %[val2] \n"
" mulhw %[val1], %[val1], %[val2] \n"
" srw %[result], %[result], %[shift] \n"
" subfic %[shift], %[shift], 0x20 \n"
" slw %[val1], %[val1], %[shift] \n"
" or %[result], %[result], %[val1] \n"
: [result] "=&r" (result)
, [shift] "+r" (shift)
, [val1] "+r" (val1)
: [val2] "r" (val2)
: "xer"
);
return result;
}
#endif
/*-------------------------------------------------
mulu_32x32_shift - perform an unsigned 32 bit x
32 bit multiply and shift the result by the
given number of bits before truncating the
result to 32 bits
-------------------------------------------------*/
#if !defined(__ppc64__) && !defined(__PPC64__) && !defined(_ARCH_PPC64)
#define mulu_32x32_shift _mulu_32x32_shift
INLINE UINT32 ATTR_CONST ATTR_FORCE_INLINE
_mulu_32x32_shift(UINT32 val1, UINT32 val2, UINT8 shift)
{
register UINT32 result;
/* Valid for (0 <= shift <= 32) */
__asm__ (
" mullw %[result], %[val1], %[val2] \n"
" mulhwu %[val1], %[val1], %[val2] \n"
" srw %[result], %[result], %[shift] \n"
" subfic %[shift], %[shift], 0x20 \n"
" slw %[val1], %[val1], %[shift] \n"
" or %[result], %[result], %[val1] \n"
: [result] "=&r" (result)
, [shift] "+r" (shift)
, [val1] "+r" (val1)
: [val2] "r" (val2)
: "xer"
);
return result;
}
#endif
/*-------------------------------------------------
div_64x32 - perform a signed 64 bit x 32 bit
divide and return the 32 bit quotient
-------------------------------------------------*/
/* TBD */
/*-------------------------------------------------
divu_64x32 - perform an unsigned 64 bit x 32 bit
divide and return the 32 bit quotient
-------------------------------------------------*/
/* TBD */
/*-------------------------------------------------
div_64x32_rem - perform a signed 64 bit x 32
bit divide and return the 32 bit quotient and
32 bit remainder
-------------------------------------------------*/
/* TBD */
/*-------------------------------------------------
divu_64x32_rem - perform an unsigned 64 bit x
32 bit divide and return the 32 bit quotient
and 32 bit remainder
-------------------------------------------------*/
/* TBD */
/*-------------------------------------------------
div_32x32_shift - perform a signed divide of
two 32 bit values, shifting the first before
division, and returning the 32 bit quotient
-------------------------------------------------*/
/* TBD */
/*-------------------------------------------------
divu_32x32_shift - perform an unsigned divide of
two 32 bit values, shifting the first before
division, and returning the 32 bit quotient
-------------------------------------------------*/
/* TBD */
/*-------------------------------------------------
mod_64x32 - perform a signed 64 bit x 32 bit
divide and return the 32 bit remainder
-------------------------------------------------*/
/* TBD */
/*-------------------------------------------------
modu_64x32 - perform an unsigned 64 bit x 32 bit
divide and return the 32 bit remainder
-------------------------------------------------*/
/* TBD */
/*-------------------------------------------------
recip_approx - compute an approximate floating
point reciprocal
-------------------------------------------------*/
#define recip_approx _recip_approx
INLINE float ATTR_CONST ATTR_FORCE_INLINE
_recip_approx(float value)
{
register float result;
__asm__ (
" fres %[result], %[value] \n"
: [result] "=f" (result)
: [value] "f" (value)
);
return result;
}
/***************************************************************************
INLINE BIT MANIPULATION FUNCTIONS
***************************************************************************/
/*-------------------------------------------------
count_leading_zeros - return the number of
leading zero bits in a 32-bit value
-------------------------------------------------*/
#define count_leading_zeros _count_leading_zeros
INLINE UINT8 ATTR_CONST ATTR_FORCE_INLINE
_count_leading_zeros(UINT32 value)
{
register UINT32 result;
__asm__ (
" cntlzw %[result], %[value] \n"
: [result] "=r" (result) /* result can be in any register */
: [value] "r" (value) /* 'value' can be in any register */
);
return result;
}
/*-------------------------------------------------
count_leading_ones - return the number of
leading one bits in a 32-bit value
-------------------------------------------------*/
#define count_leading_ones _count_leading_ones
INLINE UINT8 ATTR_CONST ATTR_FORCE_INLINE
_count_leading_ones(UINT32 value)
{
register UINT32 result;
__asm__ (
" not %[result], %[value] \n"
" cntlzw %[result], %[result] \n"
: [result] "=r" (result) /* result can be in any register */
: [value] "r" (value) /* 'value' can be in any register */
);
return result;
}
/***************************************************************************
INLINE SYNCHRONIZATION FUNCTIONS
***************************************************************************/
/*-------------------------------------------------
compare_exchange32 - compare the 'compare'
value against the memory at 'ptr'; if equal,
swap in the 'exchange' value. Regardless,
return the previous value at 'ptr'.
-------------------------------------------------*/
#define compare_exchange32 _compare_exchange32
INLINE INT32 ATTR_NONNULL(1) ATTR_FORCE_INLINE
_compare_exchange32(INT32 volatile *ptr, INT32 compare, INT32 exchange)
{
register INT32 result;
__asm__ __volatile__ (
"1: lwarx %[result], 0, %[ptr] \n"
" cmpw %[compare], %[result] \n"
" bne 2f \n"
" sync \n"
" stwcx. %[exchange], 0, %[ptr] \n"
" bne- 1b \n"
"2: "
: [dummy] "+m" (*ptr) /* Lets GCC know that *ptr will be read/written in case it's not marked volatile */
, [result] "=&r" (result)
: [ptr] "r" (ptr)
, [exchange] "r" (exchange)
, [compare] "r" (compare)
: "cr0"
);
return result;
}
/*-------------------------------------------------
compare_exchange64 - compare the 'compare'
value against the memory at 'ptr'; if equal,
swap in the 'exchange' value. Regardless,
return the previous value at 'ptr'.
-------------------------------------------------*/
#if defined(__ppc64__) || defined(__PPC64__)
#define compare_exchange64 _compare_exchange64
INLINE INT64 ATTR_NONNULL(1) ATTR_FORCE_INLINE
_compare_exchange64(INT64 volatile *ptr, INT64 compare, INT64 exchange)
{
register INT64 result;
__asm__ __volatile__ (
"1: ldarx %[result], 0, %[ptr] \n"
" cmpd %[compare], %[result] \n"
" bne 2f \n"
" stdcx. %[exchange], 0, %[ptr] \n"
" bne- 1b \n"
"2: "
: [dummy] "+m" (*ptr) /* Lets GCC know that *ptr will be read/written in case it's not marked volatile */
, [result] "=&r" (result)
: [ptr] "r" (ptr)
, [exchange] "r" (exchange)
, [compare] "r" (compare)
: "cr0"
);
return result;
}
#endif
/*-------------------------------------------------
atomic_exchange32 - atomically exchange the
exchange value with the memory at 'ptr',
returning the original value.
-------------------------------------------------*/
#define atomic_exchange32 _atomic_exchange32
INLINE INT32 ATTR_NONNULL(1) ATTR_FORCE_INLINE
_atomic_exchange32(INT32 volatile *ptr, INT32 exchange)
{
register INT32 result;
__asm__ __volatile__ (
"1: lwarx %[result], 0, %[ptr] \n"
" sync \n"
" stwcx. %[exchange], 0, %[ptr] \n"
" bne- 1b \n"
: [dummy] "+m" (*ptr) /* Lets GCC know that *ptr will be read/written in case it's not marked volatile */
, [result] "=&r" (result)
: [ptr] "r" (ptr)
, [exchange] "r" (exchange)
: "cr0"
);
return result;
}
/*-------------------------------------------------
atomic_add32 - atomically add the delta value
to the memory at 'ptr', returning the final
result.
-------------------------------------------------*/
#define atomic_add32 _atomic_add32
INLINE INT32 ATTR_NONNULL(1) ATTR_FORCE_INLINE
_atomic_add32(INT32 volatile *ptr, INT32 delta)
{
register INT32 result;
__asm__ __volatile__ (
"1: lwarx %[result], 0, %[ptr] \n"
" add %[result], %[result], %[delta] \n"
" sync \n"
" stwcx. %[result], 0, %[ptr] \n"
" bne- 1b \n"
: [dummy] "+m" (*ptr) /* Lets GCC know that *ptr will be read/written in case it's not marked volatile */
, [result] "=&b" (result)
: [ptr] "r" (ptr)
, [delta] "r" (delta)
: "cr0"
);
return result;
}
/*-------------------------------------------------
atomic_increment32 - atomically increment the
32-bit value in memory at 'ptr', returning the
final result.
-------------------------------------------------*/
#define atomic_increment32 _atomic_increment32
INLINE INT32 ATTR_NONNULL(1) ATTR_FORCE_INLINE
_atomic_increment32(INT32 volatile *ptr)
{
register INT32 result;
__asm__ __volatile__ (
"1: lwarx %[result], 0, %[ptr] \n"
" addi %[result], %[result], 1 \n"
" sync \n"
" stwcx. %[result], 0, %[ptr] \n"
" bne- 1b \n"
: [dummy] "+m" (*ptr) /* Lets GCC know that *ptr will be read/written in case it's not marked volatile */
, [result] "=&b" (result)
: [ptr] "r" (ptr)
: "cr0"
);
return result;
}
/*-------------------------------------------------
atomic_decrement32 - atomically decrement the
32-bit value in memory at 'ptr', returning the
final result.
-------------------------------------------------*/
#define atomic_decrement32 _atomic_decrement32
INLINE INT32 ATTR_NONNULL(1) ATTR_FORCE_INLINE
_atomic_decrement32(INT32 volatile *ptr)
{
register INT32 result;
__asm__ __volatile__ (
"1: lwarx %[result], 0, %[ptr] \n"
" addi %[result], %[result], -1 \n"
" sync \n"
" stwcx. %[result], 0, %[ptr] \n"
" bne- 1b \n"
: [dummy] "+m" (*ptr) /* Lets GCC know that *ptr will be read/written in case it's not marked volatile */
, [result] "=&b" (result)
: [ptr] "r" (ptr)
: "cr0"
);
return result;
}
#endif /* __EIGCCPPC__ */