#ifndef TATOMICVAR_H
#define TATOMICVAR_H
#undef DVAPI
#undef DVVAR
#ifdef TNZCORE_EXPORTS
#define DVAPI DV_EXPORT_API
#define DVVAR DV_EXPORT_VAR
#else
#define DVAPI DV_IMPORT_API
#define DVVAR DV_IMPORT_VAR
#endif
#ifdef _WIN32
#include <windows.h>
#elif defined(__sgi)
#include <sys/atomic_ops.h>
#elif defined(LINUX)
// #include <asm/atomic.h>
// it's broken, either include the kernel header
// /usr/src/linux/include/asm/atomic.h
// or copy it here
#include "/usr/src/linux-2.4/include/asm/atomic.h"
#elif defined(powerpc)
// from linux-2.4.20-20.9/include/linux/asm-ppc
typedef struct {
volatile int counter;
} atomic_t;
#define ATOMIC_INIT(i) \
{ \
(i) \
}
#define atomic_read(v) ((v)->counter)
#define atomic_set(v, i) (((v)->counter) = (i))
extern void atomic_clear_mask(unsigned long mask, unsigned long *addr);
extern void atomic_set_mask(unsigned long mask, unsigned long *addr);
// we define it as we want to run on SMP dual processors...
#define CONFIG_SMP
#ifdef CONFIG_SMP
#define SMP_ISYNC "\n\tisync"
#else
#define SMP_ISYNC
#endif
static __inline__ void atomic_add(int a, atomic_t *v)
{
int t;
__asm__ __volatile__(
"1: lwarx %0,0,%3 \n\
add %0,%2,%0\n\
stwcx. %0,0,%3\n\
bne- 1b"
: "=&r"(t), "=m"(v->counter)
: "r"(a), "r"(&v->counter), "m"(v->counter)
: "cc");
}
static __inline__ int atomic_add_return(int a, atomic_t *v)
{
int t;
__asm__ __volatile__(
"1: lwarx %0,0,%2 \n\
add %0,%1,%0\n\
stwcx. %0,0,%2\n\
bne- 1b" SMP_ISYNC
: "=&r"(t)
: "r"(a), "r"(&v->counter)
: "cc", "memory");
return t;
}
static __inline__ void atomic_sub(int a, atomic_t *v)
{
int t;
__asm__ __volatile__(
"1: lwarx %0,0,%3 \n\
subf %0,%2,%0\n\
stwcx. %0,0,%3\n\
bne- 1b"
: "=&r"(t), "=m"(v->counter)
: "r"(a), "r"(&v->counter), "m"(v->counter)
: "cc");
}
static __inline__ int atomic_sub_return(int a, atomic_t *v)
{
int t;
__asm__ __volatile__(
"1: lwarx %0,0,%2 \n\
subf %0,%1,%0\n\
stwcx. %0,0,%2\n\
bne- 1b" SMP_ISYNC
: "=&r"(t)
: "r"(a), "r"(&v->counter)
: "cc", "memory");
return t;
}
static __inline__ void atomic_inc(atomic_t *v)
{
int t;
__asm__ __volatile__(
"1: lwarx %0,0,%2 \n\
addic %0,%0,1\n\
stwcx. %0,0,%2\n\
bne- 1b"
: "=&r"(t), "=m"(v->counter)
: "r"(&v->counter), "m"(v->counter)
: "cc");
}
static __inline__ int atomic_inc_return(atomic_t *v)
{
int t;
__asm__ __volatile__(
"1: lwarx %0,0,%1 \n\
addic %0,%0,1\n\
stwcx. %0,0,%1\n\
bne- 1b" SMP_ISYNC
: "=&r"(t)
: "r"(&v->counter)
: "cc", "memory");
return t;
}
static __inline__ void atomic_dec(atomic_t *v)
{
int t;
__asm__ __volatile__(
"1: lwarx %0,0,%2 \n\
addic %0,%0,-1\n\
stwcx. %0,0,%2\n\
bne- 1b"
: "=&r"(t), "=m"(v->counter)
: "r"(&v->counter), "m"(v->counter)
: "cc");
}
static __inline__ int atomic_dec_return(atomic_t *v)
{
int t;
__asm__ __volatile__(
"1: lwarx %0,0,%1 \n\
addic %0,%0,-1\n\
stwcx. %0,0,%1\n\
bne- 1b" SMP_ISYNC
: "=&r"(t)
: "r"(&v->counter)
: "cc", "memory");
return t;
}
#elif defined(i386)
#include <atomic>
typedef std::atomic<int> atomic_t;
static __inline__ void atomic_set(atomic_t *v, const int value)
{
v->store(value);
}
static __inline__ int atomic_inc_return(atomic_t *v)
{
return v->fetch_add(1) + 1; // post increment atomic
}
static __inline__ int atomic_dec_return(atomic_t *v)
{
return v->fetch_sub(1) - 1; // post decriment atomic
}
static __inline__ int atomic_read(const atomic_t *v)
{
return v->load();
}
static __inline__ int atomic_add(int num, const atomic_t *v)
{
return const_cast<atomic_t *>(v)->fetch_add(num) + num; /* なんで const つけた? */
}
#else
@ @PLATFORM NOT SUPPORTED !@ @
#endif
/*! Platform specific
Provides class with increment & decrement absolutely done in interlocked way
*/
class DVAPI TAtomicVar
{
public:
#if defined(LINUX) || defined(MACOSX)
TAtomicVar()
{
atomic_set(&m_var, 0);
}
#else
TAtomicVar() : m_var(0)
{
}
#endif
long operator++()
{
#ifdef _WIN32
return InterlockedIncrement(&m_var);
#elif defined(__sgi)
return ++m_var;
#elif defined(LINUX)
// atomic_inc(&m_var);
// return atomic_read(&m_var);
// this is broken as it can return a value != from ++m_var
atomic_inc(&m_var);
return atomic_read(&m_var);
#elif defined(MACOSX)
return atomic_inc_return(&m_var);
#endif
}
long operator+=(long value)
{
#ifdef _WIN32
InterlockedExchangeAdd(&m_var, value);
return m_var;
#elif defined(__sgi)
m_var += value;
return m_var;
#elif defined(LINUX)
// atomic_inc(&m_var);
// return atomic_read(&m_var);
// this is broken as it can return a value != from ++m_var
assert(false);
return m_var;
#elif defined(MACOSX)
atomic_add(value, &m_var);
return atomic_read(&m_var);
#endif
}
long operator--()
{
#ifdef _WIN32
return InterlockedDecrement(&m_var);
#elif defined(__sgi)
return --m_var;
#elif defined(LINUX)
// atomic_dec(&m_var);
// return atomic_read(&m_var);
// broken as above...
atomic_dec(&m_var);
return atomic_read(&m_var);
#elif defined(MACOSX)
// atomic_dec(&m_var);
// return atomic_read(&m_var);
// broken as above...
return atomic_dec_return(&m_var);
#endif
}
bool operator<=(const long &rhs)
{
#if defined(LINUX) || defined(MACOSX)
return atomic_read(&m_var) <= rhs;
#else
return m_var <= rhs;
#endif
};
operator long() const
{
#if defined(LINUX) || defined(MACOSX)
return atomic_read(&m_var);
#else
return m_var;
#endif
};
#ifdef _WIN32
long m_var;
#elif defined(__sgi)
long m_var;
#elif defined(LINUX) || defined(MACOSX)
atomic_t m_var;
#endif
#if !defined(LINUX) || defined(LINUX) && (__GNUC__ == 3) && (__GNUC_MINOR__ > 1)
private: // to avoid well known bug in gcc3 ... fixed in later versions..
#endif
TAtomicVar &operator=(const TAtomicVar &); //not implemented
TAtomicVar(const TAtomicVar &v); //not implemented
};
#endif