diff options
author | Chris Robinson <[email protected]> | 2014-07-21 23:14:48 -0700 |
---|---|---|
committer | Chris Robinson <[email protected]> | 2014-07-22 00:20:28 -0700 |
commit | 5a339a2a5b12545c105a2a3dcfb1d8e466b0381f (patch) | |
tree | 95dab694ebe14d41e3e775dc651d07ee26cbe44a /include | |
parent | 7b41ed7ec4bfb7e8ac3daef3eebc6831b96a8fa4 (diff) |
Add macros for generic atomic functionality
Diffstat (limited to 'include')
-rw-r--r-- | include/atomic.h | 181 |
1 files changed, 145 insertions, 36 deletions
diff --git a/include/atomic.h b/include/atomic.h index 4adb7a94..174aebe9 100644 --- a/include/atomic.h +++ b/include/atomic.h @@ -40,6 +40,33 @@ inline int CompExchangeInt(volatile int *ptr, int oldval, int newval) inline void *CompExchangePtr(XchgPtr *ptr, void *oldval, void *newval) { return __sync_val_compare_and_swap(ptr, oldval, newval); } + +#define ATOMIC(T) struct { T volatile value; } + +#define ATOMIC_INIT_STATIC(_newval) {(_newval)} + +#define ATOMIC_LOAD_UNSAFE(_val) ((_val).value) +#define ATOMIC_STORE_UNSAFE(_val, _newval) do { \ + (_val).value = (_newval); \ +} while(0) + +#define ATOMIC_LOAD(_val) (__sync_synchronize(),(_val).value) +#define ATOMIC_STORE(_val, _newval) do { \ + (_val).value = (_newval); \ + __sync_synchronize(); \ +} while(0) + +#define ATOMIC_EXCHANGE(T, _val, _newval) __extension__({ \ + static_assert(sizeof(T)==sizeof((_val).value), "Type "#T" has incorrect size!"); \ + T _r = __sync_lock_test_and_set(&(_val).value, (_newval)); \ + _r; \ +}) +#define ATOMIC_COMPARE_EXCHANGE(T, _val, _oldval, _newval) __extension__({ \ + static_assert(sizeof(T)==sizeof((_val).value), "Type "#T" has incorrect size!"); \ + T _r = __sync_val_compare_and_swap(&(_val).value, (_oldval), (_newval)); \ + _r; \ +}) + #elif defined(__GNUC__) && (defined(__i386__) || defined(__x86_64__)) inline uint xaddl(volatile uint *dest, int incr) @@ -79,54 +106,91 @@ inline uint CompExchangeRef(volatile RefCount *ptr, uint oldval, uint newval) return ret; } +#define EXCHANGE(S, ret, dest, newval) __asm__ __volatile__( \ + "lock; xchg"S" %0,(%1)" \ + : "=r" (ret) \ + : "r" (dest), "0" (newval) \ + : "memory" \ +) +#define COMP_EXCHANGE(S, ret, dest, oldval, newval) __asm__ __volatile__( \ + "lock; cmpxchg"S" %2,(%1)" \ + : "=a" (ret) \ + : "r" (dest), "r" (newval), "0" (oldval) \ + : "memory" \ +) + + inline int ExchangeInt(volatile int *dest, int newval) { int ret; - __asm__ __volatile__("lock; xchgl %0,(%1)" - : "=r" (ret) - : "r" (dest), "0" (newval) - : "memory"); + EXCHANGE("l", ret, dest, newval); return ret; } +inline int CompExchangeInt(volatile int *dest, int oldval, int newval) +{ + int ret; + COMP_EXCHANGE("l", ret, dest, oldval, newval); + return ret; +} + +#ifdef __i386__ inline void *ExchangePtr(XchgPtr *dest, void *newval) { void *ret; - __asm__ __volatile__( -#ifdef __i386__ - "lock; xchgl %0,(%1)" -#else - "lock; xchgq %0,(%1)" -#endif - : "=r" (ret) - : "r" (dest), "0" (newval) - : "memory" - ); + EXCHANGE("l", ret, dest, newval); return ret; } -inline int CompExchangeInt(volatile int *dest, int oldval, int newval) +inline void *CompExchangePtr(XchgPtr *dest, void *oldval, void *newval) { - int ret; - __asm__ __volatile__("lock; cmpxchgl %2,(%1)" - : "=a" (ret) - : "r" (dest), "r" (newval), "0" (oldval) - : "memory"); + void *ret; + COMP_EXCHANGE("l", ret, dest, oldval, newval); + return ret; +} +#else +inline void *ExchangePtr(XchgPtr *dest, void *newval) +{ + void *ret; + EXCHANGE("q", ret, dest, newval); return ret; } inline void *CompExchangePtr(XchgPtr *dest, void *oldval, void *newval) { void *ret; - __asm__ __volatile__( -#ifdef __i386__ - "lock; cmpxchgl %2,(%1)" -#else - "lock; cmpxchgq %2,(%1)" -#endif - : "=a" (ret) - : "r" (dest), "r" (newval), "0" (oldval) - : "memory" - ); + COMP_EXCHANGE("q", ret, dest, oldval, newval); return ret; } +#endif + + +#define ATOMIC(T) struct { T volatile value; } + +#define ATOMIC_INIT_STATIC(_newval) {(_newval)} + +#define ATOMIC_LOAD_UNSAFE(_val) ((_val).value) +#define ATOMIC_STORE_UNSAFE(_val, _newval) do { \ + (_val).value = (_newval); \ +} while(0) + +#define ATOMIC_LOAD(_val) (__asm__ __volatile__("" ::: "memory"),(_val).value) +#define ATOMIC_STORE(_val, _newval) do { \ + (_val).value = (_newval); \ + __asm__ __volatile__("" ::: "memory"); \ +} while(0) + +#define ATOMIC_EXCHANGE(T, _val, _newval) __extension__({ \ + T _r; \ + static_assert(sizeof(T)==sizeof((_val).value), "Type "#T" has incorrect size!"); \ + if(sizeof(T) == 4) EXCHANGE("l", _r, &(_val).value, (_newval)); \ + else if(sizeof(T) == 8) EXCHANGE("q", _r, &(_val).value, (_newval)); \ + _r; \ +}) +#define ATOMIC_COMPARE_EXCHANGE(T, _val, _oldval, _newval) __extension__({ \ + T _r; \ + static_assert(sizeof(T)==sizeof((_val).value), "Type "#T" has incorrect size!"); \ + if(sizeof(T) == 4) COMP_EXCHANGE("l", _r, &(_val).value, (_oldval), (_newval)); \ + else if(sizeof(T) == 8) COMP_EXCHANGE("q", _r, &(_val).value, (_oldval), (_newval)); \ + _r; \ +}) #elif defined(_WIN32) @@ -172,7 +236,7 @@ inline uint CompExchangeRef(volatile RefCount *ptr, uint oldval, uint newval) return InterlockedCompareExchange(u.l, newval, oldval); } -inline int ExchangeInt(volatile int *ptr, int newval) +inline int ExchangeInt32(volatile int *ptr, int newval) { union { volatile int *i; @@ -180,11 +244,7 @@ inline int ExchangeInt(volatile int *ptr, int newval) } u = { ptr }; return InterlockedExchange(u.l, newval); } -inline void *ExchangePtr(XchgPtr *ptr, void *newval) -{ - return InterlockedExchangePointer(ptr, newval); -} -inline int CompExchangeInt(volatile int *ptr, int oldval, int newval) +inline int CompExchangeInt32(volatile int *ptr, int oldval, int newval) { union { volatile int *i; @@ -192,11 +252,60 @@ inline int CompExchangeInt(volatile int *ptr, int oldval, int newval) } u = { ptr }; return InterlockedCompareExchange(u.l, newval, oldval); } +inline __int64 ExchangeInt64(volatile __int64 *ptr, __int64 newval) +{ + union { + volatile __int64 *i; + volatile LONGLONG *l; + } u = { ptr }; + return InterlockedExchange64(u.l, newval); +} +inline __int64 CompExchangeInt64(volatile __int64 *ptr, __int64 oldval, __int64 newval) +{ + union { + volatile __int64 *i; + volatile LONGLONG *l; + } u = { ptr }; + return InterlockedCompareExchange64(u.l, newval, oldval); +} + +inline int ExchangeInt(volatile int *ptr, int newval) +{ return ExchangeInt32(ptr, newval); } +inline int CompExchangeInt(volatile int *ptr, int oldval, int newval) +{ return CompExchangeInt32(ptr, oldval, newval); } + +inline void *ExchangePtr(XchgPtr *ptr, void *newval) +{ + return InterlockedExchangePointer(ptr, newval); +} inline void *CompExchangePtr(XchgPtr *ptr, void *oldval, void *newval) { return InterlockedCompareExchangePointer(ptr, newval, oldval); } + +#define ATOMIC(T) struct { T volatile value; } + +#define ATOMIC_INIT_STATIC(_newval) {(_newval)} + +#define ATOMIC_LOAD_UNSAFE(_val) ((_val).value) +#define ATOMIC_STORE_UNSAFE(_val, _newval) do { \ + (_val).value = (_newval); \ +} while(0) + +#define ATOMIC_LOAD(_val) (_ReadBarrier(),(_val).value) +#define ATOMIC_STORE(_val, _newval) do { \ + (_val).value = (_newval); \ + _WriteBarrier(); \ +} while(0) + +int _al_invalid_atomic_size(); /* not defined */ + +#define ATOMIC_FUNC_SELECT(T, C, F32, F64) ((sizeof(T) == 4) ? (C)F32 : ((sizeof(T) == 8) ? (C)F64 : (C)_al_invalid_atomic_size)) + +#define ATOMIC_EXCHANGE(T, _val, _newval) (ATOMIC_FUNC_SELECT(T, T(*)(volatile T*,T), ExchangeInt32, ExchangeInt64)(&(_val).value, (_newval))) +#define ATOMIC_COMPARE_EXCHANGE(T, _val, _oldval, _newval) (ATOMIC_FUNC_SELECT(T, T(*)(volatile T*,T,T), CompExchangeInt32, CompExchangeInt64)(&(_val).value, (_oldval), (_newval))) + #else #error "No atomic functions available on this platform!" #endif |