/* * Copyright (c) 2010 Kungliga Tekniska Högskolan * (Royal Institute of Technology, Stockholm, Sweden). * All rights reserved. * * Portions Copyright (c) 2010 Apple Inc. All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * * 3. Neither the name of the Institute nor the names of its contributors * may be used to endorse or promote products derived from this software * without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE INSTITUTE AND CONTRIBUTORS ``AS IS'' AND * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE * ARE DISCLAIMED. IN NO EVENT SHALL THE INSTITUTE OR CONTRIBUTORS BE LIABLE * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. */ #ifndef HEIM_BASE_ATOMICS_H #define HEIM_BASE_ATOMICS_H 1 #include /* * Atomic operations * * (#define HEIM_BASE_ATOMICS_FALLBACK to test fallbacks.) */ #if !defined(HEIM_BASE_ATOMICS_FALLBACK) && defined(HAVE_STDATOMIC_H) #include #define heim_base_atomic_init(t, v) atomic_init(t, v) #define heim_base_atomic_load(x) atomic_load((x)) #define heim_base_atomic_store(t, v) atomic_store((t), (v)) #define heim_base_atomic(T) _Atomic(T) #define heim_base_atomic_inc_32(x) (atomic_fetch_add((x), 1) + 1) #define heim_base_atomic_dec_32(x) (atomic_fetch_sub((x), 1) - 1) #define heim_base_atomic_inc_64(x) (atomic_fetch_add((x), 1) + 1) #define heim_base_atomic_dec_64(x) (atomic_fetch_sub((x), 1) - 1) #define heim_base_exchange_pointer(t,v) atomic_exchange((t), (v)) #define heim_base_exchange_32(t,v) atomic_exchange((t), (v)) #define heim_base_exchange_64(t,v) atomic_exchange((t), (v)) /* * 's and AIX's CAS functions take a pointer to an expected value * and return a boolean, setting the pointed-to variable to the old value of * the target. * * Other CAS functions, like GCC's, Solaris'/Illumos', and Windows', return the * old value and don't take a pointer to an expected value. * * We implement the latter semantics. */ static inline void * heim_base_cas_pointer_(heim_base_atomic(void *)*t, void *e, void *d) { return atomic_compare_exchange_strong(t, &e, d), e; } static inline uint32_t heim_base_cas_32_(heim_base_atomic(uint32_t)*t, uint32_t e, uint32_t d) { return atomic_compare_exchange_strong(t, &e, d), e; } static inline uint64_t heim_base_cas_64_(heim_base_atomic(uint64_t)*t, uint64_t e, uint64_t d) { return atomic_compare_exchange_strong(t, &e, d), e; } #define heim_base_cas_pointer(t,e,d) heim_base_cas_pointer_((t), (e), (d)) #define heim_base_cas_32(t,e,d) heim_base_cas_32_((t), (e), (d)) #define heim_base_cas_64(t,e,d) heim_base_cas_64_((t), (e), (d)) #elif !defined(HEIM_BASE_ATOMICS_FALLBACK) && defined(__GNUC__) && defined(HAVE___SYNC_ADD_AND_FETCH) #define heim_base_atomic_barrier() __sync_synchronize() #define heim_base_atomic_inc_32(x) __sync_add_and_fetch((x), 1) #define heim_base_atomic_dec_32(x) __sync_sub_and_fetch((x), 1) #define heim_base_atomic_inc_64(x) __sync_add_and_fetch((x), 1) #define heim_base_atomic_dec_64(x) __sync_sub_and_fetch((x), 1) #ifndef __has_builtin #define __has_builtin(x) 0 #endif #if __has_builtin(__sync_swap) #define heim_base_exchange_pointer(t,v) __sync_swap((t), (v)) #else /* FIXME: some targets may only write the value 1 into *t */ #define heim_base_exchange_pointer(t,v) __sync_lock_test_and_set((t), (v)) #endif #define heim_base_exchange_32(t,v) heim_base_exchange_pointer((t), (v)) #define heim_base_exchange_64(t,v) heim_base_exchange_pointer((t), (v)) #define heim_base_cas_pointer(t,e,d) __sync_val_compare_and_swap((t), (e), (d)) #define heim_base_cas_32(t,e,d) __sync_val_compare_and_swap((t), (e), (d)) #define heim_base_cas_64(t,e,d) __sync_val_compare_and_swap((t), (e), (d)) #elif !defined(HEIM_BASE_ATOMICS_FALLBACK) && defined(__sun) #include #include static inline void __heim_base_atomic_barrier(void) { __machine_rw_barrier(); } #define heim_base_atomic_barrier() __heim_base_atomic_barrier() #define heim_base_atomic(T) volatile T #define heim_base_atomic_inc_32(x) atomic_inc_32_nv((x)) #define heim_base_atomic_dec_32(x) atomic_dec_32_nv((x)) #define heim_base_atomic_inc_64(x) atomic_inc_64_nv((x)) #define heim_base_atomic_dec_64(x) atomic_dec_64_nv((x)) #define heim_base_exchange_pointer(t,v) atomic_swap_ptr((t), (void *)(v)) #define heim_base_exchange_32(t,v) atomic_swap_32((t), (v)) #define heim_base_exchange_64(t,v) atomic_swap_64((t), (v)) #define heim_base_cas_pointer(t,e,d) atomic_cas_ptr((t), (e), (d)) #define heim_base_cas_32(t,e,d) atomic_cas_32((t), (e), (d)) #define heim_base_cas_64(t,e,d) atomic_cas_64((t), (e), (d)) #elif !defined(HEIM_BASE_ATOMICS_FALLBACK) && defined(_AIX) #include #define heim_base_atomic_barrier() __isync() #define heim_base_atomic_inc_32(x) (fetch_and_add((atomic_p)(x), 1) + 1) #define heim_base_atomic_dec_32(x) (fetch_and_add((atomic_p)(x), -1) - 1) #define heim_base_atomic_inc_64(x) (fetch_and_addlp((atomic_l)(x), 1) + 1) #define heim_base_atomic_dec_64(x) (fetch_and_addlp((atomic_l)(x), -1) - 1) static inline void * heim_base_exchange_pointer(void *p, void *newval) { void *val = *(void **)p; while (!compare_and_swaplp((atomic_l)p, (long *)&val, (long)newval)) ; return val; } static inline uint32_t heim_base_exchange_32(uint32_t *p, uint32_t newval) { uint32_t val = *p; while (!compare_and_swap((atomic_p)p, (int *)&val, (int)newval)) ; return val; } static inline uint64_t heim_base_exchange_64(uint64_t *p, uint64_t newval) { uint64_t val = *p; while (!compare_and_swaplp((atomic_l)p, (long *)&val, (long)newval)) ; return val; } static inline void * heim_base_cas_pointer_(heim_base_atomic(void *)*t, void *e, void *d) { return compare_and_swaplp((atomic_l)t, &e, d), e; } static inline uint32_t heim_base_cas_32_(heim_base_atomic(uint32_t)*t, uint32_t e, uint32_t d) { return compare_and_swap((atomic_p)t, &e, d), e; } static inline uint64_t heim_base_cas_64_(heim_base_atomic(uint64_t)*t, uint64_t e, uint64_t d) { return compare_and_swaplp((atomic_l)t, &e, d), e; } #define heim_base_cas_pointer(t,e,d) heim_base_cas_pointer_((t), (e), (d)) #define heim_base_cas_32(t,e,d) heim_base_cas_32_((t), (e), (d)) #define heim_base_cas_64(t,e,d) heim_base_cas_64_((t), (e), (d)) #elif !defined(HEIM_BASE_ATOMICS_FALLBACK) && defined(_WIN32) #define heim_base_atomic_barrier() MemoryBarrier() #define heim_base_atomic_inc_32(x) InterlockedIncrement(x) #define heim_base_atomic_dec_32(x) InterlockedDecrement(x) #define heim_base_atomic_inc_64(x) InterlockedIncrement64(x) #define heim_base_atomic_dec_64(x) InterlockedDecrement64(x) #define heim_base_exchange_pointer(t,v) InterlockedExchangePointer((PVOID volatile *)(t), (PVOID)(v)) #define heim_base_exchange_32(t,v) ((ULONG)InterlockedExchange((LONG volatile *)(t), (LONG)(v))) #define heim_base_exchange_64(t,v) ((ULONG64)InterlockedExchange64((ULONG64 volatile *)(t), (LONG64)(v))) #define heim_base_cas_pointer(t,e,d) InterlockedCompareExchangePointer((PVOID volatile *)(t), (d), (e)) #define heim_base_cas_32(t,e,d) InterlockedCompareExchange ((LONG volatile *)(t), (d), (e)) #define heim_base_cas_64(t,e,d) InterlockedCompareExchange64((ULONG64 volatile *)(t), (d), (e)) #else #define heim_base_atomic(T) volatile T #define heim_base_atomic_barrier() #define heim_base_atomic_load(x) (*(x)) #define heim_base_atomic_init(t, v) do { (*(t) = (v)); } while (0) #define heim_base_atomic_store(t, v) do { (*(t) = (v)); } while (0) #include #define HEIM_BASE_NEED_ATOMIC_MUTEX 1 static inline uint32_t heim_base_atomic_inc_32(heim_base_atomic(uint32_t) *x) { uint32_t t; HEIMDAL_MUTEX_lock(heim_base_mutex()); t = ++(*x); HEIMDAL_MUTEX_unlock(heim_base_mutex()); return t; } static inline uint32_t heim_base_atomic_dec_32(heim_base_atomic(uint32_t) *x) { uint32_t t; HEIMDAL_MUTEX_lock(heim_base_mutex()); t = --(*x); HEIMDAL_MUTEX_unlock(heim_base_mutex()); return t; } static inline uint64_t heim_base_atomic_inc_64(heim_base_atomic(uint64_t) *x) { uint64_t t; HEIMDAL_MUTEX_lock(heim_base_mutex()); t = ++(*x); HEIMDAL_MUTEX_unlock(heim_base_mutex()); return t; } static inline uint64_t heim_base_atomic_dec_64(heim_base_atomic(uint64_t) *x) { uint64_t t; HEIMDAL_MUTEX_lock(heim_base_mutex()); t = --(*x); HEIMDAL_MUTEX_unlock(heim_base_mutex()); return t; } static inline void * heim_base_exchange_pointer(heim_base_atomic(void *)target, void *value) { void *old; HEIMDAL_MUTEX_lock(heim_base_mutex()); old = *(void **)target; *(void **)target = value; HEIMDAL_MUTEX_unlock(heim_base_mutex()); return old; } static inline uint32_t heim_base_exchange_32(heim_base_atomic(uint32_t) *target, uint32_t newval) { uint32_t old; HEIMDAL_MUTEX_lock(heim_base_mutex()); old = *target; *target = newval; HEIMDAL_MUTEX_unlock(heim_base_mutex()); return old; } static inline uint64_t heim_base_exchange_64(heim_base_atomic(uint64_t) *target, uint64_t newval) { uint64_t old; HEIMDAL_MUTEX_lock(heim_base_mutex()); old = *target; *target = newval; HEIMDAL_MUTEX_unlock(heim_base_mutex()); return old; } static inline void * heim_base_cas_pointer(heim_base_atomic(void *)target, void *expected, void *desired) { void *old; HEIMDAL_MUTEX_lock(heim_base_mutex()); if ((old = *(void **)target) == expected) *(void **)target = desired; HEIMDAL_MUTEX_unlock(heim_base_mutex()); return old; } static inline uint32_t heim_base_cas_32(heim_base_atomic(uint32_t) *target, uint32_t expected, uint32_t desired) { uint32_t old; HEIMDAL_MUTEX_lock(heim_base_mutex()); if ((old = *(uint32_t *)target) == expected) *target = desired; HEIMDAL_MUTEX_unlock(heim_base_mutex()); return old; } static inline uint64_t heim_base_cas_64(heim_base_atomic(uint64_t) *target, uint64_t expected,uint64_t desired) { uint64_t old; HEIMDAL_MUTEX_lock(heim_base_mutex()); if ((old = *(uint64_t *)target) == expected) *target = desired; HEIMDAL_MUTEX_unlock(heim_base_mutex()); return old; } #endif /* defined(__GNUC__) && defined(HAVE___SYNC_ADD_AND_FETCH) */ #ifndef heim_base_atomic #define heim_base_atomic(T) T #endif #ifndef heim_base_atomic_barrier static inline void heim_base_atomic_barrier(void) { return; } #endif #ifndef heim_base_atomic_load #define heim_base_atomic_load(x) (heim_base_atomic_barrier(), *(x)) #endif #ifndef heim_base_atomic_init #define heim_base_atomic_init(t, v) do { (*(t) = (v)); } while (0) #endif #ifndef heim_base_atomic_store #define heim_base_atomic_store(t, v) do { \ (*(t) = (v)); \ heim_base_atomic_barrier(); \ } while (0) #endif #if SIZEOF_TIME_T == 8 #define heim_base_exchange_time_t(t,v) heim_base_exchange_64((t), (v)) #elif SIZEOF_TIME_T == 4 #define heim_base_exchange_time_t(t,v) heim_base_exchange_32((t), (v)) #else #error set SIZEOF_TIME_T for your platform #endif #endif /* HEIM_BASE_ATOMICS_H */