123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332 |
- /* ----------------------------------------------------------------------------
- Copyright (c) 2018-2021 Microsoft Research, Daan Leijen
- This is free software; you can redistribute it and/or modify it under the
- terms of the MIT license. A copy of the license can be found in the file
- "LICENSE" at the root of this distribution.
- -----------------------------------------------------------------------------*/
- #pragma once
- #ifndef MIMALLOC_ATOMIC_H
- #define MIMALLOC_ATOMIC_H
- // --------------------------------------------------------------------------------------------
- // Atomics
- // We need to be portable between C, C++, and MSVC.
- // We base the primitives on the C/C++ atomics and create a mimimal wrapper for MSVC in C compilation mode.
- // This is why we try to use only `uintptr_t` and `<type>*` as atomic types.
- // To gain better insight in the range of used atomics, we use explicitly named memory order operations
- // instead of passing the memory order as a parameter.
- // -----------------------------------------------------------------------------------------------
- #if defined(__cplusplus)
- // Use C++ atomics
- #include <atomic>
- #define _Atomic(tp) std::atomic<tp>
- #define mi_atomic(name) std::atomic_##name
- #define mi_memory_order(name) std::memory_order_##name
- #elif defined(_MSC_VER)
- // Use MSVC C wrapper for C11 atomics
- #define _Atomic(tp) tp
- #define ATOMIC_VAR_INIT(x) x
- #define mi_atomic(name) mi_atomic_##name
- #define mi_memory_order(name) mi_memory_order_##name
- #else
- // Use C11 atomics
- #include <stdatomic.h>
- #define mi_atomic(name) atomic_##name
- #define mi_memory_order(name) memory_order_##name
- #endif
- // Various defines for all used memory orders in mimalloc
- #define mi_atomic_cas_weak(p,expected,desired,mem_success,mem_fail) \
- mi_atomic(compare_exchange_weak_explicit)(p,expected,desired,mem_success,mem_fail)
- #define mi_atomic_cas_strong(p,expected,desired,mem_success,mem_fail) \
- mi_atomic(compare_exchange_strong_explicit)(p,expected,desired,mem_success,mem_fail)
- #define mi_atomic_load_acquire(p) mi_atomic(load_explicit)(p,mi_memory_order(acquire))
- #define mi_atomic_load_relaxed(p) mi_atomic(load_explicit)(p,mi_memory_order(relaxed))
- #define mi_atomic_store_release(p,x) mi_atomic(store_explicit)(p,x,mi_memory_order(release))
- #define mi_atomic_store_relaxed(p,x) mi_atomic(store_explicit)(p,x,mi_memory_order(relaxed))
- #define mi_atomic_exchange_release(p,x) mi_atomic(exchange_explicit)(p,x,mi_memory_order(release))
- #define mi_atomic_exchange_acq_rel(p,x) mi_atomic(exchange_explicit)(p,x,mi_memory_order(acq_rel))
- #define mi_atomic_cas_weak_release(p,exp,des) mi_atomic_cas_weak(p,exp,des,mi_memory_order(release),mi_memory_order(relaxed))
- #define mi_atomic_cas_weak_acq_rel(p,exp,des) mi_atomic_cas_weak(p,exp,des,mi_memory_order(acq_rel),mi_memory_order(acquire))
- #define mi_atomic_cas_strong_release(p,exp,des) mi_atomic_cas_strong(p,exp,des,mi_memory_order(release),mi_memory_order(relaxed))
- #define mi_atomic_cas_strong_acq_rel(p,exp,des) mi_atomic_cas_strong(p,exp,des,mi_memory_order(acq_rel),mi_memory_order(acquire))
- #define mi_atomic_add_relaxed(p,x) mi_atomic(fetch_add_explicit)(p,x,mi_memory_order(relaxed))
- #define mi_atomic_sub_relaxed(p,x) mi_atomic(fetch_sub_explicit)(p,x,mi_memory_order(relaxed))
- #define mi_atomic_add_acq_rel(p,x) mi_atomic(fetch_add_explicit)(p,x,mi_memory_order(acq_rel))
- #define mi_atomic_sub_acq_rel(p,x) mi_atomic(fetch_sub_explicit)(p,x,mi_memory_order(acq_rel))
- #define mi_atomic_and_acq_rel(p,x) mi_atomic(fetch_and_explicit)(p,x,mi_memory_order(acq_rel))
- #define mi_atomic_or_acq_rel(p,x) mi_atomic(fetch_or_explicit)(p,x,mi_memory_order(acq_rel))
- #define mi_atomic_increment_relaxed(p) mi_atomic_add_relaxed(p,(uintptr_t)1)
- #define mi_atomic_decrement_relaxed(p) mi_atomic_sub_relaxed(p,(uintptr_t)1)
- #define mi_atomic_increment_acq_rel(p) mi_atomic_add_acq_rel(p,(uintptr_t)1)
- #define mi_atomic_decrement_acq_rel(p) mi_atomic_sub_acq_rel(p,(uintptr_t)1)
- static inline void mi_atomic_yield(void);
- static inline intptr_t mi_atomic_addi(_Atomic(intptr_t)*p, intptr_t add);
- static inline intptr_t mi_atomic_subi(_Atomic(intptr_t)*p, intptr_t sub);
- #if defined(__cplusplus) || !defined(_MSC_VER)
- // In C++/C11 atomics we have polymorphic atomics so can use the typed `ptr` variants (where `tp` is the type of atomic value)
- // We use these macros so we can provide a typed wrapper in MSVC in C compilation mode as well
- #define mi_atomic_load_ptr_acquire(tp,p) mi_atomic_load_acquire(p)
- #define mi_atomic_load_ptr_relaxed(tp,p) mi_atomic_load_relaxed(p)
- // In C++ we need to add casts to help resolve templates if NULL is passed
- #if defined(__cplusplus)
- #define mi_atomic_store_ptr_release(tp,p,x) mi_atomic_store_release(p,(tp*)x)
- #define mi_atomic_store_ptr_relaxed(tp,p,x) mi_atomic_store_relaxed(p,(tp*)x)
- #define mi_atomic_cas_ptr_weak_release(tp,p,exp,des) mi_atomic_cas_weak_release(p,exp,(tp*)des)
- #define mi_atomic_cas_ptr_weak_acq_rel(tp,p,exp,des) mi_atomic_cas_weak_acq_rel(p,exp,(tp*)des)
- #define mi_atomic_cas_ptr_strong_release(tp,p,exp,des) mi_atomic_cas_strong_release(p,exp,(tp*)des)
- #define mi_atomic_exchange_ptr_release(tp,p,x) mi_atomic_exchange_release(p,(tp*)x)
- #define mi_atomic_exchange_ptr_acq_rel(tp,p,x) mi_atomic_exchange_acq_rel(p,(tp*)x)
- #else
- #define mi_atomic_store_ptr_release(tp,p,x) mi_atomic_store_release(p,x)
- #define mi_atomic_store_ptr_relaxed(tp,p,x) mi_atomic_store_relaxed(p,x)
- #define mi_atomic_cas_ptr_weak_release(tp,p,exp,des) mi_atomic_cas_weak_release(p,exp,des)
- #define mi_atomic_cas_ptr_weak_acq_rel(tp,p,exp,des) mi_atomic_cas_weak_acq_rel(p,exp,des)
- #define mi_atomic_cas_ptr_strong_release(tp,p,exp,des) mi_atomic_cas_strong_release(p,exp,des)
- #define mi_atomic_exchange_ptr_release(tp,p,x) mi_atomic_exchange_release(p,x)
- #define mi_atomic_exchange_ptr_acq_rel(tp,p,x) mi_atomic_exchange_acq_rel(p,x)
- #endif
- // These are used by the statistics
- static inline int64_t mi_atomic_addi64_relaxed(volatile int64_t* p, int64_t add) {
- return mi_atomic(fetch_add_explicit)((_Atomic(int64_t)*)p, add, mi_memory_order(relaxed));
- }
- static inline void mi_atomic_maxi64_relaxed(volatile int64_t* p, int64_t x) {
- int64_t current = mi_atomic_load_relaxed((_Atomic(int64_t)*)p);
- while (current < x && !mi_atomic_cas_weak_release((_Atomic(int64_t)*)p, ¤t, x)) { /* nothing */ };
- }
- // Used by timers
- #define mi_atomic_loadi64_acquire(p) mi_atomic(load_explicit)(p,mi_memory_order(acquire))
- #define mi_atomic_loadi64_relaxed(p) mi_atomic(load_explicit)(p,mi_memory_order(relaxed))
- #define mi_atomic_storei64_release(p,x) mi_atomic(store_explicit)(p,x,mi_memory_order(release))
- #define mi_atomic_storei64_relaxed(p,x) mi_atomic(store_explicit)(p,x,mi_memory_order(relaxed))
- #elif defined(_MSC_VER)
- // MSVC C compilation wrapper that uses Interlocked operations to model C11 atomics.
- #define WIN32_LEAN_AND_MEAN
- #include <windows.h>
- #include <intrin.h>
- #ifdef _WIN64
- typedef LONG64 msc_intptr_t;
- #define MI_64(f) f##64
- #else
- typedef LONG msc_intptr_t;
- #define MI_64(f) f
- #endif
- typedef enum mi_memory_order_e {
- mi_memory_order_relaxed,
- mi_memory_order_consume,
- mi_memory_order_acquire,
- mi_memory_order_release,
- mi_memory_order_acq_rel,
- mi_memory_order_seq_cst
- } mi_memory_order;
- static inline uintptr_t mi_atomic_fetch_add_explicit(_Atomic(uintptr_t)*p, uintptr_t add, mi_memory_order mo) {
- (void)(mo);
- return (uintptr_t)MI_64(_InterlockedExchangeAdd)((volatile msc_intptr_t*)p, (msc_intptr_t)add);
- }
- static inline uintptr_t mi_atomic_fetch_sub_explicit(_Atomic(uintptr_t)*p, uintptr_t sub, mi_memory_order mo) {
- (void)(mo);
- return (uintptr_t)MI_64(_InterlockedExchangeAdd)((volatile msc_intptr_t*)p, -((msc_intptr_t)sub));
- }
- static inline uintptr_t mi_atomic_fetch_and_explicit(_Atomic(uintptr_t)*p, uintptr_t x, mi_memory_order mo) {
- (void)(mo);
- return (uintptr_t)MI_64(_InterlockedAnd)((volatile msc_intptr_t*)p, (msc_intptr_t)x);
- }
- static inline uintptr_t mi_atomic_fetch_or_explicit(_Atomic(uintptr_t)*p, uintptr_t x, mi_memory_order mo) {
- (void)(mo);
- return (uintptr_t)MI_64(_InterlockedOr)((volatile msc_intptr_t*)p, (msc_intptr_t)x);
- }
- static inline bool mi_atomic_compare_exchange_strong_explicit(_Atomic(uintptr_t)*p, uintptr_t* expected, uintptr_t desired, mi_memory_order mo1, mi_memory_order mo2) {
- (void)(mo1); (void)(mo2);
- uintptr_t read = (uintptr_t)MI_64(_InterlockedCompareExchange)((volatile msc_intptr_t*)p, (msc_intptr_t)desired, (msc_intptr_t)(*expected));
- if (read == *expected) {
- return true;
- }
- else {
- *expected = read;
- return false;
- }
- }
- static inline bool mi_atomic_compare_exchange_weak_explicit(_Atomic(uintptr_t)*p, uintptr_t* expected, uintptr_t desired, mi_memory_order mo1, mi_memory_order mo2) {
- return mi_atomic_compare_exchange_strong_explicit(p, expected, desired, mo1, mo2);
- }
- static inline uintptr_t mi_atomic_exchange_explicit(_Atomic(uintptr_t)*p, uintptr_t exchange, mi_memory_order mo) {
- (void)(mo);
- return (uintptr_t)MI_64(_InterlockedExchange)((volatile msc_intptr_t*)p, (msc_intptr_t)exchange);
- }
- static inline void mi_atomic_thread_fence(mi_memory_order mo) {
- (void)(mo);
- _Atomic(uintptr_t)x = 0;
- mi_atomic_exchange_explicit(&x, 1, mo);
- }
- static inline uintptr_t mi_atomic_load_explicit(_Atomic(uintptr_t) const* p, mi_memory_order mo) {
- (void)(mo);
- #if defined(_M_IX86) || defined(_M_X64)
- return *p;
- #else
- uintptr_t x = *p;
- if (mo > mi_memory_order_relaxed) {
- while (!mi_atomic_compare_exchange_weak_explicit(p, &x, x, mo, mi_memory_order_relaxed)) { /* nothing */ };
- }
- return x;
- #endif
- }
- static inline void mi_atomic_store_explicit(_Atomic(uintptr_t)*p, uintptr_t x, mi_memory_order mo) {
- (void)(mo);
- #if defined(_M_IX86) || defined(_M_X64)
- *p = x;
- #else
- mi_atomic_exchange_explicit(p, x, mo);
- #endif
- }
- static inline int64_t mi_atomic_loadi64_explicit(_Atomic(int64_t)*p, mi_memory_order mo) {
- (void)(mo);
- #if defined(_M_X64)
- return *p;
- #else
- int64_t old = *p;
- int64_t x = old;
- while ((old = InterlockedCompareExchange64(p, x, old)) != x) {
- x = old;
- }
- return x;
- #endif
- }
- static inline void mi_atomic_storei64_explicit(_Atomic(int64_t)*p, int64_t x, mi_memory_order mo) {
- (void)(mo);
- #if defined(x_M_IX86) || defined(_M_X64)
- *p = x;
- #else
- InterlockedExchange64(p, x);
- #endif
- }
- // These are used by the statistics
- static inline int64_t mi_atomic_addi64_relaxed(volatile _Atomic(int64_t)*p, int64_t add) {
- #ifdef _WIN64
- return (int64_t)mi_atomic_addi((int64_t*)p, add);
- #else
- int64_t current;
- int64_t sum;
- do {
- current = *p;
- sum = current + add;
- } while (_InterlockedCompareExchange64(p, sum, current) != current);
- return current;
- #endif
- }
- static inline void mi_atomic_maxi64_relaxed(volatile _Atomic(int64_t)*p, int64_t x) {
- int64_t current;
- do {
- current = *p;
- } while (current < x && _InterlockedCompareExchange64(p, x, current) != current);
- }
- // The pointer macros cast to `uintptr_t`.
- #define mi_atomic_load_ptr_acquire(tp,p) (tp*)mi_atomic_load_acquire((_Atomic(uintptr_t)*)(p))
- #define mi_atomic_load_ptr_relaxed(tp,p) (tp*)mi_atomic_load_relaxed((_Atomic(uintptr_t)*)(p))
- #define mi_atomic_store_ptr_release(tp,p,x) mi_atomic_store_release((_Atomic(uintptr_t)*)(p),(uintptr_t)(x))
- #define mi_atomic_store_ptr_relaxed(tp,p,x) mi_atomic_store_relaxed((_Atomic(uintptr_t)*)(p),(uintptr_t)(x))
- #define mi_atomic_cas_ptr_weak_release(tp,p,exp,des) mi_atomic_cas_weak_release((_Atomic(uintptr_t)*)(p),(uintptr_t*)exp,(uintptr_t)des)
- #define mi_atomic_cas_ptr_weak_acq_rel(tp,p,exp,des) mi_atomic_cas_weak_acq_rel((_Atomic(uintptr_t)*)(p),(uintptr_t*)exp,(uintptr_t)des)
- #define mi_atomic_cas_ptr_strong_release(tp,p,exp,des) mi_atomic_cas_strong_release((_Atomic(uintptr_t)*)(p),(uintptr_t*)exp,(uintptr_t)des)
- #define mi_atomic_exchange_ptr_release(tp,p,x) (tp*)mi_atomic_exchange_release((_Atomic(uintptr_t)*)(p),(uintptr_t)x)
- #define mi_atomic_exchange_ptr_acq_rel(tp,p,x) (tp*)mi_atomic_exchange_acq_rel((_Atomic(uintptr_t)*)(p),(uintptr_t)x)
- #define mi_atomic_loadi64_acquire(p) mi_atomic(loadi64_explicit)(p,mi_memory_order(acquire))
- #define mi_atomic_loadi64_relaxed(p) mi_atomic(loadi64_explicit)(p,mi_memory_order(relaxed))
- #define mi_atomic_storei64_release(p,x) mi_atomic(storei64_explicit)(p,x,mi_memory_order(release))
- #define mi_atomic_storei64_relaxed(p,x) mi_atomic(storei64_explicit)(p,x,mi_memory_order(relaxed))
- #endif
- // Atomically add a signed value; returns the previous value.
- static inline intptr_t mi_atomic_addi(_Atomic(intptr_t)*p, intptr_t add) {
- return (intptr_t)mi_atomic_add_acq_rel((_Atomic(uintptr_t)*)p, (uintptr_t)add);
- }
- // Atomically subtract a signed value; returns the previous value.
- static inline intptr_t mi_atomic_subi(_Atomic(intptr_t)*p, intptr_t sub) {
- return (intptr_t)mi_atomic_addi(p, -sub);
- }
- // Yield
- #if defined(__cplusplus)
- #include <thread>
- static inline void mi_atomic_yield(void) {
- std::this_thread::yield();
- }
- #elif defined(_WIN32)
- #define WIN32_LEAN_AND_MEAN
- #include <windows.h>
- static inline void mi_atomic_yield(void) {
- YieldProcessor();
- }
- #elif defined(__SSE2__)
- #include <emmintrin.h>
- static inline void mi_atomic_yield(void) {
- _mm_pause();
- }
- #elif (defined(__GNUC__) || defined(__clang__)) && \
- (defined(__x86_64__) || defined(__i386__) || defined(__arm__) || defined(__armel__) || defined(__ARMEL__) || \
- defined(__aarch64__) || defined(__powerpc__) || defined(__ppc__) || defined(__PPC__))
- #if defined(__x86_64__) || defined(__i386__)
- static inline void mi_atomic_yield(void) {
- __asm__ volatile ("pause" ::: "memory");
- }
- #elif defined(__aarch64__)
- static inline void mi_atomic_yield(void) {
- __asm__ volatile("wfe");
- }
- #elif (defined(__arm__) && __ARM_ARCH__ >= 7)
- static inline void mi_atomic_yield(void) {
- __asm__ volatile("yield" ::: "memory");
- }
- #elif defined(__powerpc__) || defined(__ppc__) || defined(__PPC__)
- static inline void mi_atomic_yield(void) {
- __asm__ __volatile__ ("or 27,27,27" ::: "memory");
- }
- #elif defined(__armel__) || defined(__ARMEL__)
- static inline void mi_atomic_yield(void) {
- __asm__ volatile ("nop" ::: "memory");
- }
- #endif
- #elif defined(__sun)
- // Fallback for other archs
- #error #include <synch.h>
- static inline void mi_atomic_yield(void) {
- smt_pause();
- }
- #elif defined(__wasi__)
- #include <sched.h>
- static inline void mi_atomic_yield(void) {
- sched_yield();
- }
- #else
- #include <unistd.h>
- static inline void mi_atomic_yield(void) {
- sleep(0);
- }
- #endif
- #endif // __MIMALLOC_ATOMIC_H
|