/* $OpenBSD: atomic.h,v 1.9 2022/03/21 05:45:52 visa Exp $ */ /* * Copyright (c) 2014 David Gwynne * Copyright (c) 2022 Alexander Bluhm * * Permission to use, copy, modify, and distribute this software for any * purpose with or without fee is hereby granted, provided that the above * copyright notice and this permission notice appear in all copies. * * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE. */ #ifndef _SYS_ATOMIC_H_ #define _SYS_ATOMIC_H_ #include /* * an arch wanting to provide its own implementations does so by defining * macros. */ /* * atomic_cas_* */ #ifndef atomic_cas_uint static inline unsigned int atomic_cas_uint(volatile unsigned int *p, unsigned int o, unsigned int n) { return __sync_val_compare_and_swap(p, o, n); } #endif #ifndef atomic_cas_ulong static inline unsigned long atomic_cas_ulong(volatile unsigned long *p, unsigned long o, unsigned long n) { return __sync_val_compare_and_swap(p, o, n); } #endif #ifndef atomic_cas_ptr static inline void * atomic_cas_ptr(volatile void *pp, void *o, void *n) { void * volatile *p = (void * volatile *)pp; return __sync_val_compare_and_swap(p, o, n); } #endif /* * atomic_swap_* */ #ifndef atomic_swap_uint static inline unsigned int atomic_swap_uint(volatile unsigned int *p, unsigned int v) { return __sync_lock_test_and_set(p, v); } #endif #ifndef atomic_swap_ulong static inline unsigned long atomic_swap_ulong(volatile unsigned long *p, unsigned long v) { return __sync_lock_test_and_set(p, v); } #endif #ifndef atomic_swap_ptr static inline void * atomic_swap_ptr(volatile void *pp, void *v) { void * volatile *p = (void * volatile *)pp; return __sync_lock_test_and_set(p, v); } #endif /* * atomic_add_*_nv - add and fetch */ #ifndef atomic_add_int_nv static inline unsigned int atomic_add_int_nv(volatile unsigned int *p, unsigned int v) { return __sync_add_and_fetch(p, v); } #endif #ifndef atomic_add_long_nv static inline unsigned long atomic_add_long_nv(volatile unsigned long *p, unsigned long v) { return __sync_add_and_fetch(p, v); } #endif /* * atomic_add - add */ #ifndef atomic_add_int #define atomic_add_int(_p, _v) ((void)atomic_add_int_nv((_p), (_v))) #endif #ifndef atomic_add_long #define atomic_add_long(_p, _v) ((void)atomic_add_long_nv((_p), (_v))) #endif /* * atomic_inc_*_nv - increment and fetch */ #ifndef atomic_inc_int_nv #define atomic_inc_int_nv(_p) atomic_add_int_nv((_p), 1) #endif #ifndef atomic_inc_long_nv #define atomic_inc_long_nv(_p) atomic_add_long_nv((_p), 1) #endif /* * atomic_inc_* - increment */ #ifndef atomic_inc_int #define atomic_inc_int(_p) ((void)atomic_inc_int_nv(_p)) #endif #ifndef atomic_inc_long #define atomic_inc_long(_p) ((void)atomic_inc_long_nv(_p)) #endif /* * atomic_sub_*_nv - sub and fetch */ #ifndef atomic_sub_int_nv static inline unsigned int atomic_sub_int_nv(volatile unsigned int *p, unsigned int v) { return __sync_sub_and_fetch(p, v); } #endif #ifndef atomic_sub_long_nv static inline unsigned long atomic_sub_long_nv(volatile unsigned long *p, unsigned long v) { return __sync_sub_and_fetch(p, v); } #endif /* * atomic_sub_* - sub */ #ifndef atomic_sub_int #define atomic_sub_int(_p, _v) ((void)atomic_sub_int_nv((_p), (_v))) #endif #ifndef atomic_sub_long #define atomic_sub_long(_p, _v) ((void)atomic_sub_long_nv((_p), (_v))) #endif /* * atomic_dec_*_nv - decrement and fetch */ #ifndef atomic_dec_int_nv #define atomic_dec_int_nv(_p) atomic_sub_int_nv((_p), 1) #endif #ifndef atomic_dec_long_nv #define atomic_dec_long_nv(_p) atomic_sub_long_nv((_p), 1) #endif /* * atomic_dec_* - decrement */ #ifndef atomic_dec_int #define atomic_dec_int(_p) ((void)atomic_dec_int_nv(_p)) #endif #ifndef atomic_dec_long #define atomic_dec_long(_p) ((void)atomic_dec_long_nv(_p)) #endif #ifdef _KERNEL /* * atomic_load_* - read from memory */ static inline unsigned int atomic_load_int(volatile unsigned int *p) { return *p; } static inline unsigned long atomic_load_long(volatile unsigned long *p) { return *p; } /* * atomic_store_* - write to memory */ static inline void atomic_store_int(volatile unsigned int *p, unsigned int v) { *p = v; } static inline void atomic_store_long(volatile unsigned long *p, unsigned long v) { *p = v; } #endif /* _KERNEL */ /* * memory barriers */ #ifndef membar_enter #define membar_enter() __sync_synchronize() #endif #ifndef membar_exit #define membar_exit() __sync_synchronize() #endif #ifndef membar_producer #define membar_producer() __sync_synchronize() #endif #ifndef membar_consumer #define membar_consumer() __sync_synchronize() #endif #ifndef membar_sync #define membar_sync() __sync_synchronize() #endif #ifndef membar_enter_after_atomic #define membar_enter_after_atomic() membar_enter() #endif #ifndef membar_exit_before_atomic #define membar_exit_before_atomic() membar_exit() #endif #ifdef _KERNEL /* * Force any preceding reads to happen before any subsequent reads that * depend on the value returned by the preceding reads. */ static inline void membar_datadep_consumer(void) { #ifdef __alpha__ membar_consumer(); #endif } #define READ_ONCE(x) ({ \ typeof(x) __tmp = *(volatile typeof(x) *)&(x); \ membar_datadep_consumer(); \ __tmp; \ }) #define WRITE_ONCE(x, val) ({ \ typeof(x) __tmp = (val); \ *(volatile typeof(x) *)&(x) = __tmp; \ __tmp; \ }) #endif /* _KERNEL */ #endif /* _SYS_ATOMIC_H_ */