49#ifndef __ATOMIC_OPS_EXT_H__
50#define __ATOMIC_OPS_EXT_H__
60#if (LG_SIZEOF_PTR == 8)
62#elif (LG_SIZEOF_PTR == 4)
69#if (LG_SIZEOF_PTR == 8)
71#elif (LG_SIZEOF_PTR == 4)
78#if (LG_SIZEOF_PTR == 8)
80#elif (LG_SIZEOF_PTR == 4)
87#if (LG_SIZEOF_PTR == 8)
89#elif (LG_SIZEOF_PTR == 4)
96#if (LG_SIZEOF_PTR == 8)
98#elif (LG_SIZEOF_PTR == 4)
105#if (LG_SIZEOF_PTR == 8)
107#elif (LG_SIZEOF_PTR == 4)
114#if (LG_SIZEOF_PTR == 8)
116#elif (LG_SIZEOF_PTR == 4)
124 while ((prev_value = *p) < x) {
135 "sizeof(unsigned int) != LG_SIZEOF_INT");
139#if (LG_SIZEOF_INT == 8)
141#elif (LG_SIZEOF_INT == 4)
148#if (LG_SIZEOF_INT == 8)
150#elif (LG_SIZEOF_INT == 4)
157#if (LG_SIZEOF_INT == 8)
159#elif (LG_SIZEOF_INT == 4)
166#if (LG_SIZEOF_INT == 8)
168#elif (LG_SIZEOF_INT == 4)
175#if (LG_SIZEOF_INT == 8)
177#elif (LG_SIZEOF_INT == 4)
199#if (LG_SIZEOF_PTR == 8)
201#elif (LG_SIZEOF_PTR == 4)
208#if (LG_SIZEOF_PTR == 8)
210#elif (LG_SIZEOF_PTR == 4)
217#if (LG_SIZEOF_PTR == 8)
219#elif (LG_SIZEOF_PTR == 4)
231 return *(
float *)&
ret;
236 float oldval, newval;
ATOMIC_INLINE void atomic_store_uint64(uint64_t *p, uint64_t v)
ATOMIC_INLINE uint8_t atomic_fetch_and_and_uint8(uint8_t *p, uint8_t b)
ATOMIC_INLINE uint64_t atomic_load_uint64(const uint64_t *v)
ATOMIC_INLINE uint32_t atomic_fetch_and_add_uint32(uint32_t *p, uint32_t x)
ATOMIC_INLINE uint8_t atomic_fetch_and_or_uint8(uint8_t *p, uint8_t b)
ATOMIC_INLINE uint32_t atomic_add_and_fetch_uint32(uint32_t *p, uint32_t x)
ATOMIC_INLINE uint64_t atomic_cas_uint64(uint64_t *v, uint64_t old, uint64_t _new)
ATOMIC_INLINE uint64_t atomic_fetch_and_add_uint64(uint64_t *p, uint64_t x)
ATOMIC_INLINE uint64_t atomic_add_and_fetch_uint64(uint64_t *p, uint64_t x)
ATOMIC_INLINE uint32_t atomic_load_uint32(const uint32_t *v)
ATOMIC_INLINE void atomic_store_uint32(uint32_t *p, uint32_t v)
ATOMIC_INLINE uint32_t atomic_cas_uint32(uint32_t *v, uint32_t old, uint32_t _new)
ATOMIC_INLINE unsigned int atomic_fetch_and_sub_u(unsigned int *p, unsigned int x)
ATOMIC_INLINE size_t atomic_load_z(const size_t *v)
ATOMIC_INLINE size_t atomic_add_and_fetch_z(size_t *p, size_t x)
ATOMIC_INLINE void * atomic_load_ptr(void *const *v)
ATOMIC_INLINE size_t atomic_sub_and_fetch_z(size_t *p, size_t x)
ATOMIC_INLINE void * atomic_cas_ptr(void **v, void *old, void *_new)
ATOMIC_INLINE size_t atomic_fetch_and_sub_z(size_t *p, size_t x)
ATOMIC_INLINE float atomic_add_and_fetch_fl(float *p, const float x)
ATOMIC_INLINE char atomic_fetch_and_or_char(char *p, char b)
ATOMIC_INLINE unsigned int atomic_add_and_fetch_u(unsigned int *p, unsigned int x)
ATOMIC_INLINE char atomic_fetch_and_and_char(char *p, char b)
ATOMIC_INLINE unsigned int atomic_cas_u(unsigned int *v, unsigned int old, unsigned int _new)
ATOMIC_INLINE size_t atomic_cas_z(size_t *v, size_t old, size_t _new)
ATOMIC_INLINE void atomic_store_z(size_t *p, size_t v)
ATOMIC_INLINE float atomic_cas_float(float *v, float old, float _new)
ATOMIC_INLINE size_t atomic_fetch_and_add_z(size_t *p, size_t x)
ATOMIC_INLINE size_t atomic_fetch_and_update_max_z(size_t *p, size_t x)
ATOMIC_INLINE void atomic_store_ptr(void **p, void *v)
ATOMIC_INLINE unsigned int atomic_sub_and_fetch_u(unsigned int *p, unsigned int x)
ATOMIC_INLINE unsigned int atomic_fetch_and_add_u(unsigned int *p, unsigned int x)
#define _ATOMIC_UNLIKELY(x)
#define ATOMIC_STATIC_ASSERT(a, msg)
ATTR_WARN_UNUSED_RESULT const BMVert * v
local_group_size(16, 16) .push_constant(Type b
unsigned __int64 uint64_t