2 //===--------------------------- atomic -----------------------------------===//
4 // The LLVM Compiler Infrastructure
6 // This file is distributed under the University of Illinois Open Source
7 // License. See LICENSE.TXT for details.
9 //===----------------------------------------------------------------------===//
11 #ifndef _LIBCPP_ATOMIC
12 #define _LIBCPP_ATOMIC
22 #define __cpp_lib_atomic_is_always_lock_free // as specified by SG10
24 // order and consistency
26 typedef enum memory_order
29 memory_order_consume, // load-consume
30 memory_order_acquire, // load-acquire
31 memory_order_release, // store-release
32 memory_order_acq_rel, // store-release load-acquire
33 memory_order_seq_cst // store-release load-acquire
36 template <class T> T kill_dependency(T y) noexcept;
40 #define ATOMIC_BOOL_LOCK_FREE unspecified
41 #define ATOMIC_CHAR_LOCK_FREE unspecified
42 #define ATOMIC_CHAR16_T_LOCK_FREE unspecified
43 #define ATOMIC_CHAR32_T_LOCK_FREE unspecified
44 #define ATOMIC_WCHAR_T_LOCK_FREE unspecified
45 #define ATOMIC_SHORT_LOCK_FREE unspecified
46 #define ATOMIC_INT_LOCK_FREE unspecified
47 #define ATOMIC_LONG_LOCK_FREE unspecified
48 #define ATOMIC_LLONG_LOCK_FREE unspecified
49 #define ATOMIC_POINTER_LOCK_FREE unspecified
51 // flag type and operations
53 typedef struct atomic_flag
55 bool test_and_set(memory_order m = memory_order_seq_cst) volatile noexcept;
56 bool test_and_set(memory_order m = memory_order_seq_cst) noexcept;
57 void clear(memory_order m = memory_order_seq_cst) volatile noexcept;
58 void clear(memory_order m = memory_order_seq_cst) noexcept;
59 atomic_flag() noexcept = default;
60 atomic_flag(const atomic_flag&) = delete;
61 atomic_flag& operator=(const atomic_flag&) = delete;
62 atomic_flag& operator=(const atomic_flag&) volatile = delete;
66 atomic_flag_test_and_set(volatile atomic_flag* obj) noexcept;
69 atomic_flag_test_and_set(atomic_flag* obj) noexcept;
72 atomic_flag_test_and_set_explicit(volatile atomic_flag* obj,
73 memory_order m) noexcept;
76 atomic_flag_test_and_set_explicit(atomic_flag* obj, memory_order m) noexcept;
79 atomic_flag_clear(volatile atomic_flag* obj) noexcept;
82 atomic_flag_clear(atomic_flag* obj) noexcept;
85 atomic_flag_clear_explicit(volatile atomic_flag* obj, memory_order m) noexcept;
88 atomic_flag_clear_explicit(atomic_flag* obj, memory_order m) noexcept;
90 #define ATOMIC_FLAG_INIT see below
91 #define ATOMIC_VAR_INIT(value) see below
96 static constexpr bool is_always_lock_free;
97 bool is_lock_free() const volatile noexcept;
98 bool is_lock_free() const noexcept;
99 void store(T desr, memory_order m = memory_order_seq_cst) volatile noexcept;
100 void store(T desr, memory_order m = memory_order_seq_cst) noexcept;
101 T load(memory_order m = memory_order_seq_cst) const volatile noexcept;
102 T load(memory_order m = memory_order_seq_cst) const noexcept;
103 operator T() const volatile noexcept;
104 operator T() const noexcept;
105 T exchange(T desr, memory_order m = memory_order_seq_cst) volatile noexcept;
106 T exchange(T desr, memory_order m = memory_order_seq_cst) noexcept;
107 bool compare_exchange_weak(T& expc, T desr,
108 memory_order s, memory_order f) volatile noexcept;
109 bool compare_exchange_weak(T& expc, T desr, memory_order s, memory_order f) noexcept;
110 bool compare_exchange_strong(T& expc, T desr,
111 memory_order s, memory_order f) volatile noexcept;
112 bool compare_exchange_strong(T& expc, T desr,
113 memory_order s, memory_order f) noexcept;
114 bool compare_exchange_weak(T& expc, T desr,
115 memory_order m = memory_order_seq_cst) volatile noexcept;
116 bool compare_exchange_weak(T& expc, T desr,
117 memory_order m = memory_order_seq_cst) noexcept;
118 bool compare_exchange_strong(T& expc, T desr,
119 memory_order m = memory_order_seq_cst) volatile noexcept;
120 bool compare_exchange_strong(T& expc, T desr,
121 memory_order m = memory_order_seq_cst) noexcept;
123 atomic() noexcept = default;
124 constexpr atomic(T desr) noexcept;
125 atomic(const atomic&) = delete;
126 atomic& operator=(const atomic&) = delete;
127 atomic& operator=(const atomic&) volatile = delete;
128 T operator=(T) volatile noexcept;
129 T operator=(T) noexcept;
133 struct atomic<integral>
135 static constexpr bool is_always_lock_free;
136 bool is_lock_free() const volatile noexcept;
137 bool is_lock_free() const noexcept;
138 void store(integral desr, memory_order m = memory_order_seq_cst) volatile noexcept;
139 void store(integral desr, memory_order m = memory_order_seq_cst) noexcept;
140 integral load(memory_order m = memory_order_seq_cst) const volatile noexcept;
141 integral load(memory_order m = memory_order_seq_cst) const noexcept;
142 operator integral() const volatile noexcept;
143 operator integral() const noexcept;
144 integral exchange(integral desr,
145 memory_order m = memory_order_seq_cst) volatile noexcept;
146 integral exchange(integral desr, memory_order m = memory_order_seq_cst) noexcept;
147 bool compare_exchange_weak(integral& expc, integral desr,
148 memory_order s, memory_order f) volatile noexcept;
149 bool compare_exchange_weak(integral& expc, integral desr,
150 memory_order s, memory_order f) noexcept;
151 bool compare_exchange_strong(integral& expc, integral desr,
152 memory_order s, memory_order f) volatile noexcept;
153 bool compare_exchange_strong(integral& expc, integral desr,
154 memory_order s, memory_order f) noexcept;
155 bool compare_exchange_weak(integral& expc, integral desr,
156 memory_order m = memory_order_seq_cst) volatile noexcept;
157 bool compare_exchange_weak(integral& expc, integral desr,
158 memory_order m = memory_order_seq_cst) noexcept;
159 bool compare_exchange_strong(integral& expc, integral desr,
160 memory_order m = memory_order_seq_cst) volatile noexcept;
161 bool compare_exchange_strong(integral& expc, integral desr,
162 memory_order m = memory_order_seq_cst) noexcept;
165 fetch_add(integral op, memory_order m = memory_order_seq_cst) volatile noexcept;
166 integral fetch_add(integral op, memory_order m = memory_order_seq_cst) noexcept;
168 fetch_sub(integral op, memory_order m = memory_order_seq_cst) volatile noexcept;
169 integral fetch_sub(integral op, memory_order m = memory_order_seq_cst) noexcept;
171 fetch_and(integral op, memory_order m = memory_order_seq_cst) volatile noexcept;
172 integral fetch_and(integral op, memory_order m = memory_order_seq_cst) noexcept;
174 fetch_or(integral op, memory_order m = memory_order_seq_cst) volatile noexcept;
175 integral fetch_or(integral op, memory_order m = memory_order_seq_cst) noexcept;
177 fetch_xor(integral op, memory_order m = memory_order_seq_cst) volatile noexcept;
178 integral fetch_xor(integral op, memory_order m = memory_order_seq_cst) noexcept;
180 atomic() noexcept = default;
181 constexpr atomic(integral desr) noexcept;
182 atomic(const atomic&) = delete;
183 atomic& operator=(const atomic&) = delete;
184 atomic& operator=(const atomic&) volatile = delete;
185 integral operator=(integral desr) volatile noexcept;
186 integral operator=(integral desr) noexcept;
188 integral operator++(int) volatile noexcept;
189 integral operator++(int) noexcept;
190 integral operator--(int) volatile noexcept;
191 integral operator--(int) noexcept;
192 integral operator++() volatile noexcept;
193 integral operator++() noexcept;
194 integral operator--() volatile noexcept;
195 integral operator--() noexcept;
196 integral operator+=(integral op) volatile noexcept;
197 integral operator+=(integral op) noexcept;
198 integral operator-=(integral op) volatile noexcept;
199 integral operator-=(integral op) noexcept;
200 integral operator&=(integral op) volatile noexcept;
201 integral operator&=(integral op) noexcept;
202 integral operator|=(integral op) volatile noexcept;
203 integral operator|=(integral op) noexcept;
204 integral operator^=(integral op) volatile noexcept;
205 integral operator^=(integral op) noexcept;
211 static constexpr bool is_always_lock_free;
212 bool is_lock_free() const volatile noexcept;
213 bool is_lock_free() const noexcept;
214 void store(T* desr, memory_order m = memory_order_seq_cst) volatile noexcept;
215 void store(T* desr, memory_order m = memory_order_seq_cst) noexcept;
216 T* load(memory_order m = memory_order_seq_cst) const volatile noexcept;
217 T* load(memory_order m = memory_order_seq_cst) const noexcept;
218 operator T*() const volatile noexcept;
219 operator T*() const noexcept;
220 T* exchange(T* desr, memory_order m = memory_order_seq_cst) volatile noexcept;
221 T* exchange(T* desr, memory_order m = memory_order_seq_cst) noexcept;
222 bool compare_exchange_weak(T*& expc, T* desr,
223 memory_order s, memory_order f) volatile noexcept;
224 bool compare_exchange_weak(T*& expc, T* desr,
225 memory_order s, memory_order f) noexcept;
226 bool compare_exchange_strong(T*& expc, T* desr,
227 memory_order s, memory_order f) volatile noexcept;
228 bool compare_exchange_strong(T*& expc, T* desr,
229 memory_order s, memory_order f) noexcept;
230 bool compare_exchange_weak(T*& expc, T* desr,
231 memory_order m = memory_order_seq_cst) volatile noexcept;
232 bool compare_exchange_weak(T*& expc, T* desr,
233 memory_order m = memory_order_seq_cst) noexcept;
234 bool compare_exchange_strong(T*& expc, T* desr,
235 memory_order m = memory_order_seq_cst) volatile noexcept;
236 bool compare_exchange_strong(T*& expc, T* desr,
237 memory_order m = memory_order_seq_cst) noexcept;
238 T* fetch_add(ptrdiff_t op, memory_order m = memory_order_seq_cst) volatile noexcept;
239 T* fetch_add(ptrdiff_t op, memory_order m = memory_order_seq_cst) noexcept;
240 T* fetch_sub(ptrdiff_t op, memory_order m = memory_order_seq_cst) volatile noexcept;
241 T* fetch_sub(ptrdiff_t op, memory_order m = memory_order_seq_cst) noexcept;
243 atomic() noexcept = default;
244 constexpr atomic(T* desr) noexcept;
245 atomic(const atomic&) = delete;
246 atomic& operator=(const atomic&) = delete;
247 atomic& operator=(const atomic&) volatile = delete;
249 T* operator=(T*) volatile noexcept;
250 T* operator=(T*) noexcept;
251 T* operator++(int) volatile noexcept;
252 T* operator++(int) noexcept;
253 T* operator--(int) volatile noexcept;
254 T* operator--(int) noexcept;
255 T* operator++() volatile noexcept;
256 T* operator++() noexcept;
257 T* operator--() volatile noexcept;
258 T* operator--() noexcept;
259 T* operator+=(ptrdiff_t op) volatile noexcept;
260 T* operator+=(ptrdiff_t op) noexcept;
261 T* operator-=(ptrdiff_t op) volatile noexcept;
262 T* operator-=(ptrdiff_t op) noexcept;
268 atomic_is_lock_free(const volatile atomic<T>* obj) noexcept;
272 atomic_is_lock_free(const atomic<T>* obj) noexcept;
276 atomic_init(volatile atomic<T>* obj, T desr) noexcept;
280 atomic_init(atomic<T>* obj, T desr) noexcept;
284 atomic_store(volatile atomic<T>* obj, T desr) noexcept;
288 atomic_store(atomic<T>* obj, T desr) noexcept;
292 atomic_store_explicit(volatile atomic<T>* obj, T desr, memory_order m) noexcept;
296 atomic_store_explicit(atomic<T>* obj, T desr, memory_order m) noexcept;
300 atomic_load(const volatile atomic<T>* obj) noexcept;
304 atomic_load(const atomic<T>* obj) noexcept;
308 atomic_load_explicit(const volatile atomic<T>* obj, memory_order m) noexcept;
312 atomic_load_explicit(const atomic<T>* obj, memory_order m) noexcept;
316 atomic_exchange(volatile atomic<T>* obj, T desr) noexcept;
320 atomic_exchange(atomic<T>* obj, T desr) noexcept;
324 atomic_exchange_explicit(volatile atomic<T>* obj, T desr, memory_order m) noexcept;
328 atomic_exchange_explicit(atomic<T>* obj, T desr, memory_order m) noexcept;
332 atomic_compare_exchange_weak(volatile atomic<T>* obj, T* expc, T desr) noexcept;
336 atomic_compare_exchange_weak(atomic<T>* obj, T* expc, T desr) noexcept;
340 atomic_compare_exchange_strong(volatile atomic<T>* obj, T* expc, T desr) noexcept;
344 atomic_compare_exchange_strong(atomic<T>* obj, T* expc, T desr) noexcept;
348 atomic_compare_exchange_weak_explicit(volatile atomic<T>* obj, T* expc,
350 memory_order s, memory_order f) noexcept;
354 atomic_compare_exchange_weak_explicit(atomic<T>* obj, T* expc, T desr,
355 memory_order s, memory_order f) noexcept;
359 atomic_compare_exchange_strong_explicit(volatile atomic<T>* obj,
361 memory_order s, memory_order f) noexcept;
365 atomic_compare_exchange_strong_explicit(atomic<T>* obj, T* expc,
367 memory_order s, memory_order f) noexcept;
369 template <class Integral>
371 atomic_fetch_add(volatile atomic<Integral>* obj, Integral op) noexcept;
373 template <class Integral>
375 atomic_fetch_add(atomic<Integral>* obj, Integral op) noexcept;
377 template <class Integral>
379 atomic_fetch_add_explicit(volatile atomic<Integral>* obj, Integral op,
380 memory_order m) noexcept;
381 template <class Integral>
383 atomic_fetch_add_explicit(atomic<Integral>* obj, Integral op,
384 memory_order m) noexcept;
385 template <class Integral>
387 atomic_fetch_sub(volatile atomic<Integral>* obj, Integral op) noexcept;
389 template <class Integral>
391 atomic_fetch_sub(atomic<Integral>* obj, Integral op) noexcept;
393 template <class Integral>
395 atomic_fetch_sub_explicit(volatile atomic<Integral>* obj, Integral op,
396 memory_order m) noexcept;
397 template <class Integral>
399 atomic_fetch_sub_explicit(atomic<Integral>* obj, Integral op,
400 memory_order m) noexcept;
401 template <class Integral>
403 atomic_fetch_and(volatile atomic<Integral>* obj, Integral op) noexcept;
405 template <class Integral>
407 atomic_fetch_and(atomic<Integral>* obj, Integral op) noexcept;
409 template <class Integral>
411 atomic_fetch_and_explicit(volatile atomic<Integral>* obj, Integral op,
412 memory_order m) noexcept;
413 template <class Integral>
415 atomic_fetch_and_explicit(atomic<Integral>* obj, Integral op,
416 memory_order m) noexcept;
417 template <class Integral>
419 atomic_fetch_or(volatile atomic<Integral>* obj, Integral op) noexcept;
421 template <class Integral>
423 atomic_fetch_or(atomic<Integral>* obj, Integral op) noexcept;
425 template <class Integral>
427 atomic_fetch_or_explicit(volatile atomic<Integral>* obj, Integral op,
428 memory_order m) noexcept;
429 template <class Integral>
431 atomic_fetch_or_explicit(atomic<Integral>* obj, Integral op,
432 memory_order m) noexcept;
433 template <class Integral>
435 atomic_fetch_xor(volatile atomic<Integral>* obj, Integral op) noexcept;
437 template <class Integral>
439 atomic_fetch_xor(atomic<Integral>* obj, Integral op) noexcept;
441 template <class Integral>
443 atomic_fetch_xor_explicit(volatile atomic<Integral>* obj, Integral op,
444 memory_order m) noexcept;
445 template <class Integral>
447 atomic_fetch_xor_explicit(atomic<Integral>* obj, Integral op,
448 memory_order m) noexcept;
452 atomic_fetch_add(volatile atomic<T*>* obj, ptrdiff_t op) noexcept;
456 atomic_fetch_add(atomic<T*>* obj, ptrdiff_t op) noexcept;
460 atomic_fetch_add_explicit(volatile atomic<T*>* obj, ptrdiff_t op,
461 memory_order m) noexcept;
464 atomic_fetch_add_explicit(atomic<T*>* obj, ptrdiff_t op, memory_order m) noexcept;
468 atomic_fetch_sub(volatile atomic<T*>* obj, ptrdiff_t op) noexcept;
472 atomic_fetch_sub(atomic<T*>* obj, ptrdiff_t op) noexcept;
476 atomic_fetch_sub_explicit(volatile atomic<T*>* obj, ptrdiff_t op,
477 memory_order m) noexcept;
480 atomic_fetch_sub_explicit(atomic<T*>* obj, ptrdiff_t op, memory_order m) noexcept;
482 // Atomics for standard typedef types
484 typedef atomic<bool> atomic_bool;
485 typedef atomic<char> atomic_char;
486 typedef atomic<signed char> atomic_schar;
487 typedef atomic<unsigned char> atomic_uchar;
488 typedef atomic<short> atomic_short;
489 typedef atomic<unsigned short> atomic_ushort;
490 typedef atomic<int> atomic_int;
491 typedef atomic<unsigned int> atomic_uint;
492 typedef atomic<long> atomic_long;
493 typedef atomic<unsigned long> atomic_ulong;
494 typedef atomic<long long> atomic_llong;
495 typedef atomic<unsigned long long> atomic_ullong;
496 typedef atomic<char16_t> atomic_char16_t;
497 typedef atomic<char32_t> atomic_char32_t;
498 typedef atomic<wchar_t> atomic_wchar_t;
500 typedef atomic<int_least8_t> atomic_int_least8_t;
501 typedef atomic<uint_least8_t> atomic_uint_least8_t;
502 typedef atomic<int_least16_t> atomic_int_least16_t;
503 typedef atomic<uint_least16_t> atomic_uint_least16_t;
504 typedef atomic<int_least32_t> atomic_int_least32_t;
505 typedef atomic<uint_least32_t> atomic_uint_least32_t;
506 typedef atomic<int_least64_t> atomic_int_least64_t;
507 typedef atomic<uint_least64_t> atomic_uint_least64_t;
509 typedef atomic<int_fast8_t> atomic_int_fast8_t;
510 typedef atomic<uint_fast8_t> atomic_uint_fast8_t;
511 typedef atomic<int_fast16_t> atomic_int_fast16_t;
512 typedef atomic<uint_fast16_t> atomic_uint_fast16_t;
513 typedef atomic<int_fast32_t> atomic_int_fast32_t;
514 typedef atomic<uint_fast32_t> atomic_uint_fast32_t;
515 typedef atomic<int_fast64_t> atomic_int_fast64_t;
516 typedef atomic<uint_fast64_t> atomic_uint_fast64_t;
518 typedef atomic<int8_t> atomic_int8_t;
519 typedef atomic<uint8_t> atomic_uint8_t;
520 typedef atomic<int16_t> atomic_int16_t;
521 typedef atomic<uint16_t> atomic_uint16_t;
522 typedef atomic<int32_t> atomic_int32_t;
523 typedef atomic<uint32_t> atomic_uint32_t;
524 typedef atomic<int64_t> atomic_int64_t;
525 typedef atomic<uint64_t> atomic_uint64_t;
527 typedef atomic<intptr_t> atomic_intptr_t;
528 typedef atomic<uintptr_t> atomic_uintptr_t;
529 typedef atomic<size_t> atomic_size_t;
530 typedef atomic<ptrdiff_t> atomic_ptrdiff_t;
531 typedef atomic<intmax_t> atomic_intmax_t;
532 typedef atomic<uintmax_t> atomic_uintmax_t;
536 void atomic_thread_fence(memory_order m) noexcept;
537 void atomic_signal_fence(memory_order m) noexcept;
546 #include <type_traits>
548 #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER)
549 #pragma GCC system_header
552 #ifdef _LIBCPP_HAS_NO_THREADS
553 #error <atomic> is not supported on this single threaded system
555 #if !defined(_LIBCPP_HAS_C_ATOMIC_IMP) && !defined(_LIBCPP_HAS_GCC_ATOMIC_IMP)
556 #error <atomic> is not implemented
558 #ifdef kill_dependency
559 #error C++ standard library is incompatible with <stdatomic.h>
562 #if _LIBCPP_STD_VER > 14
563 # define __cpp_lib_atomic_is_always_lock_free 201603L
566 #define _LIBCPP_CHECK_STORE_MEMORY_ORDER(__m) \
567 _LIBCPP_DIAGNOSE_WARNING(__m == memory_order_consume || \
568 __m == memory_order_acquire || \
569 __m == memory_order_acq_rel, \
570 "memory order argument to atomic operation is invalid")
572 #define _LIBCPP_CHECK_LOAD_MEMORY_ORDER(__m) \
573 _LIBCPP_DIAGNOSE_WARNING(__m == memory_order_release || \
574 __m == memory_order_acq_rel, \
575 "memory order argument to atomic operation is invalid")
577 #define _LIBCPP_CHECK_EXCHANGE_MEMORY_ORDER(__m, __f) \
578 _LIBCPP_DIAGNOSE_WARNING(__f == memory_order_release || \
579 __f == memory_order_acq_rel, \
580 "memory order argument to atomic operation is invalid")
582 _LIBCPP_BEGIN_NAMESPACE_STD
584 typedef enum memory_order
586 memory_order_relaxed, memory_order_consume, memory_order_acquire,
587 memory_order_release, memory_order_acq_rel, memory_order_seq_cst
590 #if defined(_LIBCPP_HAS_GCC_ATOMIC_IMP)
591 namespace __gcc_atomic {
592 template <typename _Tp>
593 struct __gcc_atomic_t {
596 static_assert(is_trivially_copyable<_Tp>::value,
597 "std::atomic<Tp> requires that 'Tp' be a trivially copyable type");
600 _LIBCPP_INLINE_VISIBILITY
601 #ifndef _LIBCPP_CXX03_LANG
602 __gcc_atomic_t() _NOEXCEPT = default;
604 __gcc_atomic_t() _NOEXCEPT : __a_value() {}
605 #endif // _LIBCPP_CXX03_LANG
606 _LIBCPP_CONSTEXPR explicit __gcc_atomic_t(_Tp value) _NOEXCEPT
607 : __a_value(value) {}
610 #define _Atomic(x) __gcc_atomic::__gcc_atomic_t<x>
612 template <typename _Tp> _Tp __create();
614 template <typename _Tp, typename _Td>
615 typename enable_if<sizeof(_Tp()->__a_value = __create<_Td>()), char>::type
616 __test_atomic_assignable(int);
617 template <typename _Tp, typename _Up>
618 __two __test_atomic_assignable(...);
620 template <typename _Tp, typename _Td>
621 struct __can_assign {
622 static const bool value =
623 sizeof(__test_atomic_assignable<_Tp, _Td>(1)) == sizeof(char);
626 static inline _LIBCPP_CONSTEXPR int __to_gcc_order(memory_order __order) {
627 // Avoid switch statement to make this a constexpr.
628 return __order == memory_order_relaxed ? __ATOMIC_RELAXED:
629 (__order == memory_order_acquire ? __ATOMIC_ACQUIRE:
630 (__order == memory_order_release ? __ATOMIC_RELEASE:
631 (__order == memory_order_seq_cst ? __ATOMIC_SEQ_CST:
632 (__order == memory_order_acq_rel ? __ATOMIC_ACQ_REL:
633 __ATOMIC_CONSUME))));
636 static inline _LIBCPP_CONSTEXPR int __to_gcc_failure_order(memory_order __order) {
637 // Avoid switch statement to make this a constexpr.
638 return __order == memory_order_relaxed ? __ATOMIC_RELAXED:
639 (__order == memory_order_acquire ? __ATOMIC_ACQUIRE:
640 (__order == memory_order_release ? __ATOMIC_RELAXED:
641 (__order == memory_order_seq_cst ? __ATOMIC_SEQ_CST:
642 (__order == memory_order_acq_rel ? __ATOMIC_ACQUIRE:
643 __ATOMIC_CONSUME))));
646 } // namespace __gcc_atomic
648 template <typename _Tp>
651 __gcc_atomic::__can_assign<volatile _Atomic(_Tp)*, _Tp>::value>::type
652 __c11_atomic_init(volatile _Atomic(_Tp)* __a, _Tp __val) {
653 __a->__a_value = __val;
656 template <typename _Tp>
659 !__gcc_atomic::__can_assign<volatile _Atomic(_Tp)*, _Tp>::value &&
660 __gcc_atomic::__can_assign< _Atomic(_Tp)*, _Tp>::value>::type
661 __c11_atomic_init(volatile _Atomic(_Tp)* __a, _Tp __val) {
662 // [atomics.types.generic]p1 guarantees _Tp is trivially copyable. Because
663 // the default operator= in an object is not volatile, a byte-by-byte copy
665 volatile char* to = reinterpret_cast<volatile char*>(&__a->__a_value);
666 volatile char* end = to + sizeof(_Tp);
667 char* from = reinterpret_cast<char*>(&__val);
673 template <typename _Tp>
674 static inline void __c11_atomic_init(_Atomic(_Tp)* __a, _Tp __val) {
675 __a->__a_value = __val;
678 static inline void __c11_atomic_thread_fence(memory_order __order) {
679 __atomic_thread_fence(__gcc_atomic::__to_gcc_order(__order));
682 static inline void __c11_atomic_signal_fence(memory_order __order) {
683 __atomic_signal_fence(__gcc_atomic::__to_gcc_order(__order));
686 template <typename _Tp>
687 static inline void __c11_atomic_store(volatile _Atomic(_Tp)* __a, _Tp __val,
688 memory_order __order) {
689 return __atomic_store(&__a->__a_value, &__val,
690 __gcc_atomic::__to_gcc_order(__order));
693 template <typename _Tp>
694 static inline void __c11_atomic_store(_Atomic(_Tp)* __a, _Tp __val,
695 memory_order __order) {
696 __atomic_store(&__a->__a_value, &__val,
697 __gcc_atomic::__to_gcc_order(__order));
700 template <typename _Tp>
701 static inline _Tp __c11_atomic_load(const volatile _Atomic(_Tp)* __a,
702 memory_order __order) {
704 __atomic_load(&__a->__a_value, &__ret,
705 __gcc_atomic::__to_gcc_order(__order));
709 template <typename _Tp>
710 static inline _Tp __c11_atomic_load(const _Atomic(_Tp)* __a, memory_order __order) {
712 __atomic_load(&__a->__a_value, &__ret,
713 __gcc_atomic::__to_gcc_order(__order));
717 template <typename _Tp>
718 static inline _Tp __c11_atomic_exchange(volatile _Atomic(_Tp)* __a,
719 _Tp __value, memory_order __order) {
721 __atomic_exchange(&__a->__a_value, &__value, &__ret,
722 __gcc_atomic::__to_gcc_order(__order));
726 template <typename _Tp>
727 static inline _Tp __c11_atomic_exchange(_Atomic(_Tp)* __a, _Tp __value,
728 memory_order __order) {
730 __atomic_exchange(&__a->__a_value, &__value, &__ret,
731 __gcc_atomic::__to_gcc_order(__order));
735 template <typename _Tp>
736 static inline bool __c11_atomic_compare_exchange_strong(
737 volatile _Atomic(_Tp)* __a, _Tp* __expected, _Tp __value,
738 memory_order __success, memory_order __failure) {
739 return __atomic_compare_exchange(&__a->__a_value, __expected, &__value,
741 __gcc_atomic::__to_gcc_order(__success),
742 __gcc_atomic::__to_gcc_failure_order(__failure));
745 template <typename _Tp>
746 static inline bool __c11_atomic_compare_exchange_strong(
747 _Atomic(_Tp)* __a, _Tp* __expected, _Tp __value, memory_order __success,
748 memory_order __failure) {
749 return __atomic_compare_exchange(&__a->__a_value, __expected, &__value,
751 __gcc_atomic::__to_gcc_order(__success),
752 __gcc_atomic::__to_gcc_failure_order(__failure));
755 template <typename _Tp>
756 static inline bool __c11_atomic_compare_exchange_weak(
757 volatile _Atomic(_Tp)* __a, _Tp* __expected, _Tp __value,
758 memory_order __success, memory_order __failure) {
759 return __atomic_compare_exchange(&__a->__a_value, __expected, &__value,
761 __gcc_atomic::__to_gcc_order(__success),
762 __gcc_atomic::__to_gcc_failure_order(__failure));
765 template <typename _Tp>
766 static inline bool __c11_atomic_compare_exchange_weak(
767 _Atomic(_Tp)* __a, _Tp* __expected, _Tp __value, memory_order __success,
768 memory_order __failure) {
769 return __atomic_compare_exchange(&__a->__a_value, __expected, &__value,
771 __gcc_atomic::__to_gcc_order(__success),
772 __gcc_atomic::__to_gcc_failure_order(__failure));
775 template <typename _Tp>
776 struct __skip_amt { enum {value = 1}; };
778 template <typename _Tp>
779 struct __skip_amt<_Tp*> { enum {value = sizeof(_Tp)}; };
781 // FIXME: Haven't figured out what the spec says about using arrays with
782 // atomic_fetch_add. Force a failure rather than creating bad behavior.
783 template <typename _Tp>
784 struct __skip_amt<_Tp[]> { };
785 template <typename _Tp, int n>
786 struct __skip_amt<_Tp[n]> { };
788 template <typename _Tp, typename _Td>
789 static inline _Tp __c11_atomic_fetch_add(volatile _Atomic(_Tp)* __a,
790 _Td __delta, memory_order __order) {
791 return __atomic_fetch_add(&__a->__a_value, __delta * __skip_amt<_Tp>::value,
792 __gcc_atomic::__to_gcc_order(__order));
795 template <typename _Tp, typename _Td>
796 static inline _Tp __c11_atomic_fetch_add(_Atomic(_Tp)* __a, _Td __delta,
797 memory_order __order) {
798 return __atomic_fetch_add(&__a->__a_value, __delta * __skip_amt<_Tp>::value,
799 __gcc_atomic::__to_gcc_order(__order));
802 template <typename _Tp, typename _Td>
803 static inline _Tp __c11_atomic_fetch_sub(volatile _Atomic(_Tp)* __a,
804 _Td __delta, memory_order __order) {
805 return __atomic_fetch_sub(&__a->__a_value, __delta * __skip_amt<_Tp>::value,
806 __gcc_atomic::__to_gcc_order(__order));
809 template <typename _Tp, typename _Td>
810 static inline _Tp __c11_atomic_fetch_sub(_Atomic(_Tp)* __a, _Td __delta,
811 memory_order __order) {
812 return __atomic_fetch_sub(&__a->__a_value, __delta * __skip_amt<_Tp>::value,
813 __gcc_atomic::__to_gcc_order(__order));
816 template <typename _Tp>
817 static inline _Tp __c11_atomic_fetch_and(volatile _Atomic(_Tp)* __a,
818 _Tp __pattern, memory_order __order) {
819 return __atomic_fetch_and(&__a->__a_value, __pattern,
820 __gcc_atomic::__to_gcc_order(__order));
823 template <typename _Tp>
824 static inline _Tp __c11_atomic_fetch_and(_Atomic(_Tp)* __a,
825 _Tp __pattern, memory_order __order) {
826 return __atomic_fetch_and(&__a->__a_value, __pattern,
827 __gcc_atomic::__to_gcc_order(__order));
830 template <typename _Tp>
831 static inline _Tp __c11_atomic_fetch_or(volatile _Atomic(_Tp)* __a,
832 _Tp __pattern, memory_order __order) {
833 return __atomic_fetch_or(&__a->__a_value, __pattern,
834 __gcc_atomic::__to_gcc_order(__order));
837 template <typename _Tp>
838 static inline _Tp __c11_atomic_fetch_or(_Atomic(_Tp)* __a, _Tp __pattern,
839 memory_order __order) {
840 return __atomic_fetch_or(&__a->__a_value, __pattern,
841 __gcc_atomic::__to_gcc_order(__order));
844 template <typename _Tp>
845 static inline _Tp __c11_atomic_fetch_xor(volatile _Atomic(_Tp)* __a,
846 _Tp __pattern, memory_order __order) {
847 return __atomic_fetch_xor(&__a->__a_value, __pattern,
848 __gcc_atomic::__to_gcc_order(__order));
851 template <typename _Tp>
852 static inline _Tp __c11_atomic_fetch_xor(_Atomic(_Tp)* __a, _Tp __pattern,
853 memory_order __order) {
854 return __atomic_fetch_xor(&__a->__a_value, __pattern,
855 __gcc_atomic::__to_gcc_order(__order));
857 #endif // _LIBCPP_HAS_GCC_ATOMIC_IMP
860 inline _LIBCPP_INLINE_VISIBILITY
862 kill_dependency(_Tp __y) _NOEXCEPT
867 #if defined(__CLANG_ATOMIC_BOOL_LOCK_FREE)
868 # define ATOMIC_BOOL_LOCK_FREE __CLANG_ATOMIC_BOOL_LOCK_FREE
869 # define ATOMIC_CHAR_LOCK_FREE __CLANG_ATOMIC_CHAR_LOCK_FREE
870 # define ATOMIC_CHAR16_T_LOCK_FREE __CLANG_ATOMIC_CHAR16_T_LOCK_FREE
871 # define ATOMIC_CHAR32_T_LOCK_FREE __CLANG_ATOMIC_CHAR32_T_LOCK_FREE
872 # define ATOMIC_WCHAR_T_LOCK_FREE __CLANG_ATOMIC_WCHAR_T_LOCK_FREE
873 # define ATOMIC_SHORT_LOCK_FREE __CLANG_ATOMIC_SHORT_LOCK_FREE
874 # define ATOMIC_INT_LOCK_FREE __CLANG_ATOMIC_INT_LOCK_FREE
875 # define ATOMIC_LONG_LOCK_FREE __CLANG_ATOMIC_LONG_LOCK_FREE
876 # define ATOMIC_LLONG_LOCK_FREE __CLANG_ATOMIC_LLONG_LOCK_FREE
877 # define ATOMIC_POINTER_LOCK_FREE __CLANG_ATOMIC_POINTER_LOCK_FREE
879 # define ATOMIC_BOOL_LOCK_FREE __GCC_ATOMIC_BOOL_LOCK_FREE
880 # define ATOMIC_CHAR_LOCK_FREE __GCC_ATOMIC_CHAR_LOCK_FREE
881 # define ATOMIC_CHAR16_T_LOCK_FREE __GCC_ATOMIC_CHAR16_T_LOCK_FREE
882 # define ATOMIC_CHAR32_T_LOCK_FREE __GCC_ATOMIC_CHAR32_T_LOCK_FREE
883 # define ATOMIC_WCHAR_T_LOCK_FREE __GCC_ATOMIC_WCHAR_T_LOCK_FREE
884 # define ATOMIC_SHORT_LOCK_FREE __GCC_ATOMIC_SHORT_LOCK_FREE
885 # define ATOMIC_INT_LOCK_FREE __GCC_ATOMIC_INT_LOCK_FREE
886 # define ATOMIC_LONG_LOCK_FREE __GCC_ATOMIC_LONG_LOCK_FREE
887 # define ATOMIC_LLONG_LOCK_FREE __GCC_ATOMIC_LLONG_LOCK_FREE
888 # define ATOMIC_POINTER_LOCK_FREE __GCC_ATOMIC_POINTER_LOCK_FREE
893 template <class _Tp, bool = is_integral<_Tp>::value && !is_same<_Tp, bool>::value>
894 struct __atomic_base // false
896 mutable _Atomic(_Tp) __a_;
898 #if defined(__cpp_lib_atomic_is_always_lock_free)
899 static _LIBCPP_CONSTEXPR bool is_always_lock_free = __atomic_always_lock_free(sizeof(__a_), 0);
902 _LIBCPP_INLINE_VISIBILITY
903 bool is_lock_free() const volatile _NOEXCEPT
905 #if defined(_LIBCPP_HAS_C_ATOMIC_IMP)
906 return __c11_atomic_is_lock_free(sizeof(_Tp));
908 return __atomic_is_lock_free(sizeof(_Tp), 0);
911 _LIBCPP_INLINE_VISIBILITY
912 bool is_lock_free() const _NOEXCEPT
913 {return static_cast<__atomic_base const volatile*>(this)->is_lock_free();}
914 _LIBCPP_INLINE_VISIBILITY
915 void store(_Tp __d, memory_order __m = memory_order_seq_cst) volatile _NOEXCEPT
916 _LIBCPP_CHECK_STORE_MEMORY_ORDER(__m)
917 {__c11_atomic_store(&__a_, __d, __m);}
918 _LIBCPP_INLINE_VISIBILITY
919 void store(_Tp __d, memory_order __m = memory_order_seq_cst) _NOEXCEPT
920 _LIBCPP_CHECK_STORE_MEMORY_ORDER(__m)
921 {__c11_atomic_store(&__a_, __d, __m);}
922 _LIBCPP_INLINE_VISIBILITY
923 _Tp load(memory_order __m = memory_order_seq_cst) const volatile _NOEXCEPT
924 _LIBCPP_CHECK_LOAD_MEMORY_ORDER(__m)
925 {return __c11_atomic_load(&__a_, __m);}
926 _LIBCPP_INLINE_VISIBILITY
927 _Tp load(memory_order __m = memory_order_seq_cst) const _NOEXCEPT
928 _LIBCPP_CHECK_LOAD_MEMORY_ORDER(__m)
929 {return __c11_atomic_load(&__a_, __m);}
930 _LIBCPP_INLINE_VISIBILITY
931 operator _Tp() const volatile _NOEXCEPT {return load();}
932 _LIBCPP_INLINE_VISIBILITY
933 operator _Tp() const _NOEXCEPT {return load();}
934 _LIBCPP_INLINE_VISIBILITY
935 _Tp exchange(_Tp __d, memory_order __m = memory_order_seq_cst) volatile _NOEXCEPT
936 {return __c11_atomic_exchange(&__a_, __d, __m);}
937 _LIBCPP_INLINE_VISIBILITY
938 _Tp exchange(_Tp __d, memory_order __m = memory_order_seq_cst) _NOEXCEPT
939 {return __c11_atomic_exchange(&__a_, __d, __m);}
940 _LIBCPP_INLINE_VISIBILITY
941 bool compare_exchange_weak(_Tp& __e, _Tp __d,
942 memory_order __s, memory_order __f) volatile _NOEXCEPT
943 _LIBCPP_CHECK_EXCHANGE_MEMORY_ORDER(__s, __f)
944 {return __c11_atomic_compare_exchange_weak(&__a_, &__e, __d, __s, __f);}
945 _LIBCPP_INLINE_VISIBILITY
946 bool compare_exchange_weak(_Tp& __e, _Tp __d,
947 memory_order __s, memory_order __f) _NOEXCEPT
948 _LIBCPP_CHECK_EXCHANGE_MEMORY_ORDER(__s, __f)
949 {return __c11_atomic_compare_exchange_weak(&__a_, &__e, __d, __s, __f);}
950 _LIBCPP_INLINE_VISIBILITY
951 bool compare_exchange_strong(_Tp& __e, _Tp __d,
952 memory_order __s, memory_order __f) volatile _NOEXCEPT
953 _LIBCPP_CHECK_EXCHANGE_MEMORY_ORDER(__s, __f)
954 {return __c11_atomic_compare_exchange_strong(&__a_, &__e, __d, __s, __f);}
955 _LIBCPP_INLINE_VISIBILITY
956 bool compare_exchange_strong(_Tp& __e, _Tp __d,
957 memory_order __s, memory_order __f) _NOEXCEPT
958 _LIBCPP_CHECK_EXCHANGE_MEMORY_ORDER(__s, __f)
959 {return __c11_atomic_compare_exchange_strong(&__a_, &__e, __d, __s, __f);}
960 _LIBCPP_INLINE_VISIBILITY
961 bool compare_exchange_weak(_Tp& __e, _Tp __d,
962 memory_order __m = memory_order_seq_cst) volatile _NOEXCEPT
963 {return __c11_atomic_compare_exchange_weak(&__a_, &__e, __d, __m, __m);}
964 _LIBCPP_INLINE_VISIBILITY
965 bool compare_exchange_weak(_Tp& __e, _Tp __d,
966 memory_order __m = memory_order_seq_cst) _NOEXCEPT
967 {return __c11_atomic_compare_exchange_weak(&__a_, &__e, __d, __m, __m);}
968 _LIBCPP_INLINE_VISIBILITY
969 bool compare_exchange_strong(_Tp& __e, _Tp __d,
970 memory_order __m = memory_order_seq_cst) volatile _NOEXCEPT
971 {return __c11_atomic_compare_exchange_strong(&__a_, &__e, __d, __m, __m);}
972 _LIBCPP_INLINE_VISIBILITY
973 bool compare_exchange_strong(_Tp& __e, _Tp __d,
974 memory_order __m = memory_order_seq_cst) _NOEXCEPT
975 {return __c11_atomic_compare_exchange_strong(&__a_, &__e, __d, __m, __m);}
977 _LIBCPP_INLINE_VISIBILITY
978 #ifndef _LIBCPP_CXX03_LANG
979 __atomic_base() _NOEXCEPT = default;
981 __atomic_base() _NOEXCEPT : __a_() {}
982 #endif // _LIBCPP_CXX03_LANG
984 _LIBCPP_INLINE_VISIBILITY
985 _LIBCPP_CONSTEXPR __atomic_base(_Tp __d) _NOEXCEPT : __a_(__d) {}
986 #ifndef _LIBCPP_CXX03_LANG
987 __atomic_base(const __atomic_base&) = delete;
988 __atomic_base& operator=(const __atomic_base&) = delete;
989 __atomic_base& operator=(const __atomic_base&) volatile = delete;
992 __atomic_base(const __atomic_base&);
993 __atomic_base& operator=(const __atomic_base&);
994 __atomic_base& operator=(const __atomic_base&) volatile;
998 #if defined(__cpp_lib_atomic_is_always_lock_free)
999 template <class _Tp, bool __b>
1000 _LIBCPP_CONSTEXPR bool __atomic_base<_Tp, __b>::is_always_lock_free;
1005 template <class _Tp>
1006 struct __atomic_base<_Tp, true>
1007 : public __atomic_base<_Tp, false>
1009 typedef __atomic_base<_Tp, false> __base;
1010 _LIBCPP_INLINE_VISIBILITY
1011 __atomic_base() _NOEXCEPT _LIBCPP_DEFAULT
1012 _LIBCPP_INLINE_VISIBILITY
1013 _LIBCPP_CONSTEXPR __atomic_base(_Tp __d) _NOEXCEPT : __base(__d) {}
1015 _LIBCPP_INLINE_VISIBILITY
1016 _Tp fetch_add(_Tp __op, memory_order __m = memory_order_seq_cst) volatile _NOEXCEPT
1017 {return __c11_atomic_fetch_add(&this->__a_, __op, __m);}
1018 _LIBCPP_INLINE_VISIBILITY
1019 _Tp fetch_add(_Tp __op, memory_order __m = memory_order_seq_cst) _NOEXCEPT
1020 {return __c11_atomic_fetch_add(&this->__a_, __op, __m);}
1021 _LIBCPP_INLINE_VISIBILITY
1022 _Tp fetch_sub(_Tp __op, memory_order __m = memory_order_seq_cst) volatile _NOEXCEPT
1023 {return __c11_atomic_fetch_sub(&this->__a_, __op, __m);}
1024 _LIBCPP_INLINE_VISIBILITY
1025 _Tp fetch_sub(_Tp __op, memory_order __m = memory_order_seq_cst) _NOEXCEPT
1026 {return __c11_atomic_fetch_sub(&this->__a_, __op, __m);}
1027 _LIBCPP_INLINE_VISIBILITY
1028 _Tp fetch_and(_Tp __op, memory_order __m = memory_order_seq_cst) volatile _NOEXCEPT
1029 {return __c11_atomic_fetch_and(&this->__a_, __op, __m);}
1030 _LIBCPP_INLINE_VISIBILITY
1031 _Tp fetch_and(_Tp __op, memory_order __m = memory_order_seq_cst) _NOEXCEPT
1032 {return __c11_atomic_fetch_and(&this->__a_, __op, __m);}
1033 _LIBCPP_INLINE_VISIBILITY
1034 _Tp fetch_or(_Tp __op, memory_order __m = memory_order_seq_cst) volatile _NOEXCEPT
1035 {return __c11_atomic_fetch_or(&this->__a_, __op, __m);}
1036 _LIBCPP_INLINE_VISIBILITY
1037 _Tp fetch_or(_Tp __op, memory_order __m = memory_order_seq_cst) _NOEXCEPT
1038 {return __c11_atomic_fetch_or(&this->__a_, __op, __m);}
1039 _LIBCPP_INLINE_VISIBILITY
1040 _Tp fetch_xor(_Tp __op, memory_order __m = memory_order_seq_cst) volatile _NOEXCEPT
1041 {return __c11_atomic_fetch_xor(&this->__a_, __op, __m);}
1042 _LIBCPP_INLINE_VISIBILITY
1043 _Tp fetch_xor(_Tp __op, memory_order __m = memory_order_seq_cst) _NOEXCEPT
1044 {return __c11_atomic_fetch_xor(&this->__a_, __op, __m);}
1046 _LIBCPP_INLINE_VISIBILITY
1047 _Tp operator++(int) volatile _NOEXCEPT {return fetch_add(_Tp(1));}
1048 _LIBCPP_INLINE_VISIBILITY
1049 _Tp operator++(int) _NOEXCEPT {return fetch_add(_Tp(1));}
1050 _LIBCPP_INLINE_VISIBILITY
1051 _Tp operator--(int) volatile _NOEXCEPT {return fetch_sub(_Tp(1));}
1052 _LIBCPP_INLINE_VISIBILITY
1053 _Tp operator--(int) _NOEXCEPT {return fetch_sub(_Tp(1));}
1054 _LIBCPP_INLINE_VISIBILITY
1055 _Tp operator++() volatile _NOEXCEPT {return fetch_add(_Tp(1)) + _Tp(1);}
1056 _LIBCPP_INLINE_VISIBILITY
1057 _Tp operator++() _NOEXCEPT {return fetch_add(_Tp(1)) + _Tp(1);}
1058 _LIBCPP_INLINE_VISIBILITY
1059 _Tp operator--() volatile _NOEXCEPT {return fetch_sub(_Tp(1)) - _Tp(1);}
1060 _LIBCPP_INLINE_VISIBILITY
1061 _Tp operator--() _NOEXCEPT {return fetch_sub(_Tp(1)) - _Tp(1);}
1062 _LIBCPP_INLINE_VISIBILITY
1063 _Tp operator+=(_Tp __op) volatile _NOEXCEPT {return fetch_add(__op) + __op;}
1064 _LIBCPP_INLINE_VISIBILITY
1065 _Tp operator+=(_Tp __op) _NOEXCEPT {return fetch_add(__op) + __op;}
1066 _LIBCPP_INLINE_VISIBILITY
1067 _Tp operator-=(_Tp __op) volatile _NOEXCEPT {return fetch_sub(__op) - __op;}
1068 _LIBCPP_INLINE_VISIBILITY
1069 _Tp operator-=(_Tp __op) _NOEXCEPT {return fetch_sub(__op) - __op;}
1070 _LIBCPP_INLINE_VISIBILITY
1071 _Tp operator&=(_Tp __op) volatile _NOEXCEPT {return fetch_and(__op) & __op;}
1072 _LIBCPP_INLINE_VISIBILITY
1073 _Tp operator&=(_Tp __op) _NOEXCEPT {return fetch_and(__op) & __op;}
1074 _LIBCPP_INLINE_VISIBILITY
1075 _Tp operator|=(_Tp __op) volatile _NOEXCEPT {return fetch_or(__op) | __op;}
1076 _LIBCPP_INLINE_VISIBILITY
1077 _Tp operator|=(_Tp __op) _NOEXCEPT {return fetch_or(__op) | __op;}
1078 _LIBCPP_INLINE_VISIBILITY
1079 _Tp operator^=(_Tp __op) volatile _NOEXCEPT {return fetch_xor(__op) ^ __op;}
1080 _LIBCPP_INLINE_VISIBILITY
1081 _Tp operator^=(_Tp __op) _NOEXCEPT {return fetch_xor(__op) ^ __op;}
1086 template <class _Tp>
1088 : public __atomic_base<_Tp>
1090 typedef __atomic_base<_Tp> __base;
1091 _LIBCPP_INLINE_VISIBILITY
1092 atomic() _NOEXCEPT _LIBCPP_DEFAULT
1093 _LIBCPP_INLINE_VISIBILITY
1094 _LIBCPP_CONSTEXPR atomic(_Tp __d) _NOEXCEPT : __base(__d) {}
1096 _LIBCPP_INLINE_VISIBILITY
1097 _Tp operator=(_Tp __d) volatile _NOEXCEPT
1098 {__base::store(__d); return __d;}
1099 _LIBCPP_INLINE_VISIBILITY
1100 _Tp operator=(_Tp __d) _NOEXCEPT
1101 {__base::store(__d); return __d;}
1106 template <class _Tp>
1108 : public __atomic_base<_Tp*>
1110 typedef __atomic_base<_Tp*> __base;
1111 _LIBCPP_INLINE_VISIBILITY
1112 atomic() _NOEXCEPT _LIBCPP_DEFAULT
1113 _LIBCPP_INLINE_VISIBILITY
1114 _LIBCPP_CONSTEXPR atomic(_Tp* __d) _NOEXCEPT : __base(__d) {}
1116 _LIBCPP_INLINE_VISIBILITY
1117 _Tp* operator=(_Tp* __d) volatile _NOEXCEPT
1118 {__base::store(__d); return __d;}
1119 _LIBCPP_INLINE_VISIBILITY
1120 _Tp* operator=(_Tp* __d) _NOEXCEPT
1121 {__base::store(__d); return __d;}
1123 _LIBCPP_INLINE_VISIBILITY
1124 _Tp* fetch_add(ptrdiff_t __op, memory_order __m = memory_order_seq_cst)
1126 {return __c11_atomic_fetch_add(&this->__a_, __op, __m);}
1127 _LIBCPP_INLINE_VISIBILITY
1128 _Tp* fetch_add(ptrdiff_t __op, memory_order __m = memory_order_seq_cst) _NOEXCEPT
1129 {return __c11_atomic_fetch_add(&this->__a_, __op, __m);}
1130 _LIBCPP_INLINE_VISIBILITY
1131 _Tp* fetch_sub(ptrdiff_t __op, memory_order __m = memory_order_seq_cst)
1133 {return __c11_atomic_fetch_sub(&this->__a_, __op, __m);}
1134 _LIBCPP_INLINE_VISIBILITY
1135 _Tp* fetch_sub(ptrdiff_t __op, memory_order __m = memory_order_seq_cst) _NOEXCEPT
1136 {return __c11_atomic_fetch_sub(&this->__a_, __op, __m);}
1138 _LIBCPP_INLINE_VISIBILITY
1139 _Tp* operator++(int) volatile _NOEXCEPT {return fetch_add(1);}
1140 _LIBCPP_INLINE_VISIBILITY
1141 _Tp* operator++(int) _NOEXCEPT {return fetch_add(1);}
1142 _LIBCPP_INLINE_VISIBILITY
1143 _Tp* operator--(int) volatile _NOEXCEPT {return fetch_sub(1);}
1144 _LIBCPP_INLINE_VISIBILITY
1145 _Tp* operator--(int) _NOEXCEPT {return fetch_sub(1);}
1146 _LIBCPP_INLINE_VISIBILITY
1147 _Tp* operator++() volatile _NOEXCEPT {return fetch_add(1) + 1;}
1148 _LIBCPP_INLINE_VISIBILITY
1149 _Tp* operator++() _NOEXCEPT {return fetch_add(1) + 1;}
1150 _LIBCPP_INLINE_VISIBILITY
1151 _Tp* operator--() volatile _NOEXCEPT {return fetch_sub(1) - 1;}
1152 _LIBCPP_INLINE_VISIBILITY
1153 _Tp* operator--() _NOEXCEPT {return fetch_sub(1) - 1;}
1154 _LIBCPP_INLINE_VISIBILITY
1155 _Tp* operator+=(ptrdiff_t __op) volatile _NOEXCEPT {return fetch_add(__op) + __op;}
1156 _LIBCPP_INLINE_VISIBILITY
1157 _Tp* operator+=(ptrdiff_t __op) _NOEXCEPT {return fetch_add(__op) + __op;}
1158 _LIBCPP_INLINE_VISIBILITY
1159 _Tp* operator-=(ptrdiff_t __op) volatile _NOEXCEPT {return fetch_sub(__op) - __op;}
1160 _LIBCPP_INLINE_VISIBILITY
1161 _Tp* operator-=(ptrdiff_t __op) _NOEXCEPT {return fetch_sub(__op) - __op;}
1164 // atomic_is_lock_free
1166 template <class _Tp>
1167 inline _LIBCPP_INLINE_VISIBILITY
1169 atomic_is_lock_free(const volatile atomic<_Tp>* __o) _NOEXCEPT
1171 return __o->is_lock_free();
1174 template <class _Tp>
1175 inline _LIBCPP_INLINE_VISIBILITY
1177 atomic_is_lock_free(const atomic<_Tp>* __o) _NOEXCEPT
1179 return __o->is_lock_free();
1184 template <class _Tp>
1185 inline _LIBCPP_INLINE_VISIBILITY
1187 atomic_init(volatile atomic<_Tp>* __o, _Tp __d) _NOEXCEPT
1189 __c11_atomic_init(&__o->__a_, __d);
1192 template <class _Tp>
1193 inline _LIBCPP_INLINE_VISIBILITY
1195 atomic_init(atomic<_Tp>* __o, _Tp __d) _NOEXCEPT
1197 __c11_atomic_init(&__o->__a_, __d);
1202 template <class _Tp>
1203 inline _LIBCPP_INLINE_VISIBILITY
1205 atomic_store(volatile atomic<_Tp>* __o, _Tp __d) _NOEXCEPT
1210 template <class _Tp>
1211 inline _LIBCPP_INLINE_VISIBILITY
1213 atomic_store(atomic<_Tp>* __o, _Tp __d) _NOEXCEPT
1218 // atomic_store_explicit
1220 template <class _Tp>
1221 inline _LIBCPP_INLINE_VISIBILITY
1223 atomic_store_explicit(volatile atomic<_Tp>* __o, _Tp __d, memory_order __m) _NOEXCEPT
1224 _LIBCPP_CHECK_STORE_MEMORY_ORDER(__m)
1226 __o->store(__d, __m);
1229 template <class _Tp>
1230 inline _LIBCPP_INLINE_VISIBILITY
1232 atomic_store_explicit(atomic<_Tp>* __o, _Tp __d, memory_order __m) _NOEXCEPT
1233 _LIBCPP_CHECK_STORE_MEMORY_ORDER(__m)
1235 __o->store(__d, __m);
1240 template <class _Tp>
1241 inline _LIBCPP_INLINE_VISIBILITY
1243 atomic_load(const volatile atomic<_Tp>* __o) _NOEXCEPT
1248 template <class _Tp>
1249 inline _LIBCPP_INLINE_VISIBILITY
1251 atomic_load(const atomic<_Tp>* __o) _NOEXCEPT
1256 // atomic_load_explicit
1258 template <class _Tp>
1259 inline _LIBCPP_INLINE_VISIBILITY
1261 atomic_load_explicit(const volatile atomic<_Tp>* __o, memory_order __m) _NOEXCEPT
1262 _LIBCPP_CHECK_LOAD_MEMORY_ORDER(__m)
1264 return __o->load(__m);
1267 template <class _Tp>
1268 inline _LIBCPP_INLINE_VISIBILITY
1270 atomic_load_explicit(const atomic<_Tp>* __o, memory_order __m) _NOEXCEPT
1271 _LIBCPP_CHECK_LOAD_MEMORY_ORDER(__m)
1273 return __o->load(__m);
1278 template <class _Tp>
1279 inline _LIBCPP_INLINE_VISIBILITY
1281 atomic_exchange(volatile atomic<_Tp>* __o, _Tp __d) _NOEXCEPT
1283 return __o->exchange(__d);
1286 template <class _Tp>
1287 inline _LIBCPP_INLINE_VISIBILITY
1289 atomic_exchange(atomic<_Tp>* __o, _Tp __d) _NOEXCEPT
1291 return __o->exchange(__d);
1294 // atomic_exchange_explicit
1296 template <class _Tp>
1297 inline _LIBCPP_INLINE_VISIBILITY
1299 atomic_exchange_explicit(volatile atomic<_Tp>* __o, _Tp __d, memory_order __m) _NOEXCEPT
1301 return __o->exchange(__d, __m);
1304 template <class _Tp>
1305 inline _LIBCPP_INLINE_VISIBILITY
1307 atomic_exchange_explicit(atomic<_Tp>* __o, _Tp __d, memory_order __m) _NOEXCEPT
1309 return __o->exchange(__d, __m);
1312 // atomic_compare_exchange_weak
1314 template <class _Tp>
1315 inline _LIBCPP_INLINE_VISIBILITY
1317 atomic_compare_exchange_weak(volatile atomic<_Tp>* __o, _Tp* __e, _Tp __d) _NOEXCEPT
1319 return __o->compare_exchange_weak(*__e, __d);
1322 template <class _Tp>
1323 inline _LIBCPP_INLINE_VISIBILITY
1325 atomic_compare_exchange_weak(atomic<_Tp>* __o, _Tp* __e, _Tp __d) _NOEXCEPT
1327 return __o->compare_exchange_weak(*__e, __d);
1330 // atomic_compare_exchange_strong
1332 template <class _Tp>
1333 inline _LIBCPP_INLINE_VISIBILITY
1335 atomic_compare_exchange_strong(volatile atomic<_Tp>* __o, _Tp* __e, _Tp __d) _NOEXCEPT
1337 return __o->compare_exchange_strong(*__e, __d);
1340 template <class _Tp>
1341 inline _LIBCPP_INLINE_VISIBILITY
1343 atomic_compare_exchange_strong(atomic<_Tp>* __o, _Tp* __e, _Tp __d) _NOEXCEPT
1345 return __o->compare_exchange_strong(*__e, __d);
1348 // atomic_compare_exchange_weak_explicit
1350 template <class _Tp>
1351 inline _LIBCPP_INLINE_VISIBILITY
1353 atomic_compare_exchange_weak_explicit(volatile atomic<_Tp>* __o, _Tp* __e,
1355 memory_order __s, memory_order __f) _NOEXCEPT
1356 _LIBCPP_CHECK_EXCHANGE_MEMORY_ORDER(__s, __f)
1358 return __o->compare_exchange_weak(*__e, __d, __s, __f);
1361 template <class _Tp>
1362 inline _LIBCPP_INLINE_VISIBILITY
1364 atomic_compare_exchange_weak_explicit(atomic<_Tp>* __o, _Tp* __e, _Tp __d,
1365 memory_order __s, memory_order __f) _NOEXCEPT
1366 _LIBCPP_CHECK_EXCHANGE_MEMORY_ORDER(__s, __f)
1368 return __o->compare_exchange_weak(*__e, __d, __s, __f);
1371 // atomic_compare_exchange_strong_explicit
1373 template <class _Tp>
1374 inline _LIBCPP_INLINE_VISIBILITY
1376 atomic_compare_exchange_strong_explicit(volatile atomic<_Tp>* __o,
1378 memory_order __s, memory_order __f) _NOEXCEPT
1379 _LIBCPP_CHECK_EXCHANGE_MEMORY_ORDER(__s, __f)
1381 return __o->compare_exchange_strong(*__e, __d, __s, __f);
1384 template <class _Tp>
1385 inline _LIBCPP_INLINE_VISIBILITY
1387 atomic_compare_exchange_strong_explicit(atomic<_Tp>* __o, _Tp* __e,
1389 memory_order __s, memory_order __f) _NOEXCEPT
1390 _LIBCPP_CHECK_EXCHANGE_MEMORY_ORDER(__s, __f)
1392 return __o->compare_exchange_strong(*__e, __d, __s, __f);
1397 template <class _Tp>
1398 inline _LIBCPP_INLINE_VISIBILITY
1401 is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
1404 atomic_fetch_add(volatile atomic<_Tp>* __o, _Tp __op) _NOEXCEPT
1406 return __o->fetch_add(__op);
1409 template <class _Tp>
1410 inline _LIBCPP_INLINE_VISIBILITY
1413 is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
1416 atomic_fetch_add(atomic<_Tp>* __o, _Tp __op) _NOEXCEPT
1418 return __o->fetch_add(__op);
1421 template <class _Tp>
1422 inline _LIBCPP_INLINE_VISIBILITY
1424 atomic_fetch_add(volatile atomic<_Tp*>* __o, ptrdiff_t __op) _NOEXCEPT
1426 return __o->fetch_add(__op);
1429 template <class _Tp>
1430 inline _LIBCPP_INLINE_VISIBILITY
1432 atomic_fetch_add(atomic<_Tp*>* __o, ptrdiff_t __op) _NOEXCEPT
1434 return __o->fetch_add(__op);
1437 // atomic_fetch_add_explicit
1439 template <class _Tp>
1440 inline _LIBCPP_INLINE_VISIBILITY
1443 is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
1446 atomic_fetch_add_explicit(volatile atomic<_Tp>* __o, _Tp __op, memory_order __m) _NOEXCEPT
1448 return __o->fetch_add(__op, __m);
1451 template <class _Tp>
1452 inline _LIBCPP_INLINE_VISIBILITY
1455 is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
1458 atomic_fetch_add_explicit(atomic<_Tp>* __o, _Tp __op, memory_order __m) _NOEXCEPT
1460 return __o->fetch_add(__op, __m);
1463 template <class _Tp>
1464 inline _LIBCPP_INLINE_VISIBILITY
1466 atomic_fetch_add_explicit(volatile atomic<_Tp*>* __o, ptrdiff_t __op,
1467 memory_order __m) _NOEXCEPT
1469 return __o->fetch_add(__op, __m);
1472 template <class _Tp>
1473 inline _LIBCPP_INLINE_VISIBILITY
1475 atomic_fetch_add_explicit(atomic<_Tp*>* __o, ptrdiff_t __op, memory_order __m) _NOEXCEPT
1477 return __o->fetch_add(__op, __m);
1482 template <class _Tp>
1483 inline _LIBCPP_INLINE_VISIBILITY
1486 is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
1489 atomic_fetch_sub(volatile atomic<_Tp>* __o, _Tp __op) _NOEXCEPT
1491 return __o->fetch_sub(__op);
1494 template <class _Tp>
1495 inline _LIBCPP_INLINE_VISIBILITY
1498 is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
1501 atomic_fetch_sub(atomic<_Tp>* __o, _Tp __op) _NOEXCEPT
1503 return __o->fetch_sub(__op);
1506 template <class _Tp>
1507 inline _LIBCPP_INLINE_VISIBILITY
1509 atomic_fetch_sub(volatile atomic<_Tp*>* __o, ptrdiff_t __op) _NOEXCEPT
1511 return __o->fetch_sub(__op);
1514 template <class _Tp>
1515 inline _LIBCPP_INLINE_VISIBILITY
1517 atomic_fetch_sub(atomic<_Tp*>* __o, ptrdiff_t __op) _NOEXCEPT
1519 return __o->fetch_sub(__op);
1522 // atomic_fetch_sub_explicit
1524 template <class _Tp>
1525 inline _LIBCPP_INLINE_VISIBILITY
1528 is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
1531 atomic_fetch_sub_explicit(volatile atomic<_Tp>* __o, _Tp __op, memory_order __m) _NOEXCEPT
1533 return __o->fetch_sub(__op, __m);
1536 template <class _Tp>
1537 inline _LIBCPP_INLINE_VISIBILITY
1540 is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
1543 atomic_fetch_sub_explicit(atomic<_Tp>* __o, _Tp __op, memory_order __m) _NOEXCEPT
1545 return __o->fetch_sub(__op, __m);
1548 template <class _Tp>
1549 inline _LIBCPP_INLINE_VISIBILITY
1551 atomic_fetch_sub_explicit(volatile atomic<_Tp*>* __o, ptrdiff_t __op,
1552 memory_order __m) _NOEXCEPT
1554 return __o->fetch_sub(__op, __m);
1557 template <class _Tp>
1558 inline _LIBCPP_INLINE_VISIBILITY
1560 atomic_fetch_sub_explicit(atomic<_Tp*>* __o, ptrdiff_t __op, memory_order __m) _NOEXCEPT
1562 return __o->fetch_sub(__op, __m);
1567 template <class _Tp>
1568 inline _LIBCPP_INLINE_VISIBILITY
1571 is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
1574 atomic_fetch_and(volatile atomic<_Tp>* __o, _Tp __op) _NOEXCEPT
1576 return __o->fetch_and(__op);
1579 template <class _Tp>
1580 inline _LIBCPP_INLINE_VISIBILITY
1583 is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
1586 atomic_fetch_and(atomic<_Tp>* __o, _Tp __op) _NOEXCEPT
1588 return __o->fetch_and(__op);
1591 // atomic_fetch_and_explicit
1593 template <class _Tp>
1594 inline _LIBCPP_INLINE_VISIBILITY
1597 is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
1600 atomic_fetch_and_explicit(volatile atomic<_Tp>* __o, _Tp __op, memory_order __m) _NOEXCEPT
1602 return __o->fetch_and(__op, __m);
1605 template <class _Tp>
1606 inline _LIBCPP_INLINE_VISIBILITY
1609 is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
1612 atomic_fetch_and_explicit(atomic<_Tp>* __o, _Tp __op, memory_order __m) _NOEXCEPT
1614 return __o->fetch_and(__op, __m);
1619 template <class _Tp>
1620 inline _LIBCPP_INLINE_VISIBILITY
1623 is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
1626 atomic_fetch_or(volatile atomic<_Tp>* __o, _Tp __op) _NOEXCEPT
1628 return __o->fetch_or(__op);
1631 template <class _Tp>
1632 inline _LIBCPP_INLINE_VISIBILITY
1635 is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
1638 atomic_fetch_or(atomic<_Tp>* __o, _Tp __op) _NOEXCEPT
1640 return __o->fetch_or(__op);
1643 // atomic_fetch_or_explicit
1645 template <class _Tp>
1646 inline _LIBCPP_INLINE_VISIBILITY
1649 is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
1652 atomic_fetch_or_explicit(volatile atomic<_Tp>* __o, _Tp __op, memory_order __m) _NOEXCEPT
1654 return __o->fetch_or(__op, __m);
1657 template <class _Tp>
1658 inline _LIBCPP_INLINE_VISIBILITY
1661 is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
1664 atomic_fetch_or_explicit(atomic<_Tp>* __o, _Tp __op, memory_order __m) _NOEXCEPT
1666 return __o->fetch_or(__op, __m);
1671 template <class _Tp>
1672 inline _LIBCPP_INLINE_VISIBILITY
1675 is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
1678 atomic_fetch_xor(volatile atomic<_Tp>* __o, _Tp __op) _NOEXCEPT
1680 return __o->fetch_xor(__op);
1683 template <class _Tp>
1684 inline _LIBCPP_INLINE_VISIBILITY
1687 is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
1690 atomic_fetch_xor(atomic<_Tp>* __o, _Tp __op) _NOEXCEPT
1692 return __o->fetch_xor(__op);
1695 // atomic_fetch_xor_explicit
1697 template <class _Tp>
1698 inline _LIBCPP_INLINE_VISIBILITY
1701 is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
1704 atomic_fetch_xor_explicit(volatile atomic<_Tp>* __o, _Tp __op, memory_order __m) _NOEXCEPT
1706 return __o->fetch_xor(__op, __m);
1709 template <class _Tp>
1710 inline _LIBCPP_INLINE_VISIBILITY
1713 is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
1716 atomic_fetch_xor_explicit(atomic<_Tp>* __o, _Tp __op, memory_order __m) _NOEXCEPT
1718 return __o->fetch_xor(__op, __m);
1721 // flag type and operations
1723 typedef struct atomic_flag
1727 _LIBCPP_INLINE_VISIBILITY
1728 bool test_and_set(memory_order __m = memory_order_seq_cst) volatile _NOEXCEPT
1729 {return __c11_atomic_exchange(&__a_, true, __m);}
1730 _LIBCPP_INLINE_VISIBILITY
1731 bool test_and_set(memory_order __m = memory_order_seq_cst) _NOEXCEPT
1732 {return __c11_atomic_exchange(&__a_, true, __m);}
1733 _LIBCPP_INLINE_VISIBILITY
1734 void clear(memory_order __m = memory_order_seq_cst) volatile _NOEXCEPT
1735 {__c11_atomic_store(&__a_, false, __m);}
1736 _LIBCPP_INLINE_VISIBILITY
1737 void clear(memory_order __m = memory_order_seq_cst) _NOEXCEPT
1738 {__c11_atomic_store(&__a_, false, __m);}
1740 _LIBCPP_INLINE_VISIBILITY
1741 #ifndef _LIBCPP_CXX03_LANG
1742 atomic_flag() _NOEXCEPT = default;
1744 atomic_flag() _NOEXCEPT : __a_() {}
1745 #endif // _LIBCPP_CXX03_LANG
1747 _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR
1748 atomic_flag(bool __b) _NOEXCEPT : __a_(__b) {} // EXTENSION
1750 #ifndef _LIBCPP_CXX03_LANG
1751 atomic_flag(const atomic_flag&) = delete;
1752 atomic_flag& operator=(const atomic_flag&) = delete;
1753 atomic_flag& operator=(const atomic_flag&) volatile = delete;
1756 atomic_flag(const atomic_flag&);
1757 atomic_flag& operator=(const atomic_flag&);
1758 atomic_flag& operator=(const atomic_flag&) volatile;
1762 inline _LIBCPP_INLINE_VISIBILITY
1764 atomic_flag_test_and_set(volatile atomic_flag* __o) _NOEXCEPT
1766 return __o->test_and_set();
1769 inline _LIBCPP_INLINE_VISIBILITY
1771 atomic_flag_test_and_set(atomic_flag* __o) _NOEXCEPT
1773 return __o->test_and_set();
1776 inline _LIBCPP_INLINE_VISIBILITY
1778 atomic_flag_test_and_set_explicit(volatile atomic_flag* __o, memory_order __m) _NOEXCEPT
1780 return __o->test_and_set(__m);
1783 inline _LIBCPP_INLINE_VISIBILITY
1785 atomic_flag_test_and_set_explicit(atomic_flag* __o, memory_order __m) _NOEXCEPT
1787 return __o->test_and_set(__m);
1790 inline _LIBCPP_INLINE_VISIBILITY
1792 atomic_flag_clear(volatile atomic_flag* __o) _NOEXCEPT
1797 inline _LIBCPP_INLINE_VISIBILITY
1799 atomic_flag_clear(atomic_flag* __o) _NOEXCEPT
1804 inline _LIBCPP_INLINE_VISIBILITY
1806 atomic_flag_clear_explicit(volatile atomic_flag* __o, memory_order __m) _NOEXCEPT
1811 inline _LIBCPP_INLINE_VISIBILITY
1813 atomic_flag_clear_explicit(atomic_flag* __o, memory_order __m) _NOEXCEPT
1820 inline _LIBCPP_INLINE_VISIBILITY
1822 atomic_thread_fence(memory_order __m) _NOEXCEPT
1824 __c11_atomic_thread_fence(__m);
1827 inline _LIBCPP_INLINE_VISIBILITY
1829 atomic_signal_fence(memory_order __m) _NOEXCEPT
1831 __c11_atomic_signal_fence(__m);
1834 // Atomics for standard typedef types
1836 typedef atomic<bool> atomic_bool;
1837 typedef atomic<char> atomic_char;
1838 typedef atomic<signed char> atomic_schar;
1839 typedef atomic<unsigned char> atomic_uchar;
1840 typedef atomic<short> atomic_short;
1841 typedef atomic<unsigned short> atomic_ushort;
1842 typedef atomic<int> atomic_int;
1843 typedef atomic<unsigned int> atomic_uint;
1844 typedef atomic<long> atomic_long;
1845 typedef atomic<unsigned long> atomic_ulong;
1846 typedef atomic<long long> atomic_llong;
1847 typedef atomic<unsigned long long> atomic_ullong;
1848 typedef atomic<char16_t> atomic_char16_t;
1849 typedef atomic<char32_t> atomic_char32_t;
1850 typedef atomic<wchar_t> atomic_wchar_t;
1852 typedef atomic<int_least8_t> atomic_int_least8_t;
1853 typedef atomic<uint_least8_t> atomic_uint_least8_t;
1854 typedef atomic<int_least16_t> atomic_int_least16_t;
1855 typedef atomic<uint_least16_t> atomic_uint_least16_t;
1856 typedef atomic<int_least32_t> atomic_int_least32_t;
1857 typedef atomic<uint_least32_t> atomic_uint_least32_t;
1858 typedef atomic<int_least64_t> atomic_int_least64_t;
1859 typedef atomic<uint_least64_t> atomic_uint_least64_t;
1861 typedef atomic<int_fast8_t> atomic_int_fast8_t;
1862 typedef atomic<uint_fast8_t> atomic_uint_fast8_t;
1863 typedef atomic<int_fast16_t> atomic_int_fast16_t;
1864 typedef atomic<uint_fast16_t> atomic_uint_fast16_t;
1865 typedef atomic<int_fast32_t> atomic_int_fast32_t;
1866 typedef atomic<uint_fast32_t> atomic_uint_fast32_t;
1867 typedef atomic<int_fast64_t> atomic_int_fast64_t;
1868 typedef atomic<uint_fast64_t> atomic_uint_fast64_t;
1870 typedef atomic< int8_t> atomic_int8_t;
1871 typedef atomic<uint8_t> atomic_uint8_t;
1872 typedef atomic< int16_t> atomic_int16_t;
1873 typedef atomic<uint16_t> atomic_uint16_t;
1874 typedef atomic< int32_t> atomic_int32_t;
1875 typedef atomic<uint32_t> atomic_uint32_t;
1876 typedef atomic< int64_t> atomic_int64_t;
1877 typedef atomic<uint64_t> atomic_uint64_t;
1879 typedef atomic<intptr_t> atomic_intptr_t;
1880 typedef atomic<uintptr_t> atomic_uintptr_t;
1881 typedef atomic<size_t> atomic_size_t;
1882 typedef atomic<ptrdiff_t> atomic_ptrdiff_t;
1883 typedef atomic<intmax_t> atomic_intmax_t;
1884 typedef atomic<uintmax_t> atomic_uintmax_t;
1886 #define ATOMIC_FLAG_INIT {false}
1887 #define ATOMIC_VAR_INIT(__v) {__v}
1889 _LIBCPP_END_NAMESPACE_STD
1891 #endif // _LIBCPP_ATOMIC