]> CyberLeo.Net >> Repos - FreeBSD/FreeBSD.git/blob - sys/contrib/ck/include/gcc/x86_64/ck_pr.h
Import CK as of commit b19ed4c6a56ec93215ab567ba18ba61bf1cfbac8
[FreeBSD/FreeBSD.git] / sys / contrib / ck / include / gcc / x86_64 / ck_pr.h
1 /*
2  * Copyright 2009-2015 Samy Al Bahra.
3  * All rights reserved.
4  *
5  * Redistribution and use in source and binary forms, with or without
6  * modification, are permitted provided that the following conditions
7  * are met:
8  * 1. Redistributions of source code must retain the above copyright
9  *    notice, this list of conditions and the following disclaimer.
10  * 2. Redistributions in binary form must reproduce the above copyright
11  *    notice, this list of conditions and the following disclaimer in the
12  *    documentation and/or other materials provided with the distribution.
13  *
14  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
15  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
16  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
17  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
18  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
19  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
20  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
21  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
22  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
23  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
24  * SUCH DAMAGE.
25  */
26
27 #ifndef CK_PR_X86_64_H
28 #define CK_PR_X86_64_H
29
30 #ifndef CK_PR_H
31 #error Do not include this file directly, use ck_pr.h
32 #endif
33
34 #include <ck_cc.h>
35 #include <ck_md.h>
36 #include <ck_stdint.h>
37
38 /*
39  * The following represent supported atomic operations.
40  * These operations may be emulated.
41  */
42 #include "ck_f_pr.h"
43
44 /*
45  * Support for TSX extensions.
46  */
47 #ifdef CK_MD_RTM_ENABLE
48 #include "ck_pr_rtm.h"
49 #endif
50
51 /* Minimum requirements for the CK_PR interface are met. */
52 #define CK_F_PR
53
54 #ifdef CK_MD_UMP
55 #define CK_PR_LOCK_PREFIX
56 #else
57 #define CK_PR_LOCK_PREFIX "lock "
58 #endif
59
60 /*
61  * Prevent speculative execution in busy-wait loops (P4 <=) or "predefined
62  * delay".
63  */
64 CK_CC_INLINE static void
65 ck_pr_stall(void)
66 {
67         __asm__ __volatile__("pause" ::: "memory");
68         return;
69 }
70
71 #define CK_PR_FENCE(T, I)                               \
72         CK_CC_INLINE static void                        \
73         ck_pr_fence_strict_##T(void)                    \
74         {                                               \
75                 __asm__ __volatile__(I ::: "memory");   \
76         }
77
78 /* Atomic operations are always serializing. */
79 CK_PR_FENCE(atomic, "")
80 CK_PR_FENCE(atomic_store, "")
81 CK_PR_FENCE(atomic_load, "")
82 CK_PR_FENCE(store_atomic, "")
83 CK_PR_FENCE(load_atomic, "")
84
85 /* Traditional fence interface. */
86 CK_PR_FENCE(load, "lfence")
87 CK_PR_FENCE(load_store, "mfence")
88 CK_PR_FENCE(store, "sfence")
89 CK_PR_FENCE(store_load, "mfence")
90 CK_PR_FENCE(memory, "mfence")
91
92 /* Below are stdatomic-style fences. */
93
94 /*
95  * Provides load-store and store-store ordering. However, Intel specifies that
96  * the WC memory model is relaxed. It is likely an sfence *is* sufficient (in
97  * particular, stores are not re-ordered with respect to prior loads and it is
98  * really just the stores that are subject to re-ordering). However, we take
99  * the conservative route as the manuals are too ambiguous for my taste.
100  */
101 CK_PR_FENCE(release, "mfence")
102
103 /*
104  * Provides load-load and load-store ordering. The lfence instruction ensures
105  * all prior load operations are complete before any subsequent instructions
106  * actually begin execution. However, the manual also ends up going to describe
107  * WC memory as a relaxed model.
108  */
109 CK_PR_FENCE(acquire, "mfence")
110
111 CK_PR_FENCE(acqrel, "mfence")
112 CK_PR_FENCE(lock, "mfence")
113 CK_PR_FENCE(unlock, "mfence")
114
115 #undef CK_PR_FENCE
116
117 /*
118  * Read for ownership. Older compilers will generate the 32-bit
119  * 3DNow! variant which is binary compatible with x86-64 variant
120  * of prefetchw.
121  */
122 #ifndef CK_F_PR_RFO
123 #define CK_F_PR_RFO
124 CK_CC_INLINE static void
125 ck_pr_rfo(const void *m)
126 {
127
128         __asm__ __volatile__("prefetchw (%0)"
129             :
130             : "r" (m)
131             : "memory");
132
133         return;
134 }
135 #endif /* CK_F_PR_RFO */
136
137 /*
138  * Atomic fetch-and-store operations.
139  */
140 #define CK_PR_FAS(S, M, T, C, I)                                \
141         CK_CC_INLINE static T                                   \
142         ck_pr_fas_##S(M *target, T v)                           \
143         {                                                       \
144                 __asm__ __volatile__(I " %0, %1"                \
145                                         : "+m" (*(C *)target),  \
146                                           "+q" (v)              \
147                                         :                       \
148                                         : "memory");            \
149                 return v;                                       \
150         }
151
152 CK_PR_FAS(ptr, void, void *, char, "xchgq")
153
154 #define CK_PR_FAS_S(S, T, I) CK_PR_FAS(S, T, T, T, I)
155
156 #ifndef CK_PR_DISABLE_DOUBLE
157 CK_PR_FAS_S(double, double, "xchgq")
158 #endif
159 CK_PR_FAS_S(char, char, "xchgb")
160 CK_PR_FAS_S(uint, unsigned int, "xchgl")
161 CK_PR_FAS_S(int, int, "xchgl")
162 CK_PR_FAS_S(64, uint64_t, "xchgq")
163 CK_PR_FAS_S(32, uint32_t, "xchgl")
164 CK_PR_FAS_S(16, uint16_t, "xchgw")
165 CK_PR_FAS_S(8,  uint8_t,  "xchgb")
166
167 #undef CK_PR_FAS_S
168 #undef CK_PR_FAS
169
170 /*
171  * Atomic load-from-memory operations.
172  */
173 #define CK_PR_LOAD(S, M, T, C, I)                               \
174         CK_CC_INLINE static T                                   \
175         ck_pr_md_load_##S(const M *target)                      \
176         {                                                       \
177                 T r;                                            \
178                 __asm__ __volatile__(I " %1, %0"                \
179                     : "=q" (r)                                  \
180                     : "m"  (*(const C *)target)                 \
181                     : "memory");                                \
182                 return (r);                                     \
183         }
184
185 CK_PR_LOAD(ptr, void, void *, char, "movq")
186
187 #define CK_PR_LOAD_S(S, T, I) CK_PR_LOAD(S, T, T, T, I)
188
189 CK_PR_LOAD_S(char, char, "movb")
190 CK_PR_LOAD_S(uint, unsigned int, "movl")
191 CK_PR_LOAD_S(int, int, "movl")
192 #ifndef CK_PR_DISABLE_DOUBLE
193 CK_PR_LOAD_S(double, double, "movq")
194 #endif
195 CK_PR_LOAD_S(64, uint64_t, "movq")
196 CK_PR_LOAD_S(32, uint32_t, "movl")
197 CK_PR_LOAD_S(16, uint16_t, "movw")
198 CK_PR_LOAD_S(8,  uint8_t,  "movb")
199
200 #undef CK_PR_LOAD_S
201 #undef CK_PR_LOAD
202
203 CK_CC_INLINE static void
204 ck_pr_load_64_2(const uint64_t target[2], uint64_t v[2])
205 {
206         __asm__ __volatile__("movq %%rdx, %%rcx;"
207                              "movq %%rax, %%rbx;"
208                              CK_PR_LOCK_PREFIX "cmpxchg16b %2;"
209                                 : "=a" (v[0]),
210                                   "=d" (v[1])
211                                 : "m" (*(const uint64_t *)target)
212                                 : "rbx", "rcx", "memory", "cc");
213         return;
214 }
215
216 CK_CC_INLINE static void
217 ck_pr_load_ptr_2(const void *t, void *v)
218 {
219         ck_pr_load_64_2(CK_CPP_CAST(const uint64_t *, t),
220                         CK_CPP_CAST(uint64_t *, v));
221         return;
222 }
223
224 #define CK_PR_LOAD_2(S, W, T)                                                   \
225         CK_CC_INLINE static void                                                \
226         ck_pr_md_load_##S##_##W(const T t[2], T v[2])                           \
227         {                                                                       \
228                 ck_pr_load_64_2((const uint64_t *)(const void *)t,              \
229                                 (uint64_t *)(void *)v);                         \
230                 return;                                                         \
231         }
232
233 CK_PR_LOAD_2(char, 16, char)
234 CK_PR_LOAD_2(int, 4, int)
235 CK_PR_LOAD_2(uint, 4, unsigned int)
236 CK_PR_LOAD_2(32, 4, uint32_t)
237 CK_PR_LOAD_2(16, 8, uint16_t)
238 CK_PR_LOAD_2(8, 16, uint8_t)
239
240 #undef CK_PR_LOAD_2
241
242 /*
243  * Atomic store-to-memory operations.
244  */
245 #define CK_PR_STORE_IMM(S, M, T, C, I, K)                               \
246         CK_CC_INLINE static void                                        \
247         ck_pr_md_store_##S(M *target, T v)                              \
248         {                                                               \
249                 __asm__ __volatile__(I " %1, %0"                        \
250                                         : "=m" (*(C *)target)           \
251                                         : K "q" (v)                     \
252                                         : "memory");                    \
253                 return;                                                 \
254         }
255
256 #define CK_PR_STORE(S, M, T, C, I)                              \
257         CK_CC_INLINE static void                                \
258         ck_pr_md_store_##S(M *target, T v)                      \
259         {                                                       \
260                 __asm__ __volatile__(I " %1, %0"                \
261                                         : "=m" (*(C *)target)   \
262                                         : "q" (v)               \
263                                         : "memory");            \
264                 return;                                         \
265         }
266
267 CK_PR_STORE_IMM(ptr, void, const void *, char, "movq", CK_CC_IMM_U32)
268 #ifndef CK_PR_DISABLE_DOUBLE
269 CK_PR_STORE(double, double, double, double, "movq")
270 #endif
271
272 #define CK_PR_STORE_S(S, T, I, K) CK_PR_STORE_IMM(S, T, T, T, I, K)
273
274 CK_PR_STORE_S(char, char, "movb", CK_CC_IMM_S32)
275 CK_PR_STORE_S(int, int, "movl", CK_CC_IMM_S32)
276 CK_PR_STORE_S(uint, unsigned int, "movl", CK_CC_IMM_U32)
277 CK_PR_STORE_S(64, uint64_t, "movq", CK_CC_IMM_U32)
278 CK_PR_STORE_S(32, uint32_t, "movl", CK_CC_IMM_U32)
279 CK_PR_STORE_S(16, uint16_t, "movw", CK_CC_IMM_U32)
280 CK_PR_STORE_S(8,  uint8_t, "movb", CK_CC_IMM_U32)
281
282 #undef CK_PR_STORE_S
283 #undef CK_PR_STORE_IMM
284 #undef CK_PR_STORE
285
286 /*
287  * Atomic fetch-and-add operations.
288  */
289 #define CK_PR_FAA(S, M, T, C, I)                                        \
290         CK_CC_INLINE static T                                           \
291         ck_pr_faa_##S(M *target, T d)                                   \
292         {                                                               \
293                 __asm__ __volatile__(CK_PR_LOCK_PREFIX I " %1, %0"      \
294                                         : "+m" (*(C *)target),          \
295                                           "+q" (d)                      \
296                                         :                               \
297                                         : "memory", "cc");              \
298                 return (d);                                             \
299         }
300
301 CK_PR_FAA(ptr, void, uintptr_t, char, "xaddq")
302
303 #define CK_PR_FAA_S(S, T, I) CK_PR_FAA(S, T, T, T, I)
304
305 CK_PR_FAA_S(char, char, "xaddb")
306 CK_PR_FAA_S(uint, unsigned int, "xaddl")
307 CK_PR_FAA_S(int, int, "xaddl")
308 CK_PR_FAA_S(64, uint64_t, "xaddq")
309 CK_PR_FAA_S(32, uint32_t, "xaddl")
310 CK_PR_FAA_S(16, uint16_t, "xaddw")
311 CK_PR_FAA_S(8,  uint8_t,  "xaddb")
312
313 #undef CK_PR_FAA_S
314 #undef CK_PR_FAA
315
316 /*
317  * Atomic store-only unary operations.
318  */
319 #define CK_PR_UNARY(K, S, T, C, I)                              \
320         CK_PR_UNARY_R(K, S, T, C, I)                            \
321         CK_PR_UNARY_V(K, S, T, C, I)
322
323 #define CK_PR_UNARY_R(K, S, T, C, I)                            \
324         CK_CC_INLINE static void                                \
325         ck_pr_##K##_##S(T *target)                              \
326         {                                                       \
327                 __asm__ __volatile__(CK_PR_LOCK_PREFIX I " %0"  \
328                                         : "+m" (*(C *)target)   \
329                                         :                       \
330                                         : "memory", "cc");      \
331                 return;                                         \
332         }
333
334 #define CK_PR_UNARY_V(K, S, T, C, I)                                    \
335         CK_CC_INLINE static void                                        \
336         ck_pr_##K##_##S##_zero(T *target, bool *r)                      \
337         {                                                               \
338                 __asm__ __volatile__(CK_PR_LOCK_PREFIX I " %0; setz %1" \
339                                         : "+m" (*(C *)target),          \
340                                           "=m" (*r)                     \
341                                         :                               \
342                                         : "memory", "cc");              \
343                 return;                                                 \
344         }
345
346
347 #define CK_PR_UNARY_S(K, S, T, I) CK_PR_UNARY(K, S, T, T, I)
348
349 #define CK_PR_GENERATE(K)                               \
350         CK_PR_UNARY(K, ptr, void, char, #K "q")         \
351         CK_PR_UNARY_S(K, char, char, #K "b")            \
352         CK_PR_UNARY_S(K, int, int, #K "l")              \
353         CK_PR_UNARY_S(K, uint, unsigned int, #K "l")    \
354         CK_PR_UNARY_S(K, 64, uint64_t, #K "q")          \
355         CK_PR_UNARY_S(K, 32, uint32_t, #K "l")          \
356         CK_PR_UNARY_S(K, 16, uint16_t, #K "w")          \
357         CK_PR_UNARY_S(K, 8, uint8_t, #K "b")
358
359 CK_PR_GENERATE(inc)
360 CK_PR_GENERATE(dec)
361 CK_PR_GENERATE(neg)
362
363 /* not does not affect condition flags. */
364 #undef CK_PR_UNARY_V
365 #define CK_PR_UNARY_V(a, b, c, d, e)
366 CK_PR_GENERATE(not)
367
368 #undef CK_PR_GENERATE
369 #undef CK_PR_UNARY_S
370 #undef CK_PR_UNARY_V
371 #undef CK_PR_UNARY_R
372 #undef CK_PR_UNARY
373
374 /*
375  * Atomic store-only binary operations.
376  */
377 #define CK_PR_BINARY(K, S, M, T, C, I, O)                               \
378         CK_CC_INLINE static void                                        \
379         ck_pr_##K##_##S(M *target, T d)                                 \
380         {                                                               \
381                 __asm__ __volatile__(CK_PR_LOCK_PREFIX I " %1, %0"      \
382                                         : "+m" (*(C *)target)           \
383                                         : O "q" (d)                     \
384                                         : "memory", "cc");              \
385                 return;                                                 \
386         }
387
388 #define CK_PR_BINARY_S(K, S, T, I, O) CK_PR_BINARY(K, S, T, T, T, I, O)
389
390 #define CK_PR_GENERATE(K)                                                       \
391         CK_PR_BINARY(K, ptr, void, uintptr_t, char, #K "q", CK_CC_IMM_U32)      \
392         CK_PR_BINARY_S(K, char, char, #K "b", CK_CC_IMM_S32)                    \
393         CK_PR_BINARY_S(K, int, int, #K "l", CK_CC_IMM_S32)                      \
394         CK_PR_BINARY_S(K, uint, unsigned int, #K "l", CK_CC_IMM_U32)            \
395         CK_PR_BINARY_S(K, 64, uint64_t, #K "q", CK_CC_IMM_U32)                  \
396         CK_PR_BINARY_S(K, 32, uint32_t, #K "l", CK_CC_IMM_U32)                  \
397         CK_PR_BINARY_S(K, 16, uint16_t, #K "w", CK_CC_IMM_U32)                  \
398         CK_PR_BINARY_S(K, 8, uint8_t, #K "b", CK_CC_IMM_U32)
399
400 CK_PR_GENERATE(add)
401 CK_PR_GENERATE(sub)
402 CK_PR_GENERATE(and)
403 CK_PR_GENERATE(or)
404 CK_PR_GENERATE(xor)
405
406 #undef CK_PR_GENERATE
407 #undef CK_PR_BINARY_S
408 #undef CK_PR_BINARY
409
410 /*
411  * Atomic compare and swap.
412  */
413 #define CK_PR_CAS(S, M, T, C, I)                                                \
414         CK_CC_INLINE static bool                                                \
415         ck_pr_cas_##S(M *target, T compare, T set)                              \
416         {                                                                       \
417                 bool z;                                                         \
418                 __asm__ __volatile__(CK_PR_LOCK_PREFIX I " %2, %0; setz %1"     \
419                                         : "+m"  (*(C *)target),                 \
420                                           "=a"  (z)                             \
421                                         : "q"   (set),                          \
422                                           "a"   (compare)                       \
423                                         : "memory", "cc");                      \
424                 return z;                                                       \
425         }
426
427 CK_PR_CAS(ptr, void, void *, char, "cmpxchgq")
428
429 #define CK_PR_CAS_S(S, T, I) CK_PR_CAS(S, T, T, T, I)
430
431 CK_PR_CAS_S(char, char, "cmpxchgb")
432 CK_PR_CAS_S(int, int, "cmpxchgl")
433 CK_PR_CAS_S(uint, unsigned int, "cmpxchgl")
434 #ifndef CK_PR_DISABLE_DOUBLE
435 CK_PR_CAS_S(double, double, "cmpxchgq")
436 #endif
437 CK_PR_CAS_S(64, uint64_t, "cmpxchgq")
438 CK_PR_CAS_S(32, uint32_t, "cmpxchgl")
439 CK_PR_CAS_S(16, uint16_t, "cmpxchgw")
440 CK_PR_CAS_S(8,  uint8_t,  "cmpxchgb")
441
442 #undef CK_PR_CAS_S
443 #undef CK_PR_CAS
444
445 /*
446  * Compare and swap, set *v to old value of target.
447  */
448 #define CK_PR_CAS_O(S, M, T, C, I, R)                                           \
449         CK_CC_INLINE static bool                                                \
450         ck_pr_cas_##S##_value(M *target, T compare, T set, M *v)                \
451         {                                                                       \
452                 bool z;                                                         \
453                 __asm__ __volatile__(CK_PR_LOCK_PREFIX "cmpxchg" I " %3, %0;"   \
454                                      "mov %% " R ", %2;"                        \
455                                      "setz %1;"                                 \
456                                         : "+m"  (*(C *)target),                 \
457                                           "=a"  (z),                            \
458                                           "=m"  (*(C *)v)                       \
459                                         : "q"   (set),                          \
460                                           "a"   (compare)                       \
461                                         : "memory", "cc");                      \
462                 return z;                                                       \
463         }
464
465 CK_PR_CAS_O(ptr, void, void *, char, "q", "rax")
466
467 #define CK_PR_CAS_O_S(S, T, I, R)       \
468         CK_PR_CAS_O(S, T, T, T, I, R)
469
470 CK_PR_CAS_O_S(char, char, "b", "al")
471 CK_PR_CAS_O_S(int, int, "l", "eax")
472 CK_PR_CAS_O_S(uint, unsigned int, "l", "eax")
473 #ifndef CK_PR_DISABLE_DOUBLE
474 CK_PR_CAS_O_S(double, double, "q", "rax")
475 #endif
476 CK_PR_CAS_O_S(64, uint64_t, "q", "rax")
477 CK_PR_CAS_O_S(32, uint32_t, "l", "eax")
478 CK_PR_CAS_O_S(16, uint16_t, "w", "ax")
479 CK_PR_CAS_O_S(8,  uint8_t,  "b", "al")
480
481 #undef CK_PR_CAS_O_S
482 #undef CK_PR_CAS_O
483
484 /*
485  * Contrary to C-interface, alignment requirements are that of uint64_t[2].
486  */
487 CK_CC_INLINE static bool
488 ck_pr_cas_64_2(uint64_t target[2], uint64_t compare[2], uint64_t set[2])
489 {
490         bool z;
491
492         __asm__ __volatile__("movq 0(%4), %%rax;"
493                              "movq 8(%4), %%rdx;"
494                              CK_PR_LOCK_PREFIX "cmpxchg16b %0; setz %1"
495                                 : "+m" (*target),
496                                   "=q" (z)
497                                 : "b"  (set[0]),
498                                   "c"  (set[1]),
499                                   "q"  (compare)
500                                 : "memory", "cc", "%rax", "%rdx");
501         return z;
502 }
503
504 CK_CC_INLINE static bool
505 ck_pr_cas_ptr_2(void *t, void *c, void *s)
506 {
507         return ck_pr_cas_64_2(CK_CPP_CAST(uint64_t *, t),
508                               CK_CPP_CAST(uint64_t *, c),
509                               CK_CPP_CAST(uint64_t *, s));
510 }
511
512 CK_CC_INLINE static bool
513 ck_pr_cas_64_2_value(uint64_t target[2],
514                      uint64_t compare[2],
515                      uint64_t set[2],
516                      uint64_t v[2])
517 {
518         bool z;
519
520         __asm__ __volatile__(CK_PR_LOCK_PREFIX "cmpxchg16b %0;"
521                              "setz %3"
522                                 : "+m" (*target),
523                                   "=a" (v[0]),
524                                   "=d" (v[1]),
525                                   "=q" (z)
526                                 : "a" (compare[0]),
527                                   "d" (compare[1]),
528                                   "b" (set[0]),
529                                   "c" (set[1])
530                                 : "memory", "cc");
531         return z;
532 }
533
534 CK_CC_INLINE static bool
535 ck_pr_cas_ptr_2_value(void *t, void *c, void *s, void *v)
536 {
537         return ck_pr_cas_64_2_value(CK_CPP_CAST(uint64_t *,t),
538                                     CK_CPP_CAST(uint64_t *,c),
539                                     CK_CPP_CAST(uint64_t *,s),
540                                     CK_CPP_CAST(uint64_t *,v));
541 }
542
543 #define CK_PR_CAS_V(S, W, T)                                    \
544 CK_CC_INLINE static bool                                        \
545 ck_pr_cas_##S##_##W(T t[W], T c[W], T s[W])                     \
546 {                                                               \
547         return ck_pr_cas_64_2((uint64_t *)(void *)t,            \
548                               (uint64_t *)(void *)c,            \
549                               (uint64_t *)(void *)s);           \
550 }                                                               \
551 CK_CC_INLINE static bool                                        \
552 ck_pr_cas_##S##_##W##_value(T *t, T c[W], T s[W], T *v)         \
553 {                                                               \
554         return ck_pr_cas_64_2_value((uint64_t *)(void *)t,      \
555                                     (uint64_t *)(void *)c,      \
556                                     (uint64_t *)(void *)s,      \
557                                     (uint64_t *)(void *)v);     \
558 }
559
560 #ifndef CK_PR_DISABLE_DOUBLE
561 CK_PR_CAS_V(double, 2, double)
562 #endif
563 CK_PR_CAS_V(char, 16, char)
564 CK_PR_CAS_V(int, 4, int)
565 CK_PR_CAS_V(uint, 4, unsigned int)
566 CK_PR_CAS_V(32, 4, uint32_t)
567 CK_PR_CAS_V(16, 8, uint16_t)
568 CK_PR_CAS_V(8, 16, uint8_t)
569
570 #undef CK_PR_CAS_V
571
572 /*
573  * Atomic bit test operations.
574  */
575 #define CK_PR_BT(K, S, T, P, C, I)                                      \
576         CK_CC_INLINE static bool                                        \
577         ck_pr_##K##_##S(T *target, unsigned int b)                      \
578         {                                                               \
579                 bool c;                                                 \
580                 __asm__ __volatile__(CK_PR_LOCK_PREFIX I "; setc %1"    \
581                                         : "+m" (*(C *)target),          \
582                                           "=q" (c)                      \
583                                         : "q"  ((P)b)                   \
584                                         : "memory", "cc");              \
585                 return c;                                               \
586         }
587
588 #define CK_PR_BT_S(K, S, T, I) CK_PR_BT(K, S, T, T, T, I)
589
590 #define CK_PR_GENERATE(K)                                       \
591         CK_PR_BT(K, ptr, void, uint64_t, char, #K "q %2, %0")   \
592         CK_PR_BT_S(K, uint, unsigned int, #K "l %2, %0")        \
593         CK_PR_BT_S(K, int, int, #K "l %2, %0")                  \
594         CK_PR_BT_S(K, 64, uint64_t, #K "q %2, %0")              \
595         CK_PR_BT_S(K, 32, uint32_t, #K "l %2, %0")              \
596         CK_PR_BT_S(K, 16, uint16_t, #K "w %w2, %0")
597
598 CK_PR_GENERATE(btc)
599 CK_PR_GENERATE(bts)
600 CK_PR_GENERATE(btr)
601
602 #undef CK_PR_GENERATE
603 #undef CK_PR_BT
604
605 #endif /* CK_PR_X86_64_H */
606