]> CyberLeo.Net >> Repos - FreeBSD/releng/10.2.git/blob - contrib/netbsd-tests/lib/libc/sync/all_sync_ops_linkable.c
- Copy stable/10@285827 to releng/10.2 in preparation for 10.2-RC1
[FreeBSD/releng/10.2.git] / contrib / netbsd-tests / lib / libc / sync / all_sync_ops_linkable.c
1 /* $NetBSD: all_sync_ops_linkable.c,v 1.4 2014/02/21 10:26:25 martin Exp $ */
2
3 /*-
4  * Copyright (c) 2014 The NetBSD Foundation, Inc.
5  * All rights reserved.
6  *
7  * This code is derived from software contributed to The NetBSD Foundation
8  * by Martin Husemann <martin@NetBSD.org>.
9  *
10  * Redistribution and use in source and binary forms, with or without
11  * modification, are permitted provided that the following conditions
12  * are met:
13  * 1. Redistributions of source code must retain the above copyright
14  *    notice, this list of conditions and the following disclaimer.
15  * 2. Redistributions in binary form must reproduce the above copyright
16  *    notice, this list of conditions and the following disclaimer in the
17  *    documentation and/or other materials provided with the distribution.
18  *
19  * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
20  * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
21  * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
22  * PURPOSE ARE DISCLAIMED.  IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
23  * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
24  * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
25  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
26  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
27  * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
28  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
29  * POSSIBILITY OF SUCH DAMAGE.
30  */
31
32 /*
33  * This is a simple link-time test to verify all builtin atomic sync
34  * operations are available. Depending on the exact cpu/arch code generator
35  * options, some of these need support functions (which on NetBSD we
36  * typically provide in src/common/lib/libc/atomic).
37  *
38  * The list of operations has been extracted from sync-builtins.def file
39  * in the gcc distribution (as of gcc 4.8.2).
40  */
41
42 #include <machine/types.h>
43 #include <sys/inttypes.h>
44
45 volatile uint8_t u8 = 0;
46 volatile uint16_t u16 = 0;
47 volatile uint32_t u32 = 0;
48
49 #ifdef __HAVE_ATOMIC64_OPS
50 volatile uint64_t u64 = 0;
51 #endif
52
53 int
54 main(int argc, char **argv)
55 {
56         __sync_synchronize();
57         __sync_add_and_fetch(&u8, 1);
58         __sync_add_and_fetch_1(&u8, 1);
59         __sync_add_and_fetch_2(&u16, 1);
60         __sync_add_and_fetch_4(&u32, 1);
61 #ifdef __HAVE_ATOMIC64_OPS
62         __sync_add_and_fetch_8(&u64, 1);
63 #endif
64         __sync_bool_compare_and_swap(&u8, 1, 2);
65         __sync_bool_compare_and_swap_1(&u8, 1, 2);
66         __sync_bool_compare_and_swap_2(&u16, 1, 2);
67         __sync_bool_compare_and_swap_4(&u32, 1, 2);
68 #ifdef __HAVE_ATOMIC64_OPS
69         __sync_bool_compare_and_swap_8(&u64, 1, 2);
70 #endif
71         __sync_fetch_and_add(&u8, 1);
72         __sync_fetch_and_add_1(&u8, 1);
73         __sync_fetch_and_add_2(&u16, 1);
74         __sync_fetch_and_add_4(&u32, 1);
75 #ifdef __HAVE_ATOMIC64_OPS
76         __sync_fetch_and_add_8(&u64, 1);
77 #endif
78         __sync_fetch_and_and(&u8, 0x80);
79         __sync_fetch_and_and_1(&u8, 0x80);
80         __sync_fetch_and_and_2(&u16, 0x80);
81         __sync_fetch_and_and_4(&u32, 0x80);
82 #ifdef __HAVE_ATOMIC64_OPS
83         __sync_fetch_and_and_8(&u64, 0x80);
84 #endif
85 #ifndef __clang__
86         __sync_fetch_and_nand(&u8, 0x80);
87         __sync_fetch_and_nand_1(&u8, 0x80);
88         __sync_fetch_and_nand_2(&u16, 0x80);
89         __sync_fetch_and_nand_4(&u32, 0x80);
90 #ifdef __HAVE_ATOMIC64_OPS
91         __sync_fetch_and_nand_8(&u64, 0x80);
92 #endif
93 #endif
94         __sync_fetch_and_or(&u8, 0x80);
95         __sync_fetch_and_or_1(&u8, 0x80);
96         __sync_fetch_and_or_2(&u16, 0x80);
97         __sync_fetch_and_or_4(&u32, 0x80);
98 #ifdef __HAVE_ATOMIC64_OPS
99         __sync_fetch_and_or_8(&u64, 0x80);
100 #endif
101         __sync_fetch_and_sub(&u8, 0x80);
102         __sync_fetch_and_sub_1(&u8, 0x80);
103         __sync_fetch_and_sub_2(&u16, 0x80);
104         __sync_fetch_and_sub_4(&u32, 0x80);
105 #ifdef __HAVE_ATOMIC64_OPS
106         __sync_fetch_and_sub_8(&u64, 0x80);
107 #endif
108         __sync_fetch_and_xor(&u8, 0x80);
109         __sync_fetch_and_xor_1(&u8, 0x80);
110         __sync_fetch_and_xor_2(&u16, 0x80);
111         __sync_fetch_and_xor_4(&u32, 0x80);
112 #ifdef __HAVE_ATOMIC64_OPS
113         __sync_fetch_and_xor_8(&u64, 0x80);
114 #endif
115         __sync_lock_release(&u8);
116         __sync_lock_release_1(&u8);
117         __sync_lock_release_2(&u16);
118         __sync_lock_release_4(&u32);
119 #ifdef __HAVE_ATOMIC64_OPS
120         __sync_lock_release_8(&u64);
121 #endif
122         __sync_lock_test_and_set(&u8, 5);
123         __sync_lock_test_and_set_1(&u8, 5);
124         __sync_lock_test_and_set_2(&u16, 5);
125         __sync_lock_test_and_set_4(&u32, 5);
126 #ifdef __HAVE_ATOMIC64_OPS
127         __sync_lock_test_and_set_8(&u64, 5);
128 #endif
129 #ifndef __clang__
130         __sync_nand_and_fetch(&u8, 5);
131         __sync_nand_and_fetch_1(&u8, 5);
132         __sync_nand_and_fetch_2(&u16, 5);
133         __sync_nand_and_fetch_4(&u32, 5);
134 #ifdef __HAVE_ATOMIC64_OPS
135         __sync_nand_and_fetch_8(&u64, 5);
136 #endif
137 #endif
138         __sync_or_and_fetch(&u8, 5);
139         __sync_or_and_fetch_1(&u8, 5);
140         __sync_or_and_fetch_2(&u16, 5);
141         __sync_or_and_fetch_4(&u32, 5);
142 #ifdef __HAVE_ATOMIC64_OPS
143         __sync_or_and_fetch_8(&u64, 5);
144 #endif
145         __sync_sub_and_fetch(&u8, 5);
146         __sync_sub_and_fetch_1(&u8, 5);
147         __sync_sub_and_fetch_2(&u16, 5);
148         __sync_sub_and_fetch_4(&u32, 5);
149 #ifdef __HAVE_ATOMIC64_OPS
150         __sync_sub_and_fetch_8(&u64, 5);
151 #endif
152         __sync_val_compare_and_swap(&u8, 5, 9);
153         __sync_val_compare_and_swap_1(&u8, 5, 9);
154         __sync_val_compare_and_swap_2(&u16, 5, 9);
155         __sync_val_compare_and_swap_4(&u32, 5, 9);
156 #ifdef __HAVE_ATOMIC64_OPS
157         __sync_val_compare_and_swap_8(&u64, 5, 9);
158 #endif
159         __sync_xor_and_fetch(&u8, 5);
160         __sync_xor_and_fetch_1(&u8, 5);
161         __sync_xor_and_fetch_2(&u16, 5);
162         __sync_xor_and_fetch_4(&u32, 5);
163 #ifdef __HAVE_ATOMIC64_OPS
164         __sync_xor_and_fetch_8(&u64, 5);
165 #endif
166
167         return 0;
168 }