1 /*
2 * Copyright (C) 2014 The Android Open Source Project
3 *
4 * Licensed under the Apache License, Version 2.0 (the "License");
5 * you may not use this file except in compliance with the License.
6 * You may obtain a copy of the License at
7 *
8 * http://www.apache.org/licenses/LICENSE-2.0
9 *
10 * Unless required by applicable law or agreed to in writing, software
11 * distributed under the License is distributed on an "AS IS" BASIS,
12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13 * See the License for the specific language governing permissions and
14 * limitations under the License.
15 */
16
17 #include <gtest/gtest.h>
18
19 // The real <stdatomic.h> checks for the availability of C++'s atomics and uses them if present. Since
20 // we want to test the libc versions, we instead include <bits/stdatomic.h> where they're actually defined.
21 #include <bits/stdatomic.h>
22
23 #include <pthread.h>
24 #include <stdint.h>
25
TEST(stdatomic,LOCK_FREE)26 TEST(stdatomic, LOCK_FREE) {
27 ASSERT_TRUE(ATOMIC_BOOL_LOCK_FREE);
28 ASSERT_TRUE(ATOMIC_CHAR16_T_LOCK_FREE);
29 ASSERT_TRUE(ATOMIC_CHAR32_T_LOCK_FREE);
30 ASSERT_TRUE(ATOMIC_CHAR_LOCK_FREE);
31 ASSERT_TRUE(ATOMIC_INT_LOCK_FREE);
32 ASSERT_TRUE(ATOMIC_LLONG_LOCK_FREE);
33 ASSERT_TRUE(ATOMIC_LONG_LOCK_FREE);
34 ASSERT_TRUE(ATOMIC_POINTER_LOCK_FREE);
35 ASSERT_TRUE(ATOMIC_SHORT_LOCK_FREE);
36 ASSERT_TRUE(ATOMIC_WCHAR_T_LOCK_FREE);
37 }
38
TEST(stdatomic,init)39 TEST(stdatomic, init) {
40 atomic_int v = ATOMIC_VAR_INIT(123);
41 ASSERT_EQ(123, atomic_load(&v));
42
43 atomic_init(&v, 456);
44 ASSERT_EQ(456, atomic_load(&v));
45
46 atomic_flag f = ATOMIC_FLAG_INIT;
47 ASSERT_FALSE(atomic_flag_test_and_set(&f));
48 }
49
TEST(stdatomic,atomic_thread_fence)50 TEST(stdatomic, atomic_thread_fence) {
51 atomic_thread_fence(memory_order_relaxed);
52 atomic_thread_fence(memory_order_consume);
53 atomic_thread_fence(memory_order_acquire);
54 atomic_thread_fence(memory_order_release);
55 atomic_thread_fence(memory_order_acq_rel);
56 atomic_thread_fence(memory_order_seq_cst);
57 }
58
TEST(stdatomic,atomic_signal_fence)59 TEST(stdatomic, atomic_signal_fence) {
60 atomic_signal_fence(memory_order_relaxed);
61 atomic_signal_fence(memory_order_consume);
62 atomic_signal_fence(memory_order_acquire);
63 atomic_signal_fence(memory_order_release);
64 atomic_signal_fence(memory_order_acq_rel);
65 atomic_signal_fence(memory_order_seq_cst);
66 }
67
TEST(stdatomic,atomic_is_lock_free)68 TEST(stdatomic, atomic_is_lock_free) {
69 atomic_char small;
70 ASSERT_TRUE(atomic_is_lock_free(&small));
71 atomic_intmax_t big;
72 ASSERT_TRUE(atomic_is_lock_free(&big));
73 }
74
TEST(stdatomic,atomic_flag)75 TEST(stdatomic, atomic_flag) {
76 atomic_flag f = ATOMIC_FLAG_INIT;
77 ASSERT_FALSE(atomic_flag_test_and_set(&f));
78 ASSERT_TRUE(atomic_flag_test_and_set(&f));
79
80 atomic_flag_clear(&f);
81
82 ASSERT_FALSE(atomic_flag_test_and_set_explicit(&f, memory_order_relaxed));
83 ASSERT_TRUE(atomic_flag_test_and_set_explicit(&f, memory_order_relaxed));
84
85 atomic_flag_clear_explicit(&f, memory_order_relaxed);
86 ASSERT_FALSE(atomic_flag_test_and_set_explicit(&f, memory_order_relaxed));
87 }
88
TEST(stdatomic,atomic_store)89 TEST(stdatomic, atomic_store) {
90 atomic_int i;
91 atomic_store(&i, 123);
92 ASSERT_EQ(123, atomic_load(&i));
93 atomic_store_explicit(&i, 123, memory_order_relaxed);
94 ASSERT_EQ(123, atomic_load_explicit(&i, memory_order_relaxed));
95 }
96
TEST(stdatomic,atomic_exchange)97 TEST(stdatomic, atomic_exchange) {
98 atomic_int i;
99 atomic_store(&i, 123);
100 ASSERT_EQ(123, atomic_exchange(&i, 456));
101 ASSERT_EQ(456, atomic_exchange_explicit(&i, 123, memory_order_relaxed));
102 }
103
TEST(stdatomic,atomic_compare_exchange)104 TEST(stdatomic, atomic_compare_exchange) {
105 atomic_int i;
106 int expected;
107
108 atomic_store(&i, 123);
109 expected = 123;
110 ASSERT_TRUE(atomic_compare_exchange_strong(&i, &expected, 456));
111 ASSERT_FALSE(atomic_compare_exchange_strong(&i, &expected, 456));
112 ASSERT_EQ(456, expected);
113
114 atomic_store(&i, 123);
115 expected = 123;
116 ASSERT_TRUE(atomic_compare_exchange_strong_explicit(&i, &expected, 456, memory_order_relaxed,
117 memory_order_relaxed));
118 ASSERT_FALSE(atomic_compare_exchange_strong_explicit(&i, &expected, 456, memory_order_relaxed,
119 memory_order_relaxed));
120 ASSERT_EQ(456, expected);
121
122 atomic_store(&i, 123);
123 expected = 123;
124 int iter_count = 0;
125 do {
126 ++iter_count;
127 ASSERT_LT(iter_count, 100); // Arbitrary limit on spurious compare_exchange failures.
128 ASSERT_EQ(expected, 123);
129 } while(!atomic_compare_exchange_weak(&i, &expected, 456));
130 ASSERT_FALSE(atomic_compare_exchange_weak(&i, &expected, 456));
131 ASSERT_EQ(456, expected);
132
133 atomic_store(&i, 123);
134 expected = 123;
135 iter_count = 0;
136 do {
137 ++iter_count;
138 ASSERT_LT(iter_count, 100);
139 ASSERT_EQ(expected, 123);
140 } while(!atomic_compare_exchange_weak_explicit(&i, &expected, 456, memory_order_relaxed,
141 memory_order_relaxed));
142 ASSERT_FALSE(atomic_compare_exchange_weak_explicit(&i, &expected, 456, memory_order_relaxed,
143 memory_order_relaxed));
144 ASSERT_EQ(456, expected);
145 }
146
TEST(stdatomic,atomic_fetch_add)147 TEST(stdatomic, atomic_fetch_add) {
148 atomic_int i = ATOMIC_VAR_INIT(123);
149 ASSERT_EQ(123, atomic_fetch_add(&i, 1));
150 ASSERT_EQ(124, atomic_fetch_add_explicit(&i, 1, memory_order_relaxed));
151 ASSERT_EQ(125, atomic_load(&i));
152 }
153
TEST(stdatomic,atomic_fetch_sub)154 TEST(stdatomic, atomic_fetch_sub) {
155 atomic_int i = ATOMIC_VAR_INIT(123);
156 ASSERT_EQ(123, atomic_fetch_sub(&i, 1));
157 ASSERT_EQ(122, atomic_fetch_sub_explicit(&i, 1, memory_order_relaxed));
158 ASSERT_EQ(121, atomic_load(&i));
159 }
160
TEST(stdatomic,atomic_fetch_or)161 TEST(stdatomic, atomic_fetch_or) {
162 atomic_int i = ATOMIC_VAR_INIT(0x100);
163 ASSERT_EQ(0x100, atomic_fetch_or(&i, 0x020));
164 ASSERT_EQ(0x120, atomic_fetch_or_explicit(&i, 0x003, memory_order_relaxed));
165 ASSERT_EQ(0x123, atomic_load(&i));
166 }
167
TEST(stdatomic,atomic_fetch_xor)168 TEST(stdatomic, atomic_fetch_xor) {
169 atomic_int i = ATOMIC_VAR_INIT(0x100);
170 ASSERT_EQ(0x100, atomic_fetch_xor(&i, 0x120));
171 ASSERT_EQ(0x020, atomic_fetch_xor_explicit(&i, 0x103, memory_order_relaxed));
172 ASSERT_EQ(0x123, atomic_load(&i));
173 }
174
TEST(stdatomic,atomic_fetch_and)175 TEST(stdatomic, atomic_fetch_and) {
176 atomic_int i = ATOMIC_VAR_INIT(0x123);
177 ASSERT_EQ(0x123, atomic_fetch_and(&i, 0x00f));
178 ASSERT_EQ(0x003, atomic_fetch_and_explicit(&i, 0x2, memory_order_relaxed));
179 ASSERT_EQ(0x002, atomic_load(&i));
180 }
181
182 // And a rudimentary test of acquire-release memory ordering:
183
184 constexpr static uint_least32_t BIG = 30'000'000ul; // Assumed even below.
185
186 struct three_atomics {
187 atomic_uint_least32_t x;
188 char a[123]; // Everything in different cache lines,
189 // increase chance of compiler getting alignment wrong.
190 atomic_uint_least32_t y;
191 char b[4013];
192 atomic_uint_least32_t z;
193 };
194
195 // Very simple acquire/release memory ordering smoke test.
writer(void * arg)196 static void* writer(void* arg) {
197 three_atomics* a = reinterpret_cast<three_atomics*>(arg);
198 for (uint_least32_t i = 0; i <= BIG; i+=2) {
199 atomic_store_explicit(&a->x, i, memory_order_relaxed);
200 atomic_store_explicit(&a->z, i, memory_order_relaxed);
201 atomic_store_explicit(&a->y, i, memory_order_release);
202 atomic_store_explicit(&a->x, i+1, memory_order_relaxed);
203 atomic_store_explicit(&a->z, i+1, memory_order_relaxed);
204 atomic_store_explicit(&a->y, i+1, memory_order_release);
205 }
206 return nullptr;
207 }
208
reader(void * arg)209 static void* reader(void* arg) {
210 three_atomics* a = reinterpret_cast<three_atomics*>(arg);
211 uint_least32_t xval = 0, yval = 0, zval = 0;
212 size_t repeat = 0;
213 size_t repeat_limit = 1000;
214 while (yval != BIG + 1) {
215 yval = atomic_load_explicit(&a->y, memory_order_acquire);
216 zval = atomic_load_explicit(&a->z, memory_order_relaxed);
217 xval = atomic_load_explicit(&a->x, memory_order_relaxed);
218 // If we see a given value of y, the immediately preceding
219 // stores to z and x, or later ones, should also be visible.
220 if (zval < yval) {
221 // Cant just ASSERT, since we are in a non-void function.
222 ADD_FAILURE() << "acquire-release ordering violation: "
223 << zval << " < " << yval << ", " << xval << "\n";
224 return nullptr; // Only report once.
225 }
226 if (xval < yval) {
227 // Cant just ASSERT, since we are in a non-void function.
228 ADD_FAILURE() << "acquire-release ordering violation: "
229 << xval << " < " << yval << ", " << zval << "\n";
230 return nullptr; // Only report once.
231 }
232 if (repeat < repeat_limit) ++repeat;
233 }
234 // The following assertion is not technically guaranteed to hold.
235 // But if it fails to hold, this test was useless, and we have a
236 // serious scheduling issue that we should probably know about.
237 EXPECT_EQ(repeat, repeat_limit);
238 return nullptr;
239 }
240
TEST(stdatomic,ordering)241 TEST(stdatomic, ordering) {
242 // Run a memory ordering smoke test.
243 void* result;
244 three_atomics a;
245 atomic_init(&a.x, 0ul);
246 atomic_init(&a.y, 0ul);
247 atomic_init(&a.z, 0ul);
248 pthread_t t1,t2;
249 ASSERT_EQ(0, pthread_create(&t1, nullptr, reader, &a));
250 ASSERT_EQ(0, pthread_create(&t2, nullptr, writer, &a));
251 ASSERT_EQ(0, pthread_join(t1, &result));
252 EXPECT_EQ(nullptr, result);
253 ASSERT_EQ(0, pthread_join(t2, &result));
254 EXPECT_EQ(nullptr, result);
255 EXPECT_EQ(atomic_load_explicit(&a.x, memory_order_consume), BIG + 1);
256 EXPECT_EQ(atomic_load_explicit(&a.y, memory_order_seq_cst), BIG + 1);
257 EXPECT_EQ(atomic_load(&a.z), BIG + 1);
258 }
259