aboutsummaryrefslogtreecommitdiff
path: root/vendor/portable-atomic/src/tests
diff options
context:
space:
mode:
authorValentin Popov <valentin@popov.link>2024-01-08 00:21:28 +0300
committerValentin Popov <valentin@popov.link>2024-01-08 00:21:28 +0300
commit1b6a04ca5504955c571d1c97504fb45ea0befee4 (patch)
tree7579f518b23313e8a9748a88ab6173d5e030b227 /vendor/portable-atomic/src/tests
parent5ecd8cf2cba827454317368b68571df0d13d7842 (diff)
downloadfparkan-1b6a04ca5504955c571d1c97504fb45ea0befee4.tar.xz
fparkan-1b6a04ca5504955c571d1c97504fb45ea0befee4.zip
Initial vendor packages
Signed-off-by: Valentin Popov <valentin@popov.link>
Diffstat (limited to 'vendor/portable-atomic/src/tests')
-rw-r--r--vendor/portable-atomic/src/tests/helper.rs2420
-rw-r--r--vendor/portable-atomic/src/tests/mod.rs357
2 files changed, 2777 insertions, 0 deletions
diff --git a/vendor/portable-atomic/src/tests/helper.rs b/vendor/portable-atomic/src/tests/helper.rs
new file mode 100644
index 0000000..65831c6
--- /dev/null
+++ b/vendor/portable-atomic/src/tests/helper.rs
@@ -0,0 +1,2420 @@
+// SPDX-License-Identifier: Apache-2.0 OR MIT
+
+#![allow(unused_macros)]
+
+use core::sync::atomic::Ordering;
+
+macro_rules! __test_atomic_common {
+ ($atomic_type:ty, $value_type:ty) => {
+ #[test]
+ fn assert_auto_traits() {
+ fn _assert<T: Send + Sync + Unpin + std::panic::UnwindSafe>() {}
+ _assert::<$atomic_type>();
+ }
+ #[test]
+ fn alignment() {
+ // https://github.com/rust-lang/rust/blob/1.70.0/library/core/tests/atomic.rs#L250
+ assert_eq!(core::mem::align_of::<$atomic_type>(), core::mem::size_of::<$atomic_type>());
+ assert_eq!(core::mem::size_of::<$atomic_type>(), core::mem::size_of::<$value_type>());
+ }
+ #[test]
+ fn is_lock_free() {
+ const IS_ALWAYS_LOCK_FREE: bool = <$atomic_type>::is_always_lock_free();
+ assert_eq!(IS_ALWAYS_LOCK_FREE, <$atomic_type>::is_always_lock_free());
+ let is_lock_free = <$atomic_type>::is_lock_free();
+ if IS_ALWAYS_LOCK_FREE {
+ // If is_always_lock_free is true, then is_lock_free must always be true.
+ assert!(is_lock_free);
+ }
+ }
+ };
+}
+macro_rules! __test_atomic_pub_common {
+ ($atomic_type:ty, $value_type:ty) => {
+ #[test]
+ fn assert_ref_unwind_safe() {
+ #[cfg(not(all(portable_atomic_no_core_unwind_safe, not(feature = "std"))))]
+ static_assertions::assert_impl_all!($atomic_type: std::panic::RefUnwindSafe);
+ #[cfg(all(portable_atomic_no_core_unwind_safe, not(feature = "std")))]
+ static_assertions::assert_not_impl_all!($atomic_type: std::panic::RefUnwindSafe);
+ }
+ };
+}
+
+macro_rules! __test_atomic_int_load_store {
+ ($atomic_type:ty, $int_type:ident, single_thread) => {
+ __test_atomic_common!($atomic_type, $int_type);
+ use crate::tests::helper::*;
+ #[test]
+ fn accessor() {
+ let mut a = <$atomic_type>::new(10);
+ assert_eq!(*a.get_mut(), 10);
+ *a.get_mut() = 5;
+ assert_eq!(a.as_ptr() as *const (), &a as *const _ as *const ());
+ assert_eq!(a.into_inner(), 5);
+ }
+ // https://bugs.llvm.org/show_bug.cgi?id=37061
+ #[test]
+ fn static_load_only() {
+ static VAR: $atomic_type = <$atomic_type>::new(10);
+ for &order in &test_helper::LOAD_ORDERINGS {
+ assert_eq!(VAR.load(order), 10);
+ }
+ }
+ #[test]
+ fn load_store() {
+ static VAR: $atomic_type = <$atomic_type>::new(10);
+ test_load_ordering(|order| VAR.load(order));
+ test_store_ordering(|order| VAR.store(10, order));
+ for (&load_order, &store_order) in
+ test_helper::LOAD_ORDERINGS.iter().zip(&test_helper::STORE_ORDERINGS)
+ {
+ assert_eq!(VAR.load(load_order), 10);
+ VAR.store(5, store_order);
+ assert_eq!(VAR.load(load_order), 5);
+ VAR.store(10, store_order);
+ let a = <$atomic_type>::new(1);
+ assert_eq!(a.load(load_order), 1);
+ a.store(2, store_order);
+ assert_eq!(a.load(load_order), 2);
+ }
+ }
+ };
+ ($atomic_type:ty, $int_type:ident) => {
+ __test_atomic_int_load_store!($atomic_type, $int_type, single_thread);
+ use crossbeam_utils::thread;
+ use std::{collections::BTreeSet, vec, vec::Vec};
+ #[test]
+ fn stress_load_store() {
+ let (iterations, threads) = stress_test_config();
+ let data1 = (0..iterations).map(|_| fastrand::$int_type(..)).collect::<Vec<_>>();
+ let set = data1.iter().copied().collect::<BTreeSet<_>>();
+ let a = <$atomic_type>::new(data1[fastrand::usize(0..iterations)]);
+ let now = &std::time::Instant::now();
+ thread::scope(|s| {
+ for _ in 0..threads {
+ s.spawn(|_| {
+ let now = *now;
+ for i in 0..iterations {
+ a.store(data1[i], rand_store_ordering());
+ }
+ std::eprintln!("store end={:?}", now.elapsed());
+ });
+ s.spawn(|_| {
+ let now = *now;
+ let mut v = vec![0; iterations];
+ for i in 0..iterations {
+ v[i] = a.load(rand_load_ordering());
+ }
+ std::eprintln!("load end={:?}", now.elapsed());
+ for v in v {
+ assert!(set.contains(&v), "v={}", v);
+ }
+ });
+ }
+ })
+ .unwrap();
+ }
+ };
+}
+macro_rules! __test_atomic_float_load_store {
+ ($atomic_type:ty, $float_type:ident, single_thread) => {
+ __test_atomic_common!($atomic_type, $float_type);
+ use crate::tests::helper::*;
+ #[test]
+ fn accessor() {
+ let mut a = <$atomic_type>::new(10.0);
+ assert_eq!(*a.get_mut(), 10.0);
+ *a.get_mut() = 5.0;
+ assert_eq!(a.as_ptr() as *const (), &a as *const _ as *const ());
+ assert_eq!(a.into_inner(), 5.0);
+ }
+ // https://bugs.llvm.org/show_bug.cgi?id=37061
+ #[test]
+ fn static_load_only() {
+ static VAR: $atomic_type = <$atomic_type>::new(10.0);
+ for &order in &test_helper::LOAD_ORDERINGS {
+ assert_eq!(VAR.load(order), 10.0);
+ }
+ }
+ #[test]
+ fn load_store() {
+ static VAR: $atomic_type = <$atomic_type>::new(10.0);
+ test_load_ordering(|order| VAR.load(order));
+ test_store_ordering(|order| VAR.store(10.0, order));
+ for (&load_order, &store_order) in
+ test_helper::LOAD_ORDERINGS.iter().zip(&test_helper::STORE_ORDERINGS)
+ {
+ assert_eq!(VAR.load(load_order), 10.0);
+ VAR.store(5.0, store_order);
+ assert_eq!(VAR.load(load_order), 5.0);
+ VAR.store(10.0, store_order);
+ let a = <$atomic_type>::new(1.0);
+ assert_eq!(a.load(load_order), 1.0);
+ a.store(2.0, store_order);
+ assert_eq!(a.load(load_order), 2.0);
+ }
+ }
+ };
+ ($atomic_type:ty, $float_type:ident) => {
+ __test_atomic_float_load_store!($atomic_type, $float_type, single_thread);
+ // TODO: multi thread
+ };
+}
+macro_rules! __test_atomic_bool_load_store {
+ ($atomic_type:ty, single_thread) => {
+ __test_atomic_common!($atomic_type, bool);
+ use crate::tests::helper::*;
+ #[test]
+ fn accessor() {
+ let mut a = <$atomic_type>::new(false);
+ assert_eq!(*a.get_mut(), false);
+ *a.get_mut() = true;
+ assert_eq!(a.as_ptr() as *const (), &a as *const _ as *const ());
+ assert_eq!(a.into_inner(), true);
+ }
+ // https://bugs.llvm.org/show_bug.cgi?id=37061
+ #[test]
+ fn static_load_only() {
+ static VAR: $atomic_type = <$atomic_type>::new(false);
+ for &order in &test_helper::LOAD_ORDERINGS {
+ assert_eq!(VAR.load(order), false);
+ }
+ }
+ #[test]
+ fn load_store() {
+ static VAR: $atomic_type = <$atomic_type>::new(false);
+ test_load_ordering(|order| VAR.load(order));
+ test_store_ordering(|order| VAR.store(false, order));
+ for (&load_order, &store_order) in
+ test_helper::LOAD_ORDERINGS.iter().zip(&test_helper::STORE_ORDERINGS)
+ {
+ assert_eq!(VAR.load(load_order), false);
+ VAR.store(true, store_order);
+ assert_eq!(VAR.load(load_order), true);
+ VAR.store(false, store_order);
+ let a = <$atomic_type>::new(true);
+ assert_eq!(a.load(load_order), true);
+ a.store(false, store_order);
+ assert_eq!(a.load(load_order), false);
+ }
+ }
+ };
+ ($atomic_type:ty) => {
+ __test_atomic_bool_load_store!($atomic_type, single_thread);
+ // TODO: multi thread
+ };
+}
+macro_rules! __test_atomic_ptr_load_store {
+ ($atomic_type:ty, single_thread) => {
+ __test_atomic_common!($atomic_type, *mut u8);
+ use crate::tests::helper::*;
+ use std::ptr;
+ #[test]
+ fn accessor() {
+ let mut v = 1;
+ let mut a = <$atomic_type>::new(ptr::null_mut());
+ assert!(a.get_mut().is_null());
+ *a.get_mut() = &mut v;
+ assert_eq!(a.as_ptr() as *const (), &a as *const _ as *const ());
+ assert!(!a.into_inner().is_null());
+ }
+ // https://bugs.llvm.org/show_bug.cgi?id=37061
+ #[test]
+ fn static_load_only() {
+ static VAR: $atomic_type = <$atomic_type>::new(ptr::null_mut());
+ for &order in &test_helper::LOAD_ORDERINGS {
+ assert_eq!(VAR.load(order), ptr::null_mut());
+ }
+ }
+ #[test]
+ fn load_store() {
+ static VAR: $atomic_type = <$atomic_type>::new(ptr::null_mut());
+ test_load_ordering(|order| VAR.load(order));
+ test_store_ordering(|order| VAR.store(ptr::null_mut(), order));
+ let mut v = 1_u8;
+ let p = &mut v as *mut u8;
+ for (&load_order, &store_order) in
+ test_helper::LOAD_ORDERINGS.iter().zip(&test_helper::STORE_ORDERINGS)
+ {
+ assert_eq!(VAR.load(load_order), ptr::null_mut());
+ VAR.store(p, store_order);
+ assert_eq!(VAR.load(load_order), p);
+ VAR.store(ptr::null_mut(), store_order);
+ let a = <$atomic_type>::new(p);
+ assert_eq!(a.load(load_order), p);
+ a.store(ptr::null_mut(), store_order);
+ assert_eq!(a.load(load_order), ptr::null_mut());
+ }
+ }
+ };
+ ($atomic_type:ty) => {
+ __test_atomic_ptr_load_store!($atomic_type, single_thread);
+ // TODO: multi thread
+ };
+}
+
+macro_rules! __test_atomic_int {
+ ($atomic_type:ty, $int_type:ident, single_thread) => {
+ use core::$int_type;
+ #[test]
+ fn swap() {
+ let a = <$atomic_type>::new(5);
+ test_swap_ordering(|order| a.swap(5, order));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ assert_eq!(a.swap(10, order), 5);
+ assert_eq!(a.swap(5, order), 10);
+ }
+ }
+ #[test]
+ fn compare_exchange() {
+ let a = <$atomic_type>::new(5);
+ test_compare_exchange_ordering(|success, failure| {
+ a.compare_exchange(5, 5, success, failure)
+ });
+ for &(success, failure) in &test_helper::COMPARE_EXCHANGE_ORDERINGS {
+ let a = <$atomic_type>::new(5);
+ assert_eq!(a.compare_exchange(5, 10, success, failure), Ok(5));
+ assert_eq!(a.load(Ordering::Relaxed), 10);
+ assert_eq!(a.compare_exchange(6, 12, success, failure), Err(10));
+ assert_eq!(a.load(Ordering::Relaxed), 10);
+ }
+ }
+ #[test]
+ fn compare_exchange_weak() {
+ let a = <$atomic_type>::new(4);
+ test_compare_exchange_ordering(|success, failure| {
+ a.compare_exchange_weak(4, 4, success, failure)
+ });
+ for &(success, failure) in &test_helper::COMPARE_EXCHANGE_ORDERINGS {
+ let a = <$atomic_type>::new(4);
+ assert_eq!(a.compare_exchange_weak(6, 8, success, failure), Err(4));
+ let mut old = a.load(Ordering::Relaxed);
+ loop {
+ let new = old * 2;
+ match a.compare_exchange_weak(old, new, success, failure) {
+ Ok(_) => break,
+ Err(x) => old = x,
+ }
+ }
+ assert_eq!(a.load(Ordering::Relaxed), 8);
+ }
+ }
+ #[test]
+ fn fetch_add() {
+ let a = <$atomic_type>::new(0);
+ test_swap_ordering(|order| a.fetch_add(0, order));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(0);
+ assert_eq!(a.fetch_add(10, order), 0);
+ assert_eq!(a.load(Ordering::Relaxed), 10);
+ let a = <$atomic_type>::new($int_type::MAX);
+ assert_eq!(a.fetch_add(1, order), $int_type::MAX);
+ assert_eq!(a.load(Ordering::Relaxed), $int_type::MAX.wrapping_add(1));
+ }
+ }
+ #[test]
+ fn add() {
+ let a = <$atomic_type>::new(0);
+ test_swap_ordering(|order| a.add(0, order));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(0);
+ a.add(10, order);
+ assert_eq!(a.load(Ordering::Relaxed), 10);
+ let a = <$atomic_type>::new($int_type::MAX);
+ a.add(1, order);
+ assert_eq!(a.load(Ordering::Relaxed), $int_type::MAX.wrapping_add(1));
+ }
+ }
+ #[test]
+ fn fetch_sub() {
+ let a = <$atomic_type>::new(20);
+ test_swap_ordering(|order| a.fetch_sub(0, order));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(20);
+ assert_eq!(a.fetch_sub(10, order), 20);
+ assert_eq!(a.load(Ordering::Relaxed), 10);
+ let a = <$atomic_type>::new($int_type::MIN);
+ assert_eq!(a.fetch_sub(1, order), $int_type::MIN);
+ assert_eq!(a.load(Ordering::Relaxed), $int_type::MIN.wrapping_sub(1));
+ }
+ }
+ #[test]
+ fn sub() {
+ let a = <$atomic_type>::new(20);
+ test_swap_ordering(|order| a.sub(0, order));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(20);
+ a.sub(10, order);
+ assert_eq!(a.load(Ordering::Relaxed), 10);
+ let a = <$atomic_type>::new($int_type::MIN);
+ a.sub(1, order);
+ assert_eq!(a.load(Ordering::Relaxed), $int_type::MIN.wrapping_sub(1));
+ }
+ }
+ #[test]
+ fn fetch_and() {
+ let a = <$atomic_type>::new(0b101101);
+ test_swap_ordering(|order| a.fetch_and(0b101101, order));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(0b101101);
+ assert_eq!(a.fetch_and(0b110011, order), 0b101101);
+ assert_eq!(a.load(Ordering::Relaxed), 0b100001);
+ }
+ }
+ #[test]
+ fn and() {
+ let a = <$atomic_type>::new(0b101101);
+ test_swap_ordering(|order| a.and(0b101101, order));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(0b101101);
+ a.and(0b110011, order);
+ assert_eq!(a.load(Ordering::Relaxed), 0b100001);
+ }
+ }
+ #[test]
+ fn fetch_nand() {
+ let a = <$atomic_type>::new(0x13);
+ test_swap_ordering(|order| a.fetch_nand(0x31, order));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(0x13);
+ assert_eq!(a.fetch_nand(0x31, order), 0x13);
+ assert_eq!(a.load(Ordering::Relaxed), !(0x13 & 0x31));
+ }
+ }
+ #[test]
+ fn fetch_or() {
+ let a = <$atomic_type>::new(0b101101);
+ test_swap_ordering(|order| a.fetch_or(0, order));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(0b101101);
+ assert_eq!(a.fetch_or(0b110011, order), 0b101101);
+ assert_eq!(a.load(Ordering::Relaxed), 0b111111);
+ }
+ }
+ #[test]
+ fn or() {
+ let a = <$atomic_type>::new(0b101101);
+ test_swap_ordering(|order| a.or(0, order));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(0b101101);
+ a.or(0b110011, order);
+ assert_eq!(a.load(Ordering::Relaxed), 0b111111);
+ }
+ }
+ #[test]
+ fn fetch_xor() {
+ let a = <$atomic_type>::new(0b101101);
+ test_swap_ordering(|order| a.fetch_xor(0, order));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(0b101101);
+ assert_eq!(a.fetch_xor(0b110011, order), 0b101101);
+ assert_eq!(a.load(Ordering::Relaxed), 0b011110);
+ }
+ }
+ #[test]
+ fn xor() {
+ let a = <$atomic_type>::new(0b101101);
+ test_swap_ordering(|order| a.xor(0, order));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(0b101101);
+ a.xor(0b110011, order);
+ assert_eq!(a.load(Ordering::Relaxed), 0b011110);
+ }
+ }
+ #[test]
+ fn fetch_max() {
+ let a = <$atomic_type>::new(23);
+ test_swap_ordering(|order| a.fetch_max(23, order));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(23);
+ assert_eq!(a.fetch_max(22, order), 23);
+ assert_eq!(a.load(Ordering::Relaxed), 23);
+ assert_eq!(a.fetch_max(24, order), 23);
+ assert_eq!(a.load(Ordering::Relaxed), 24);
+ let a = <$atomic_type>::new(0);
+ assert_eq!(a.fetch_max(1, order), 0);
+ assert_eq!(a.load(Ordering::Relaxed), 1);
+ assert_eq!(a.fetch_max(0, order), 1);
+ assert_eq!(a.load(Ordering::Relaxed), 1);
+ let a = <$atomic_type>::new(!0);
+ assert_eq!(a.fetch_max(0, order), !0);
+ assert_eq!(a.load(Ordering::Relaxed), core::cmp::max(!0, 0));
+ }
+ }
+ #[test]
+ fn fetch_min() {
+ let a = <$atomic_type>::new(23);
+ test_swap_ordering(|order| a.fetch_min(23, order));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(23);
+ assert_eq!(a.fetch_min(24, order), 23);
+ assert_eq!(a.load(Ordering::Relaxed), 23);
+ assert_eq!(a.fetch_min(22, order), 23);
+ assert_eq!(a.load(Ordering::Relaxed), 22);
+ let a = <$atomic_type>::new(1);
+ assert_eq!(a.fetch_min(0, order), 1);
+ assert_eq!(a.load(Ordering::Relaxed), 0);
+ assert_eq!(a.fetch_min(1, order), 0);
+ assert_eq!(a.load(Ordering::Relaxed), 0);
+ let a = <$atomic_type>::new(!0);
+ assert_eq!(a.fetch_min(0, order), !0);
+ assert_eq!(a.load(Ordering::Relaxed), core::cmp::min(!0, 0));
+ }
+ }
+ #[test]
+ fn fetch_not() {
+ let a = <$atomic_type>::new(1);
+ test_swap_ordering(|order| a.fetch_not(order));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(1);
+ assert_eq!(a.fetch_not(order), 1);
+ assert_eq!(a.load(Ordering::Relaxed), !1);
+ }
+ }
+ #[test]
+ fn not() {
+ let a = <$atomic_type>::new(1);
+ test_swap_ordering(|order| a.not(order));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(1);
+ a.not(order);
+ assert_eq!(a.load(Ordering::Relaxed), !1);
+ }
+ }
+ #[test]
+ fn fetch_neg() {
+ let a = <$atomic_type>::new(5);
+ test_swap_ordering(|order| a.fetch_neg(order));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(5);
+ assert_eq!(a.fetch_neg(order), 5);
+ assert_eq!(a.load(Ordering::Relaxed), <$int_type>::wrapping_neg(5));
+ assert_eq!(a.fetch_neg(order), <$int_type>::wrapping_neg(5));
+ assert_eq!(a.load(Ordering::Relaxed), 5);
+ let a = <$atomic_type>::new(<$int_type>::MIN);
+ assert_eq!(a.fetch_neg(order), <$int_type>::MIN);
+ assert_eq!(a.load(Ordering::Relaxed), <$int_type>::MIN.wrapping_neg());
+ assert_eq!(a.fetch_neg(order), <$int_type>::MIN.wrapping_neg());
+ assert_eq!(a.load(Ordering::Relaxed), <$int_type>::MIN);
+ }
+ }
+ #[test]
+ fn neg() {
+ let a = <$atomic_type>::new(5);
+ test_swap_ordering(|order| a.neg(order));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(5);
+ a.neg(order);
+ assert_eq!(a.load(Ordering::Relaxed), <$int_type>::wrapping_neg(5));
+ a.neg(order);
+ assert_eq!(a.load(Ordering::Relaxed), 5);
+ let a = <$atomic_type>::new(<$int_type>::MIN);
+ a.neg(order);
+ assert_eq!(a.load(Ordering::Relaxed), <$int_type>::MIN.wrapping_neg());
+ a.neg(order);
+ assert_eq!(a.load(Ordering::Relaxed), <$int_type>::MIN);
+ }
+ }
+ #[test]
+ fn bit_set() {
+ let a = <$atomic_type>::new(0b0001);
+ test_swap_ordering(|order| assert!(a.bit_set(0, order)));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(0b0000);
+ assert!(!a.bit_set(0, order));
+ assert_eq!(a.load(Ordering::Relaxed), 0b0001);
+ assert!(a.bit_set(0, order));
+ assert_eq!(a.load(Ordering::Relaxed), 0b0001);
+ }
+ }
+ #[test]
+ fn bit_clear() {
+ let a = <$atomic_type>::new(0b0000);
+ test_swap_ordering(|order| assert!(!a.bit_clear(0, order)));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(0b0001);
+ assert!(a.bit_clear(0, order));
+ assert_eq!(a.load(Ordering::Relaxed), 0b0000);
+ assert!(!a.bit_clear(0, order));
+ assert_eq!(a.load(Ordering::Relaxed), 0b0000);
+ }
+ }
+ #[test]
+ fn bit_toggle() {
+ let a = <$atomic_type>::new(0b0000);
+ test_swap_ordering(|order| a.bit_toggle(0, order));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(0b0000);
+ assert!(!a.bit_toggle(0, order));
+ assert_eq!(a.load(Ordering::Relaxed), 0b0001);
+ assert!(a.bit_toggle(0, order));
+ assert_eq!(a.load(Ordering::Relaxed), 0b0000);
+ }
+ }
+ ::quickcheck::quickcheck! {
+ fn quickcheck_swap(x: $int_type, y: $int_type) -> bool {
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(x);
+ assert_eq!(a.swap(y, order), x);
+ assert_eq!(a.swap(x, order), y);
+ }
+ true
+ }
+ fn quickcheck_compare_exchange(x: $int_type, y: $int_type) -> bool {
+ #[cfg(all(
+ target_arch = "arm",
+ not(any(target_feature = "v6", portable_atomic_target_feature = "v6")),
+ ))]
+ {
+ // TODO: LLVM bug:
+ // https://github.com/llvm/llvm-project/issues/61880
+ // https://github.com/taiki-e/portable-atomic/issues/2
+ if core::mem::size_of::<$int_type>() <= 2 {
+ return true;
+ }
+ }
+ let z = loop {
+ let z = fastrand::$int_type(..);
+ if z != y {
+ break z;
+ }
+ };
+ for &(success, failure) in &test_helper::COMPARE_EXCHANGE_ORDERINGS {
+ let a = <$atomic_type>::new(x);
+ assert_eq!(a.compare_exchange(x, y, success, failure).unwrap(), x);
+ assert_eq!(a.load(Ordering::Relaxed), y);
+ assert_eq!(a.compare_exchange(z, x, success, failure).unwrap_err(), y);
+ assert_eq!(a.load(Ordering::Relaxed), y);
+ }
+ true
+ }
+ fn quickcheck_fetch_add(x: $int_type, y: $int_type) -> bool {
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(x);
+ assert_eq!(a.fetch_add(y, order), x);
+ assert_eq!(a.load(Ordering::Relaxed), x.wrapping_add(y));
+ let a = <$atomic_type>::new(y);
+ assert_eq!(a.fetch_add(x, order), y);
+ assert_eq!(a.load(Ordering::Relaxed), y.wrapping_add(x));
+ }
+ true
+ }
+ fn quickcheck_add(x: $int_type, y: $int_type) -> bool {
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(x);
+ a.add(y, order);
+ assert_eq!(a.load(Ordering::Relaxed), x.wrapping_add(y));
+ let a = <$atomic_type>::new(y);
+ a.add(x, order);
+ assert_eq!(a.load(Ordering::Relaxed), y.wrapping_add(x));
+ }
+ true
+ }
+ fn quickcheck_fetch_sub(x: $int_type, y: $int_type) -> bool {
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(x);
+ assert_eq!(a.fetch_sub(y, order), x);
+ assert_eq!(a.load(Ordering::Relaxed), x.wrapping_sub(y));
+ let a = <$atomic_type>::new(y);
+ assert_eq!(a.fetch_sub(x, order), y);
+ assert_eq!(a.load(Ordering::Relaxed), y.wrapping_sub(x));
+ }
+ true
+ }
+ fn quickcheck_sub(x: $int_type, y: $int_type) -> bool {
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(x);
+ a.sub(y, order);
+ assert_eq!(a.load(Ordering::Relaxed), x.wrapping_sub(y));
+ let a = <$atomic_type>::new(y);
+ a.sub(x, order);
+ assert_eq!(a.load(Ordering::Relaxed), y.wrapping_sub(x));
+ }
+ true
+ }
+ fn quickcheck_fetch_and(x: $int_type, y: $int_type) -> bool {
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(x);
+ assert_eq!(a.fetch_and(y, order), x);
+ assert_eq!(a.load(Ordering::Relaxed), x & y);
+ let a = <$atomic_type>::new(y);
+ assert_eq!(a.fetch_and(x, order), y);
+ assert_eq!(a.load(Ordering::Relaxed), y & x);
+ }
+ true
+ }
+ fn quickcheck_and(x: $int_type, y: $int_type) -> bool {
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(x);
+ a.and(y, order);
+ assert_eq!(a.load(Ordering::Relaxed), x & y);
+ let a = <$atomic_type>::new(y);
+ a.and(x, order);
+ assert_eq!(a.load(Ordering::Relaxed), y & x);
+ }
+ true
+ }
+ fn quickcheck_fetch_nand(x: $int_type, y: $int_type) -> bool {
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(x);
+ assert_eq!(a.fetch_nand(y, order), x);
+ assert_eq!(a.load(Ordering::Relaxed), !(x & y));
+ let a = <$atomic_type>::new(y);
+ assert_eq!(a.fetch_nand(x, order), y);
+ assert_eq!(a.load(Ordering::Relaxed), !(y & x));
+ }
+ true
+ }
+ fn quickcheck_fetch_or(x: $int_type, y: $int_type) -> bool {
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(x);
+ assert_eq!(a.fetch_or(y, order), x);
+ assert_eq!(a.load(Ordering::Relaxed), x | y);
+ let a = <$atomic_type>::new(y);
+ assert_eq!(a.fetch_or(x, order), y);
+ assert_eq!(a.load(Ordering::Relaxed), y | x);
+ }
+ true
+ }
+ fn quickcheck_or(x: $int_type, y: $int_type) -> bool {
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(x);
+ a.or(y, order);
+ assert_eq!(a.load(Ordering::Relaxed), x | y);
+ let a = <$atomic_type>::new(y);
+ a.or(x, order);
+ assert_eq!(a.load(Ordering::Relaxed), y | x);
+ }
+ true
+ }
+ fn quickcheck_fetch_xor(x: $int_type, y: $int_type) -> bool {
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(x);
+ assert_eq!(a.fetch_xor(y, order), x);
+ assert_eq!(a.load(Ordering::Relaxed), x ^ y);
+ let a = <$atomic_type>::new(y);
+ assert_eq!(a.fetch_xor(x, order), y);
+ assert_eq!(a.load(Ordering::Relaxed), y ^ x);
+ }
+ true
+ }
+ fn quickcheck_xor(x: $int_type, y: $int_type) -> bool {
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(x);
+ a.xor(y, order);
+ assert_eq!(a.load(Ordering::Relaxed), x ^ y);
+ let a = <$atomic_type>::new(y);
+ a.xor(x, order);
+ assert_eq!(a.load(Ordering::Relaxed), y ^ x);
+ }
+ true
+ }
+ fn quickcheck_fetch_max(x: $int_type, y: $int_type) -> bool {
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(x);
+ assert_eq!(a.fetch_max(y, order), x);
+ assert_eq!(a.load(Ordering::Relaxed), core::cmp::max(x, y));
+ let a = <$atomic_type>::new(y);
+ assert_eq!(a.fetch_max(x, order), y);
+ assert_eq!(a.load(Ordering::Relaxed), core::cmp::max(y, x));
+ }
+ true
+ }
+ fn quickcheck_fetch_min(x: $int_type, y: $int_type) -> bool {
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(x);
+ assert_eq!(a.fetch_min(y, order), x);
+ assert_eq!(a.load(Ordering::Relaxed), core::cmp::min(x, y));
+ let a = <$atomic_type>::new(y);
+ assert_eq!(a.fetch_min(x, order), y);
+ assert_eq!(a.load(Ordering::Relaxed), core::cmp::min(y, x));
+ }
+ true
+ }
+ fn quickcheck_fetch_not(x: $int_type) -> bool {
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(x);
+ assert_eq!(a.fetch_not(order), x);
+ assert_eq!(a.load(Ordering::Relaxed), !x);
+ assert_eq!(a.fetch_not(order), !x);
+ assert_eq!(a.load(Ordering::Relaxed), x);
+ }
+ true
+ }
+ fn quickcheck_not(x: $int_type) -> bool {
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(x);
+ a.not(order);
+ assert_eq!(a.load(Ordering::Relaxed), !x);
+ a.not(order);
+ assert_eq!(a.load(Ordering::Relaxed), x);
+ }
+ true
+ }
+ fn quickcheck_fetch_neg(x: $int_type) -> bool {
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(x);
+ assert_eq!(a.fetch_neg(order), x);
+ assert_eq!(a.load(Ordering::Relaxed), x.wrapping_neg());
+ assert_eq!(a.fetch_neg(order), x.wrapping_neg());
+ assert_eq!(a.load(Ordering::Relaxed), x);
+ }
+ true
+ }
+ fn quickcheck_neg(x: $int_type) -> bool {
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(x);
+ a.neg(order);
+ assert_eq!(a.load(Ordering::Relaxed), x.wrapping_neg());
+ a.neg(order);
+ assert_eq!(a.load(Ordering::Relaxed), x);
+ }
+ true
+ }
+ fn quickcheck_bit_set(x: $int_type, bit: u32) -> bool {
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(x);
+ let b = a.bit_set(bit, order);
+ let mask = <$int_type>::wrapping_shl(1, bit);
+ assert_eq!(a.load(Ordering::Relaxed), x | mask);
+ assert_eq!(b, x & mask != 0);
+ }
+ true
+ }
+ fn quickcheck_bit_clear(x: $int_type, bit: u32) -> bool {
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(x);
+ let b = a.bit_clear(bit, order);
+ let mask = <$int_type>::wrapping_shl(1, bit);
+ assert_eq!(a.load(Ordering::Relaxed), x & !mask);
+ assert_eq!(b, x & mask != 0);
+ }
+ true
+ }
+ fn quickcheck_bit_toggle(x: $int_type, bit: u32) -> bool {
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(x);
+ let b = a.bit_toggle(bit, order);
+ let mask = <$int_type>::wrapping_shl(1, bit);
+ assert_eq!(a.load(Ordering::Relaxed), x ^ mask);
+ assert_eq!(b, x & mask != 0);
+ }
+ true
+ }
+ }
+ };
+ ($atomic_type:ty, $int_type:ident) => {
+ __test_atomic_int!($atomic_type, $int_type, single_thread);
+
+ #[test]
+ fn stress_swap() {
+ let (iterations, threads) = stress_test_config();
+ let data1 = &(0..threads)
+ .map(|_| (0..iterations).map(|_| fastrand::$int_type(..)).collect::<Vec<_>>())
+ .collect::<Vec<_>>();
+ let data2 = &(0..threads)
+ .map(|_| (0..iterations).map(|_| fastrand::$int_type(..)).collect::<Vec<_>>())
+ .collect::<Vec<_>>();
+ let set = &data1
+ .iter()
+ .flat_map(|v| v.iter().copied())
+ .chain(data2.iter().flat_map(|v| v.iter().copied()))
+ .collect::<BTreeSet<_>>();
+ let a = &<$atomic_type>::new(data2[0][fastrand::usize(0..iterations)]);
+ let now = &std::time::Instant::now();
+ thread::scope(|s| {
+ for thread in 0..threads {
+ if thread % 2 == 0 {
+ s.spawn(move |_| {
+ let now = *now;
+ for i in 0..iterations {
+ a.store(data1[thread][i], rand_store_ordering());
+ }
+ std::eprintln!("store end={:?}", now.elapsed());
+ });
+ } else {
+ s.spawn(|_| {
+ let now = *now;
+ let mut v = vec![0; iterations];
+ for i in 0..iterations {
+ v[i] = a.load(rand_load_ordering());
+ }
+ std::eprintln!("load end={:?}", now.elapsed());
+ for v in v {
+ assert!(set.contains(&v), "v={}", v);
+ }
+ });
+ }
+ s.spawn(move |_| {
+ let now = *now;
+ let mut v = vec![0; iterations];
+ for i in 0..iterations {
+ v[i] = a.swap(data2[thread][i], rand_swap_ordering());
+ }
+ std::eprintln!("swap end={:?}", now.elapsed());
+ for v in v {
+ assert!(set.contains(&v), "v={}", v);
+ }
+ });
+ }
+ })
+ .unwrap();
+ }
+ #[test]
+ fn stress_compare_exchange() {
+ let (iterations, threads) = stress_test_config();
+ let data1 = &(0..threads)
+ .map(|_| (0..iterations).map(|_| fastrand::$int_type(..)).collect::<Vec<_>>())
+ .collect::<Vec<_>>();
+ let data2 = &(0..threads)
+ .map(|_| (0..iterations).map(|_| fastrand::$int_type(..)).collect::<Vec<_>>())
+ .collect::<Vec<_>>();
+ let set = &data1
+ .iter()
+ .flat_map(|v| v.iter().copied())
+ .chain(data2.iter().flat_map(|v| v.iter().copied()))
+ .collect::<BTreeSet<_>>();
+ let a = &<$atomic_type>::new(data2[0][fastrand::usize(0..iterations)]);
+ let now = &std::time::Instant::now();
+ thread::scope(|s| {
+ for thread in 0..threads {
+ s.spawn(move |_| {
+ let now = *now;
+ for i in 0..iterations {
+ a.store(data1[thread][i], rand_store_ordering());
+ }
+ std::eprintln!("store end={:?}", now.elapsed());
+ });
+ s.spawn(|_| {
+ let now = *now;
+ let mut v = vec![data2[0][0]; iterations];
+ for i in 0..iterations {
+ v[i] = a.load(rand_load_ordering());
+ }
+ std::eprintln!("load end={:?}", now.elapsed());
+ for v in v {
+ assert!(set.contains(&v), "v={}", v);
+ }
+ });
+ s.spawn(move |_| {
+ let now = *now;
+ let mut v = vec![data2[0][0]; iterations];
+ for i in 0..iterations {
+ let old = if i % 2 == 0 {
+ fastrand::$int_type(..)
+ } else {
+ a.load(Ordering::Relaxed)
+ };
+ let new = data2[thread][i];
+ let o = rand_compare_exchange_ordering();
+ match a.compare_exchange(old, new, o.0, o.1) {
+ Ok(r) => assert_eq!(old, r),
+ Err(r) => v[i] = r,
+ }
+ }
+ std::eprintln!("compare_exchange end={:?}", now.elapsed());
+ for v in v {
+ assert!(set.contains(&v), "v={}", v);
+ }
+ });
+ }
+ })
+ .unwrap();
+ }
+ };
+}
+macro_rules! __test_atomic_float {
+ ($atomic_type:ty, $float_type:ident, single_thread) => {
+ use core::$float_type;
+ #[test]
+ fn swap() {
+ let a = <$atomic_type>::new(5.0);
+ test_swap_ordering(|order| a.swap(5.0, order));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ assert_eq!(a.swap(10.0, order), 5.0);
+ assert_eq!(a.swap(5.0, order), 10.0);
+ }
+ }
+ #[test]
+ fn compare_exchange() {
+ let a = <$atomic_type>::new(5.0);
+ test_compare_exchange_ordering(|success, failure| {
+ a.compare_exchange(5.0, 5.0, success, failure)
+ });
+ for &(success, failure) in &test_helper::COMPARE_EXCHANGE_ORDERINGS {
+ let a = <$atomic_type>::new(5.0);
+ assert_eq!(a.compare_exchange(5.0, 10.0, success, failure), Ok(5.0));
+ assert_eq!(a.load(Ordering::Relaxed), 10.0);
+ assert_eq!(a.compare_exchange(6.0, 12.0, success, failure), Err(10.0));
+ assert_eq!(a.load(Ordering::Relaxed), 10.0);
+ }
+ }
+ #[test]
+ fn compare_exchange_weak() {
+ let a = <$atomic_type>::new(4.0);
+ test_compare_exchange_ordering(|success, failure| {
+ a.compare_exchange_weak(4.0, 4.0, success, failure)
+ });
+ for &(success, failure) in &test_helper::COMPARE_EXCHANGE_ORDERINGS {
+ let a = <$atomic_type>::new(4.0);
+ assert_eq!(a.compare_exchange_weak(6.0, 8.0, success, failure), Err(4.0));
+ let mut old = a.load(Ordering::Relaxed);
+ loop {
+ let new = old * 2.0;
+ match a.compare_exchange_weak(old, new, success, failure) {
+ Ok(_) => break,
+ Err(x) => old = x,
+ }
+ }
+ assert_eq!(a.load(Ordering::Relaxed), 8.0);
+ }
+ }
+ #[test]
+ fn fetch_add() {
+ let a = <$atomic_type>::new(0.0);
+ test_swap_ordering(|order| a.fetch_add(0.0, order));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(0.0);
+ assert_eq!(a.fetch_add(10.0, order), 0.0);
+ assert_eq!(a.load(Ordering::Relaxed), 10.0);
+ let a = <$atomic_type>::new($float_type::MAX);
+ assert_eq!(a.fetch_add(1.0, order), $float_type::MAX);
+ assert_eq!(a.load(Ordering::Relaxed), $float_type::MAX + 1.0);
+ }
+ }
+ #[test]
+ fn fetch_sub() {
+ let a = <$atomic_type>::new(20.0);
+ test_swap_ordering(|order| a.fetch_sub(0.0, order));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(20.0);
+ assert_eq!(a.fetch_sub(10.0, order), 20.0);
+ assert_eq!(a.load(Ordering::Relaxed), 10.0);
+ let a = <$atomic_type>::new($float_type::MIN);
+ assert_eq!(a.fetch_sub(1.0, order), $float_type::MIN);
+ assert_eq!(a.load(Ordering::Relaxed), $float_type::MIN - 1.0);
+ }
+ }
+ #[test]
+ fn fetch_max() {
+ let a = <$atomic_type>::new(23.0);
+ test_swap_ordering(|order| a.fetch_max(23.0, order));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(23.0);
+ assert_eq!(a.fetch_max(22.0, order), 23.0);
+ assert_eq!(a.load(Ordering::Relaxed), 23.0);
+ assert_eq!(a.fetch_max(24.0, order), 23.0);
+ assert_eq!(a.load(Ordering::Relaxed), 24.0);
+ }
+ }
+ #[test]
+ fn fetch_min() {
+ let a = <$atomic_type>::new(23.0);
+ test_swap_ordering(|order| a.fetch_min(23.0, order));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(23.0);
+ assert_eq!(a.fetch_min(24.0, order), 23.0);
+ assert_eq!(a.load(Ordering::Relaxed), 23.0);
+ assert_eq!(a.fetch_min(22.0, order), 23.0);
+ assert_eq!(a.load(Ordering::Relaxed), 22.0);
+ }
+ }
+ #[test]
+ fn fetch_neg() {
+ let a = <$atomic_type>::new(5.0);
+ test_swap_ordering(|order| a.fetch_neg(order));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(5.0);
+ assert_eq!(a.fetch_neg(order), 5.0);
+ assert_eq!(a.load(Ordering::Relaxed), -5.0);
+ assert_eq!(a.fetch_neg(order), -5.0);
+ assert_eq!(a.load(Ordering::Relaxed), 5.0);
+ }
+ }
+ #[test]
+ fn fetch_abs() {
+ let a = <$atomic_type>::new(23.0);
+ test_swap_ordering(|order| a.fetch_abs(order));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(-23.0);
+ assert_eq!(a.fetch_abs(order), -23.0);
+ assert_eq!(a.load(Ordering::Relaxed), 23.0);
+ assert_eq!(a.fetch_abs(order), 23.0);
+ assert_eq!(a.load(Ordering::Relaxed), 23.0);
+ }
+ }
+ ::quickcheck::quickcheck! {
+ fn quickcheck_swap(x: $float_type, y: $float_type) -> bool {
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(x);
+ assert_float_op_eq!(a.swap(y, order), x);
+ assert_float_op_eq!(a.swap(x, order), y);
+ }
+ true
+ }
+ fn quickcheck_compare_exchange(x: $float_type, y: $float_type) -> bool {
+ let z = loop {
+ let z = fastrand::$float_type();
+ if z != y {
+ break z;
+ }
+ };
+ for &(success, failure) in &test_helper::COMPARE_EXCHANGE_ORDERINGS {
+ let a = <$atomic_type>::new(x);
+ assert_float_op_eq!(a.compare_exchange(x, y, success, failure).unwrap(), x);
+ assert_float_op_eq!(a.load(Ordering::Relaxed), y);
+ assert_float_op_eq!(
+ a.compare_exchange(z, x, success, failure).unwrap_err(),
+ y,
+ );
+ assert_float_op_eq!(a.load(Ordering::Relaxed), y);
+ }
+ true
+ }
+ fn quickcheck_fetch_add(x: $float_type, y: $float_type) -> bool {
+ if cfg!(all(not(debug_assertions), target_arch = "x86", not(target_feature = "sse2"))) {
+ // TODO: rustc bug:
+ // https://github.com/rust-lang/rust/issues/72327
+ // https://github.com/rust-lang/rust/issues/73288
+ return true;
+ }
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(x);
+ assert_float_op_eq!(a.fetch_add(y, order), x);
+ assert_float_op_eq!(a.load(Ordering::Relaxed), x + y);
+ let a = <$atomic_type>::new(y);
+ assert_float_op_eq!(a.fetch_add(x, order), y);
+ assert_float_op_eq!(a.load(Ordering::Relaxed), y + x);
+ }
+ true
+ }
+ fn quickcheck_fetch_sub(x: $float_type, y: $float_type) -> bool {
+ if cfg!(all(not(debug_assertions), target_arch = "x86", not(target_feature = "sse2"))) {
+ // TODO: rustc bug:
+ // https://github.com/rust-lang/rust/issues/72327
+ // https://github.com/rust-lang/rust/issues/73288
+ return true;
+ }
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(x);
+ assert_float_op_eq!(a.fetch_sub(y, order), x);
+ assert_float_op_eq!(a.load(Ordering::Relaxed), x - y);
+ let a = <$atomic_type>::new(y);
+ assert_float_op_eq!(a.fetch_sub(x, order), y);
+ assert_float_op_eq!(a.load(Ordering::Relaxed), y - x);
+ }
+ true
+ }
+ fn quickcheck_fetch_max(x: $float_type, y: $float_type) -> bool {
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(x);
+ assert_float_op_eq!(a.fetch_max(y, order), x);
+ assert_float_op_eq!(a.load(Ordering::Relaxed), x.max(y));
+ let a = <$atomic_type>::new(y);
+ assert_float_op_eq!(a.fetch_max(x, order), y);
+ assert_float_op_eq!(a.load(Ordering::Relaxed), y.max(x));
+ }
+ true
+ }
+ fn quickcheck_fetch_min(x: $float_type, y: $float_type) -> bool {
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(x);
+ assert_float_op_eq!(a.fetch_min(y, order), x);
+ assert_float_op_eq!(a.load(Ordering::Relaxed), x.min(y));
+ let a = <$atomic_type>::new(y);
+ assert_float_op_eq!(a.fetch_min(x, order), y);
+ assert_float_op_eq!(a.load(Ordering::Relaxed), y.min(x));
+ }
+ true
+ }
+ fn quickcheck_fetch_neg(x: $float_type) -> bool {
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(x);
+ assert_float_op_eq!(a.fetch_neg(order), x);
+ assert_float_op_eq!(a.load(Ordering::Relaxed), -x);
+ assert_float_op_eq!(a.fetch_neg(order), -x);
+ assert_float_op_eq!(a.load(Ordering::Relaxed), x);
+ }
+ true
+ }
+ fn quickcheck_fetch_abs(x: $float_type) -> bool {
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(x);
+ assert_float_op_eq!(a.fetch_abs(order), x);
+ assert_float_op_eq!(a.fetch_abs(order), x.abs());
+ assert_float_op_eq!(a.load(Ordering::Relaxed), x.abs());
+ }
+ true
+ }
+ }
+ };
+ ($atomic_type:ty, $float_type:ident) => {
+ __test_atomic_float!($atomic_type, $float_type, single_thread);
+ // TODO: multi thread
+ };
+}
+macro_rules! __test_atomic_bool {
+ ($atomic_type:ty, single_thread) => {
+ #[test]
+ fn swap() {
+ let a = <$atomic_type>::new(true);
+ test_swap_ordering(|order| a.swap(true, order));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ assert_eq!(a.swap(true, order), true);
+ assert_eq!(a.swap(false, order), true);
+ assert_eq!(a.swap(false, order), false);
+ assert_eq!(a.swap(true, order), false);
+ }
+ }
+ #[test]
+ fn compare_exchange() {
+ let a = <$atomic_type>::new(true);
+ test_compare_exchange_ordering(|success, failure| {
+ a.compare_exchange(true, true, success, failure)
+ });
+ for &(success, failure) in &test_helper::COMPARE_EXCHANGE_ORDERINGS {
+ let a = <$atomic_type>::new(true);
+ assert_eq!(a.compare_exchange(true, false, success, failure), Ok(true));
+ assert_eq!(a.load(Ordering::Relaxed), false);
+ assert_eq!(a.compare_exchange(true, true, success, failure), Err(false));
+ assert_eq!(a.load(Ordering::Relaxed), false);
+ }
+ }
+ #[test]
+ fn compare_exchange_weak() {
+ let a = <$atomic_type>::new(false);
+ test_compare_exchange_ordering(|success, failure| {
+ a.compare_exchange_weak(false, false, success, failure)
+ });
+ for &(success, failure) in &test_helper::COMPARE_EXCHANGE_ORDERINGS {
+ let a = <$atomic_type>::new(false);
+ assert_eq!(a.compare_exchange_weak(true, true, success, failure), Err(false));
+ let mut old = a.load(Ordering::Relaxed);
+ let new = true;
+ loop {
+ match a.compare_exchange_weak(old, new, success, failure) {
+ Ok(_) => break,
+ Err(x) => old = x,
+ }
+ }
+ assert_eq!(a.load(Ordering::Relaxed), true);
+ }
+ }
+ #[test]
+ fn fetch_and() {
+ let a = <$atomic_type>::new(true);
+ test_swap_ordering(|order| assert_eq!(a.fetch_and(true, order), true));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(true);
+ assert_eq!(a.fetch_and(false, order), true);
+ assert_eq!(a.load(Ordering::Relaxed), false);
+ let a = <$atomic_type>::new(true);
+ assert_eq!(a.fetch_and(true, order), true);
+ assert_eq!(a.load(Ordering::Relaxed), true);
+ let a = <$atomic_type>::new(false);
+ assert_eq!(a.fetch_and(false, order), false);
+ assert_eq!(a.load(Ordering::Relaxed), false);
+ let a = <$atomic_type>::new(false);
+ assert_eq!(a.fetch_and(true, order), false);
+ assert_eq!(a.load(Ordering::Relaxed), false);
+ }
+ }
+ #[test]
+ fn and() {
+ let a = <$atomic_type>::new(true);
+ test_swap_ordering(|order| a.and(true, order));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(true);
+ a.and(false, order);
+ assert_eq!(a.load(Ordering::Relaxed), false);
+ let a = <$atomic_type>::new(true);
+ a.and(true, order);
+ assert_eq!(a.load(Ordering::Relaxed), true);
+ let a = <$atomic_type>::new(false);
+ a.and(false, order);
+ assert_eq!(a.load(Ordering::Relaxed), false);
+ let a = <$atomic_type>::new(false);
+ a.and(true, order);
+ assert_eq!(a.load(Ordering::Relaxed), false);
+ }
+ }
+ #[test]
+ fn fetch_or() {
+ let a = <$atomic_type>::new(true);
+ test_swap_ordering(|order| assert_eq!(a.fetch_or(false, order), true));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(true);
+ assert_eq!(a.fetch_or(false, order), true);
+ assert_eq!(a.load(Ordering::Relaxed), true);
+ let a = <$atomic_type>::new(true);
+ assert_eq!(a.fetch_or(true, order), true);
+ assert_eq!(a.load(Ordering::Relaxed), true);
+ let a = <$atomic_type>::new(false);
+ assert_eq!(a.fetch_or(false, order), false);
+ assert_eq!(a.load(Ordering::Relaxed), false);
+ let a = <$atomic_type>::new(false);
+ assert_eq!(a.fetch_or(true, order), false);
+ assert_eq!(a.load(Ordering::Relaxed), true);
+ }
+ }
+ #[test]
+ fn or() {
+ let a = <$atomic_type>::new(true);
+ test_swap_ordering(|order| a.or(false, order));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(true);
+ a.or(false, order);
+ assert_eq!(a.load(Ordering::Relaxed), true);
+ let a = <$atomic_type>::new(true);
+ a.or(true, order);
+ assert_eq!(a.load(Ordering::Relaxed), true);
+ let a = <$atomic_type>::new(false);
+ a.or(false, order);
+ assert_eq!(a.load(Ordering::Relaxed), false);
+ let a = <$atomic_type>::new(false);
+ a.or(true, order);
+ assert_eq!(a.load(Ordering::Relaxed), true);
+ }
+ }
+ #[test]
+ fn fetch_xor() {
+ let a = <$atomic_type>::new(true);
+ test_swap_ordering(|order| assert_eq!(a.fetch_xor(false, order), true));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(true);
+ assert_eq!(a.fetch_xor(false, order), true);
+ assert_eq!(a.load(Ordering::Relaxed), true);
+ let a = <$atomic_type>::new(true);
+ assert_eq!(a.fetch_xor(true, order), true);
+ assert_eq!(a.load(Ordering::Relaxed), false);
+ let a = <$atomic_type>::new(false);
+ assert_eq!(a.fetch_xor(false, order), false);
+ assert_eq!(a.load(Ordering::Relaxed), false);
+ let a = <$atomic_type>::new(false);
+ assert_eq!(a.fetch_xor(true, order), false);
+ assert_eq!(a.load(Ordering::Relaxed), true);
+ }
+ }
+ #[test]
+ fn xor() {
+ let a = <$atomic_type>::new(true);
+ test_swap_ordering(|order| a.xor(false, order));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(true);
+ a.xor(false, order);
+ assert_eq!(a.load(Ordering::Relaxed), true);
+ let a = <$atomic_type>::new(true);
+ a.xor(true, order);
+ assert_eq!(a.load(Ordering::Relaxed), false);
+ let a = <$atomic_type>::new(false);
+ a.xor(false, order);
+ assert_eq!(a.load(Ordering::Relaxed), false);
+ let a = <$atomic_type>::new(false);
+ a.xor(true, order);
+ assert_eq!(a.load(Ordering::Relaxed), true);
+ }
+ }
+ ::quickcheck::quickcheck! {
+ fn quickcheck_compare_exchange(x: bool, y: bool) -> bool {
+ let z = !y;
+ for &(success, failure) in &test_helper::COMPARE_EXCHANGE_ORDERINGS {
+ let a = <$atomic_type>::new(x);
+ assert_eq!(a.compare_exchange(x, y, success, failure).unwrap(), x);
+ assert_eq!(a.load(Ordering::Relaxed), y);
+ assert_eq!(a.compare_exchange(z, x, success, failure).unwrap_err(), y);
+ assert_eq!(a.load(Ordering::Relaxed), y);
+ }
+ true
+ }
+ }
+ };
+ ($atomic_type:ty) => {
+ __test_atomic_bool!($atomic_type, single_thread);
+ // TODO: multi thread
+ };
+}
+macro_rules! __test_atomic_ptr {
+ ($atomic_type:ty, single_thread) => {
+ #[test]
+ fn swap() {
+ let a = <$atomic_type>::new(ptr::null_mut());
+ test_swap_ordering(|order| a.swap(ptr::null_mut(), order));
+ let x = &mut 1;
+ for &order in &test_helper::SWAP_ORDERINGS {
+ assert_eq!(a.swap(x, order), ptr::null_mut());
+ assert_eq!(a.swap(ptr::null_mut(), order), x as _);
+ }
+ }
+ #[test]
+ fn compare_exchange() {
+ let a = <$atomic_type>::new(ptr::null_mut());
+ test_compare_exchange_ordering(|success, failure| {
+ a.compare_exchange(ptr::null_mut(), ptr::null_mut(), success, failure)
+ });
+ for &(success, failure) in &test_helper::COMPARE_EXCHANGE_ORDERINGS {
+ let a = <$atomic_type>::new(ptr::null_mut());
+ let x = &mut 1;
+ assert_eq!(
+ a.compare_exchange(ptr::null_mut(), x, success, failure),
+ Ok(ptr::null_mut()),
+ );
+ assert_eq!(a.load(Ordering::Relaxed), x as _);
+ assert_eq!(
+ a.compare_exchange(ptr::null_mut(), ptr::null_mut(), success, failure),
+ Err(x as _),
+ );
+ assert_eq!(a.load(Ordering::Relaxed), x as _);
+ }
+ }
+ #[test]
+ fn compare_exchange_weak() {
+ let a = <$atomic_type>::new(ptr::null_mut());
+ test_compare_exchange_ordering(|success, failure| {
+ a.compare_exchange_weak(ptr::null_mut(), ptr::null_mut(), success, failure)
+ });
+ for &(success, failure) in &test_helper::COMPARE_EXCHANGE_ORDERINGS {
+ let a = <$atomic_type>::new(ptr::null_mut());
+ let x = &mut 1;
+ assert_eq!(a.compare_exchange_weak(x, x, success, failure), Err(ptr::null_mut()));
+ let mut old = a.load(Ordering::Relaxed);
+ loop {
+ match a.compare_exchange_weak(old, x, success, failure) {
+ Ok(_) => break,
+ Err(x) => old = x,
+ }
+ }
+ assert_eq!(a.load(Ordering::Relaxed), x as _);
+ }
+ }
+ };
+ ($atomic_type:ty) => {
+ __test_atomic_ptr!($atomic_type, single_thread);
+ // TODO: multi thread
+ };
+}
+
+macro_rules! __test_atomic_int_load_store_pub {
+ ($atomic_type:ty, $int_type:ident) => {
+ __test_atomic_pub_common!($atomic_type, $int_type);
+ use std::{boxed::Box, mem};
+ #[test]
+ fn impls() {
+ let a = <$atomic_type>::default();
+ let b = <$atomic_type>::from(0);
+ assert_eq!(a.load(Ordering::SeqCst), b.load(Ordering::SeqCst));
+ assert_eq!(std::format!("{:?}", a), std::format!("{:?}", a.load(Ordering::SeqCst)));
+
+ unsafe {
+ let ptr: *mut Align16<$int_type> = Box::into_raw(Box::new(Align16(0)));
+ assert!(ptr as usize % mem::align_of::<$atomic_type>() == 0);
+ {
+ let a = <$atomic_type>::from_ptr(ptr.cast::<$int_type>());
+ *a.as_ptr() = 1;
+ }
+ assert_eq!((*ptr).0, 1);
+ drop(Box::from_raw(ptr));
+ }
+ }
+ };
+}
+macro_rules! __test_atomic_int_pub {
+ ($atomic_type:ty, $int_type:ident) => {
+ #[test]
+ fn fetch_update() {
+ let a = <$atomic_type>::new(7);
+ test_compare_exchange_ordering(|set, fetch| a.fetch_update(set, fetch, |x| Some(x)));
+ for &(success, failure) in &test_helper::COMPARE_EXCHANGE_ORDERINGS {
+ let a = <$atomic_type>::new(7);
+ assert_eq!(a.fetch_update(success, failure, |_| None), Err(7));
+ assert_eq!(a.fetch_update(success, failure, |x| Some(x + 1)), Ok(7));
+ assert_eq!(a.fetch_update(success, failure, |x| Some(x + 1)), Ok(8));
+ assert_eq!(a.load(Ordering::SeqCst), 9);
+ }
+ }
+ ::quickcheck::quickcheck! {
+ fn quickcheck_fetch_update(x: $int_type, y: $int_type) -> bool {
+ let z = loop {
+ let z = fastrand::$int_type(..);
+ if z != y {
+ break z;
+ }
+ };
+ for &(success, failure) in &test_helper::COMPARE_EXCHANGE_ORDERINGS {
+ let a = <$atomic_type>::new(x);
+ assert_eq!(
+ a.fetch_update(success, failure, |_| Some(y))
+ .unwrap(),
+ x
+ );
+ assert_eq!(
+ a.fetch_update(success, failure, |_| Some(z))
+ .unwrap(),
+ y
+ );
+ assert_eq!(a.load(Ordering::Relaxed), z);
+ assert_eq!(
+ a.fetch_update(success, failure, |z| if z == y { Some(z) } else { None })
+ .unwrap_err(),
+ z
+ );
+ assert_eq!(a.load(Ordering::Relaxed), z);
+ }
+ true
+ }
+ }
+ };
+}
+macro_rules! __test_atomic_float_pub {
+ ($atomic_type:ty, $float_type:ident) => {
+ __test_atomic_pub_common!($atomic_type, $float_type);
+ use std::{boxed::Box, mem};
+ #[test]
+ fn fetch_update() {
+ let a = <$atomic_type>::new(7.0);
+ test_compare_exchange_ordering(|set, fetch| a.fetch_update(set, fetch, |x| Some(x)));
+ for &(success, failure) in &test_helper::COMPARE_EXCHANGE_ORDERINGS {
+ let a = <$atomic_type>::new(7.0);
+ assert_eq!(a.fetch_update(success, failure, |_| None), Err(7.0));
+ assert_eq!(a.fetch_update(success, failure, |x| Some(x + 1.0)), Ok(7.0));
+ assert_eq!(a.fetch_update(success, failure, |x| Some(x + 1.0)), Ok(8.0));
+ assert_eq!(a.load(Ordering::SeqCst), 9.0);
+ }
+ }
+ #[test]
+ fn impls() {
+ let a = <$atomic_type>::default();
+ let b = <$atomic_type>::from(0.0);
+ assert_eq!(a.load(Ordering::SeqCst), b.load(Ordering::SeqCst));
+ assert_eq!(std::format!("{:?}", a), std::format!("{:?}", a.load(Ordering::SeqCst)));
+
+ unsafe {
+ let ptr: *mut Align16<$float_type> = Box::into_raw(Box::new(Align16(0.0)));
+ assert!(ptr as usize % mem::align_of::<$atomic_type>() == 0);
+ {
+ let a = <$atomic_type>::from_ptr(ptr.cast::<$float_type>());
+ *a.as_ptr() = 1.0;
+ }
+ assert_eq!((*ptr).0, 1.0);
+ drop(Box::from_raw(ptr));
+ }
+ }
+ };
+}
+macro_rules! __test_atomic_bool_pub {
+ ($atomic_type:ty) => {
+ __test_atomic_pub_common!($atomic_type, bool);
+ use std::{boxed::Box, mem};
+ #[test]
+ fn fetch_nand() {
+ let a = <$atomic_type>::new(true);
+ test_swap_ordering(|order| assert_eq!(a.fetch_nand(false, order), true));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(true);
+ assert_eq!(a.fetch_nand(false, order), true);
+ assert_eq!(a.load(Ordering::Relaxed), true);
+ let a = <$atomic_type>::new(true);
+ assert_eq!(a.fetch_nand(true, order), true);
+ assert_eq!(a.load(Ordering::Relaxed) as usize, 0);
+ assert_eq!(a.load(Ordering::Relaxed), false);
+ let a = <$atomic_type>::new(false);
+ assert_eq!(a.fetch_nand(false, order), false);
+ assert_eq!(a.load(Ordering::Relaxed), true);
+ let a = <$atomic_type>::new(false);
+ assert_eq!(a.fetch_nand(true, order), false);
+ assert_eq!(a.load(Ordering::Relaxed), true);
+ }
+ }
+ #[test]
+ fn fetch_not() {
+ let a = <$atomic_type>::new(true);
+ test_swap_ordering(|order| a.fetch_not(order));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(true);
+ assert_eq!(a.fetch_not(order), true);
+ assert_eq!(a.load(Ordering::Relaxed), false);
+ let a = <$atomic_type>::new(false);
+ assert_eq!(a.fetch_not(order), false);
+ assert_eq!(a.load(Ordering::Relaxed), true);
+ }
+ }
+ #[test]
+ fn not() {
+ let a = <$atomic_type>::new(true);
+ test_swap_ordering(|order| a.fetch_not(order));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let a = <$atomic_type>::new(true);
+ a.not(order);
+ assert_eq!(a.load(Ordering::Relaxed), false);
+ let a = <$atomic_type>::new(false);
+ a.not(order);
+ assert_eq!(a.load(Ordering::Relaxed), true);
+ }
+ }
+ #[test]
+ fn fetch_update() {
+ let a = <$atomic_type>::new(false);
+ test_compare_exchange_ordering(|set, fetch| a.fetch_update(set, fetch, |x| Some(x)));
+ for &(success, failure) in &test_helper::COMPARE_EXCHANGE_ORDERINGS {
+ let a = <$atomic_type>::new(false);
+ assert_eq!(a.fetch_update(success, failure, |_| None), Err(false));
+ assert_eq!(a.fetch_update(success, failure, |x| Some(!x)), Ok(false));
+ assert_eq!(a.fetch_update(success, failure, |x| Some(!x)), Ok(true));
+ assert_eq!(a.load(Ordering::SeqCst), false);
+ }
+ }
+ #[test]
+ fn impls() {
+ let a = <$atomic_type>::default();
+ let b = <$atomic_type>::from(false);
+ assert_eq!(a.load(Ordering::SeqCst), b.load(Ordering::SeqCst));
+ assert_eq!(std::format!("{:?}", a), std::format!("{:?}", a.load(Ordering::SeqCst)));
+
+ unsafe {
+ let ptr: *mut bool = Box::into_raw(Box::new(false));
+ assert!(ptr as usize % mem::align_of::<$atomic_type>() == 0);
+ {
+ let a = <$atomic_type>::from_ptr(ptr);
+ *a.as_ptr() = true;
+ }
+ assert_eq!((*ptr), true);
+ drop(Box::from_raw(ptr));
+ }
+ }
+ };
+}
+macro_rules! __test_atomic_ptr_pub {
+ ($atomic_type:ty) => {
+ __test_atomic_pub_common!($atomic_type, *mut u8);
+ use sptr::Strict;
+ use std::{boxed::Box, mem};
+ #[test]
+ fn fetch_update() {
+ let a = <$atomic_type>::new(ptr::null_mut());
+ test_compare_exchange_ordering(|set, fetch| a.fetch_update(set, fetch, |x| Some(x)));
+ for &(success, failure) in &test_helper::COMPARE_EXCHANGE_ORDERINGS {
+ let a = <$atomic_type>::new(ptr::null_mut());
+ assert_eq!(a.fetch_update(success, failure, |_| None), Err(ptr::null_mut()));
+ assert_eq!(
+ a.fetch_update(success, failure, |_| Some(&a as *const _ as *mut _)),
+ Ok(ptr::null_mut())
+ );
+ assert_eq!(a.load(Ordering::SeqCst), &a as *const _ as *mut _);
+ }
+ }
+ #[test]
+ fn impls() {
+ let a = <$atomic_type>::default();
+ let b = <$atomic_type>::from(ptr::null_mut());
+ assert_eq!(a.load(Ordering::SeqCst), b.load(Ordering::SeqCst));
+ assert_eq!(std::format!("{:?}", a), std::format!("{:?}", a.load(Ordering::SeqCst)));
+ assert_eq!(std::format!("{:p}", a), std::format!("{:p}", a.load(Ordering::SeqCst)));
+
+ unsafe {
+ let ptr: *mut Align16<*mut u8> = Box::into_raw(Box::new(Align16(ptr::null_mut())));
+ assert!(ptr as usize % mem::align_of::<$atomic_type>() == 0);
+ {
+ let a = <$atomic_type>::from_ptr(ptr.cast::<*mut u8>());
+ *a.as_ptr() = ptr::null_mut::<u8>().wrapping_add(1);
+ }
+ assert_eq!((*ptr).0, ptr::null_mut::<u8>().wrapping_add(1));
+ drop(Box::from_raw(ptr));
+ }
+ }
+ // https://github.com/rust-lang/rust/blob/1.70.0/library/core/tests/atomic.rs#L130-L213
+ #[test]
+ fn ptr_add_null() {
+ let atom = AtomicPtr::<i64>::new(core::ptr::null_mut());
+ assert_eq!(atom.fetch_ptr_add(1, Ordering::SeqCst).addr(), 0);
+ assert_eq!(atom.load(Ordering::SeqCst).addr(), 8);
+
+ assert_eq!(atom.fetch_byte_add(1, Ordering::SeqCst).addr(), 8);
+ assert_eq!(atom.load(Ordering::SeqCst).addr(), 9);
+
+ assert_eq!(atom.fetch_ptr_sub(1, Ordering::SeqCst).addr(), 9);
+ assert_eq!(atom.load(Ordering::SeqCst).addr(), 1);
+
+ assert_eq!(atom.fetch_byte_sub(1, Ordering::SeqCst).addr(), 1);
+ assert_eq!(atom.load(Ordering::SeqCst).addr(), 0);
+ }
+ #[test]
+ fn ptr_add_data() {
+ let num = 0i64;
+ let n = &num as *const i64 as *mut _;
+ let atom = AtomicPtr::<i64>::new(n);
+ assert_eq!(atom.fetch_ptr_add(1, Ordering::SeqCst), n);
+ assert_eq!(atom.load(Ordering::SeqCst), n.wrapping_add(1));
+
+ assert_eq!(atom.fetch_ptr_sub(1, Ordering::SeqCst), n.wrapping_add(1));
+ assert_eq!(atom.load(Ordering::SeqCst), n);
+ let bytes_from_n = |b| n.cast::<u8>().wrapping_add(b).cast::<i64>();
+
+ assert_eq!(atom.fetch_byte_add(1, Ordering::SeqCst), n);
+ assert_eq!(atom.load(Ordering::SeqCst), bytes_from_n(1));
+
+ assert_eq!(atom.fetch_byte_add(5, Ordering::SeqCst), bytes_from_n(1));
+ assert_eq!(atom.load(Ordering::SeqCst), bytes_from_n(6));
+
+ assert_eq!(atom.fetch_byte_sub(1, Ordering::SeqCst), bytes_from_n(6));
+ assert_eq!(atom.load(Ordering::SeqCst), bytes_from_n(5));
+
+ assert_eq!(atom.fetch_byte_sub(5, Ordering::SeqCst), bytes_from_n(5));
+ assert_eq!(atom.load(Ordering::SeqCst), n);
+ }
+ #[test]
+ fn ptr_bitops() {
+ let atom = AtomicPtr::<i64>::new(core::ptr::null_mut());
+ assert_eq!(atom.fetch_or(0b0111, Ordering::SeqCst).addr(), 0);
+ assert_eq!(atom.load(Ordering::SeqCst).addr(), 0b0111);
+
+ assert_eq!(atom.fetch_and(0b1101, Ordering::SeqCst).addr(), 0b0111);
+ assert_eq!(atom.load(Ordering::SeqCst).addr(), 0b0101);
+
+ assert_eq!(atom.fetch_xor(0b1111, Ordering::SeqCst).addr(), 0b0101);
+ assert_eq!(atom.load(Ordering::SeqCst).addr(), 0b1010);
+ }
+ #[test]
+ fn ptr_bitops_tagging() {
+ const MASK_TAG: usize = 0b1111;
+ const MASK_PTR: usize = !MASK_TAG;
+
+ #[repr(align(16))]
+ struct Tagme(u128);
+
+ let tagme = Tagme(1000);
+ let ptr = &tagme as *const Tagme as *mut Tagme;
+ let atom: AtomicPtr<Tagme> = AtomicPtr::new(ptr);
+
+ assert_eq!(ptr.addr() & MASK_TAG, 0);
+
+ assert_eq!(atom.fetch_or(0b0111, Ordering::SeqCst), ptr);
+ assert_eq!(atom.load(Ordering::SeqCst), ptr.map_addr(|a| a | 0b111));
+
+ assert_eq!(
+ atom.fetch_and(MASK_PTR | 0b0010, Ordering::SeqCst),
+ ptr.map_addr(|a| a | 0b111)
+ );
+ assert_eq!(atom.load(Ordering::SeqCst), ptr.map_addr(|a| a | 0b0010));
+
+ assert_eq!(atom.fetch_xor(0b1011, Ordering::SeqCst), ptr.map_addr(|a| a | 0b0010));
+ assert_eq!(atom.load(Ordering::SeqCst), ptr.map_addr(|a| a | 0b1001));
+
+ assert_eq!(atom.fetch_and(MASK_PTR, Ordering::SeqCst), ptr.map_addr(|a| a | 0b1001));
+ assert_eq!(atom.load(Ordering::SeqCst), ptr);
+ }
+ #[test]
+ fn bit_set() {
+ let a = <$atomic_type>::new(ptr::null_mut::<u64>().cast::<u8>().map_addr(|a| a | 1));
+ test_swap_ordering(|order| assert!(a.bit_set(0, order)));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let pointer = &mut 1u64 as *mut u64 as *mut u8;
+ let atom = <$atomic_type>::new(pointer);
+ // Tag the bottom bit of the pointer.
+ assert!(!atom.bit_set(0, order));
+ // Extract and untag.
+ let tagged = atom.load(Ordering::Relaxed);
+ assert_eq!(tagged.addr() & 1, 1);
+ assert_eq!(tagged.map_addr(|p| p & !1), pointer);
+ }
+ }
+ #[test]
+ fn bit_clear() {
+ let a = <$atomic_type>::new(ptr::null_mut::<u64>().cast::<u8>());
+ test_swap_ordering(|order| assert!(!a.bit_clear(0, order)));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let pointer = &mut 1u64 as *mut u64 as *mut u8;
+ // A tagged pointer
+ let atom = <$atomic_type>::new(pointer.map_addr(|a| a | 1));
+ assert!(atom.bit_set(0, order));
+ // Untag
+ assert!(atom.bit_clear(0, order));
+ }
+ }
+ #[test]
+ fn bit_toggle() {
+ let a = <$atomic_type>::new(ptr::null_mut::<u64>().cast::<u8>());
+ test_swap_ordering(|order| a.bit_toggle(0, order));
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let pointer = &mut 1u64 as *mut u64 as *mut u8;
+ let atom = <$atomic_type>::new(pointer);
+ // Toggle a tag bit on the pointer.
+ atom.bit_toggle(0, order);
+ assert_eq!(atom.load(Ordering::Relaxed).addr() & 1, 1);
+ }
+ }
+ };
+}
+
+macro_rules! test_atomic_int_load_store {
+ ($int_type:ident) => {
+ paste::paste! {
+ #[allow(
+ clippy::alloc_instead_of_core,
+ clippy::std_instead_of_alloc,
+ clippy::std_instead_of_core,
+ clippy::undocumented_unsafe_blocks
+ )]
+ mod [<test_atomic_ $int_type>] {
+ use super::*;
+ __test_atomic_int_load_store!([<Atomic $int_type:camel>], $int_type);
+ }
+ }
+ };
+}
+macro_rules! test_atomic_bool_load_store {
+ () => {
+ #[allow(
+ clippy::alloc_instead_of_core,
+ clippy::std_instead_of_alloc,
+ clippy::std_instead_of_core,
+ clippy::undocumented_unsafe_blocks
+ )]
+ mod test_atomic_bool {
+ use super::*;
+ __test_atomic_bool_load_store!(AtomicBool);
+ }
+ };
+}
+macro_rules! test_atomic_ptr_load_store {
+ () => {
+ #[allow(
+ clippy::alloc_instead_of_core,
+ clippy::std_instead_of_alloc,
+ clippy::std_instead_of_core,
+ clippy::undocumented_unsafe_blocks
+ )]
+ mod test_atomic_ptr {
+ use super::*;
+ __test_atomic_ptr_load_store!(AtomicPtr<u8>);
+ }
+ };
+}
+
+macro_rules! test_atomic_int_single_thread {
+ ($int_type:ident) => {
+ paste::paste! {
+ #[allow(
+ clippy::alloc_instead_of_core,
+ clippy::std_instead_of_alloc,
+ clippy::std_instead_of_core,
+ clippy::undocumented_unsafe_blocks
+ )]
+ mod [<test_atomic_ $int_type>] {
+ use super::*;
+ __test_atomic_int_load_store!([<Atomic $int_type:camel>], $int_type, single_thread);
+ __test_atomic_int!([<Atomic $int_type:camel>], $int_type, single_thread);
+ }
+ }
+ };
+}
+macro_rules! test_atomic_bool_single_thread {
+ () => {
+ #[allow(
+ clippy::alloc_instead_of_core,
+ clippy::std_instead_of_alloc,
+ clippy::std_instead_of_core,
+ clippy::undocumented_unsafe_blocks
+ )]
+ mod test_atomic_bool {
+ use super::*;
+ __test_atomic_bool_load_store!(AtomicBool, single_thread);
+ __test_atomic_bool!(AtomicBool, single_thread);
+ }
+ };
+}
+macro_rules! test_atomic_ptr_single_thread {
+ () => {
+ #[allow(
+ clippy::alloc_instead_of_core,
+ clippy::std_instead_of_alloc,
+ clippy::std_instead_of_core,
+ clippy::undocumented_unsafe_blocks
+ )]
+ mod test_atomic_ptr {
+ use super::*;
+ __test_atomic_ptr_load_store!(AtomicPtr<u8>, single_thread);
+ __test_atomic_ptr!(AtomicPtr<u8>, single_thread);
+ }
+ };
+}
+
+macro_rules! test_atomic_int {
+ ($int_type:ident) => {
+ paste::paste! {
+ #[allow(
+ clippy::alloc_instead_of_core,
+ clippy::std_instead_of_alloc,
+ clippy::std_instead_of_core,
+ clippy::undocumented_unsafe_blocks
+ )]
+ mod [<test_atomic_ $int_type>] {
+ use super::*;
+ __test_atomic_int_load_store!([<Atomic $int_type:camel>], $int_type);
+ __test_atomic_int!([<Atomic $int_type:camel>], $int_type);
+ }
+ }
+ };
+}
+macro_rules! test_atomic_bool {
+ () => {
+ #[allow(
+ clippy::alloc_instead_of_core,
+ clippy::std_instead_of_alloc,
+ clippy::std_instead_of_core,
+ clippy::undocumented_unsafe_blocks
+ )]
+ mod test_atomic_bool {
+ use super::*;
+ __test_atomic_bool_load_store!(AtomicBool);
+ __test_atomic_bool!(AtomicBool);
+ }
+ };
+}
+macro_rules! test_atomic_ptr {
+ () => {
+ #[allow(
+ clippy::alloc_instead_of_core,
+ clippy::std_instead_of_alloc,
+ clippy::std_instead_of_core,
+ clippy::undocumented_unsafe_blocks
+ )]
+ #[allow(unstable_name_collisions)] // for sptr crate
+ mod test_atomic_ptr {
+ use super::*;
+ __test_atomic_ptr_load_store!(AtomicPtr<u8>);
+ __test_atomic_ptr!(AtomicPtr<u8>);
+ }
+ };
+}
+
+macro_rules! test_atomic_int_pub {
+ ($int_type:ident) => {
+ paste::paste! {
+ #[allow(
+ clippy::alloc_instead_of_core,
+ clippy::std_instead_of_alloc,
+ clippy::std_instead_of_core,
+ clippy::undocumented_unsafe_blocks
+ )]
+ mod [<test_atomic_ $int_type>] {
+ use super::*;
+ __test_atomic_int_load_store!([<Atomic $int_type:camel>], $int_type);
+ __test_atomic_int!([<Atomic $int_type:camel>], $int_type);
+ __test_atomic_int_load_store_pub!([<Atomic $int_type:camel>], $int_type);
+ __test_atomic_int_pub!([<Atomic $int_type:camel>], $int_type);
+ }
+ }
+ };
+}
+macro_rules! test_atomic_int_load_store_pub {
+ ($int_type:ident) => {
+ paste::paste! {
+ #[allow(
+ clippy::alloc_instead_of_core,
+ clippy::std_instead_of_alloc,
+ clippy::std_instead_of_core,
+ clippy::undocumented_unsafe_blocks
+ )]
+ mod [<test_atomic_ $int_type>] {
+ use super::*;
+ __test_atomic_int_load_store!([<Atomic $int_type:camel>], $int_type);
+ __test_atomic_int_load_store_pub!([<Atomic $int_type:camel>], $int_type);
+ }
+ }
+ };
+}
+#[cfg(feature = "float")]
+macro_rules! test_atomic_float_pub {
+ ($float_type:ident) => {
+ paste::paste! {
+ #[allow(
+ clippy::alloc_instead_of_core,
+ clippy::std_instead_of_alloc,
+ clippy::std_instead_of_core,
+ clippy::undocumented_unsafe_blocks
+ )]
+ mod [<test_atomic_ $float_type>] {
+ use super::*;
+ __test_atomic_float_load_store!([<Atomic $float_type:camel>], $float_type);
+ __test_atomic_float!([<Atomic $float_type:camel>], $float_type);
+ __test_atomic_float_pub!([<Atomic $float_type:camel>], $float_type);
+ }
+ }
+ };
+}
+macro_rules! test_atomic_bool_pub {
+ () => {
+ #[allow(
+ clippy::alloc_instead_of_core,
+ clippy::std_instead_of_alloc,
+ clippy::std_instead_of_core,
+ clippy::undocumented_unsafe_blocks
+ )]
+ mod test_atomic_bool {
+ use super::*;
+ __test_atomic_bool_load_store!(AtomicBool);
+ __test_atomic_bool!(AtomicBool);
+ __test_atomic_bool_pub!(AtomicBool);
+ }
+ };
+}
+macro_rules! test_atomic_ptr_pub {
+ () => {
+ #[allow(
+ clippy::alloc_instead_of_core,
+ clippy::std_instead_of_alloc,
+ clippy::std_instead_of_core,
+ clippy::undocumented_unsafe_blocks
+ )]
+ #[allow(unstable_name_collisions)] // for sptr crate
+ mod test_atomic_ptr {
+ use super::*;
+ __test_atomic_ptr_load_store!(AtomicPtr<u8>);
+ __test_atomic_ptr!(AtomicPtr<u8>);
+ __test_atomic_ptr_pub!(AtomicPtr<u8>);
+ }
+ };
+}
+
+// Asserts that `$a` and `$b` have performed equivalent operations.
+#[cfg(feature = "float")]
+macro_rules! assert_float_op_eq {
+ ($a:expr, $b:expr $(,)?) => {{
+ // See also:
+ // - https://github.com/rust-lang/unsafe-code-guidelines/issues/237.
+ // - https://github.com/rust-lang/portable-simd/issues/39.
+ let a = $a;
+ let b = $b;
+ if a.is_nan() && b.is_nan() // don't check sign of NaN: https://github.com/rust-lang/rust/issues/55131
+ || a.is_infinite()
+ && b.is_infinite()
+ && a.is_sign_positive() == b.is_sign_positive()
+ && a.is_sign_negative() == b.is_sign_negative()
+ {
+ // ok
+ } else {
+ assert_eq!(a, b);
+ }
+ }};
+}
+
+#[allow(clippy::disallowed_methods)] // set_var/remove_var is fine as we run tests with RUST_TEST_THREADS=1
+#[cfg_attr(not(portable_atomic_no_track_caller), track_caller)]
+pub(crate) fn assert_panic<T: std::fmt::Debug>(f: impl FnOnce() -> T) -> std::string::String {
+ let backtrace = std::env::var_os("RUST_BACKTRACE");
+ let hook = std::panic::take_hook();
+ std::env::set_var("RUST_BACKTRACE", "0"); // Suppress backtrace
+ std::panic::set_hook(std::boxed::Box::new(|_| {})); // Suppress panic msg
+ let res = std::panic::catch_unwind(std::panic::AssertUnwindSafe(f));
+ std::panic::set_hook(hook);
+ match backtrace {
+ Some(v) => std::env::set_var("RUST_BACKTRACE", v),
+ None => std::env::remove_var("RUST_BACKTRACE"),
+ }
+ let msg = res.unwrap_err();
+ msg.downcast_ref::<std::string::String>()
+ .cloned()
+ .unwrap_or_else(|| msg.downcast_ref::<&'static str>().copied().unwrap().into())
+}
+pub(crate) fn rand_load_ordering() -> Ordering {
+ test_helper::LOAD_ORDERINGS[fastrand::usize(0..test_helper::LOAD_ORDERINGS.len())]
+}
+pub(crate) fn test_load_ordering<T: std::fmt::Debug>(f: impl Fn(Ordering) -> T) {
+ for &order in &test_helper::LOAD_ORDERINGS {
+ f(order);
+ }
+
+ if !skip_should_panic_test() {
+ assert_eq!(
+ assert_panic(|| f(Ordering::Release)),
+ "there is no such thing as a release load"
+ );
+ assert_eq!(
+ assert_panic(|| f(Ordering::AcqRel)),
+ "there is no such thing as an acquire-release load"
+ );
+ }
+}
+pub(crate) fn rand_store_ordering() -> Ordering {
+ test_helper::STORE_ORDERINGS[fastrand::usize(0..test_helper::STORE_ORDERINGS.len())]
+}
+pub(crate) fn test_store_ordering<T: std::fmt::Debug>(f: impl Fn(Ordering) -> T) {
+ for &order in &test_helper::STORE_ORDERINGS {
+ f(order);
+ }
+
+ if !skip_should_panic_test() {
+ assert_eq!(
+ assert_panic(|| f(Ordering::Acquire)),
+ "there is no such thing as an acquire store"
+ );
+ assert_eq!(
+ assert_panic(|| f(Ordering::AcqRel)),
+ "there is no such thing as an acquire-release store"
+ );
+ }
+}
+pub(crate) fn rand_compare_exchange_ordering() -> (Ordering, Ordering) {
+ test_helper::COMPARE_EXCHANGE_ORDERINGS
+ [fastrand::usize(0..test_helper::COMPARE_EXCHANGE_ORDERINGS.len())]
+}
+pub(crate) fn test_compare_exchange_ordering<T: std::fmt::Debug>(
+ f: impl Fn(Ordering, Ordering) -> T,
+) {
+ for &(success, failure) in &test_helper::COMPARE_EXCHANGE_ORDERINGS {
+ f(success, failure);
+ }
+
+ if !skip_should_panic_test() {
+ for &order in &test_helper::SWAP_ORDERINGS {
+ let msg = assert_panic(|| f(order, Ordering::AcqRel));
+ assert!(
+ msg == "there is no such thing as an acquire-release failure ordering"
+ || msg == "there is no such thing as an acquire-release load",
+ "{}",
+ msg
+ );
+ let msg = assert_panic(|| f(order, Ordering::Release));
+ assert!(
+ msg == "there is no such thing as a release failure ordering"
+ || msg == "there is no such thing as a release load",
+ "{}",
+ msg
+ );
+ }
+ }
+}
+pub(crate) fn rand_swap_ordering() -> Ordering {
+ test_helper::SWAP_ORDERINGS[fastrand::usize(0..test_helper::SWAP_ORDERINGS.len())]
+}
+pub(crate) fn test_swap_ordering<T: std::fmt::Debug>(f: impl Fn(Ordering) -> T) {
+ for &order in &test_helper::SWAP_ORDERINGS {
+ f(order);
+ }
+}
+// for stress test generated by __test_atomic_* macros
+pub(crate) fn stress_test_config() -> (usize, usize) {
+ let iterations = if cfg!(miri) {
+ 50
+ } else if cfg!(debug_assertions) {
+ 5_000
+ } else {
+ 25_000
+ };
+ let threads = if cfg!(debug_assertions) { 2 } else { fastrand::usize(2..=8) };
+ std::eprintln!("threads={}", threads);
+ (iterations, threads)
+}
+fn skip_should_panic_test() -> bool {
+ // Miri's panic handling is slow
+ // MSAN false positive: https://gist.github.com/taiki-e/dd6269a8ffec46284fdc764a4849f884
+ is_panic_abort()
+ || cfg!(miri)
+ || option_env!("CARGO_PROFILE_RELEASE_LTO").map_or(false, |v| v == "fat")
+ && build_context::SANITIZE.contains("memory")
+}
+
+// For -C panic=abort -Z panic_abort_tests: https://github.com/rust-lang/rust/issues/67650
+fn is_panic_abort() -> bool {
+ build_context::PANIC.contains("abort")
+}
+
+#[repr(C, align(16))]
+pub(crate) struct Align16<T>(pub(crate) T);
+
+// Test the cases that should not fail if the memory ordering is implemented correctly.
+// This is still not exhaustive and only tests a few cases.
+// This currently only supports 32-bit or more integers.
+macro_rules! __stress_test_acquire_release {
+ (should_pass, $int_type:ident, $write:ident, $load_order:ident, $store_order:ident) => {
+ paste::paste! {
+ #[test]
+ fn [<load_ $load_order:lower _ $write _ $store_order:lower>]() {
+ __stress_test_acquire_release!([<Atomic $int_type:camel>],
+ $int_type, $write, $load_order, $store_order);
+ }
+ }
+ };
+ (can_panic, $int_type:ident, $write:ident, $load_order:ident, $store_order:ident) => {
+ paste::paste! {
+ // Currently, to make this test work well enough outside of Miri, tens of thousands
+ // of iterations are needed, but this test is slow in some environments.
+ // So, ignore on non-Miri environments by default. See also catch_unwind_on_weak_memory_arch.
+ #[test]
+ #[cfg_attr(not(miri), ignore)]
+ fn [<load_ $load_order:lower _ $write _ $store_order:lower>]() {
+ can_panic("a=", || __stress_test_acquire_release!([<Atomic $int_type:camel>],
+ $int_type, $write, $load_order, $store_order));
+ }
+ }
+ };
+ ($atomic_type:ident, $int_type:ident, $write:ident, $load_order:ident, $store_order:ident) => {{
+ use super::*;
+ use crossbeam_utils::thread;
+ use std::{
+ convert::TryFrom,
+ sync::atomic::{AtomicUsize, Ordering},
+ };
+ let mut n: usize = if cfg!(miri) { 10 } else { 50_000 };
+ // This test is relatively fast because it spawns only one thread, but
+ // the iterations are limited to a maximum value of integers.
+ if $int_type::try_from(n).is_err() {
+ n = $int_type::MAX as usize;
+ }
+ let a = &$atomic_type::new(0);
+ let b = &AtomicUsize::new(0);
+ thread::scope(|s| {
+ s.spawn(|_| {
+ for i in 0..n {
+ b.store(i, Ordering::Relaxed);
+ a.$write(i as _, Ordering::$store_order);
+ }
+ });
+ loop {
+ let a = a.load(Ordering::$load_order);
+ let b = b.load(Ordering::Relaxed);
+ assert!(a as usize <= b, "a={},b={}", a, b);
+ if a as usize == n - 1 {
+ break;
+ }
+ }
+ })
+ .unwrap();
+ }};
+}
+macro_rules! __stress_test_seqcst {
+ (should_pass, $int_type:ident, $write:ident, $load_order:ident, $store_order:ident) => {
+ paste::paste! {
+ // Currently, to make this test work well enough outside of Miri, tens of thousands
+ // of iterations are needed, but this test is very slow in some environments because
+ // it creates two threads for each iteration.
+ // So, ignore on QEMU by default.
+ #[test]
+ #[cfg_attr(qemu, ignore)]
+ fn [<load_ $load_order:lower _ $write _ $store_order:lower>]() {
+ __stress_test_seqcst!([<Atomic $int_type:camel>],
+ $write, $load_order, $store_order);
+ }
+ }
+ };
+ (can_panic, $int_type:ident, $write:ident, $load_order:ident, $store_order:ident) => {
+ paste::paste! {
+ // Currently, to make this test work well enough outside of Miri, tens of thousands
+ // of iterations are needed, but this test is very slow in some environments because
+ // it creates two threads for each iteration.
+ // So, ignore on non-Miri environments by default. See also catch_unwind_on_non_seqcst_arch.
+ #[test]
+ #[cfg_attr(not(miri), ignore)]
+ fn [<load_ $load_order:lower _ $write _ $store_order:lower>]() {
+ can_panic("c=2", || __stress_test_seqcst!([<Atomic $int_type:camel>],
+ $write, $load_order, $store_order));
+ }
+ }
+ };
+ ($atomic_type:ident, $write:ident, $load_order:ident, $store_order:ident) => {{
+ use super::*;
+ use crossbeam_utils::thread;
+ use std::sync::atomic::{AtomicUsize, Ordering};
+ let n: usize = if cfg!(miri) {
+ 8
+ } else if cfg!(valgrind)
+ || build_context::SANITIZE.contains("address")
+ || build_context::SANITIZE.contains("memory")
+ {
+ 50
+ } else if option_env!("GITHUB_ACTIONS").is_some() && cfg!(not(target_os = "linux")) {
+ // GitHub Actions' macOS and Windows runners are slow.
+ 5_000
+ } else {
+ 50_000
+ };
+ let a = &$atomic_type::new(0);
+ let b = &$atomic_type::new(0);
+ let c = &AtomicUsize::new(0);
+ let ready = &AtomicUsize::new(0);
+ thread::scope(|s| {
+ for n in 0..n {
+ a.store(0, Ordering::Relaxed);
+ b.store(0, Ordering::Relaxed);
+ c.store(0, Ordering::Relaxed);
+ let h_a = s.spawn(|_| {
+ while ready.load(Ordering::Relaxed) == 0 {}
+ a.$write(1, Ordering::$store_order);
+ if b.load(Ordering::$load_order) == 0 {
+ c.fetch_add(1, Ordering::Relaxed);
+ }
+ });
+ let h_b = s.spawn(|_| {
+ while ready.load(Ordering::Relaxed) == 0 {}
+ b.$write(1, Ordering::$store_order);
+ if a.load(Ordering::$load_order) == 0 {
+ c.fetch_add(1, Ordering::Relaxed);
+ }
+ });
+ ready.store(1, Ordering::Relaxed);
+ h_a.join().unwrap();
+ h_b.join().unwrap();
+ let c = c.load(Ordering::Relaxed);
+ assert!(c == 0 || c == 1, "c={},n={}", c, n);
+ }
+ })
+ .unwrap();
+ }};
+}
+// Catches unwinding panic on architectures with weak memory models.
+#[allow(dead_code, clippy::used_underscore_binding)]
+pub(crate) fn catch_unwind_on_weak_memory_arch(pat: &str, f: impl Fn()) {
+ // With x86 TSO, RISC-V TSO (optional, not default), SPARC TSO (optional, default),
+ // and IBM-370 memory models should never be a panic here.
+ // Miri emulates weak memory models regardless of target architectures.
+ if cfg!(all(
+ any(
+ target_arch = "x86",
+ target_arch = "x86_64",
+ target_arch = "s390x",
+ target_arch = "sparc",
+ target_arch = "sparc64",
+ ),
+ not(any(miri)),
+ )) {
+ f();
+ } else if !is_panic_abort() {
+ // This could be is_err on architectures with weak memory models.
+ // However, this does not necessarily mean that it will always be panic,
+ // and implementing it with stronger orderings is also okay.
+ match std::panic::catch_unwind(std::panic::AssertUnwindSafe(f)) {
+ Ok(()) => {
+ // panic!();
+ }
+ Err(msg) => {
+ let msg = msg
+ .downcast_ref::<std::string::String>()
+ .cloned()
+ .unwrap_or_else(|| msg.downcast_ref::<&'static str>().copied().unwrap().into());
+ assert!(msg.contains(pat), "{}", msg);
+ }
+ }
+ }
+}
+// Catches unwinding panic on architectures with non-sequentially consistent memory models.
+#[allow(dead_code, clippy::used_underscore_binding)]
+pub(crate) fn catch_unwind_on_non_seqcst_arch(pat: &str, f: impl Fn()) {
+ if !is_panic_abort() {
+ // This could be Err on architectures with non-sequentially consistent memory models.
+ // However, this does not necessarily mean that it will always be panic,
+ // and implementing it with stronger orderings is also okay.
+ match std::panic::catch_unwind(std::panic::AssertUnwindSafe(f)) {
+ Ok(()) => {
+ // panic!();
+ }
+ Err(msg) => {
+ let msg = msg
+ .downcast_ref::<std::string::String>()
+ .cloned()
+ .unwrap_or_else(|| msg.downcast_ref::<&'static str>().copied().unwrap().into());
+ assert!(msg.contains(pat), "{}", msg);
+ }
+ }
+ }
+}
+macro_rules! stress_test_load_store {
+ ($int_type:ident) => {
+ // debug mode is slow.
+ #[cfg(any(not(debug_assertions), miri))]
+ paste::paste! {
+ #[allow(
+ clippy::alloc_instead_of_core,
+ clippy::std_instead_of_alloc,
+ clippy::std_instead_of_core,
+ clippy::undocumented_unsafe_blocks
+ )]
+ mod [<stress_acquire_release_load_store_ $int_type>] {
+ use crate::tests::helper::catch_unwind_on_weak_memory_arch as can_panic;
+ __stress_test_acquire_release!(can_panic, $int_type, store, Relaxed, Relaxed);
+ __stress_test_acquire_release!(can_panic, $int_type, store, Relaxed, Release);
+ __stress_test_acquire_release!(can_panic, $int_type, store, Relaxed, SeqCst);
+ __stress_test_acquire_release!(can_panic, $int_type, store, Acquire, Relaxed);
+ __stress_test_acquire_release!(should_pass, $int_type, store, Acquire, Release);
+ __stress_test_acquire_release!(should_pass, $int_type, store, Acquire, SeqCst);
+ __stress_test_acquire_release!(can_panic, $int_type, store, SeqCst, Relaxed);
+ __stress_test_acquire_release!(should_pass, $int_type, store, SeqCst, Release);
+ __stress_test_acquire_release!(should_pass, $int_type, store, SeqCst, SeqCst);
+ }
+ #[allow(
+ clippy::alloc_instead_of_core,
+ clippy::std_instead_of_alloc,
+ clippy::std_instead_of_core,
+ clippy::undocumented_unsafe_blocks
+ )]
+ mod [<stress_seqcst_load_store_ $int_type>] {
+ use crate::tests::helper::catch_unwind_on_non_seqcst_arch as can_panic;
+ __stress_test_seqcst!(can_panic, $int_type, store, Relaxed, Relaxed);
+ __stress_test_seqcst!(can_panic, $int_type, store, Relaxed, Release);
+ __stress_test_seqcst!(can_panic, $int_type, store, Relaxed, SeqCst);
+ __stress_test_seqcst!(can_panic, $int_type, store, Acquire, Relaxed);
+ __stress_test_seqcst!(can_panic, $int_type, store, Acquire, Release);
+ __stress_test_seqcst!(can_panic, $int_type, store, Acquire, SeqCst);
+ __stress_test_seqcst!(can_panic, $int_type, store, SeqCst, Relaxed);
+ __stress_test_seqcst!(can_panic, $int_type, store, SeqCst, Release);
+ __stress_test_seqcst!(should_pass, $int_type, store, SeqCst, SeqCst);
+ }
+ }
+ };
+}
+macro_rules! stress_test {
+ ($int_type:ident) => {
+ stress_test_load_store!($int_type);
+ // debug mode is slow.
+ #[cfg(any(not(debug_assertions), miri))]
+ paste::paste! {
+ #[allow(
+ clippy::alloc_instead_of_core,
+ clippy::std_instead_of_alloc,
+ clippy::std_instead_of_core,
+ clippy::undocumented_unsafe_blocks
+ )]
+ mod [<stress_acquire_release_load_swap_ $int_type>] {
+ use crate::tests::helper::catch_unwind_on_weak_memory_arch as can_panic;
+ __stress_test_acquire_release!(can_panic, $int_type, swap, Relaxed, Relaxed);
+ __stress_test_acquire_release!(can_panic, $int_type, swap, Relaxed, Acquire);
+ __stress_test_acquire_release!(can_panic, $int_type, swap, Relaxed, Release);
+ __stress_test_acquire_release!(can_panic, $int_type, swap, Relaxed, AcqRel);
+ __stress_test_acquire_release!(can_panic, $int_type, swap, Relaxed, SeqCst);
+ __stress_test_acquire_release!(can_panic, $int_type, swap, Acquire, Relaxed);
+ __stress_test_acquire_release!(can_panic, $int_type, swap, Acquire, Acquire);
+ __stress_test_acquire_release!(should_pass, $int_type, swap, Acquire, Release);
+ __stress_test_acquire_release!(should_pass, $int_type, swap, Acquire, AcqRel);
+ __stress_test_acquire_release!(should_pass, $int_type, swap, Acquire, SeqCst);
+ __stress_test_acquire_release!(can_panic, $int_type, swap, SeqCst, Relaxed);
+ __stress_test_acquire_release!(can_panic, $int_type, swap, SeqCst, Acquire);
+ __stress_test_acquire_release!(should_pass, $int_type, swap, SeqCst, Release);
+ __stress_test_acquire_release!(should_pass, $int_type, swap, SeqCst, AcqRel);
+ __stress_test_acquire_release!(should_pass, $int_type, swap, SeqCst, SeqCst);
+ }
+ #[allow(
+ clippy::alloc_instead_of_core,
+ clippy::std_instead_of_alloc,
+ clippy::std_instead_of_core,
+ clippy::undocumented_unsafe_blocks
+ )]
+ mod [<stress_seqcst_load_swap_ $int_type>] {
+ use crate::tests::helper::catch_unwind_on_non_seqcst_arch as can_panic;
+ __stress_test_seqcst!(can_panic, $int_type, swap, Relaxed, Relaxed);
+ __stress_test_seqcst!(can_panic, $int_type, swap, Relaxed, Acquire);
+ __stress_test_seqcst!(can_panic, $int_type, swap, Relaxed, Release);
+ __stress_test_seqcst!(can_panic, $int_type, swap, Relaxed, AcqRel);
+ __stress_test_seqcst!(can_panic, $int_type, swap, Relaxed, SeqCst);
+ __stress_test_seqcst!(can_panic, $int_type, swap, Acquire, Relaxed);
+ __stress_test_seqcst!(can_panic, $int_type, swap, Acquire, Acquire);
+ __stress_test_seqcst!(can_panic, $int_type, swap, Acquire, Release);
+ __stress_test_seqcst!(can_panic, $int_type, swap, Acquire, AcqRel);
+ __stress_test_seqcst!(can_panic, $int_type, swap, Acquire, SeqCst);
+ __stress_test_seqcst!(can_panic, $int_type, swap, SeqCst, Relaxed);
+ __stress_test_seqcst!(can_panic, $int_type, swap, SeqCst, Acquire);
+ __stress_test_seqcst!(can_panic, $int_type, swap, SeqCst, Release);
+ __stress_test_seqcst!(can_panic, $int_type, swap, SeqCst, AcqRel);
+ __stress_test_seqcst!(should_pass, $int_type, swap, SeqCst, SeqCst);
+ }
+ }
+ };
+}
diff --git a/vendor/portable-atomic/src/tests/mod.rs b/vendor/portable-atomic/src/tests/mod.rs
new file mode 100644
index 0000000..63cdbbd
--- /dev/null
+++ b/vendor/portable-atomic/src/tests/mod.rs
@@ -0,0 +1,357 @@
+// SPDX-License-Identifier: Apache-2.0 OR MIT
+
+#![allow(
+ clippy::alloc_instead_of_core,
+ clippy::std_instead_of_alloc,
+ clippy::std_instead_of_core,
+ clippy::undocumented_unsafe_blocks,
+ clippy::wildcard_imports
+)]
+
+#[macro_use]
+pub(crate) mod helper;
+
+#[allow(dead_code)]
+#[path = "../../version.rs"]
+mod version;
+
+use super::*;
+
+test_atomic_bool_pub!();
+test_atomic_ptr_pub!();
+
+test_atomic_int_pub!(isize);
+test_atomic_int_pub!(usize);
+test_atomic_int_pub!(i8);
+test_atomic_int_pub!(u8);
+test_atomic_int_pub!(i16);
+test_atomic_int_pub!(u16);
+test_atomic_int_pub!(i32);
+test_atomic_int_pub!(u32);
+test_atomic_int_pub!(i64);
+test_atomic_int_pub!(u64);
+test_atomic_int_pub!(i128);
+test_atomic_int_pub!(u128);
+
+#[cfg(feature = "float")]
+test_atomic_float_pub!(f32);
+#[cfg(feature = "float")]
+test_atomic_float_pub!(f64);
+
+#[deny(improper_ctypes)]
+extern "C" {
+ fn _atomic_bool_ffi_safety(_: AtomicBool);
+ fn _atomic_ptr_ffi_safety(_: AtomicPtr<u8>);
+ fn _atomic_isize_ffi_safety(_: AtomicIsize);
+ fn _atomic_usize_ffi_safety(_: AtomicUsize);
+ fn _atomic_i8_ffi_safety(_: AtomicI8);
+ fn _atomic_u8_ffi_safety(_: AtomicU8);
+ fn _atomic_i16_ffi_safety(_: AtomicI16);
+ fn _atomic_u16_ffi_safety(_: AtomicU16);
+ fn _atomic_i32_ffi_safety(_: AtomicI32);
+ fn _atomic_u32_ffi_safety(_: AtomicU32);
+ fn _atomic_i64_ffi_safety(_: AtomicI64);
+ fn _atomic_u64_ffi_safety(_: AtomicU64);
+ // TODO: 128-bit integers are not FFI safe
+ // https://github.com/rust-lang/unsafe-code-guidelines/issues/119
+ // https://github.com/rust-lang/rust/issues/54341
+ // fn _atomic_i128_ffi_safety(_: AtomicI128);
+ // fn _atomic_u128_ffi_safety(_: AtomicU128);
+ #[cfg(feature = "float")]
+ fn _atomic_f32_ffi_safety(_: AtomicF32);
+ #[cfg(feature = "float")]
+ fn _atomic_f64_ffi_safety(_: AtomicF64);
+}
+
+#[test]
+fn test_is_lock_free() {
+ assert!(AtomicI8::is_always_lock_free());
+ assert!(AtomicI8::is_lock_free());
+ assert!(AtomicU8::is_always_lock_free());
+ assert!(AtomicU8::is_lock_free());
+ assert!(AtomicI16::is_always_lock_free());
+ assert!(AtomicI16::is_lock_free());
+ assert!(AtomicU16::is_always_lock_free());
+ assert!(AtomicU16::is_lock_free());
+ assert!(AtomicI32::is_always_lock_free());
+ assert!(AtomicI32::is_lock_free());
+ assert!(AtomicU32::is_always_lock_free());
+ assert!(AtomicU32::is_lock_free());
+ #[cfg(not(portable_atomic_no_cfg_target_has_atomic))]
+ {
+ if cfg!(all(
+ feature = "fallback",
+ target_arch = "arm",
+ not(any(miri, portable_atomic_sanitize_thread)),
+ not(portable_atomic_no_asm),
+ any(target_os = "linux", target_os = "android"),
+ not(any(target_feature = "v6", portable_atomic_target_feature = "v6")),
+ not(portable_atomic_no_outline_atomics),
+ not(target_has_atomic = "64"),
+ not(portable_atomic_test_outline_atomics_detect_false),
+ )) {
+ assert!(!AtomicI64::is_always_lock_free());
+ assert!(AtomicI64::is_lock_free());
+ assert!(!AtomicU64::is_always_lock_free());
+ assert!(AtomicU64::is_lock_free());
+ } else if cfg!(target_has_atomic = "64") {
+ assert!(AtomicI64::is_always_lock_free());
+ assert!(AtomicI64::is_lock_free());
+ assert!(AtomicU64::is_always_lock_free());
+ assert!(AtomicU64::is_lock_free());
+ } else {
+ assert!(!AtomicI64::is_always_lock_free());
+ assert!(!AtomicI64::is_lock_free());
+ assert!(!AtomicU64::is_always_lock_free());
+ assert!(!AtomicU64::is_lock_free());
+ }
+ }
+ if cfg!(portable_atomic_no_asm) && cfg!(not(portable_atomic_unstable_asm)) {
+ assert!(!AtomicI128::is_always_lock_free());
+ assert!(!AtomicI128::is_lock_free());
+ assert!(!AtomicU128::is_always_lock_free());
+ assert!(!AtomicU128::is_lock_free());
+ } else if cfg!(any(
+ target_arch = "aarch64",
+ all(
+ target_arch = "x86_64",
+ any(target_feature = "cmpxchg16b", portable_atomic_target_feature = "cmpxchg16b"),
+ ),
+ all(
+ target_arch = "powerpc64",
+ portable_atomic_unstable_asm_experimental_arch,
+ any(
+ target_feature = "quadword-atomics",
+ portable_atomic_target_feature = "quadword-atomics",
+ ),
+ ),
+ all(target_arch = "s390x", portable_atomic_unstable_asm_experimental_arch),
+ )) {
+ assert!(AtomicI128::is_always_lock_free());
+ assert!(AtomicI128::is_lock_free());
+ assert!(AtomicU128::is_always_lock_free());
+ assert!(AtomicU128::is_lock_free());
+ } else {
+ assert!(!AtomicI128::is_always_lock_free());
+ assert!(!AtomicU128::is_always_lock_free());
+ #[cfg(not(any(target_arch = "x86_64", target_arch = "powerpc64")))]
+ {
+ assert!(!AtomicI128::is_lock_free());
+ assert!(!AtomicU128::is_lock_free());
+ }
+ #[cfg(target_arch = "x86_64")]
+ {
+ let has_cmpxchg16b = cfg!(all(
+ feature = "fallback",
+ not(portable_atomic_no_cmpxchg16b_target_feature),
+ not(portable_atomic_no_outline_atomics),
+ not(any(target_env = "sgx", miri)),
+ not(portable_atomic_test_outline_atomics_detect_false),
+ )) && std::is_x86_feature_detected!("cmpxchg16b");
+ assert_eq!(AtomicI128::is_lock_free(), has_cmpxchg16b);
+ assert_eq!(AtomicU128::is_lock_free(), has_cmpxchg16b);
+ }
+ #[cfg(target_arch = "powerpc64")]
+ {
+ // TODO(powerpc64): is_powerpc_feature_detected is unstable
+ }
+ }
+}
+
+// test version parsing code used in the build script.
+#[test]
+fn test_rustc_version() {
+ use version::Version;
+
+ // rustc 1.34 (rustup)
+ let v = Version::parse(
+ "rustc 1.34.2 (6c2484dc3 2019-05-13)
+binary: rustc
+commit-hash: 6c2484dc3c532c052f159264e970278d8b77cdc9
+commit-date: 2019-05-13
+host: x86_64-apple-darwin
+release: 1.34.2
+LLVM version: 8.0",
+ )
+ .unwrap();
+ assert_eq!(v, Version::stable(34, 8));
+
+ // rustc 1.67 (rustup)
+ let v = Version::parse(
+ "rustc 1.67.0 (fc594f156 2023-01-24)
+binary: rustc
+commit-hash: fc594f15669680fa70d255faec3ca3fb507c3405
+commit-date: 2023-01-24
+host: aarch64-apple-darwin
+release: 1.67.0
+LLVM version: 15.0.6",
+ )
+ .unwrap();
+ assert_eq!(v, Version::stable(67, 15));
+
+ // rustc 1.68-beta (rustup)
+ let v = Version::parse(
+ "rustc 1.68.0-beta.2 (10b73bf73 2023-02-01)
+binary: rustc
+commit-hash: 10b73bf73a6b770cd92ad8ff538173bc3298411c
+commit-date: 2023-02-01
+host: aarch64-apple-darwin
+release: 1.68.0-beta.2
+LLVM version: 15.0.6",
+ )
+ .unwrap();
+ // We do not distinguish between stable and beta because we are only
+ // interested in whether unstable features are potentially available.
+ assert_eq!(v, Version::stable(68, 15));
+
+ // rustc nightly-2019-01-27 (rustup)
+ let v = Version::parse(
+ "rustc 1.33.0-nightly (20c2cba61 2019-01-26)
+binary: rustc
+commit-hash: 20c2cba61dc83e612d25ed496025171caa3db30f
+commit-date: 2019-01-26
+host: x86_64-apple-darwin
+release: 1.33.0-nightly
+LLVM version: 8.0",
+ )
+ .unwrap();
+ assert_eq!(v.minor, 33);
+ assert!(v.nightly);
+ assert_eq!(v.llvm, 8);
+ assert_eq!(v.commit_date().year, 2019);
+ assert_eq!(v.commit_date().month, 1);
+ assert_eq!(v.commit_date().day, 26);
+
+ // rustc 1.69-nightly (rustup)
+ let v = Version::parse(
+ "rustc 1.69.0-nightly (bd39bbb4b 2023-02-07)
+binary: rustc
+commit-hash: bd39bbb4bb92df439bf6d85470e296cc6a47ffbd
+commit-date: 2023-02-07
+host: aarch64-apple-darwin
+release: 1.69.0-nightly
+LLVM version: 15.0.7",
+ )
+ .unwrap();
+ assert_eq!(v.minor, 69);
+ assert!(v.nightly);
+ assert_eq!(v.llvm, 15);
+ assert_eq!(v.commit_date().year, 2023);
+ assert_eq!(v.commit_date().month, 2);
+ assert_eq!(v.commit_date().day, 7);
+
+ // clippy-driver 1.69-nightly (rustup)
+ let v = Version::parse(
+ "rustc 1.69.0-nightly (bd39bbb4b 2023-02-07)
+binary: rustc
+commit-hash: bd39bbb4bb92df439bf6d85470e296cc6a47ffbd
+commit-date: 2023-02-07
+host: aarch64-apple-darwin
+release: 1.69.0-nightly
+LLVM version: 15.0.7",
+ )
+ .unwrap();
+ assert_eq!(v.minor, 69);
+ assert!(v.nightly);
+ assert_eq!(v.llvm, 15);
+ assert_eq!(v.commit_date().year, 2023);
+ assert_eq!(v.commit_date().month, 2);
+ assert_eq!(v.commit_date().day, 7);
+
+ // rustc 1.69-dev (from source: ./x.py build)
+ let v = Version::parse(
+ "rustc 1.69.0-dev
+binary: rustc
+commit-hash: unknown
+commit-date: unknown
+host: aarch64-unknown-linux-gnu
+release: 1.69.0-dev
+LLVM version: 16.0.0",
+ )
+ .unwrap();
+ assert_eq!(v.minor, 69);
+ assert!(v.nightly);
+ assert_eq!(v.llvm, 16);
+ assert_eq!(v.commit_date().year, 0);
+ assert_eq!(v.commit_date().month, 0);
+ assert_eq!(v.commit_date().day, 0);
+
+ // rustc 1.64 (debian 11: apt-get install cargo)
+ let v = Version::parse(
+ "rustc 1.48.0
+binary: rustc
+commit-hash: unknown
+commit-date: unknown
+host: aarch64-unknown-linux-gnu
+release: 1.48.0
+LLVM version: 11.0",
+ )
+ .unwrap();
+ assert_eq!(v, Version::stable(48, 11));
+
+ // rustc 1.67 (fedora: dnf install cargo)
+ let v = Version::parse(
+ "rustc 1.67.0 (fc594f156 2023-01-24) (Fedora 1.67.0-2.fc37)
+binary: rustc
+commit-hash: fc594f15669680fa70d255faec3ca3fb507c3405
+commit-date: 2023-01-24
+host: aarch64-unknown-linux-gnu
+release: 1.67.0
+LLVM version: 15.0.7",
+ )
+ .unwrap();
+ assert_eq!(v, Version::stable(67, 15));
+
+ // rustc 1.64 (alpine: apk add cargo)
+ let v = Version::parse(
+ "rustc 1.64.0
+binary: rustc
+commit-hash: unknown
+commit-date: unknown
+host: aarch64-alpine-linux-musl
+release: 1.64.0
+LLVM version: 15.0.3",
+ )
+ .unwrap();
+ assert_eq!(v, Version::stable(64, 15));
+}
+
+#[cfg(feature = "serde")]
+#[test]
+fn test_serde() {
+ use test_helper::serde::{assert_tokens, DebugPartialEq, Token};
+
+ macro_rules! t {
+ ($atomic_type:ty, $value_type:ident, $token_type:ident) => {
+ std::eprint!("test_serde {} ... ", stringify!($value_type));
+ assert_tokens(&DebugPartialEq(<$atomic_type>::new($value_type::MAX)), &[
+ Token::$token_type($value_type::MAX as _),
+ ]);
+ assert_tokens(&DebugPartialEq(<$atomic_type>::new($value_type::MIN)), &[
+ Token::$token_type($value_type::MIN as _),
+ ]);
+ std::eprintln!("ok");
+ };
+ }
+
+ assert_tokens(&DebugPartialEq(AtomicBool::new(true)), &[Token::Bool(true)]);
+ assert_tokens(&DebugPartialEq(AtomicBool::new(false)), &[Token::Bool(false)]);
+ t!(AtomicIsize, isize, I64);
+ t!(AtomicUsize, usize, U64);
+ t!(AtomicI8, i8, I8);
+ t!(AtomicU8, u8, U8);
+ t!(AtomicI16, i16, I16);
+ t!(AtomicU16, u16, U16);
+ t!(AtomicI32, i32, I32);
+ t!(AtomicU32, u32, U32);
+ t!(AtomicI64, i64, I64);
+ t!(AtomicU64, u64, U64);
+ // TODO: serde_test doesn't support Token::{I128,U128}: https://github.com/serde-rs/test/pull/6
+ // t!(AtomicI128, i128, I128);
+ // t!(AtomicU128, u128, U128);
+ #[cfg(feature = "float")]
+ t!(AtomicF32, f32, F32);
+ #[cfg(feature = "float")]
+ #[cfg(not(target_arch = "mips"))] // LLVM 17 (nightly-2023-08-09) bug: assertion failed at core/src/num/diy_float.rs:78:9
+ t!(AtomicF64, f64, F64);
+}