aboutsummaryrefslogtreecommitdiff
path: root/vendor/rayon-core/src/latch.rs
diff options
context:
space:
mode:
Diffstat (limited to 'vendor/rayon-core/src/latch.rs')
-rw-r--r--vendor/rayon-core/src/latch.rs460
1 files changed, 0 insertions, 460 deletions
diff --git a/vendor/rayon-core/src/latch.rs b/vendor/rayon-core/src/latch.rs
deleted file mode 100644
index b0cbbd8..0000000
--- a/vendor/rayon-core/src/latch.rs
+++ /dev/null
@@ -1,460 +0,0 @@
-use std::marker::PhantomData;
-use std::ops::Deref;
-use std::sync::atomic::{AtomicUsize, Ordering};
-use std::sync::{Arc, Condvar, Mutex};
-use std::usize;
-
-use crate::registry::{Registry, WorkerThread};
-
-/// We define various kinds of latches, which are all a primitive signaling
-/// mechanism. A latch starts as false. Eventually someone calls `set()` and
-/// it becomes true. You can test if it has been set by calling `probe()`.
-///
-/// Some kinds of latches, but not all, support a `wait()` operation
-/// that will wait until the latch is set, blocking efficiently. That
-/// is not part of the trait since it is not possibly to do with all
-/// latches.
-///
-/// The intention is that `set()` is called once, but `probe()` may be
-/// called any number of times. Once `probe()` returns true, the memory
-/// effects that occurred before `set()` become visible.
-///
-/// It'd probably be better to refactor the API into two paired types,
-/// but that's a bit of work, and this is not a public API.
-///
-/// ## Memory ordering
-///
-/// Latches need to guarantee two things:
-///
-/// - Once `probe()` returns true, all memory effects from the `set()`
-/// are visible (in other words, the set should synchronize-with
-/// the probe).
-/// - Once `set()` occurs, the next `probe()` *will* observe it. This
-/// typically requires a seq-cst ordering. See [the "tickle-then-get-sleepy" scenario in the sleep
-/// README](/src/sleep/README.md#tickle-then-get-sleepy) for details.
-pub(super) trait Latch {
- /// Set the latch, signalling others.
- ///
- /// # WARNING
- ///
- /// Setting a latch triggers other threads to wake up and (in some
- /// cases) complete. This may, in turn, cause memory to be
- /// deallocated and so forth. One must be very careful about this,
- /// and it's typically better to read all the fields you will need
- /// to access *before* a latch is set!
- ///
- /// This function operates on `*const Self` instead of `&self` to allow it
- /// to become dangling during this call. The caller must ensure that the
- /// pointer is valid upon entry, and not invalidated during the call by any
- /// actions other than `set` itself.
- unsafe fn set(this: *const Self);
-}
-
-pub(super) trait AsCoreLatch {
- fn as_core_latch(&self) -> &CoreLatch;
-}
-
-/// Latch is not set, owning thread is awake
-const UNSET: usize = 0;
-
-/// Latch is not set, owning thread is going to sleep on this latch
-/// (but has not yet fallen asleep).
-const SLEEPY: usize = 1;
-
-/// Latch is not set, owning thread is asleep on this latch and
-/// must be awoken.
-const SLEEPING: usize = 2;
-
-/// Latch is set.
-const SET: usize = 3;
-
-/// Spin latches are the simplest, most efficient kind, but they do
-/// not support a `wait()` operation. They just have a boolean flag
-/// that becomes true when `set()` is called.
-#[derive(Debug)]
-pub(super) struct CoreLatch {
- state: AtomicUsize,
-}
-
-impl CoreLatch {
- #[inline]
- fn new() -> Self {
- Self {
- state: AtomicUsize::new(0),
- }
- }
-
- /// Invoked by owning thread as it prepares to sleep. Returns true
- /// if the owning thread may proceed to fall asleep, false if the
- /// latch was set in the meantime.
- #[inline]
- pub(super) fn get_sleepy(&self) -> bool {
- self.state
- .compare_exchange(UNSET, SLEEPY, Ordering::SeqCst, Ordering::Relaxed)
- .is_ok()
- }
-
- /// Invoked by owning thread as it falls asleep sleep. Returns
- /// true if the owning thread should block, or false if the latch
- /// was set in the meantime.
- #[inline]
- pub(super) fn fall_asleep(&self) -> bool {
- self.state
- .compare_exchange(SLEEPY, SLEEPING, Ordering::SeqCst, Ordering::Relaxed)
- .is_ok()
- }
-
- /// Invoked by owning thread as it falls asleep sleep. Returns
- /// true if the owning thread should block, or false if the latch
- /// was set in the meantime.
- #[inline]
- pub(super) fn wake_up(&self) {
- if !self.probe() {
- let _ =
- self.state
- .compare_exchange(SLEEPING, UNSET, Ordering::SeqCst, Ordering::Relaxed);
- }
- }
-
- /// Set the latch. If this returns true, the owning thread was sleeping
- /// and must be awoken.
- ///
- /// This is private because, typically, setting a latch involves
- /// doing some wakeups; those are encapsulated in the surrounding
- /// latch code.
- #[inline]
- unsafe fn set(this: *const Self) -> bool {
- let old_state = (*this).state.swap(SET, Ordering::AcqRel);
- old_state == SLEEPING
- }
-
- /// Test if this latch has been set.
- #[inline]
- pub(super) fn probe(&self) -> bool {
- self.state.load(Ordering::Acquire) == SET
- }
-}
-
-impl AsCoreLatch for CoreLatch {
- #[inline]
- fn as_core_latch(&self) -> &CoreLatch {
- self
- }
-}
-
-/// Spin latches are the simplest, most efficient kind, but they do
-/// not support a `wait()` operation. They just have a boolean flag
-/// that becomes true when `set()` is called.
-pub(super) struct SpinLatch<'r> {
- core_latch: CoreLatch,
- registry: &'r Arc<Registry>,
- target_worker_index: usize,
- cross: bool,
-}
-
-impl<'r> SpinLatch<'r> {
- /// Creates a new spin latch that is owned by `thread`. This means
- /// that `thread` is the only thread that should be blocking on
- /// this latch -- it also means that when the latch is set, we
- /// will wake `thread` if it is sleeping.
- #[inline]
- pub(super) fn new(thread: &'r WorkerThread) -> SpinLatch<'r> {
- SpinLatch {
- core_latch: CoreLatch::new(),
- registry: thread.registry(),
- target_worker_index: thread.index(),
- cross: false,
- }
- }
-
- /// Creates a new spin latch for cross-threadpool blocking. Notably, we
- /// need to make sure the registry is kept alive after setting, so we can
- /// safely call the notification.
- #[inline]
- pub(super) fn cross(thread: &'r WorkerThread) -> SpinLatch<'r> {
- SpinLatch {
- cross: true,
- ..SpinLatch::new(thread)
- }
- }
-
- #[inline]
- pub(super) fn probe(&self) -> bool {
- self.core_latch.probe()
- }
-}
-
-impl<'r> AsCoreLatch for SpinLatch<'r> {
- #[inline]
- fn as_core_latch(&self) -> &CoreLatch {
- &self.core_latch
- }
-}
-
-impl<'r> Latch for SpinLatch<'r> {
- #[inline]
- unsafe fn set(this: *const Self) {
- let cross_registry;
-
- let registry: &Registry = if (*this).cross {
- // Ensure the registry stays alive while we notify it.
- // Otherwise, it would be possible that we set the spin
- // latch and the other thread sees it and exits, causing
- // the registry to be deallocated, all before we get a
- // chance to invoke `registry.notify_worker_latch_is_set`.
- cross_registry = Arc::clone((*this).registry);
- &cross_registry
- } else {
- // If this is not a "cross-registry" spin-latch, then the
- // thread which is performing `set` is itself ensuring
- // that the registry stays alive. However, that doesn't
- // include this *particular* `Arc` handle if the waiting
- // thread then exits, so we must completely dereference it.
- (*this).registry
- };
- let target_worker_index = (*this).target_worker_index;
-
- // NOTE: Once we `set`, the target may proceed and invalidate `this`!
- if CoreLatch::set(&(*this).core_latch) {
- // Subtle: at this point, we can no longer read from
- // `self`, because the thread owning this spin latch may
- // have awoken and deallocated the latch. Therefore, we
- // only use fields whose values we already read.
- registry.notify_worker_latch_is_set(target_worker_index);
- }
- }
-}
-
-/// A Latch starts as false and eventually becomes true. You can block
-/// until it becomes true.
-#[derive(Debug)]
-pub(super) struct LockLatch {
- m: Mutex<bool>,
- v: Condvar,
-}
-
-impl LockLatch {
- #[inline]
- pub(super) fn new() -> LockLatch {
- LockLatch {
- m: Mutex::new(false),
- v: Condvar::new(),
- }
- }
-
- /// Block until latch is set, then resets this lock latch so it can be reused again.
- pub(super) fn wait_and_reset(&self) {
- let mut guard = self.m.lock().unwrap();
- while !*guard {
- guard = self.v.wait(guard).unwrap();
- }
- *guard = false;
- }
-
- /// Block until latch is set.
- pub(super) fn wait(&self) {
- let mut guard = self.m.lock().unwrap();
- while !*guard {
- guard = self.v.wait(guard).unwrap();
- }
- }
-}
-
-impl Latch for LockLatch {
- #[inline]
- unsafe fn set(this: *const Self) {
- let mut guard = (*this).m.lock().unwrap();
- *guard = true;
- (*this).v.notify_all();
- }
-}
-
-/// Once latches are used to implement one-time blocking, primarily
-/// for the termination flag of the threads in the pool.
-///
-/// Note: like a `SpinLatch`, once-latches are always associated with
-/// some registry that is probing them, which must be tickled when
-/// they are set. *Unlike* a `SpinLatch`, they don't themselves hold a
-/// reference to that registry. This is because in some cases the
-/// registry owns the once-latch, and that would create a cycle. So a
-/// `OnceLatch` must be given a reference to its owning registry when
-/// it is set. For this reason, it does not implement the `Latch`
-/// trait (but it doesn't have to, as it is not used in those generic
-/// contexts).
-#[derive(Debug)]
-pub(super) struct OnceLatch {
- core_latch: CoreLatch,
-}
-
-impl OnceLatch {
- #[inline]
- pub(super) fn new() -> OnceLatch {
- Self {
- core_latch: CoreLatch::new(),
- }
- }
-
- /// Set the latch, then tickle the specific worker thread,
- /// which should be the one that owns this latch.
- #[inline]
- pub(super) unsafe fn set_and_tickle_one(
- this: *const Self,
- registry: &Registry,
- target_worker_index: usize,
- ) {
- if CoreLatch::set(&(*this).core_latch) {
- registry.notify_worker_latch_is_set(target_worker_index);
- }
- }
-}
-
-impl AsCoreLatch for OnceLatch {
- #[inline]
- fn as_core_latch(&self) -> &CoreLatch {
- &self.core_latch
- }
-}
-
-/// Counting latches are used to implement scopes. They track a
-/// counter. Unlike other latches, calling `set()` does not
-/// necessarily make the latch be considered `set()`; instead, it just
-/// decrements the counter. The latch is only "set" (in the sense that
-/// `probe()` returns true) once the counter reaches zero.
-#[derive(Debug)]
-pub(super) struct CountLatch {
- counter: AtomicUsize,
- kind: CountLatchKind,
-}
-
-enum CountLatchKind {
- /// A latch for scopes created on a rayon thread which will participate in work-
- /// stealing while it waits for completion. This thread is not necessarily part
- /// of the same registry as the scope itself!
- Stealing {
- latch: CoreLatch,
- /// If a worker thread in registry A calls `in_place_scope` on a ThreadPool
- /// with registry B, when a job completes in a thread of registry B, we may
- /// need to call `notify_worker_latch_is_set()` to wake the thread in registry A.
- /// That means we need a reference to registry A (since at that point we will
- /// only have a reference to registry B), so we stash it here.
- registry: Arc<Registry>,
- /// The index of the worker to wake in `registry`
- worker_index: usize,
- },
-
- /// A latch for scopes created on a non-rayon thread which will block to wait.
- Blocking { latch: LockLatch },
-}
-
-impl std::fmt::Debug for CountLatchKind {
- fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
- match self {
- CountLatchKind::Stealing { latch, .. } => {
- f.debug_tuple("Stealing").field(latch).finish()
- }
- CountLatchKind::Blocking { latch, .. } => {
- f.debug_tuple("Blocking").field(latch).finish()
- }
- }
- }
-}
-
-impl CountLatch {
- pub(super) fn new(owner: Option<&WorkerThread>) -> Self {
- Self::with_count(1, owner)
- }
-
- pub(super) fn with_count(count: usize, owner: Option<&WorkerThread>) -> Self {
- Self {
- counter: AtomicUsize::new(count),
- kind: match owner {
- Some(owner) => CountLatchKind::Stealing {
- latch: CoreLatch::new(),
- registry: Arc::clone(owner.registry()),
- worker_index: owner.index(),
- },
- None => CountLatchKind::Blocking {
- latch: LockLatch::new(),
- },
- },
- }
- }
-
- #[inline]
- pub(super) fn increment(&self) {
- let old_counter = self.counter.fetch_add(1, Ordering::Relaxed);
- debug_assert!(old_counter != 0);
- }
-
- pub(super) fn wait(&self, owner: Option<&WorkerThread>) {
- match &self.kind {
- CountLatchKind::Stealing {
- latch,
- registry,
- worker_index,
- } => unsafe {
- let owner = owner.expect("owner thread");
- debug_assert_eq!(registry.id(), owner.registry().id());
- debug_assert_eq!(*worker_index, owner.index());
- owner.wait_until(latch);
- },
- CountLatchKind::Blocking { latch } => latch.wait(),
- }
- }
-}
-
-impl Latch for CountLatch {
- #[inline]
- unsafe fn set(this: *const Self) {
- if (*this).counter.fetch_sub(1, Ordering::SeqCst) == 1 {
- // NOTE: Once we call `set` on the internal `latch`,
- // the target may proceed and invalidate `this`!
- match (*this).kind {
- CountLatchKind::Stealing {
- ref latch,
- ref registry,
- worker_index,
- } => {
- let registry = Arc::clone(registry);
- if CoreLatch::set(latch) {
- registry.notify_worker_latch_is_set(worker_index);
- }
- }
- CountLatchKind::Blocking { ref latch } => LockLatch::set(latch),
- }
- }
- }
-}
-
-/// `&L` without any implication of `dereferenceable` for `Latch::set`
-pub(super) struct LatchRef<'a, L> {
- inner: *const L,
- marker: PhantomData<&'a L>,
-}
-
-impl<L> LatchRef<'_, L> {
- pub(super) fn new(inner: &L) -> LatchRef<'_, L> {
- LatchRef {
- inner,
- marker: PhantomData,
- }
- }
-}
-
-unsafe impl<L: Sync> Sync for LatchRef<'_, L> {}
-
-impl<L> Deref for LatchRef<'_, L> {
- type Target = L;
-
- fn deref(&self) -> &L {
- // SAFETY: if we have &self, the inner latch is still alive
- unsafe { &*self.inner }
- }
-}
-
-impl<L: Latch> Latch for LatchRef<'_, L> {
- #[inline]
- unsafe fn set(this: *const Self) {
- L::set((*this).inner);
- }
-}