//! Wrapper implementations for [`lock_api`]. //! //! This module does not provide any particular mutex implementation by itself, but rather can be //! used to add dependency tracking to mutexes that already exist. It implements all of the traits //! in `lock_api` based on the one it wraps. Crates such as `spin` and `parking_lot` provide base //! primitives that can be wrapped. //! //! Wrapped mutexes are at least one `usize` larger than the types they wrapped, and must be aligned //! to `usize` boundaries. As such, libraries with many mutexes may want to consider the additional //! required memory. use lock_api::GuardNoSend; use lock_api::RawMutex; use lock_api::RawMutexFair; use lock_api::RawMutexTimed; use lock_api::RawRwLock; use lock_api::RawRwLockDowngrade; use lock_api::RawRwLockFair; use lock_api::RawRwLockRecursive; use lock_api::RawRwLockRecursiveTimed; use lock_api::RawRwLockTimed; use lock_api::RawRwLockUpgrade; use lock_api::RawRwLockUpgradeDowngrade; use lock_api::RawRwLockUpgradeFair; use lock_api::RawRwLockUpgradeTimed; use crate::LazyMutexId; /// Tracing wrapper for all [`lock_api`] traits. /// /// This wrapper implements any of the locking traits available, given that the wrapped type /// implements them. As such, this wrapper can be used both for normal mutexes and rwlocks. #[derive(Debug, Default)] pub struct TracingWrapper { inner: T, // Need to use a lazy mutex ID to intialize statically. id: LazyMutexId, } impl TracingWrapper { /// Mark this lock as held in the dependency graph. fn mark_held(&self) { self.id.mark_held(); } /// Mark this lock as released in the dependency graph. /// /// # Safety /// /// This function should only be called when the lock has been previously acquired by this /// thread. unsafe fn mark_released(&self) { self.id.mark_released(); } /// First mark ourselves as held, then call the locking function. fn lock(&self, f: impl FnOnce()) { self.mark_held(); f(); } /// First call the unlocking function, then mark ourselves as realeased. unsafe fn unlock(&self, f: impl FnOnce()) { f(); self.mark_released(); } /// Conditionally lock the mutex. /// /// First acquires the lock, then runs the provided function. If that function returns true, /// then the lock is kept, otherwise the mutex is immediately marked as relased. /// /// # Returns /// /// The value returned from the callback. fn conditionally_lock(&self, f: impl FnOnce() -> bool) -> bool { // Mark as locked while we try to do the thing self.mark_held(); if f() { true } else { // Safety: we just locked it above. unsafe { self.mark_released() } false } } } unsafe impl RawMutex for TracingWrapper where T: RawMutex, { // Known issue with legacy initialisers, allow #[allow(clippy::declare_interior_mutable_const)] const INIT: Self = Self { inner: T::INIT, id: LazyMutexId::new(), }; /// Always equal to [`GuardNoSend`], as an implementation detail in the tracking system requires /// this behaviour. May change in the future to reflect the actual guard type from the wrapped /// primitive. type GuardMarker = GuardNoSend; fn lock(&self) { self.lock(|| self.inner.lock()); } fn try_lock(&self) -> bool { self.conditionally_lock(|| self.inner.try_lock()) } unsafe fn unlock(&self) { self.unlock(|| self.inner.unlock()); } fn is_locked(&self) -> bool { // Can't use the default implementation as the inner type might've overwritten it. self.inner.is_locked() } } unsafe impl RawMutexFair for TracingWrapper where T: RawMutexFair, { unsafe fn unlock_fair(&self) { self.unlock(|| self.inner.unlock_fair()) } unsafe fn bump(&self) { // Bumping effectively doesn't change which locks are held, so we don't need to manage the // lock state. self.inner.bump(); } } unsafe impl RawMutexTimed for TracingWrapper where T: RawMutexTimed, { type Duration = T::Duration; type Instant = T::Instant; fn try_lock_for(&self, timeout: Self::Duration) -> bool { self.conditionally_lock(|| self.inner.try_lock_for(timeout)) } fn try_lock_until(&self, timeout: Self::Instant) -> bool { self.conditionally_lock(|| self.inner.try_lock_until(timeout)) } } unsafe impl RawRwLock for TracingWrapper where T: RawRwLock, { // Known issue with legacy initialisers, allow #[allow(clippy::declare_interior_mutable_const)] const INIT: Self = Self { inner: T::INIT, id: LazyMutexId::new(), }; /// Always equal to [`GuardNoSend`], as an implementation detail in the tracking system requires /// this behaviour. May change in the future to reflect the actual guard type from the wrapped /// primitive. type GuardMarker = GuardNoSend; fn lock_shared(&self) { self.lock(|| self.inner.lock_shared()); } fn try_lock_shared(&self) -> bool { self.conditionally_lock(|| self.inner.try_lock_shared()) } unsafe fn unlock_shared(&self) { self.unlock(|| self.inner.unlock_shared()); } fn lock_exclusive(&self) { self.lock(|| self.inner.lock_exclusive()); } fn try_lock_exclusive(&self) -> bool { self.conditionally_lock(|| self.inner.try_lock_exclusive()) } unsafe fn unlock_exclusive(&self) { self.unlock(|| self.inner.unlock_exclusive()); } fn is_locked(&self) -> bool { self.inner.is_locked() } } unsafe impl RawRwLockDowngrade for TracingWrapper where T: RawRwLockDowngrade, { unsafe fn downgrade(&self) { // Downgrading does not require tracking self.inner.downgrade() } } unsafe impl RawRwLockUpgrade for TracingWrapper where T: RawRwLockUpgrade, { fn lock_upgradable(&self) { self.lock(|| self.inner.lock_upgradable()); } fn try_lock_upgradable(&self) -> bool { self.conditionally_lock(|| self.inner.try_lock_upgradable()) } unsafe fn unlock_upgradable(&self) { self.unlock(|| self.inner.unlock_upgradable()); } unsafe fn upgrade(&self) { self.inner.upgrade(); } unsafe fn try_upgrade(&self) -> bool { self.inner.try_upgrade() } } unsafe impl RawRwLockFair for TracingWrapper where T: RawRwLockFair, { unsafe fn unlock_shared_fair(&self) { self.unlock(|| self.inner.unlock_shared_fair()); } unsafe fn unlock_exclusive_fair(&self) { self.unlock(|| self.inner.unlock_exclusive_fair()); } unsafe fn bump_shared(&self) { self.inner.bump_shared(); } unsafe fn bump_exclusive(&self) { self.inner.bump_exclusive(); } } unsafe impl RawRwLockRecursive for TracingWrapper where T: RawRwLockRecursive, { fn lock_shared_recursive(&self) { self.lock(|| self.inner.lock_shared_recursive()); } fn try_lock_shared_recursive(&self) -> bool { self.conditionally_lock(|| self.inner.try_lock_shared_recursive()) } } unsafe impl RawRwLockRecursiveTimed for TracingWrapper where T: RawRwLockRecursiveTimed, { fn try_lock_shared_recursive_for(&self, timeout: Self::Duration) -> bool { self.conditionally_lock(|| self.inner.try_lock_shared_recursive_for(timeout)) } fn try_lock_shared_recursive_until(&self, timeout: Self::Instant) -> bool { self.conditionally_lock(|| self.inner.try_lock_shared_recursive_until(timeout)) } } unsafe impl RawRwLockTimed for TracingWrapper where T: RawRwLockTimed, { type Duration = T::Duration; type Instant = T::Instant; fn try_lock_shared_for(&self, timeout: Self::Duration) -> bool { self.conditionally_lock(|| self.inner.try_lock_shared_for(timeout)) } fn try_lock_shared_until(&self, timeout: Self::Instant) -> bool { self.conditionally_lock(|| self.inner.try_lock_shared_until(timeout)) } fn try_lock_exclusive_for(&self, timeout: Self::Duration) -> bool { self.conditionally_lock(|| self.inner.try_lock_exclusive_for(timeout)) } fn try_lock_exclusive_until(&self, timeout: Self::Instant) -> bool { self.conditionally_lock(|| self.inner.try_lock_exclusive_until(timeout)) } } unsafe impl RawRwLockUpgradeDowngrade for TracingWrapper where T: RawRwLockUpgradeDowngrade, { unsafe fn downgrade_upgradable(&self) { self.inner.downgrade_upgradable() } unsafe fn downgrade_to_upgradable(&self) { self.inner.downgrade_to_upgradable() } } unsafe impl RawRwLockUpgradeFair for TracingWrapper where T: RawRwLockUpgradeFair, { unsafe fn unlock_upgradable_fair(&self) { self.unlock(|| self.inner.unlock_upgradable_fair()) } unsafe fn bump_upgradable(&self) { self.inner.bump_upgradable() } } unsafe impl RawRwLockUpgradeTimed for TracingWrapper where T: RawRwLockUpgradeTimed, { fn try_lock_upgradable_for(&self, timeout: Self::Duration) -> bool { self.conditionally_lock(|| self.inner.try_lock_upgradable_for(timeout)) } fn try_lock_upgradable_until(&self, timeout: Self::Instant) -> bool { self.conditionally_lock(|| self.inner.try_lock_upgradable_until(timeout)) } unsafe fn try_upgrade_for(&self, timeout: Self::Duration) -> bool { self.inner.try_upgrade_for(timeout) } unsafe fn try_upgrade_until(&self, timeout: Self::Instant) -> bool { self.inner.try_upgrade_until(timeout) } }