/linux-6.6.21/include/linux/ |
D | spinlock_up.h | 27 #define arch_spin_is_locked(x) ((x)->slock == 0) macro 62 #define arch_spin_is_locked(lock) ((void)(lock), 0) macro
|
D | spinlock.h | 116 #define raw_spin_is_locked(lock) arch_spin_is_locked(&(lock)->raw_lock)
|
/linux-6.6.21/include/asm-generic/ |
D | spinlock.h | 71 static __always_inline int arch_spin_is_locked(arch_spinlock_t *lock) in arch_spin_is_locked() function 87 return !arch_spin_is_locked(&lock); in arch_spin_value_unlocked()
|
D | qspinlock.h | 143 #define arch_spin_is_locked(l) queued_spin_is_locked(l) macro
|
D | qrwlock.h | 132 return arch_spin_is_locked(&lock->wait_lock); in queued_rwlock_is_contended()
|
/linux-6.6.21/tools/include/linux/ |
D | spinlock.h | 35 static inline bool arch_spin_is_locked(arch_spinlock_t *mutex) in arch_spin_is_locked() function
|
/linux-6.6.21/arch/hexagon/include/asm/ |
D | spinlock.h | 156 #define arch_spin_is_locked(x) ((x)->lock != 0) macro
|
/linux-6.6.21/arch/alpha/include/asm/ |
D | spinlock.h | 17 #define arch_spin_is_locked(x) ((x)->lock != 0) macro
|
/linux-6.6.21/arch/sh/include/asm/ |
D | spinlock-cas.h | 26 #define arch_spin_is_locked(x) ((x)->lock <= 0) macro
|
D | spinlock-llsc.h | 18 #define arch_spin_is_locked(x) ((x)->lock <= 0) macro
|
/linux-6.6.21/arch/sparc/include/asm/ |
D | spinlock_32.h | 16 #define arch_spin_is_locked(lock) (*((volatile unsigned char *)(lock)) != 0) macro
|
/linux-6.6.21/arch/powerpc/include/asm/ |
D | qspinlock.h | 161 #define arch_spin_is_locked(l) queued_spin_is_locked(l) macro
|
D | simple_spinlock.h | 40 static inline int arch_spin_is_locked(arch_spinlock_t *lock) in arch_spin_is_locked() function
|
/linux-6.6.21/arch/s390/include/asm/ |
D | spinlock.h | 53 static inline int arch_spin_is_locked(arch_spinlock_t *lp) in arch_spin_is_locked() function
|
/linux-6.6.21/arch/parisc/include/asm/ |
D | spinlock.h | 19 static inline int arch_spin_is_locked(arch_spinlock_t *x) in arch_spin_is_locked() function
|
/linux-6.6.21/arch/arm/include/asm/ |
D | spinlock.h | 119 static inline int arch_spin_is_locked(arch_spinlock_t *lock) in arch_spin_is_locked() function
|
/linux-6.6.21/arch/arc/include/asm/ |
D | spinlock.h | 13 #define arch_spin_is_locked(x) ((x)->slock != __ARCH_SPIN_LOCK_UNLOCKED__) macro
|
/linux-6.6.21/arch/ia64/include/asm/ |
D | spinlock.h | 101 static inline int arch_spin_is_locked(arch_spinlock_t *lock) in arch_spin_is_locked() function
|
/linux-6.6.21/arch/x86/kernel/ |
D | hpet.c | 812 if (arch_spin_is_locked(&old.lock)) in read_hpet() 844 } while ((new.value == old.value) && arch_spin_is_locked(&new.lock)); in read_hpet()
|
/linux-6.6.21/kernel/rcu/ |
D | tree.c | 4212 if (rcu_rdp_cpu_online(rdp) || arch_spin_is_locked(&rcu_state.ofl_lock)) in rcu_lockdep_current_cpu_online()
|