Home
last modified time | relevance | path

Searched refs:arch_spin_is_locked (Results 1 – 19 of 19) sorted by relevance

/linux/include/linux/
A Dspinlock_up.h27 #define arch_spin_is_locked(x) ((x)->slock == 0) macro
62 #define arch_spin_is_locked(lock) ((void)(lock), 0) macro
A Dspinlock.h114 #define raw_spin_is_locked(lock) arch_spin_is_locked(&(lock)->raw_lock)
/linux/arch/riscv/include/asm/
A Dspinlock.h20 #define arch_spin_is_locked(x) (READ_ONCE((x)->lock) != 0) macro
44 if (arch_spin_is_locked(lock)) in arch_spin_lock()
/linux/tools/include/linux/
A Dspinlock.h35 static inline bool arch_spin_is_locked(arch_spinlock_t *mutex) in arch_spin_is_locked() function
/linux/arch/hexagon/include/asm/
A Dspinlock.h156 #define arch_spin_is_locked(x) ((x)->lock != 0) macro
/linux/arch/alpha/include/asm/
A Dspinlock.h17 #define arch_spin_is_locked(x) ((x)->lock != 0) macro
/linux/include/asm-generic/
A Dqspinlock.h114 #define arch_spin_is_locked(l) queued_spin_is_locked(l) macro
A Dqrwlock.h128 return arch_spin_is_locked(&lock->wait_lock); in queued_rwlock_is_contended()
/linux/arch/sh/include/asm/
A Dspinlock-cas.h26 #define arch_spin_is_locked(x) ((x)->lock <= 0) macro
A Dspinlock-llsc.h18 #define arch_spin_is_locked(x) ((x)->lock <= 0) macro
/linux/arch/csky/include/asm/
A Dspinlock.h74 static inline int arch_spin_is_locked(arch_spinlock_t *lock) in arch_spin_is_locked() function
/linux/arch/parisc/include/asm/
A Dspinlock.h10 static inline int arch_spin_is_locked(arch_spinlock_t *x) in arch_spin_is_locked() function
/linux/arch/sparc/include/asm/
A Dspinlock_32.h16 #define arch_spin_is_locked(lock) (*((volatile unsigned char *)(lock)) != 0) macro
/linux/arch/s390/include/asm/
A Dspinlock.h53 static inline int arch_spin_is_locked(arch_spinlock_t *lp) in arch_spin_is_locked() function
/linux/arch/powerpc/include/asm/
A Dsimple_spinlock.h39 static inline int arch_spin_is_locked(arch_spinlock_t *lock) in arch_spin_is_locked() function
/linux/arch/arm/include/asm/
A Dspinlock.h119 static inline int arch_spin_is_locked(arch_spinlock_t *lock) in arch_spin_is_locked() function
/linux/arch/ia64/include/asm/
A Dspinlock.h101 static inline int arch_spin_is_locked(arch_spinlock_t *lock) in arch_spin_is_locked() function
/linux/arch/arc/include/asm/
A Dspinlock.h13 #define arch_spin_is_locked(x) ((x)->slock != __ARCH_SPIN_LOCK_UNLOCKED__) macro
/linux/arch/x86/kernel/
A Dhpet.c812 if (arch_spin_is_locked(&old.lock)) in read_hpet()
844 } while ((new.value == old.value) && arch_spin_is_locked(&new.lock)); in read_hpet()

Completed in 18 milliseconds