arch_spinlock_t    47 arch/x86/include/asm/spinlock.h static inline void __ticket_enter_slowpath(arch_spinlock_t *lock)
arch_spinlock_t    53 arch/x86/include/asm/spinlock.h static __always_inline void __ticket_lock_spinning(arch_spinlock_t *lock,
arch_spinlock_t    57 arch/x86/include/asm/spinlock.h static inline void __ticket_unlock_kick(arch_spinlock_t *lock,
arch_spinlock_t    64 arch/x86/include/asm/spinlock.h static __always_inline int arch_spin_value_unlocked(arch_spinlock_t lock)
arch_spinlock_t    82 arch/x86/include/asm/spinlock.h static __always_inline void arch_spin_lock(arch_spinlock_t *lock)
arch_spinlock_t   104 arch/x86/include/asm/spinlock.h static __always_inline int arch_spin_trylock(arch_spinlock_t *lock)
arch_spinlock_t   106 arch/x86/include/asm/spinlock.h 	arch_spinlock_t old, new;
arch_spinlock_t   118 arch/x86/include/asm/spinlock.h static inline void __ticket_unlock_slowpath(arch_spinlock_t *lock,
arch_spinlock_t   119 arch/x86/include/asm/spinlock.h 					    arch_spinlock_t old)
arch_spinlock_t   121 arch/x86/include/asm/spinlock.h 	arch_spinlock_t new;
arch_spinlock_t   146 arch/x86/include/asm/spinlock.h static __always_inline void arch_spin_unlock(arch_spinlock_t *lock)
arch_spinlock_t   150 arch/x86/include/asm/spinlock.h 		arch_spinlock_t prev;
arch_spinlock_t   163 arch/x86/include/asm/spinlock.h static inline int arch_spin_is_locked(arch_spinlock_t *lock)
arch_spinlock_t   170 arch/x86/include/asm/spinlock.h static inline int arch_spin_is_contended(arch_spinlock_t *lock)
arch_spinlock_t   178 arch/x86/include/asm/spinlock.h static __always_inline void arch_spin_lock_flags(arch_spinlock_t *lock,
arch_spinlock_t   184 arch/x86/include/asm/spinlock.h static inline void arch_spin_unlock_wait(arch_spinlock_t *lock)
arch_spinlock_t    18 include/asm-generic/bitops/atomic.h extern arch_spinlock_t __atomic_hash[ATOMIC_HASH_SIZE] __lock_aligned;
arch_spinlock_t    23 include/asm-generic/bitops/atomic.h 	arch_spinlock_t *s = ATOMIC_HASH(l);	\
arch_spinlock_t    29 include/asm-generic/bitops/atomic.h 	arch_spinlock_t *s = ATOMIC_HASH(l);		\
arch_spinlock_t    13 include/asm-generic/qrwlock_types.h 	arch_spinlock_t		lock;
arch_spinlock_t    37 include/linux/lglock.h 	arch_spinlock_t __percpu *lock;
arch_spinlock_t    45 include/linux/lglock.h 	static DEFINE_PER_CPU(arch_spinlock_t, name ## _lock)		\
arch_spinlock_t    50 include/linux/lglock.h 	static DEFINE_PER_CPU(arch_spinlock_t, name ## _lock)		\
arch_spinlock_t    21 include/linux/spinlock_types.h 	arch_spinlock_t raw_lock;
arch_spinlock_t    28 include/linux/spinlock_up.h static inline void arch_spin_lock(arch_spinlock_t *lock)
arch_spinlock_t    35 include/linux/spinlock_up.h arch_spin_lock_flags(arch_spinlock_t *lock, unsigned long flags)
arch_spinlock_t    42 include/linux/spinlock_up.h static inline int arch_spin_trylock(arch_spinlock_t *lock)
arch_spinlock_t    52 include/linux/spinlock_up.h static inline void arch_spin_unlock(arch_spinlock_t *lock)