Searched refs:arch_spin_lock (Results 1 - 25 of 47) sorted by relevance

12

/linux-master/arch/parisc/include/asm/
H A Dspinlock.h30 static inline void arch_spin_lock(arch_spinlock_t *x) function
85 arch_spin_lock(&(rw->lock_mutex));
109 arch_spin_lock(&(rw->lock_mutex));
144 arch_spin_lock(&(rw->lock_mutex));
155 arch_spin_lock(&(rw->lock_mutex));
H A Dfutex.h24 arch_spin_lock(s);
/linux-master/include/linux/
H A Dspinlock_up.h29 static inline void arch_spin_lock(arch_spinlock_t *lock) function
64 # define arch_spin_lock(lock) do { barrier(); (void)(lock); } while (0) macro
/linux-master/arch/arc/include/asm/
H A Dspinlock.h17 static inline void arch_spin_lock(arch_spinlock_t *lock) function
222 static inline void arch_spin_lock(arch_spinlock_t *lock) function
284 * see pairing version/comment in arch_spin_lock above
304 arch_spin_lock(&(rw->lock_mutex));
328 arch_spin_lock(&(rw->lock_mutex));
363 arch_spin_lock(&(rw->lock_mutex));
374 arch_spin_lock(&(rw->lock_mutex));
H A Dsmp.h113 arch_spin_lock(&smp_atomic_ops_lock); \
/linux-master/kernel/locking/
H A Dqrwlock.c43 arch_spin_lock(&lock->wait_lock);
73 arch_spin_lock(&lock->wait_lock);
/linux-master/tools/include/linux/
H A Dspinlock.h26 static inline void arch_spin_lock(arch_spinlock_t *mutex) function
/linux-master/arch/alpha/include/asm/
H A Dspinlock.h30 static inline void arch_spin_lock(arch_spinlock_t * lock) function
/linux-master/arch/sh/include/asm/
H A Dspinlock-llsc.h26 static inline void arch_spin_lock(arch_spinlock_t *lock) function
33 "movli.l @%2, %0 ! arch_spin_lock \n\t"
H A Dspinlock-cas.h28 static inline void arch_spin_lock(arch_spinlock_t *lock) function
/linux-master/arch/hexagon/include/asm/
H A Dspinlock.h115 static inline void arch_spin_lock(arch_spinlock_t *lock) function
/linux-master/include/asm-generic/
H A Dspinlock.h33 static __always_inline void arch_spin_lock(arch_spinlock_t *lock) function
H A Dqspinlock.h146 #define arch_spin_lock(l) queued_spin_lock(l) macro
/linux-master/arch/sparc/include/asm/
H A Dspinlock_32.h18 static inline void arch_spin_lock(arch_spinlock_t *lock) function
/linux-master/arch/s390/include/asm/
H A Dspinlock.h64 static inline void arch_spin_lock(arch_spinlock_t *lp) function
/linux-master/arch/powerpc/include/asm/
H A Dqspinlock.h164 #define arch_spin_lock(l) queued_spin_lock(l) macro
H A Dsimple_spinlock.h114 static inline void arch_spin_lock(arch_spinlock_t *lock) function
/linux-master/arch/arm/common/
H A Dmcpm_entry.c210 arch_spin_lock(&mcpm_lock);
256 arch_spin_lock(&mcpm_lock);
333 arch_spin_lock(&mcpm_lock);
353 arch_spin_lock(&mcpm_lock);
/linux-master/kernel/kcsan/
H A Dselftest.c165 arch_spin_lock(&arch_spinlock);
194 arch_spin_lock(&arch_spinlock);
226 arch_spin_lock(&arch_spinlock);
/linux-master/kernel/trace/
H A Dtrace_stack.c179 arch_spin_lock(&stack_trace_max_lock);
352 * In case we trace inside arch_spin_lock() or after (NMI),
358 arch_spin_lock(&stack_trace_max_lock);
400 arch_spin_lock(&stack_trace_max_lock);
H A Dtrace_sched_switch.c312 arch_spin_lock(&trace_cmdline_lock);
544 arch_spin_lock(&trace_cmdline_lock);
605 arch_spin_lock(&trace_cmdline_lock);
627 arch_spin_lock(&trace_cmdline_lock);
/linux-master/arch/arm/include/asm/
H A Dspinlock.h56 static inline void arch_spin_lock(arch_spinlock_t *lock) function
/linux-master/arch/x86/mm/
H A Dkmmio.c353 arch_spin_lock(&kmmio_lock);
447 arch_spin_lock(&kmmio_lock);
504 arch_spin_lock(&kmmio_lock);
552 arch_spin_lock(&kmmio_lock);
/linux-master/arch/s390/lib/
H A Dspinlock.c280 arch_spin_lock(&rw->wait);
298 arch_spin_lock(&rw->wait);
/linux-master/arch/x86/kernel/
H A Dtsc_sync.c284 arch_spin_lock(&sync_lock);
307 arch_spin_lock(&sync_lock);

Completed in 309 milliseconds

12