Searched refs:smp_mb (Results 1 - 25 of 325) sorted by relevance

1234567891011>>

/linux-master/arch/mips/include/asm/vdso/
H A Dprocessor.h20 #define cpu_relax() smp_mb()
/linux-master/arch/csky/include/asm/
H A Dspinlock.h10 #define smp_mb__after_spinlock() smp_mb()
/linux-master/arch/xtensa/include/asm/
H A Dspinlock.h18 #define smp_mb__after_spinlock() smp_mb()
/linux-master/arch/arm/include/asm/vdso/
H A Dprocessor.h13 smp_mb(); \
/linux-master/arch/arm64/include/asm/
H A Dspinlock.h12 #define smp_mb__after_spinlock() smp_mb()
/linux-master/arch/powerpc/include/asm/
H A Dspinlock.h14 #define smp_mb__after_spinlock() smp_mb()
H A Dmembarrier.h25 smp_mb();
/linux-master/arch/arm/include/asm/
H A Dmcs_spinlock.h12 smp_mb(); \
H A Dspinlock.h78 smp_mb();
100 smp_mb();
109 smp_mb();
155 smp_mb();
175 smp_mb();
184 smp_mb();
224 smp_mb();
231 smp_mb();
266 smp_mb();
H A Dfutex.h28 smp_mb(); \
56 smp_mb();
73 smp_mb();
/linux-master/tools/include/asm/
H A Dbarrier.h44 #ifndef smp_mb
45 # define smp_mb() mb() macro
51 smp_mb(); \
60 smp_mb(); \
/linux-master/arch/riscv/include/asm/
H A Dmembarrier.h47 smp_mb();
/linux-master/arch/arc/include/asm/
H A Dspinlock.h40 smp_mb();
62 smp_mb();
69 smp_mb();
105 smp_mb();
129 smp_mb();
163 smp_mb();
188 smp_mb();
197 smp_mb();
215 smp_mb();
227 * Per lkmm, smp_mb() i
[all...]
H A Dfutex.h20 smp_mb(); \
42 smp_mb() \
48 smp_mb(); \
69 smp_mb() \
133 smp_mb();
160 smp_mb();
H A Datomic64-arcv2.h145 smp_mb();
156 : "cc"); /* memory clobber comes from smp_mb() */
158 smp_mb();
168 smp_mb();
177 : "cc"); /* memory clobber comes from smp_mb() */
179 smp_mb();
189 smp_mb();
201 : "cc"); /* memory clobber comes from smp_mb() */
203 smp_mb();
213 smp_mb();
[all...]
/linux-master/arch/arm64/include/asm/vdso/
H A Dcompat_barrier.h27 #undef smp_mb macro
31 #define smp_mb() aarch32_smp_mb() macro
/linux-master/tools/arch/x86/include/asm/
H A Dbarrier.h29 #define smp_mb() asm volatile("lock; addl $0,-132(%%rsp)" ::: "memory", "cc") macro
/linux-master/arch/alpha/include/asm/
H A Dcmpxchg.h49 smp_mb(); \
52 smp_mb(); \
61 smp_mb(); \
64 smp_mb(); \
H A Datomic.h19 * _release atomics, an smp_mb() is unconditionally inserted into the
72 smp_mb(); \
90 smp_mb(); \
126 smp_mb(); \
145 smp_mb(); \
206 smp_mb();
221 smp_mb();
229 smp_mb();
244 smp_mb();
252 smp_mb();
[all...]
/linux-master/kernel/sched/
H A Dmembarrier.c28 * a: smp_mb()
30 * c: smp_mb()
67 * a: smp_mb()
69 * c: smp_mb()
87 * a: smp_mb()
91 * c: smp_mb()
107 * a: smp_mb()
109 * d: smp_mb()
112 * c: smp_mb()
121 * a: smp_mb()
[all...]
/linux-master/tools/virtio/ringtest/
H A Dmain.h117 #define smp_mb() asm volatile("lock; addl $0,-132(%%rsp)" ::: "memory", "cc") macro
119 #define smp_mb() asm volatile("dmb ish" ::: "memory") macro
125 #define smp_mb() __sync_synchronize() macro
188 smp_mb(); /* Enforce dependency ordering from x */ \
H A Dring.c183 smp_mb();
193 smp_mb();
215 smp_mb();
260 smp_mb();
/linux-master/arch/hexagon/include/asm/
H A Dspinlock.h111 smp_mb();
132 smp_mb();
/linux-master/arch/sh/kernel/
H A Dftrace.c137 smp_mb();
143 smp_mb();
174 smp_mb();
179 smp_mb();
184 smp_mb();
/linux-master/tools/arch/arm64/include/asm/
H A Dbarrier.h23 #define smp_mb() asm volatile("dmb ish" ::: "memory") macro

Completed in 473 milliseconds

1234567891011>>