/linux-master/arch/powerpc/lib/ |
H A D | qspinlock.c | 125 u32 prev, tmp; local 144 : "=&r" (prev), "=&r" (tmp) 151 return prev; 163 u32 prev, tmp; local 174 : "=&r" (prev), "=&r"(tmp) 178 return prev; 183 u32 prev; local 190 : "=&r" (prev) 194 return prev; 199 u32 prev; local 215 u32 prev; local 576 struct qnode *prev = get_tail_qnode(lock, prev_cpu); local [all...] |
/linux-master/arch/powerpc/mm/book3s32/ |
H A D | mmu_context.c | 114 void switch_mmu_context(struct mm_struct *prev, struct mm_struct *next, struct task_struct *tsk) argument
|
/linux-master/arch/powerpc/mm/book3s64/ |
H A D | mmu_context.c | 325 void radix__switch_mmu_context(struct mm_struct *prev, struct mm_struct *next) argument
|
H A D | slice.c | 326 unsigned long found, prev; local 357 else if (slice_scan_available(addr - 1, available, 0, &prev)) { 358 addr = prev;
|
/linux-master/arch/powerpc/mm/ |
H A D | mem.c | 193 unsigned long spfn, epfn, prev = 0; local 197 if (prev && prev < spfn) 198 register_nosave_region(prev, spfn); 200 prev = epfn;
|
H A D | mmu_context.c | 43 void switch_mm_irqs_off(struct mm_struct *prev, struct mm_struct *next, argument 87 if (prev == next) 98 membarrier_arch_switch_mm(prev, next, tsk); 104 switch_mmu_context(prev, next, tsk); 106 VM_WARN_ON_ONCE(!cpumask_test_cpu(cpu, mm_cpumask(prev)));
|
/linux-master/arch/powerpc/mm/nohash/ |
H A D | mmu_context.c | 230 void switch_mmu_context(struct mm_struct *prev, struct mm_struct *next, argument 243 if (prev) { 244 WARN_ON(prev->context.active < 1); 245 prev->context.active--; 383 * we'll get called with prev == init_mm the first time
|
/linux-master/arch/powerpc/perf/ |
H A D | 8xx-pmu.c | 122 s64 prev, val = 0, delta = 0; local 128 prev = local64_read(&event->hw.prev_count); 132 delta = 16 * (val - prev); 136 delta = prev - val; 142 delta = (s64)((s32)val - (s32)prev); 146 delta = (s64)((s32)val - (s32)prev); 149 } while (local64_cmpxchg(&event->hw.prev_count, prev, val) != prev);
|
H A D | core-book3s.c | 1138 static u64 check_and_compute_delta(u64 prev, u64 val) argument 1140 u64 delta = (val - prev) & 0xfffffffful; 1151 if (prev > val && (prev - val) < 256) 1159 s64 val, delta, prev; local 1179 prev = local64_read(&event->hw.prev_count); 1182 delta = check_and_compute_delta(prev, val); 1185 } while (local64_cmpxchg(&event->hw.prev_count, prev, val) != prev); 1199 prev 1221 u64 val, prev, delta; local 1241 u64 val, prev; local 2241 s64 prev, delta, left; local [all...] |
H A D | core-fsl-emb.c | 164 s64 val, delta, prev; local 175 prev = local64_read(&event->hw.prev_count); 178 } while (local64_cmpxchg(&event->hw.prev_count, prev, val) != prev); 181 delta = (val - prev) & 0xfffffffful; 593 s64 prev, delta, left; local 602 prev = local64_read(&event->hw.prev_count); 603 delta = (val - prev) & 0xfffffffful;
|
H A D | hv-24x7.c | 1464 s64 prev; local 1466 prev = local64_xchg(&event->hw.prev_count, now); 1467 local64_add(now - prev, &event->count);
|
H A D | hv-gpci.c | 748 s64 prev; local 750 prev = local64_xchg(&event->hw.prev_count, now); 751 local64_add(now - prev, &event->count);
|
/linux-master/arch/powerpc/platforms/85xx/ |
H A D | smp.c | 71 u64 prev; local 77 prev = timebase; 80 } while (prev != timebase);
|
/linux-master/arch/powerpc/platforms/cell/spufs/ |
H A D | sched.c | 279 list_for_each_entry(ctx, gang->aff_ref_ctx->aff_list.prev, aff_list) { 369 list_for_each_entry(spu, ref->aff_list.prev, aff_list) {
|
H A D | spufs.h | 346 extern int spu_save(struct spu_state *prev, struct spu *spu); 348 extern int spu_switch(struct spu_state *prev, struct spu_state *new,
|
H A D | switch.c | 1775 static int quiece_spu(struct spu_state *prev, struct spu *spu) argument 1787 if (check_spu_isolate(prev, spu)) { /* Step 2. */ 1790 disable_interrupts(prev, spu); /* Step 3. */ 1791 set_watchdog_timer(prev, spu); /* Step 4. */ 1792 inhibit_user_access(prev, spu); /* Step 5. */ 1793 if (check_spu_isolate(prev, spu)) { /* Step 6. */ 1796 set_switch_pending(prev, spu); /* Step 7. */ 1797 save_mfc_cntl(prev, spu); /* Step 8. */ 1798 save_spu_runcntl(prev, spu); /* Step 9. */ 1799 save_mfc_sr1(prev, sp 1812 save_csa(struct spu_state *prev, struct spu *spu) argument 1848 save_lscsa(struct spu_state *prev, struct spu *spu) argument 1916 harvest(struct spu_state *prev, struct spu *spu) argument 2022 __do_spu_save(struct spu_state *prev, struct spu *spu) argument 2092 spu_save(struct spu_state *prev, struct spu *spu) argument [all...] |
/linux-master/arch/powerpc/platforms/powermac/ |
H A D | low_i2c.c | 963 struct device_node *prev = NULL; local 969 if (prev && bus->flags & pmac_i2c_multibus) { 971 reg = of_get_property(prev, "reg", 979 of_node_put(prev); 983 of_node_put(prev); 984 prev = p;
|
/linux-master/arch/powerpc/platforms/pseries/ |
H A D | mobility.c | 518 u16 prev = mmu_slb_size; local 522 return prev;
|
H A D | papr_scm.c | 422 u64 prev, now; local 433 prev = local64_xchg(&event->hw.prev_count, now); 434 local64_add(now - prev, &event->count);
|
/linux-master/arch/powerpc/sysdev/ |
H A D | fsl_msi.c | 328 if (msi->list.prev != NULL)
|
/linux-master/arch/powerpc/xmon/ |
H A D | xmon.c | 423 int last_speaker = 0, prev; local 446 prev = cmpxchg(&xmon_speaker, last_speaker, me); 447 if (prev == last_speaker)
|
/linux-master/arch/riscv/include/asm/ |
H A D | atomic.h | 201 int prev, rc; local 211 : [p]"=&r" (prev), [rc]"=&r" (rc), [c]"+A" (v->counter) 214 return prev; 221 s64 prev; local 232 : [p]"=&r" (prev), [rc]"=&r" (rc), [c]"+A" (v->counter) 235 return prev; 242 int prev, rc; local 252 : [p]"=&r" (prev), [rc]"=&r" (rc), [c]"+A" (v->counter) 255 return !(prev < 0); 262 int prev, r local 282 int prev, rc; local 303 s64 prev; local 324 s64 prev; local 345 s64 prev; local [all...] |
H A D | membarrier.h | 5 static inline void membarrier_arch_switch_mm(struct mm_struct *prev, argument 19 MEMBARRIER_STATE_GLOBAL_EXPEDITED)) || !prev))
|
H A D | mmu_context.h | 16 void switch_mm(struct mm_struct *prev, struct mm_struct *next, 20 static inline void activate_mm(struct mm_struct *prev, argument 23 switch_mm(prev, next, NULL);
|
H A D | scs.h | 27 .macro scs_load_current_if_task_changed prev 28 beq \prev, tp, _skip_scs 46 .macro scs_load_current_if_task_changed prev
|