/linux-master/arch/x86/include/asm/ |
H A D | switch_to.h | 9 struct task_struct *__switch_to_asm(struct task_struct *prev, 12 __visible struct task_struct *__switch_to(struct task_struct *prev, 16 __visible void ret_from_fork(struct task_struct *prev, struct pt_regs *regs, 49 #define switch_to(prev, next, last) \ 51 ((last) = __switch_to_asm((prev), (next))); \
|
H A D | cmpxchg_32.h | 24 u64 prev; local 26 : "=A" (prev), 32 return prev; 37 u64 prev; local 39 : "=A" (prev), 45 return prev;
|
/linux-master/arch/s390/include/asm/ |
H A D | uaccess.h | 404 unsigned int prev, shift, mask, _old, _new; local 416 "0: l %[prev],%[address]\n" 417 "1: nr %[prev],%[mask]\n" 419 " or %[new],%[prev]\n" 420 " or %[prev],%[tmp]\n" 421 "2: lr %[tmp],%[prev]\n" 422 "3: cs %[prev],%[new],%[address]\n" 424 " xr %[tmp],%[prev]\n" 431 EX_TABLE_UA_LOAD_REG(0b, 5b, %[rc], %[prev]) 432 EX_TABLE_UA_LOAD_REG(1b, 5b, %[rc], %[prev]) 452 unsigned int prev, shift, mask, _old, _new; local 500 unsigned int prev = old; local 521 unsigned long prev = old; local 542 __uint128_t prev = old; local [all...] |
H A D | mmu_context.h | 72 static inline void switch_mm_irqs_off(struct mm_struct *prev, struct mm_struct *next, argument 84 if (prev != next) 85 cpumask_clear_cpu(cpu, &prev->context.cpu_attach_mask); 89 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, argument 95 switch_mm_irqs_off(prev, next, tsk); 117 static inline void activate_mm(struct mm_struct *prev, argument 120 switch_mm(prev, next, current);
|
/linux-master/arch/um/include/asm/ |
H A D | mmu_context.h | 28 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, argument 33 if(prev != next){ 34 cpumask_clear_cpu(cpu, mm_cpumask(prev));
|
/linux-master/arch/riscv/include/asm/ |
H A D | scs.h | 27 .macro scs_load_current_if_task_changed prev 28 beq \prev, tp, _skip_scs 46 .macro scs_load_current_if_task_changed prev
|
H A D | atomic.h | 201 int prev, rc; local 211 : [p]"=&r" (prev), [rc]"=&r" (rc), [c]"+A" (v->counter) 214 return prev; 221 s64 prev; local 232 : [p]"=&r" (prev), [rc]"=&r" (rc), [c]"+A" (v->counter) 235 return prev; 242 int prev, rc; local 252 : [p]"=&r" (prev), [rc]"=&r" (rc), [c]"+A" (v->counter) 255 return !(prev < 0); 262 int prev, r local 282 int prev, rc; local 303 s64 prev; local 324 s64 prev; local 345 s64 prev; local [all...] |
H A D | switch_to.h | 50 static inline void __switch_to_fpu(struct task_struct *prev, argument 55 regs = task_pt_regs(prev); 56 fstate_save(prev, regs); 75 #define switch_to(prev, next, last) \ 77 struct task_struct *__prev = (prev); \
|
/linux-master/arch/powerpc/include/asm/ |
H A D | qspinlock.h | 95 u32 prev; local 106 : "=&r" (prev) 111 return likely(prev == 0); 117 u32 prev, tmp; local 130 : "=&r" (prev), "=&r" (tmp) 135 return likely(!(prev & ~_Q_TAIL_CPU_MASK));
|
/linux-master/lib/ |
H A D | cpumask.c | 133 * for_each_cpu_andnot(cpu, mask, prev) 135 * prev = mask; 172 unsigned int next, prev; local 175 prev = __this_cpu_read(distribute_cpu_mask_prev); 178 nr_cpumask_bits, prev + 1); 194 unsigned int next, prev; local 197 prev = __this_cpu_read(distribute_cpu_mask_prev); 198 next = find_next_bit_wrap(cpumask_bits(srcp), nr_cpumask_bits, prev + 1);
|
/linux-master/include/linux/ |
H A D | list.h | 18 * sometimes we already know the next/prev entries and we can 38 WRITE_ONCE(list->prev, list); 54 struct list_head *prev, 66 struct list_head *prev, 73 * With the hardening version, elide checking if next and prev 83 if (likely(next->prev == prev && prev->next == next && new != prev && new != next)) 88 ret &= __list_add_valid_or_report(new, prev, nex 65 __list_add_valid(struct list_head *new, struct list_head *prev, struct list_head *next) argument 111 struct list_head *prev = entry->prev; local 128 __list_add_valid(struct list_head *new, struct list_head *prev, struct list_head *next) argument 146 __list_add(struct list_head *new, struct list_head *prev, struct list_head *next) argument 193 __list_del(struct list_head * prev, struct list_head * next) argument 523 __list_splice(const struct list_head *list, struct list_head *prev, struct list_head *next) argument 1058 hlist_add_behind(struct hlist_node *n, struct hlist_node *prev) argument [all...] |
H A D | rculist.h | 25 WRITE_ONCE(list->prev, list); 35 * list_tail_rcu - returns the prev pointer of the head of the list 42 #define list_tail_rcu(head) (*((struct list_head __rcu **)(&(head)->prev))) 74 * the prev/next entries already! 77 struct list_head *prev, struct list_head *next) 79 if (!__list_add_valid(new, prev, next)) 83 new->prev = prev; 84 rcu_assign_pointer(list_next_rcu(prev), new); 85 next->prev 76 __list_add_rcu(struct list_head *new, struct list_head *prev, struct list_head *next) argument 226 __list_splice_init_rcu(struct list_head *list, struct list_head *prev, struct list_head *next, void (*sync)(void)) argument 678 hlist_add_behind_rcu(struct hlist_node *n, struct hlist_node *prev) argument [all...] |
/linux-master/drivers/video/fbdev/omap2/omapfb/dss/ |
H A D | dss-of.c | 20 struct device_node *prev) 27 if (!prev) { 44 ports = of_get_parent(prev); 49 port = of_get_next_child(ports, prev); 54 prev = port; 66 struct device_node *prev) 74 ep = of_get_next_child(parent, prev); 77 prev = ep; 19 omapdss_of_get_next_port(const struct device_node *parent, struct device_node *prev) argument 65 omapdss_of_get_next_endpoint(const struct device_node *parent, struct device_node *prev) argument
|
/linux-master/drivers/gpu/drm/i915/gt/ |
H A D | selftest_ring.c | 32 u32 next, u32 prev, 37 result = intel_ring_direction(ring, next, prev); 45 next, prev, result, expected); 54 u32 prev = x, next = intel_ring_wrap(ring, x + step); local 58 err |= check_ring_direction(ring, prev, prev, 0); 59 err |= check_ring_direction(ring, next, prev, 1); 60 err |= check_ring_direction(ring, prev, next, -1); 31 check_ring_direction(struct intel_ring *ring, u32 next, u32 prev, int expected) argument
|
/linux-master/tools/include/linux/ |
H A D | list.h | 15 * sometimes we already know the next/prev entries and we can 28 list->prev = list; 35 * the prev/next entries already! 39 struct list_head *prev, 42 next->prev = new; 44 new->prev = prev; 45 prev->next = new; 49 struct list_head *prev, 77 __list_add(new, head->prev, hea 38 __list_add(struct list_head *new, struct list_head *prev, struct list_head *next) argument 87 __list_del(struct list_head * prev, struct list_head * next) argument 274 __list_splice(const struct list_head *list, struct list_head *prev, struct list_head *next) argument 670 hlist_add_behind(struct hlist_node *n, struct hlist_node *prev) argument [all...] |
/linux-master/drivers/gpu/drm/nouveau/nvkm/core/ |
H A D | mm.c | 53 struct nvkm_mm_node *prev = node(this, prev); local 56 if (prev && prev->type == NVKM_MM_TYPE_NONE) { 57 prev->length += this->length; 59 kfree(this); this = prev; 72 list_for_each_entry(prev, &mm->free, fl_entry) { 73 if (this->offset < prev->offset) 77 list_add_tail(&this->fl_entry, &prev->fl_entry); 114 struct nvkm_mm_node *prev, *thi local 189 struct nvkm_mm_node *prev, *this, *next; local 242 struct nvkm_mm_node *node, *prev; local [all...] |
/linux-master/kernel/trace/ |
H A D | trace_stat.h | 17 void *(*stat_next)(void *prev, int idx);
|
/linux-master/arch/mips/kernel/ |
H A D | csrc-r4k.c | 48 unsigned int prev, curr, i; local 56 for (i = 0, prev = rdhwr_count(); i < 100; i++) { 59 if (curr != prev) 62 prev = curr;
|
/linux-master/arch/arc/kernel/ |
H A D | fpu.c | 32 void fpu_save_restore(struct task_struct *prev, struct task_struct *next) argument 34 unsigned int *saveto = &prev->thread.fpu.aux_dpfp[0].l; 69 void fpu_save_restore(struct task_struct *prev, struct task_struct *next) argument 71 struct arc_fpu *save = &prev->thread.fpu;
|
/linux-master/drivers/usb/host/ |
H A D | ohci-mem.c | 113 struct td **prev = &hc->td_hash [TD_HASH_FUNC (td->td_dma)]; local 116 while (*prev && *prev != td) 117 prev = &(*prev)->td_hash; 118 if (*prev) 119 *prev = td->td_hash;
|
/linux-master/sound/pci/ctxfi/ |
H A D | ctimap.c | 42 pre = pos->prev; 44 pre = head->prev; 46 __list_add(&entry->list, pos->prev, pos); 48 pre = head->prev; 75 pre = (entry->list.prev == head) ? head->prev : entry->list.prev;
|
/linux-master/fs/iomap/ |
H A D | fiemap.c | 43 struct fiemap_extent_info *fi, struct iomap *prev) 50 ret = iomap_to_fiemap(fi, prev, 0); 51 *prev = iter->iomap; 71 struct iomap prev = { local 81 iter.processed = iomap_fiemap_iter(&iter, fi, &prev); 83 if (prev.type != IOMAP_HOLE) { 84 ret = iomap_to_fiemap(fi, &prev, FIEMAP_EXTENT_LAST); 42 iomap_fiemap_iter(const struct iomap_iter *iter, struct fiemap_extent_info *fi, struct iomap *prev) argument
|
/linux-master/drivers/net/fddi/skfp/ |
H A D | smttimer.c | 33 struct smt_timer **prev ; local 43 for (prev = &smc->t.st_queue ; (tm = *prev) ; prev = &tm->tm_next ) { 45 *prev = tm->tm_next ; 57 struct smt_timer **prev ; local 84 for (prev = &smc->t.st_queue ; (tm = *prev) ; prev = &tm->tm_next ) { 91 *prev [all...] |
/linux-master/net/bridge/ |
H A D | br_forward.c | 120 static int deliver_clone(const struct net_bridge_port *prev, argument 131 __br_forward(prev, skb, local_orig); 176 struct net_bridge_port *prev, struct net_bridge_port *p, 183 return prev; 187 if (!prev) 190 err = deliver_clone(prev, skb, local_orig); 204 struct net_bridge_port *prev = NULL; local 236 prev = maybe_deliver(prev, p, skb, local_orig); 237 if (IS_ERR(prev)) 175 maybe_deliver( struct net_bridge_port *prev, struct net_bridge_port *p, struct sk_buff *skb, bool local_orig) argument 287 struct net_bridge_port *prev = NULL; local [all...] |
/linux-master/mm/ |
H A D | interval_tree.c | 27 /* Insert node immediately after prev in the interval tree */ 29 struct vm_area_struct *prev, 36 VM_BUG_ON_VMA(vma_start_pgoff(node) != vma_start_pgoff(prev), node); 38 if (!prev->shared.rb.rb_right) { 39 parent = prev; 40 link = &prev->shared.rb.rb_right; 42 parent = rb_entry(prev->shared.rb.rb_right, 28 vma_interval_tree_insert_after(struct vm_area_struct *node, struct vm_area_struct *prev, struct rb_root_cached *root) argument
|