/linux-master/scripts/kconfig/ |
H A D | list_types.h | 6 struct list_head *next, *prev; member in struct:list_head
|
H A D | list.h | 34 * sometimes we already know the next/prev entries and we can 54 list->prev = list; 61 * the prev/next entries already! 64 struct list_head *prev, 67 next->prev = new; 69 new->prev = prev; 70 prev->next = new; 96 __list_add(new, head->prev, head); 100 * Delete a list entry by making the prev/nex 63 __list_add(struct list_head *new, struct list_head *prev, struct list_head *next) argument 106 __list_del(struct list_head *prev, struct list_head *next) argument [all...] |
/linux-master/tools/testing/selftests/bpf/progs/ |
H A D | cpumask_failure.c | 111 struct bpf_cpumask *local, *prev; local 117 prev = bpf_kptr_xchg(&global_mask, local); 118 if (prev) { 119 bpf_cpumask_release(prev); 145 struct bpf_cpumask *local, *prev; local 151 prev = bpf_kptr_xchg(&global_mask, local); 152 if (prev) { 153 bpf_cpumask_release(prev); 173 struct bpf_cpumask *prev, *curr; local 179 prev [all...] |
/linux-master/kernel/locking/ |
H A D | osq_lock.c | 16 struct optimistic_spin_node *next, *prev; member in struct:optimistic_spin_node 64 * We were the last queued, we moved @lock back. @prev 74 * @node->next might complete Step-A and think its @prev is 96 struct optimistic_spin_node *prev, *next; local 114 prev = decode_cpu(old); 115 node->prev = prev; 120 * node->prev = prev osq_wait_next() 122 * prev [all...] |
H A D | mcs_spinlock.h | 67 struct mcs_spinlock *prev; local 79 prev = xchg(lock, node); 80 if (likely(prev == NULL)) { 91 WRITE_ONCE(prev->next, node);
|
/linux-master/arch/m68k/include/asm/ |
H A D | switch_to.h | 28 * pass prev in a0, next in a1 31 #define switch_to(prev,next,last) do { \ 32 register void *_prev __asm__ ("a0") = (prev); \
|
/linux-master/drivers/dma-buf/ |
H A D | dma-fence-chain.c | 23 struct dma_fence *prev; local 26 prev = dma_fence_get_rcu_safe(&chain->prev); 28 return prev; 42 struct dma_fence *prev, *replacement, *tmp; local 50 while ((prev = dma_fence_chain_get_prev(chain))) { 52 prev_chain = to_dma_fence_chain(prev); 59 if (!dma_fence_is_signaled(prev)) 65 tmp = unrcu_pointer(cmpxchg(&chain->prev, RCU_INITIALIZER(prev), 181 struct dma_fence *prev; local 241 dma_fence_chain_init(struct dma_fence_chain *chain, struct dma_fence *prev, struct dma_fence *fence, uint64_t seqno) argument [all...] |
/linux-master/tools/power/cpupower/ |
H A D | cpupower-completion.sh | 10 local prev="${COMP_WORDS[COMP_CWORD-1]}" 12 case "$prev" in 20 local prev="${COMP_WORDS[COMP_CWORD-1]}" 22 case "$prev" in 38 local prev="${COMP_WORDS[COMP_CWORD-1]}" 40 case "$prev" in 48 local prev="${COMP_WORDS[COMP_CWORD-1]}" 50 case "$prev" in 58 local prev="${COMP_WORDS[COMP_CWORD-1]}" 60 case "$prev" i [all...] |
/linux-master/scripts/mod/ |
H A D | list.h | 33 * sometimes we already know the next/prev entries and we can 39 struct list_head *next, *prev; member in struct:list_head 57 list->prev = list; 64 * the prev/next entries already! 67 struct list_head *prev, 70 next->prev = new; 72 new->prev = prev; 73 prev->next = new; 99 __list_add(new, head->prev, hea 66 __list_add(struct list_head *new, struct list_head *prev, struct list_head *next) argument 109 __list_del(struct list_head *prev, struct list_head *next) argument [all...] |
/linux-master/arch/hexagon/include/asm/ |
H A D | mmu_context.h | 29 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, argument 38 if (next->context.generation < prev->context.generation) { 42 next->context.generation = prev->context.generation; 52 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) argument 57 switch_mm(prev, next, current_thread_info()->task);
|
/linux-master/tools/lib/ |
H A D | list_sort.c | 12 * sentinel head node, "prev" links not maintained. 47 * prev-link restoration pass, or maintaining the prev links 61 a->prev = tail; 68 b->prev = tail; 90 b->prev = tail; 97 head->prev = tail; 189 if (list == head->prev) /* Zero or one elements */ 193 head->prev->next = NULL; 197 * - All lists are singly linked and null-terminated; prev [all...] |
/linux-master/arch/sh/include/asm/ |
H A D | switch_to_32.h | 64 struct task_struct *__switch_to(struct task_struct *prev, 70 #define switch_to(prev, next, last) \ 80 if (is_dsp_enabled(prev)) \ 81 __save_dsp(prev); \ 85 __ts1 = (u32 *)&prev->thread.sp; \ 86 __ts2 = (u32 *)&prev->thread.pc; \ 87 __ts4 = (u32 *)prev; \
|
/linux-master/arch/parisc/lib/ |
H A D | bitops.c | 62 u64 prev; local 65 if ((prev = *ptr) == old) 68 return prev; 74 unsigned int prev; local 77 if ((prev = *ptr) == old) 80 return (unsigned long)prev; 86 u8 prev; local 89 if ((prev = *ptr) == old) 92 return prev;
|
/linux-master/arch/alpha/include/asm/ |
H A D | xchg.h | 131 unsigned long prev, tmp, cmp, addr64; local 148 : "=&r" (prev), "=&r" (new), "=&r" (tmp), "=&r" (cmp), "=&r" (addr64) 151 return prev; 157 unsigned long prev, tmp, cmp, addr64; local 174 : "=&r" (prev), "=&r" (new), "=&r" (tmp), "=&r" (cmp), "=&r" (addr64) 177 return prev; 183 unsigned long prev, cmp; local 196 : "=&r"(prev), "=&r"(cmp), "=m"(*m) 199 return prev; 205 unsigned long prev, cm local [all...] |
/linux-master/arch/riscv/include/asm/ |
H A D | mmu_context.h | 16 void switch_mm(struct mm_struct *prev, struct mm_struct *next, 20 static inline void activate_mm(struct mm_struct *prev, argument 23 switch_mm(prev, next, NULL);
|
/linux-master/arch/x86/kernel/cpu/ |
H A D | rdrand.c | 27 unsigned long sample, prev; local 39 changed += i && sample != prev; 40 prev = sample;
|
/linux-master/include/uapi/linux/netfilter_bridge/ |
H A D | ebt_limit.h | 20 unsigned long prev; member in struct:ebt_limit_info
|
/linux-master/include/uapi/linux/netfilter/ |
H A D | xt_limit.h | 19 unsigned long prev; /* moved to xt_limit_priv */ member in struct:xt_rateinfo
|
/linux-master/include/asm-generic/ |
H A D | nommu_context.h | 11 static inline void switch_mm(struct mm_struct *prev, argument
|
/linux-master/arch/nios2/include/asm/ |
H A D | mmu_context.h | 43 void switch_mm(struct mm_struct *prev, struct mm_struct *next, 51 void activate_mm(struct mm_struct *prev, struct mm_struct *next);
|
/linux-master/fs/xfs/ |
H A D | xfs_bio_io.c | 39 struct bio *prev = bio; local 41 bio = bio_alloc(prev->bi_bdev, bio_max_vecs(left), 42 prev->bi_opf, GFP_KERNEL); 43 bio->bi_iter.bi_sector = bio_end_sector(prev); 44 bio_chain(prev, bio); 46 submit_bio(prev);
|
/linux-master/arch/parisc/include/asm/ |
H A D | mmu_context.h | 50 static inline void switch_mm_irqs_off(struct mm_struct *prev, argument 53 if (prev != next) { 65 static inline void switch_mm(struct mm_struct *prev, argument 70 if (prev == next) 74 switch_mm_irqs_off(prev, next, tsk); 80 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) argument 95 switch_mm(prev,next,current);
|
/linux-master/lib/ |
H A D | list_sort.c | 13 * sentinel head node, "prev" links not maintained. 48 * prev-link restoration pass, or maintaining the prev links 62 a->prev = tail; 69 b->prev = tail; 91 b->prev = tail; 98 head->prev = tail; 190 if (list == head->prev) /* Zero or one elements */ 194 head->prev->next = NULL; 198 * - All lists are singly linked and null-terminated; prev [all...] |
/linux-master/drivers/net/ethernet/freescale/dpaa/ |
H A D | dpaa_eth_sysfs.c | 30 struct dpaa_fq *prev = NULL; local 67 if (prev && (abs(fq->fqid - prev->fqid) != 1 || 71 "%s: %d\n", prevstr, prev->fqid); 78 if (prev && abs(fq->fqid - prev->fqid) == 1 && 86 prev = fq; 91 if (prev) { 94 prev->fqid);
|
/linux-master/arch/powerpc/mm/ |
H A D | mmu_context.c | 43 void switch_mm_irqs_off(struct mm_struct *prev, struct mm_struct *next, argument 87 if (prev == next) 98 membarrier_arch_switch_mm(prev, next, tsk); 104 switch_mmu_context(prev, next, tsk); 106 VM_WARN_ON_ONCE(!cpumask_test_cpu(cpu, mm_cpumask(prev)));
|