/linux-master/scripts/kconfig/ |
H A D | list.h | 34 * sometimes we already know the next/prev entries and we can 53 list->next = list; 61 * the prev/next entries already! 65 struct list_head *next) 67 next->prev = new; 68 new->next = next; 70 prev->next = new; 83 __list_add(new, head, head->next); 100 * Delete a list entry by making the prev/next entrie 63 __list_add(struct list_head *new, struct list_head *prev, struct list_head *next) argument 106 __list_del(struct list_head *prev, struct list_head *next) argument [all...] |
/linux-master/sound/pci/ctxfi/ |
H A D | ctimap.c | 27 entry->next = entry->addr; 49 pos = head->next; 56 entry->next = pos_ent->addr; 58 pre_ent->next = entry->addr; 67 struct list_head *next, *pre, *head; local 76 next = (entry->list.next == head) ? head->next : entry->list.next; 80 entry->next [all...] |
/linux-master/drivers/net/wireguard/ |
H A D | queueing.c | 60 offsetof(struct sk_buff, next) != offsetof(struct prev_queue, empty.next) - 82 struct sk_buff *tail = queue->tail, *next = smp_load_acquire(&NEXT(tail)); local 85 if (!next) 87 queue->tail = next; 88 tail = next; 89 next = smp_load_acquire(&NEXT(next)); 91 if (next) { 92 queue->tail = next; [all...] |
/linux-master/drivers/net/ethernet/microchip/sparx5/ |
H A D | sparx5_sdlb.c | 115 u32 itr, next; local 120 next = sparx5_sdlb_group_get_next(sparx5, group, itr); 121 if (itr == next) 124 itr = next; 138 u32 idx, u32 *prev, u32 *next, 145 *next = *first; 149 *next = sparx5_sdlb_group_get_next(sparx5, group, itr); 154 if (itr == *next) 158 itr = *next; 164 u32 itr, next; local 137 sparx5_sdlb_group_get_adjacent(struct sparx5 *sparx5, u32 group, u32 idx, u32 *prev, u32 *next, u32 *first) argument 208 u32 itr, next; local 234 sparx5_sdlb_group_link(struct sparx5 *sparx5, u32 group, u32 idx, u32 first, u32 next, bool empty) argument 260 u32 first, next; local 275 u32 first, next, prev; local [all...] |
/linux-master/lib/ |
H A D | list_sort.c | 25 tail = &a->next; 26 a = a->next; 33 tail = &b->next; 34 b = b->next; 61 tail->next = a; 64 a = a->next; 68 tail->next = b; 71 b = b->next; 80 tail->next = b; 93 b = b->next; 243 struct list_head *next = pending->prev; local [all...] |
/linux-master/arch/sh/include/asm/ |
H A D | switch_to_32.h | 65 struct task_struct *next); 70 #define switch_to(prev, next, last) \ 82 if (is_dsp_enabled(next)) \ 83 __restore_dsp(next); \ 88 __ts5 = (u32 *)next; \ 89 __ts6 = (u32 *)&next->thread.sp; \ 90 __ts7 = next->thread.pc; \
|
/linux-master/io_uring/ |
H A D | alloc_cache.h | 27 return !cache->list.next; 32 if (cache->list.next) { 35 entry = container_of(cache->list.next, struct io_cache_entry, node); 37 cache->list.next = cache->list.next->next; 48 cache->list.next = NULL;
|
/linux-master/include/linux/ |
H A D | list.h | 18 * sometimes we already know the next/prev entries and we can 37 WRITE_ONCE(list->next, list); 55 struct list_head *next); 67 struct list_head *next) 73 * With the hardening version, elide checking if next and prev 83 if (likely(next->prev == prev && prev->next == next && new != prev && new != next)) 88 ret &= __list_add_valid_or_report(new, prev, next); 65 __list_add_valid(struct list_head *new, struct list_head *prev, struct list_head *next) argument 112 struct list_head *next = entry->next; local 128 __list_add_valid(struct list_head *new, struct list_head *prev, struct list_head *next) argument 146 __list_add(struct list_head *new, struct list_head *prev, struct list_head *next) argument 193 __list_del(struct list_head * prev, struct list_head * next) argument 409 struct list_head *next = smp_load_acquire(&head->next); local 523 __list_splice(const struct list_head *list, struct list_head *prev, struct list_head *next) argument 985 struct hlist_node *next = n->next; local 1044 hlist_add_before(struct hlist_node *n, struct hlist_node *next) argument [all...] |
H A D | rculist_nulls.h | 53 (*((struct hlist_nulls_node __rcu __force **)&(node)->next)) 104 WRITE_ONCE(n->next, first); 108 WRITE_ONCE(first->pprev, &n->next); 136 for (i = h->first; !is_a_nulls(i); i = i->next) 140 WRITE_ONCE(n->next, last->next); 141 n->pprev = &last->next; 151 n->pprev = &n->next; 152 n->next = (struct hlist_nulls_node *)NULLS_MARKER(NULL);
|
H A D | user-return-notifier.h | 20 struct task_struct *next) 24 set_tsk_thread_flag(next, TIF_USER_RETURN_NOTIFY); 40 struct task_struct *next) 19 propagate_user_return_notify(struct task_struct *prev, struct task_struct *next) argument 39 propagate_user_return_notify(struct task_struct *prev, struct task_struct *next) argument
|
H A D | msg.h | 13 struct msg_msgseg *next; member in struct:msg_msg
|
H A D | pagewalk.h | 60 unsigned long next, struct mm_walk *walk); 62 unsigned long next, struct mm_walk *walk); 64 unsigned long next, struct mm_walk *walk); 66 unsigned long next, struct mm_walk *walk); 68 unsigned long next, struct mm_walk *walk); 69 int (*pte_hole)(unsigned long addr, unsigned long next, 72 unsigned long addr, unsigned long next, 74 int (*test_walk)(unsigned long addr, unsigned long next, 87 /* Descend to next level, splitting huge pages if needed and possible */ 89 /* Continue to next entr [all...] |
/linux-master/tools/testing/selftests/bpf/progs/ |
H A D | btf_dump_test_case_ordering.c | 32 struct list_head *next; member in struct:list_head 37 struct hlist_node *next; member in struct:hlist_node 46 struct callback_head *next; member in struct:callback_head
|
/linux-master/arch/riscv/include/asm/ |
H A D | mmu_context.h | 16 void switch_mm(struct mm_struct *prev, struct mm_struct *next, 21 struct mm_struct *next) 23 switch_mm(prev, next, NULL); 20 activate_mm(struct mm_struct *prev, struct mm_struct *next) argument
|
/linux-master/drivers/acpi/acpica/ |
H A D | utaddress.c | 65 range_info->next = acpi_gbl_address_range_list[space_id]; 115 range_info->next; 117 prev->next = range_info->next; 134 range_info = range_info->next; 218 range_info = range_info->next; 239 struct acpi_address_range *next; local 246 next = acpi_gbl_address_range_list[i]; 248 while (next) { 249 range_info = next; [all...] |
/linux-master/tools/testing/selftests/bpf/ |
H A D | bpf_arena_list.h | 11 arena_list_node_t *next; member in struct:arena_list_node 39 pos && ({ ___tmp = (void *)pos->member.next; 1; }); \ 49 WRITE_ONCE(n->next, first); 52 tmp = &n->next; 67 arena_list_node_t *next = n->next, *tmp; local 70 cast_user(next); 74 WRITE_ONCE(tmp, next); 75 if (next) { 77 cast_kern(next); [all...] |
/linux-master/include/asm-generic/ |
H A D | nommu_context.h | 12 struct mm_struct *next, 11 switch_mm(struct mm_struct *prev, struct mm_struct *next, struct task_struct *tsk) argument
|
/linux-master/arch/arm/mm/ |
H A D | idmap.c | 28 unsigned long next; local 49 next = pmd_addr_end(addr, end); 52 } while (pmd++, addr = next, addr != end); 73 unsigned long next; local 76 next = pud_addr_end(addr, end); 77 idmap_add_pmd(pud, addr, next, prot); 78 } while (pud++, addr = next, addr != end); 85 unsigned long next; local 98 next = pgd_addr_end(addr, end); 99 idmap_add_pud(pgd, addr, next, pro [all...] |
/linux-master/arch/loongarch/include/asm/ |
H A D | mmu_context.h | 77 static inline void switch_mm_irqs_off(struct mm_struct *prev, struct mm_struct *next, argument 83 if (!asid_valid(next, cpu)) 84 get_new_mmu_context(next, cpu); 86 write_csr_asid(cpu_asid(cpu, next)); 88 if (next != &init_mm) 89 csr_write64((unsigned long)next->pgd, LOONGARCH_CSR_PGDL); 97 cpumask_set_cpu(cpu, mm_cpumask(next)); 102 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, argument 108 switch_mm_irqs_off(prev, next, tsk); 120 #define activate_mm(prev, next) switch_m [all...] |
/linux-master/arch/nios2/include/asm/ |
H A D | mmu_context.h | 43 void switch_mm(struct mm_struct *prev, struct mm_struct *next, 51 void activate_mm(struct mm_struct *prev, struct mm_struct *next);
|
/linux-master/drivers/crypto/intel/qat/qat_common/ |
H A D | adf_cfg_user.h | 13 struct adf_user_cfg_key_val *next; member in union:adf_user_cfg_key_val::__anon213 26 struct adf_user_cfg_section *next; member in union:adf_user_cfg_section::__anon215
|
/linux-master/drivers/memory/tegra/ |
H A D | tegra210-emc-cc-r21021.c | 78 ({ next->ptfv_list[PTFV_DQSOSC_MOVAVG_ ## dev ## _INDEX] = \ 79 next->ptfv_list[PTFV_DQSOSC_MOVAVG_ ## dev ## _INDEX] / \ 80 next->ptfv_list[PTFV_DVFS_SAMPLES_INDEX]; }) 86 ({ next->ptfv_list[PTFV_DQSOSC_MOVAVG_ ## dev ## _INDEX] += \ 102 next->ptfv_list[dqs] = \ 104 (next->ptfv_list[dqs] * \ 105 next->ptfv_list[w])) / \ 106 (next->ptfv_list[w] + 1); \ 109 __stringify(dev), nval, next->ptfv_list[dqs]); \ 120 struct tegra210_emc_timing *next local 403 periodic_compensation_handler(struct tegra210_emc *emc, u32 type, struct tegra210_emc_timing *last, struct tegra210_emc_timing *next) argument 607 struct tegra210_emc_timing *fake, *last = emc->last, *next = emc->next; local [all...] |
/linux-master/drivers/infiniband/hw/hfi1/ |
H A D | rc.h | 15 unsigned int next; local 17 next = n + 1; 18 if (next > rvt_size_atomic(ib_to_rvt(qp->ibqp.device))) 19 next = 0; 20 qp->s_tail_ack_queue = next; 21 qp->s_acked_ack_queue = next;
|
/linux-master/arch/sparc/include/asm/ |
H A D | switch_to_32.h | 28 #define SWITCH_DO_LAZY_FPU(next) /* */ 38 #define prepare_arch_switch(next) do { \ 58 #define switch_to(prev, next, last) do { \ 60 SWITCH_DO_LAZY_FPU(next); \ 61 cpumask_set_cpu(smp_processor_id(), mm_cpumask(next->active_mm)); \ 93 "r" (task_thread_info(next)), \
|
/linux-master/arch/csky/kernel/ |
H A D | cpu-probe.c | 14 unsigned int cur, next, i; local 24 next = mfcr("cr13"); 27 if (cur == next) 30 cur = next; 33 if (!(next >> 28)) { 76 .next = c_next,
|