/linux-master/arch/hexagon/include/asm/ |
H A D | mmu_context.h | 29 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, argument 38 if (next->context.generation < prev->context.generation) { 40 next->pgd[l1] = init_mm.pgd[l1]; 42 next->context.generation = prev->context.generation; 45 __vmnewmap((void *)next->context.ptbase); 52 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) argument 57 switch_mm(prev, next, current_thread_info()->task);
|
/linux-master/tools/lib/ |
H A D | list_sort.c | 24 tail = &a->next; 25 a = a->next; 32 tail = &b->next; 33 b = b->next; 60 tail->next = a; 63 a = a->next; 67 tail->next = b; 70 b = b->next; 79 tail->next = b; 92 b = b->next; 242 struct list_head *next = pending->prev; local [all...] |
/linux-master/arch/parisc/include/asm/ |
H A D | mmu_context.h | 51 struct mm_struct *next, struct task_struct *tsk) 53 if (prev != next) { 57 spinlock_t *pgd_lock = &next->page_table_lock; 60 mtctl(__pa(next->pgd), 25); 61 load_context(next->context); 66 struct mm_struct *next, struct task_struct *tsk) 70 if (prev == next) 74 switch_mm_irqs_off(prev, next, tsk); 80 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) argument 90 BUG_ON(next 50 switch_mm_irqs_off(struct mm_struct *prev, struct mm_struct *next, struct task_struct *tsk) argument 65 switch_mm(struct mm_struct *prev, struct mm_struct *next, struct task_struct *tsk) argument [all...] |
/linux-master/include/linux/ |
H A D | list_bl.h | 39 struct hlist_bl_node *next, **pprev; member in struct:hlist_bl_node 46 h->next = NULL; 82 n->next = first; 84 first->pprev = &n->next; 90 struct hlist_bl_node *next) 92 struct hlist_bl_node **pprev = next->pprev; 95 n->next = next; 96 next->pprev = &n->next; 89 hlist_bl_add_before(struct hlist_bl_node *n, struct hlist_bl_node *next) argument 117 struct hlist_bl_node *next = n->next; local [all...] |
H A D | list_nulls.h | 26 struct hlist_nulls_node *next, **pprev; member in struct:hlist_nulls_node 96 n->next = first; 100 WRITE_ONCE(first->pprev, &n->next); 105 struct hlist_nulls_node *next = n->next; local 108 WRITE_ONCE(*pprev, next); 109 if (!is_a_nulls(next)) 110 WRITE_ONCE(next->pprev, pprev); 131 pos = pos->next) 143 pos = pos->next) [all...] |
H A D | rculist.h | 24 WRITE_ONCE(list->next, list); 29 * return the ->next pointer of a list_head in an rcu safe 32 #define list_next_rcu(list) (*((struct list_head __rcu **)(&(list)->next))) 74 * the prev/next entries already! 77 struct list_head *prev, struct list_head *next) 79 if (!__list_add_valid(new, prev, next)) 82 new->next = next; 85 next->prev = new; 106 __list_add_rcu(new, head, head->next); 76 __list_add_rcu(struct list_head *new, struct list_head *prev, struct list_head *next) argument 226 __list_splice_init_rcu(struct list_head *list, struct list_head *prev, struct list_head *next, void (*sync)(void)) argument 527 struct hlist_node *next = old->next; local 651 hlist_add_before_rcu(struct hlist_node *n, struct hlist_node *next) argument [all...] |
/linux-master/scripts/mod/ |
H A D | list.h | 33 * sometimes we already know the next/prev entries and we can 39 struct list_head *next, *prev; member in struct:list_head 56 list->next = list; 64 * the prev/next entries already! 68 struct list_head *next) 70 next->prev = new; 71 new->next = next; 73 prev->next = new; 86 __list_add(new, head, head->next); 66 __list_add(struct list_head *new, struct list_head *prev, struct list_head *next) argument 109 __list_del(struct list_head *prev, struct list_head *next) argument [all...] |
/linux-master/arch/csky/include/asm/ |
H A D | mmu_context.h | 24 switch_mm(struct mm_struct *prev, struct mm_struct *next, argument 29 if (prev != next) 30 check_and_switch_context(next, cpu); 32 setup_pgd(next->pgd, next->context.asid.counter); 34 flush_icache_deferred(next);
|
/linux-master/drivers/scsi/sym53c8xx_2/ |
H A D | sym_malloc.c | 64 while (!h[j].next) { 66 h[j].next = (m_link_p) M_GET_MEM_CLUSTER(); 67 if (h[j].next) 68 h[j].next->next = NULL; 74 a = h[j].next; 76 h[j].next = h[j].next->next; 80 h[j].next [all...] |
/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/fb/ |
H A D | gddr3.c | 75 switch (ram->next->bios.timing_ver) { 77 CWL = ram->next->bios.timing_10_CWL; 78 CL = ram->next->bios.timing_10_CL; 79 WR = ram->next->bios.timing_10_WR; 80 DLL = !ram->next->bios.ramcfg_DLLoff; 81 ODT = ram->next->bios.timing_10_ODT; 82 RON = ram->next->bios.ramcfg_RON; 85 CWL = (ram->next->bios.timing[1] & 0x00000f80) >> 7; 86 CL = (ram->next->bios.timing[1] & 0x0000001f) >> 0; 87 WR = (ram->next [all...] |
H A D | gddr5.c | 41 xd = !ram->next->bios.ramcfg_DLLoff; 43 switch (ram->next->bios.ramcfg_ver) { 45 pd = ram->next->bios.ramcfg_11_01_80; 46 lf = ram->next->bios.ramcfg_11_01_40; 47 vh = ram->next->bios.ramcfg_11_02_10; 48 vr = ram->next->bios.ramcfg_11_02_04; 49 vo = ram->next->bios.ramcfg_11_06; 50 l3 = !ram->next->bios.ramcfg_11_07_02; 56 switch (ram->next->bios.timing_ver) { 58 WL = (ram->next [all...] |
H A D | sddr2.c | 65 switch (ram->next->bios.timing_ver) { 67 CL = ram->next->bios.timing_10_CL; 68 WR = ram->next->bios.timing_10_WR; 69 DLL = !ram->next->bios.ramcfg_DLLoff; 70 ODT = ram->next->bios.timing_10_ODT & 3; 73 CL = (ram->next->bios.timing[1] & 0x0000001f); 74 WR = (ram->next->bios.timing[2] & 0x007f0000) >> 16; 80 if (ram->next->bios.timing_ver == 0x20 || 81 ram->next->bios.ramcfg_timing == 0xff) {
|
H A D | sddr3.c | 74 DLL = !ram->next->bios.ramcfg_DLLoff; 76 switch (ram->next->bios.timing_ver) { 78 if (ram->next->bios.timing_hdr < 0x17) { 82 CWL = ram->next->bios.timing_10_CWL; 83 CL = ram->next->bios.timing_10_CL; 84 WR = ram->next->bios.timing_10_WR; 85 ODT = ram->next->bios.timing_10_ODT; 88 CWL = (ram->next->bios.timing[1] & 0x00000f80) >> 7; 89 CL = (ram->next->bios.timing[1] & 0x0000001f) >> 0; 90 WR = (ram->next [all...] |
/linux-master/arch/openrisc/include/asm/ |
H A D | mmu_context.h | 24 extern void switch_mm(struct mm_struct *prev, struct mm_struct *next, 27 #define activate_mm(prev, next) switch_mm((prev), (next), NULL)
|
/linux-master/arch/x86/kernel/ |
H A D | process.h | 14 struct task_struct *next) 16 unsigned long next_tif = read_task_thread_flags(next); 38 __switch_to_xtra(prev, next); 13 switch_to_extra(struct task_struct *prev, struct task_struct *next) argument
|
/linux-master/arch/arm/include/asm/ |
H A D | switch_to.h | 21 * switch_to(prev, next) should switch from task `prev' to `next' 22 * `prev' will never be the same as `next'. schedule() itself 27 #define switch_to(prev,next,last) \ 31 __this_cpu_write(__entry_task, next); \ 32 last = __switch_to(prev,task_thread_info(prev), task_thread_info(next)); \
|
H A D | spinlock_types.h | 16 u16 next; member in struct:__anon3::__anon4::__raw_tickets 20 u16 next;
|
/linux-master/arch/nios2/mm/ |
H A D | mmu_context.c | 60 /* Return the next pid */ 80 void switch_mm(struct mm_struct *prev, struct mm_struct *next, argument 89 if (unlikely(CTX_VERSION(next->context) != 91 next->context = get_new_context(); 94 pgd_current = next->pgd; 97 set_context(next->context); 106 void activate_mm(struct mm_struct *prev, struct mm_struct *next) argument 108 next->context = get_new_context(); 109 set_context(next->context); 110 pgd_current = next [all...] |
/linux-master/arch/powerpc/mm/ |
H A D | mmu_context.c | 43 void switch_mm_irqs_off(struct mm_struct *prev, struct mm_struct *next, argument 50 if (!cpumask_test_cpu(cpu, mm_cpumask(next))) { 51 VM_WARN_ON_ONCE(next == &init_mm); 52 cpumask_set_cpu(cpu, mm_cpumask(next)); 53 inc_mm_active_cpus(next); 58 * translations for 'next' from page table PTEs into the TLB. 84 switch_mm_pgdir(tsk, next); 87 if (prev == next) 98 membarrier_arch_switch_mm(prev, next, tsk); 104 switch_mmu_context(prev, next, ts [all...] |
/linux-master/arch/x86/mm/ |
H A D | ident_map.c | 24 unsigned long next; local 26 for (; addr < end; addr = next) { 31 next = (addr & PUD_MASK) + PUD_SIZE; 32 if (next > end) 33 next = end; 46 use_gbpage &= ((next & ~PUD_MASK) == 0); 61 ident_pmd_init(info, pmd, addr, next); 67 ident_pmd_init(info, pmd, addr, next); 77 unsigned long next; local 80 for (; addr < end; addr = next) { 115 unsigned long next; local [all...] |
/linux-master/tools/include/linux/ |
H A D | list.h | 15 * sometimes we already know the next/prev entries and we can 27 list->next = list; 35 * the prev/next entries already! 40 struct list_head *next) 42 next->prev = new; 43 new->next = next; 45 prev->next = new; 50 struct list_head *next); 63 __list_add(new, head, head->next); 38 __list_add(struct list_head *new, struct list_head *prev, struct list_head *next) argument 87 __list_del(struct list_head * prev, struct list_head * next) argument 207 struct list_head *next = head->next; local 274 __list_splice(const struct list_head *list, struct list_head *prev, struct list_head *next) argument 627 struct hlist_node *next = n->next; local 661 hlist_add_before(struct hlist_node *n, struct hlist_node *next) argument [all...] |
/linux-master/drivers/acpi/acpica/ |
H A D | pstree.c | 74 arg = arg->common.next; 133 while (prev_arg->common.next) { 134 prev_arg = prev_arg->common.next; 136 prev_arg->common.next = arg; 147 arg = arg->common.next; 164 * DESCRIPTION: Get next op in tree (walking the tree in depth-first order) 172 union acpi_parse_object *next = NULL; local 184 next = acpi_ps_get_arg(op, 0); 185 if (next) { 186 ASL_CV_LABEL_FILENODE(next); [all...] |
H A D | pswalk.c | 32 union acpi_parse_object *next = NULL; local 72 next = acpi_ps_get_arg(op, 0); 73 if (next) { 77 op = next; 85 next = op->common.next; 96 if (next) { 97 op = next;
|
/linux-master/arch/m68k/include/asm/ |
H A D | switch_to.h | 21 * Beware that resume now expects *next to be in d1 and the offset of 28 * pass prev in a0, next in a1 31 #define switch_to(prev,next,last) do { \ 33 register void *_next __asm__ ("a1") = (next); \
|
/linux-master/arch/um/include/asm/ |
H A D | mmu_context.h | 28 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, argument 33 if(prev != next){ 35 cpumask_set_cpu(cpu, mm_cpumask(next)); 36 if(next != &init_mm) 37 __switch_mm(&next->context.id);
|