/linux-master/arch/arm/mm/ |
H A D | kasan_init.c | 46 unsigned long next; local 53 next = addr + PAGE_SIZE; 87 } while (ptep++, addr = next, addr != end); 96 unsigned long next; local 118 next = pmd_addr_end(addr, end); 119 kasan_pte_populate(pmdp, addr, next, early); 120 } while (pmdp++, addr = next, addr != end); 126 unsigned long next; local 149 next = pgd_addr_end(addr, end); 158 kasan_pmd_populate(pudp, addr, next, earl [all...] |
/linux-master/lib/zlib_inflate/ |
H A D | inftrees.c | 18 on return points to the next available entry's address. bits is the 40 code *next; /* next available space in table */ local 137 filled is at next and has curr index bits. The code being used is huff 190 next = *table; /* current table to fill in */ 221 min = fill; /* save offset to next table */ 224 next[(huff >> drop) + fill] = this; 238 /* go to next symbol, update count, len */ 252 next += min; /* here min is 1 << curr */ 254 /* determine length of next tabl [all...] |
/linux-master/fs/nilfs2/ |
H A D | segbuf.h | 29 * @next: Block number of the next full segment 41 sector_t next; member in struct:nilfs_segsum_info 50 * @sb_nextnum: Index number of the next full segment 87 #define NILFS_NEXT_SEGBUF(segbuf) NILFS_LIST_SEGBUF((segbuf)->sb_list.next) 90 #define NILFS_FIRST_SEGBUF(head) NILFS_LIST_SEGBUF((head)->next) 91 #define NILFS_SEGBUF_IS_LAST(segbuf, head) ((segbuf)->sb_list.next == (head)) 98 (list_entry((head)->next, struct buffer_head, b_assoc_buffers)) 100 (list_entry((bh)->b_assoc_buffers.next, struct buffer_head, \ 102 #define NILFS_SEGBUF_BH_IS_LAST(bh, head) ((bh)->b_assoc_buffers.next [all...] |
/linux-master/arch/x86/include/asm/ |
H A D | mmu_context.h | 76 extern void switch_ldt(struct mm_struct *prev, struct mm_struct *next); 82 static inline void switch_ldt(struct mm_struct *prev, struct mm_struct *next) argument 168 extern void switch_mm(struct mm_struct *prev, struct mm_struct *next, 171 extern void switch_mm_irqs_off(struct mm_struct *prev, struct mm_struct *next, 175 #define activate_mm(prev, next) \ 177 paravirt_enter_mmap(next); \ 178 switch_mm((prev), (next), NULL); \
|
/linux-master/drivers/dma-buf/ |
H A D | dma-fence-unwrap.c | 42 * dma_fence_unwrap_next - return the next fence from a fence containers 46 * the next fence from them. 115 struct dma_fence *next; local 120 next = fences[i]; 121 if (!next) 131 if (!tmp || tmp->context > next->context) { 132 tmp = next; 135 } else if (tmp->context < next->context) { 138 } else if (dma_fence_is_later(tmp, next)) {
|
/linux-master/kernel/rcu/ |
H A D | tiny.c | 33 struct rcu_head **donetail; /* ->next pointer of last "done" CB. */ 34 struct rcu_head **curtail; /* ->next pointer of last CB. */ 110 struct rcu_head *next, *list; local 130 next = list->next; 131 prefetch(next); 136 list = next; 188 head->next = NULL; 192 rcu_ctrlblk.curtail = &head->next;
|
/linux-master/kernel/trace/ |
H A D | pid_list.c | 22 pid_list->lower_list = chunk->next; 25 chunk->next = NULL; 46 pid_list->upper_list = chunk->next; 49 chunk->next = NULL; 65 chunk->next = pid_list->lower_list; 75 chunk->next = pid_list->upper_list; 262 * trace_pid_list_next - return the next pid in the list 265 * @next: The pointer to place the pid that is set starting from @pid. 267 * Looks for the next consecutive pid that is in @pid_list starting 269 * that pid is placed into @next 273 trace_pid_list_next(struct trace_pid_list *pid_list, unsigned int pid, unsigned int *next) argument [all...] |
/linux-master/drivers/acpi/acpica/ |
H A D | uttrack.c | 299 if (!element->next) { 303 element = element->next; 383 allocation->next = mem_list->list_head; 390 allocation->next = element->next; 393 if (element->next) { 394 (element->next)->previous = allocation; 397 element->next = allocation; 452 (allocation->previous)->next = allocation->next; [all...] |
/linux-master/fs/omfs/ |
H A D | file.c | 37 u64 next, last; local 45 next = inode->i_ino; 52 bh = omfs_bread(inode->i_sb, next); 61 if (omfs_is_bad(sbi, (struct omfs_header *) bh->b_data, next)) 69 last = next; 70 next = be64_to_cpu(oe->e_next); 89 if (next == ~0) 92 bh = omfs_bread(inode->i_sb, next); 141 /* trivially grow current extent, if next block is not taken */ 221 sector_t next, offse local [all...] |
/linux-master/drivers/crypto/intel/qat/qat_common/ |
H A D | adf_dev_mgr.c | 191 struct vf_id_map *next; local 197 next = list_next_entry(map, list); 198 while (next && &next->list != &vfs_table) { 199 next->fake_id++; 200 next = list_next_entry(next, list); 257 struct vf_id_map *map, *next; local 266 next = list_next_entry(map, list); 267 while (next [all...] |
/linux-master/tools/testing/selftests/kvm/lib/ |
H A D | sparsebit.c | 224 * If current node has a right child, next node is the left-most 235 * That parent is then the next node. 252 * If current node has a left child, next node is the right-most 263 * That parent is then the next node. 559 * reductions are possible with the new previous and next nodes. Note, 605 struct node *prev, *next, *tmp; local 614 * for the next pass through the reduction loop, 617 * by first remembering the location of the next 620 * no other nodes between prev and next. 623 * both prev and next bot 1373 struct node *nodep, *next; local 1455 struct node *nodep, *next; local 1916 sparsebit_idx_t next; local [all...] |
/linux-master/arch/arm64/mm/ |
H A D | kasan_init.c | 118 unsigned long next; local 127 next = addr + PAGE_SIZE; 129 } while (ptep++, addr = next, addr != end && pte_none(__ptep_get(ptep))); 135 unsigned long next; local 139 next = pmd_addr_end(addr, end); 140 kasan_pte_populate(pmdp, addr, next, node, early); 141 } while (pmdp++, addr = next, addr != end && pmd_none(READ_ONCE(*pmdp))); 147 unsigned long next; local 151 next = pud_addr_end(addr, end); 152 kasan_pmd_populate(pudp, addr, next, nod 159 unsigned long next; local 171 unsigned long next; local [all...] |
/linux-master/block/ |
H A D | blk-merge.c | 53 struct request *prev_rq, struct bio *prev, struct bio *next) 77 * - if 'pb' ends unaligned, the next bio must include 82 bio_get_first_bvec(next, &nb); 132 * If the next starting sector would be misaligned, stop the discard at 670 struct request *next) 676 if (blk_rq_sectors(req) + bio_sectors(next->bio) > 680 req->nr_phys_segments = segments + blk_rq_nr_discard_segments(next); 688 struct request *next) 692 if (req_gap_back_merge(req, next->bio)) 698 if ((blk_rq_sectors(req) + blk_rq_sectors(next)) > 52 bio_will_gap(struct request_queue *q, struct request *prev_rq, struct bio *prev, struct bio *next) argument 669 req_attempt_discard_merge(struct request_queue *q, struct request *req, struct request *next) argument 687 ll_merge_requests_fn(struct request_queue *q, struct request *req, struct request *next) argument 786 blk_try_req_merge(struct request *req, struct request *next) argument 801 attempt_merge(struct request_queue *q, struct request *req, struct request *next) argument 890 struct request *next = elv_latter_request(q, rq); local 914 blk_attempt_req_merge(struct request_queue *q, struct request *rq, struct request *next) argument [all...] |
/linux-master/fs/ |
H A D | char_dev.c | 35 struct char_device_struct *next; member in struct:char_device_struct 56 for (cd = chrdevs[major_to_index(offset)]; cd; cd = cd->next) { 77 for (cd = chrdevs[major_to_index(i)]; cd; cd = cd->next) 134 for (curr = chrdevs[i]; curr; prev = curr, curr = curr->next) { 156 cd->next = curr; 159 cd->next = prev->next; 160 prev->next = cd; 178 for (cp = &chrdevs[i]; *cp; cp = &(*cp)->next) 185 *cp = cd->next; 204 dev_t n, next; local 314 dev_t n, next; local [all...] |
H A D | filesystems.c | 52 for (p = &file_systems; *p; p = &(*p)->next) 82 if (fs->next) 116 *tmp = fs->next; 117 fs->next = NULL; 122 tmp = &(*tmp)->next; 145 for (tmp=file_systems, index=0 ; tmp ; tmp=tmp->next, index++) { 162 for (tmp = file_systems; tmp; tmp = tmp->next, index--) 182 for (tmp = file_systems, index = 0 ; tmp ; tmp = tmp->next, index++) 219 for (p = file_systems; p; p = p->next) { 247 tmp = tmp->next; [all...] |
/linux-master/arch/s390/mm/ |
H A D | pageattr.c | 178 unsigned long next; local 187 next = pmd_addr_end(addr, end); 191 need_split |= !!(addr + PMD_SIZE > next); 200 rc = walk_pte_level(pmdp, addr, next, flags); 205 addr = next; 267 unsigned long next; local 276 next = pud_addr_end(addr, end); 280 need_split |= !!(addr + PUD_SIZE > next); 289 rc = walk_pmd_level(pudp, addr, next, flags); 292 addr = next; 301 unsigned long next; local 323 unsigned long next; local [all...] |
/linux-master/drivers/tty/ |
H A D | tty_buffer.c | 106 p->next = NULL; 123 struct tty_buffer *p, *next; local 129 buf->head = p->next; 135 llist_for_each_entry_safe(p, next, llist, free) 225 struct tty_buffer *next; local 233 while ((next = smp_load_acquire(&buf->head->next)) != NULL) { 235 buf->head = next; 287 * is advanced to the next buffer. 289 smp_store_release(&b->next, 402 struct tty_buffer *next; local 471 struct tty_buffer *next; local [all...] |
/linux-master/drivers/gpu/drm/radeon/ |
H A D | mkregtable.c | 37 * sometimes we already know the next/prev entries and we can 43 struct list_head *next, *prev; member in struct:list_head 49 list->next = list; 57 * the prev/next entries already! 61 struct list_head *prev, struct list_head *next) 63 next->prev = new; 64 new->next = next; 66 prev->next = new; 70 struct list_head *prev, struct list_head *next); 60 __list_add(struct list_head *new, struct list_head *prev, struct list_head *next) argument [all...] |
/linux-master/drivers/pcmcia/ |
H A D | rsrc_nonstatic.c | 50 struct resource_map *next; member in struct:resource_map 112 for (p = map; ; p = p->next) { 117 if ((p->next == map) || (p->next->base > base+num-1)) 126 q->next = p->next; p->next = q; 137 q = p->next; 144 p->next = q->next; [all...] |
/linux-master/arch/powerpc/mm/book3s64/ |
H A D | subpage_prot.c | 20 * reinitialized for the next user. 97 unsigned long next, limit; local 108 for (; addr < limit; addr = next) { 109 next = pmd_addr_end(addr, limit); 124 if (addr + (nw << PAGE_SHIFT) > next) 125 nw = (next - addr) >> PAGE_SHIFT; 192 unsigned long next, limit; local 233 for (limit = addr + len; addr < limit; addr = next) { 234 next = pmd_addr_end(addr, limit); 263 if (addr + (nw << PAGE_SHIFT) > next) [all...] |
/linux-master/arch/powerpc/include/asm/ |
H A D | mmu_context.h | 56 struct mm_struct *next); 58 struct mm_struct *next, 62 return radix__switch_mmu_context(prev, next); 63 return switch_slb(tsk, next); 100 extern void switch_mmu_context(struct mm_struct *prev, struct mm_struct *next, 151 * flush of the full mm, as the next TLBI may be local and the 226 extern void switch_mm_irqs_off(struct mm_struct *prev, struct mm_struct *next, 229 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, 235 switch_mm_irqs_off(prev, next, tsk); 245 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) 57 switch_mmu_context(struct mm_struct *prev, struct mm_struct *next, struct task_struct *tsk) argument [all...] |
/linux-master/arch/riscv/kernel/ |
H A D | vector.c | 214 int cur, next; local 219 next = riscv_v_ctrl_get_next(tsk); 220 if (!next) { 226 cur = next; 228 /* Clear next mask if inherit-bit is not set */ 231 next = PR_RISCV_V_VSTATE_CTRL_DEFAULT; 233 riscv_v_ctrl_set(tsk, cur, next, inherit); 247 int cur, next; local 272 next = VSTATE_CTRL_GET_NEXT(arg); 274 switch (next) { [all...] |
/linux-master/arch/riscv/mm/ |
H A D | context.c | 91 /* Queue a TLB invalidation for each CPU on next context-switch */ 166 * of active_context as used and next context switch we will 209 struct mm_struct *next, unsigned int cpu) 219 cpumask_set_cpu(cpu, mm_cpumask(next)); 221 set_mm_asid(next, cpu); 224 set_mm_noasid(next); 280 struct mm_struct *next, unsigned int cpu) 318 void switch_mm(struct mm_struct *prev, struct mm_struct *next, argument 323 if (unlikely(prev == next)) 326 membarrier_arch_switch_mm(prev, next, tas 208 set_mm(struct mm_struct *prev, struct mm_struct *next, unsigned int cpu) argument 279 set_mm(struct mm_struct *prev, struct mm_struct *next, unsigned int cpu) argument [all...] |
/linux-master/net/ax25/ |
H A D | ax25_dev.c | 37 for (ax25_dev = ax25_dev_list; ax25_dev != NULL; ax25_dev = ax25_dev->next) 88 ax25_dev->next = ax25_dev_list; 114 for (s = ax25_dev_list; s != NULL; s = s->next) 119 ax25_dev_list = s->next; 123 while (s != NULL && s->next != NULL) { 124 if (s->next == ax25_dev) { 125 s->next = ax25_dev->next; 129 s = s->next; 210 ax25_dev = ax25_dev->next; [all...] |
/linux-master/net/unix/ |
H A D | garbage.c | 163 struct sk_buff *next; local 166 skb_queue_walk_safe(&x->sk_receive_queue, skb, next) { 203 struct sk_buff *next; local 211 skb_queue_walk_safe(&x->sk_receive_queue, skb, next) { 223 u = list_entry(embryos.next, struct unix_sock, link); 257 struct unix_sock *u, *next; local 288 list_for_each_entry_safe(u, next, &gc_inflight_list, link) { 322 while (cursor.next != &gc_candidates) { 323 u = list_entry(cursor.next, struct unix_sock, link); 356 u = list_entry(not_cycle_list.next, struc [all...] |