/linux-master/mm/ |
H A D | huge_memory.c | 1803 pmd_t *pmd, unsigned long addr, unsigned long next) 1842 if (next - addr != HPAGE_PMD_SIZE) { 2715 /* Check if we need to split end next. */ 2719 * If we're also updating the next vma vm_start, 2723 struct vm_area_struct *next = find_vma(vma->vm_mm, vma->vm_end); local 2724 unsigned long nstart = next->vm_start; 2726 split_huge_pmd_if_needed(next, nstart); 3320 struct folio *folio, *next; local 3330 list_for_each_entry_safe(folio, next, &ds_queue->split_queue, 3344 list_for_each_entry_safe(folio, next, 1802 madvise_free_huge_pmd(struct mmu_gather *tlb, struct vm_area_struct *vma, pmd_t *pmd, unsigned long addr, unsigned long next) argument [all...] |
H A D | gup.c | 309 struct page *next = nth_page(start, i); local 310 struct folio *folio = page_folio(next); 315 folio_nr_pages(folio) - folio_page_idx(folio, next)); 391 * next writeback cycle. This is harmless. 1105 unsigned long now, next; 1118 now = jiffies; next = next_warn; 1119 if (next && time_before(now, next)) 1804 continue; /* continue at next VMA */ 2856 unsigned long next; local 3008 unsigned long next; local 3047 unsigned long next; local 3075 unsigned long next; local 3100 unsigned long next; local [all...] |
/linux-master/lib/ |
H A D | bootconfig.c | 43 * simply a key word or a value. A key node may have a next key node or/and 44 * a child node (both key and value). A value node may have a next value 163 * xbc_node_get_next() - Get the next sibling XBC node 166 * Return the NEXT sibling node of @node. If the node has no next sibling, 173 return node->next ? &xbc_nodes[node->next] : NULL; 334 * xbc_node_find_next_leaf() - Find the next leaf node under given node 338 * Search the next leaf node (which means the terminal key node) of @node 340 * Return the next node or NULL if next lea 345 struct xbc_node *next; local 615 char *next; local 699 char *next; local [all...] |
/linux-master/kernel/ |
H A D | fork.c | 1163 tsk->wake_q.next = NULL;
|
/linux-master/init/ |
H A D | main.c | 428 /* parse_args() stops at the next param of '--' and returns an address */ 1111 struct list_head next; member in struct:blacklist_entry 1138 list_add(&entry->next, &blacklisted_initcalls); 1163 list_for_each_entry(entry, &blacklisted_initcalls, next) {
|
/linux-master/include/net/ |
H A D | sch_generic.h | 33 struct qdisc_rate_table *next; member in struct:qdisc_rate_table 86 * dequeueing next packet. 287 struct Qdisc_ops *next; member in struct:Qdisc_ops 407 struct tcf_proto __rcu *next; member in struct:tcf_proto 988 skb->next = NULL; 989 last->next = skb; 1008 skb->next = qh->head; 1021 qh->head = skb->next; 1025 skb->next = NULL; 1079 skb->next [all...] |
/linux-master/include/linux/ |
H A D | bootconfig.h | 52 uint16_t next; member in struct:xbc_node
|
/linux-master/fs/fuse/ |
H A D | file.c | 420 struct fuse_writepage_args *next; member in struct:fuse_writepage_args 1799 struct fuse_writepage_args *aux, *next; variable in typeref:struct: 1840 for (aux = wpa->next; aux; aux = next) { 1841 next = aux->next; 1842 aux->next = NULL; 1866 wpa = list_entry(fi->queued_writes.next, 1930 while (wpa->next) { 1933 struct fuse_writepage_args *next local [all...] |
H A D | fuse_i.h | 63 struct fuse_forget_link *next; member in struct:fuse_forget_link 157 /* position at end of cache (position of next entry) */ 488 /** The next unique request id */ 639 /** The next unique kernel file handle */ 1365 * Get the next unique ID for a request
|
/linux-master/drivers/scsi/ |
H A D | scsi_lib.c | 496 sdev = list_entry(starved_list.next, 531 * scsi_run_queue - Select a proper request queue to serve next. 714 * scsi_rq_err_bytes - determine number of bytes till the next failure boundary 1031 * are not done then we have to figure out what to do next:
|
/linux-master/drivers/net/ |
H A D | tun.c | 146 struct list_head next; member in struct:tun_file 596 list_add_tail(&tfile->next, &tun->disabled); 605 list_del_init(&tfile->next); 725 list_for_each_entry(tfile, &tun->disabled, next) { 741 list_for_each_entry_safe(tfile, tmp, &tun->disabled, next) { 1210 list_for_each_entry(tfile, &tun->disabled, next) { 3487 INIT_LIST_HEAD(&tfile->next); 3681 list_for_each_entry(tfile, &tun->disabled, next)
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | en_main.c | 606 /* prepare next */ 1251 &wqe->next.next_wqe_index); 2272 /* remove current rl index to free space to next ones */ 3773 /* update HW stats in background for next time */
|
/linux-master/drivers/infiniband/core/ |
H A D | cm.c | 877 work = list_entry(cm_id_priv->work_list.next, struct cm_work, list);
|
/linux-master/drivers/gpu/drm/vmwgfx/ |
H A D | vmwgfx_ttm_buffer.c | 15 * next paragraph) shall be included in all copies or substantial portions 111 viter->next = &__vmw_piter_non_sg_next; 117 viter->next = &__vmw_piter_sg_next;
|
H A D | vmwgfx_drv.h | 15 * next paragraph) shall be included in all copies or substantial portions 348 * @next: Function to advance the iterator. Returns false if past the list 358 bool (*next)(struct vmw_piter *); member in struct:vmw_piter 967 return viter->next(viter);
|
/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/instmem/ |
H A D | nv50.c | 205 if (likely(iobj->lru.next) && iobj->map) { 249 if (likely(iobj->lru.next)) 275 if (likely(iobj->lru.next)) { 277 iobj->lru.next = NULL; 303 iobj->lru.next = NULL; /* Exclude from eviction. */ 325 if (likely(iobj->lru.next))
|
/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_vm.c | 328 base->next = NULL; 333 base->next = bo->vm_bo; 354 * is validated on next vm use to avoid fault. 438 /* Add one if the page tables will be re-generated on next CS */ 768 for (base = bo->vm_bo; base; base = base->next) { 1213 * next command submission. 1833 struct amdgpu_bo_va_mapping *before, *after, *tmp, *next; local 1889 list_for_each_entry_safe(tmp, next, &removed, list) { 1997 struct amdgpu_bo_va_mapping *mapping, *next; local 2010 base = &(*base)->next) { [all...] |
/linux-master/arch/s390/kernel/ |
H A D | entry.S | 142 * The following nop exists only in order to avoid that the next 154 * gpr3 = (task_struct *)next 164 lg %r15,0(%r4,%r3) # start of kernel stack of next 165 agr %r15,%r5 # end of kernel stack of next 166 stg %r3,__LC_CURRENT # store task struct of next 168 lg %r15,__THREAD_ksp(%r1,%r3) # load kernel stack of next 170 mvc __LC_CURRENT_PID(4,%r0),0(%r3) # store pid of next 171 lmg %r6,%r15,__SF_GPRS(%r15) # load gprs of next task 206 # Let the next instruction be NOP to avoid triggering a machine check 565 lhi %r3,0 # next CP [all...] |
/linux-master/fs/btrfs/ |
H A D | extent_io.c | 70 if (!fs_info->allocated_ebs.next) 1261 * Return the next dirty range in [@start, @end). 1349 * the next time the page gets dirtied, we will try to 1402 * So clear subpage dirty bit here so next time we won't submit 1583 * non-log tree extents, and the next filemap_fdatawait_range() call 2398 goto next; 2420 goto next; 2433 next: 2467 * - Cache the next entry to be emitted to the fiemap buffer, so that we can 2486 /* Index of the next entr [all...] |
H A D | extent-tree.c | 3767 * trying to use, and go to the next block group. 4394 * | |- Jump to next block group 4584 /* move on to the next group */ 4674 * next to the block it returns. This is just a hint and 5468 struct extent_buffer *next; local 5496 next = find_extent_buffer(fs_info, bytenr); 5497 if (!next) { 5498 next = btrfs_find_create_tree_block(fs_info, bytenr, 5500 if (IS_ERR(next)) 5501 return PTR_ERR(next); [all...] |
/linux-master/kernel/time/ |
H A D | tick-common.c | 31 * Tick next event: keeps track of the tick time. It's updated by the 47 * TICK_DO_TIMER_NONE, i.e. a non existing CPU. So the next cpu which looks 92 /* Keep track of the next tick event */ 111 ktime_t next = dev->next_event; local 127 * Setup the next period for devices, which do not have 130 next = ktime_add_ns(next, TICK_NSEC); 132 if (!clockevents_program_event(dev, next, false)) 164 ktime_t next; local 168 next [all...] |
/linux-master/fs/kernfs/ |
H A D | file.c | 35 * kernfs_node is on the list or not can be determined by testing the next 160 void *next = ops->seq_start(sf, ppos); local 162 if (next == ERR_PTR(-ENODEV)) 163 kernfs_seq_stop_active(sf, next); 164 return next; 175 void *next = ops->seq_next(sf, v, ppos); local 177 if (next == ERR_PTR(-ENODEV)) 178 kernfs_seq_stop_active(sf, next); 179 return next; 210 .next [all...] |
/linux-master/fs/bcachefs/ |
H A D | extents.c | 715 union bch_extent_entry *next = extent_entry_next(entry); local 717 memmove_u64s(entry, next, (u64 *) end - (u64 *) next); 766 * Returns pointer to the next entry after the one being dropped: 772 union bch_extent_entry *entry = to_entry(ptr), *next; local 780 for (next = extent_entry_next(entry); 781 next != ptrs.end; 782 next = extent_entry_next(next)) { 783 if (extent_entry_is_crc(next)) { [all...] |
H A D | btree_trans_commit.c | 684 h = h->next;
|
/linux-master/drivers/vhost/ |
H A D | vhost.c | 2378 * function returns the next descriptor in the chain, 2382 unsigned int next; local 2389 next = vhost16_to_cpu(vq, READ_ONCE(desc->next)); 2390 return next; 2423 /* Buffers are chained via a 16 bit next field, so 2535 /* Grab the next descriptor number they're advertising, and increment
|