/linux-master/fs/bcachefs/ |
H A D | btree_locking.c | 309 next: 363 goto next; 390 goto next; 400 goto next;
|
/linux-master/drivers/scsi/ |
H A D | sg.c | 154 unsigned char next_cmd_len; /* 0: automatic, >0: use on next write() */ 1883 /* round request up to next highest SG_SECTOR_SZ byte boundary */ 2345 .next = dev_seq_next, 2353 .next = dev_seq_next, 2361 .next = dev_seq_next,
|
/linux-master/drivers/scsi/qla2xxx/ |
H A D | qla_edif.c | 87 list_for_each_entry_safe(entry, tentry, indx_list, next) { 180 INIT_LIST_HEAD(&entry->next); 188 list_add_tail(&entry->next, &fcport->edif.edif_indx_list); 199 list_del(&entry->next); 324 list_del(&sa_ctl->next); 456 list_for_each_entry_safe(entry, tmp, &ha->sadb_rx_index_list, next) { 458 list_del(&entry->next); 467 list_for_each_entry_safe(entry, tmp, &ha->sadb_tx_index_list, next) { 469 list_del(&entry->next); 1276 * when the next d [all...] |
/linux-master/drivers/net/ |
H A D | virtio_net.c | 349 * with this header sg. This padding makes next sg 16 byte aligned 1740 curr_skb->next = nskb;
|
H A D | geneve.c | 74 struct list_head next; /* geneve's per namespace list */ member in struct:geneve_dev 1285 list_for_each_entry(geneve, &gn->geneve_list, next) { 1377 list_add(&geneve->next, &gn->geneve_list); 1715 list_del(&geneve->next); 1892 struct geneve_dev *geneve, *next; local 1901 list_for_each_entry_safe(geneve, next, &gn->geneve_list, next) {
|
/linux-master/drivers/net/hyperv/ |
H A D | netvsc.c | 1321 while (mrc->first != mrc->next) { 1353 if (mrc->next >= mrc->first) 1354 *filled = mrc->next - mrc->first; 1356 *filled = (count - mrc->first) + mrc->next; 1384 rcd = mrc->slots + mrc->next; 1388 if (++mrc->next == nvdev->recv_completion_cnt) 1389 mrc->next = 0;
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/steering/ |
H A D | dr_dbg.c | 1152 .next = dr_dump_next,
|
/linux-master/drivers/net/ethernet/marvell/octeontx2/nic/ |
H A D | qos.c | 439 for (tmp = head->next; tmp != head; tmp = tmp->next) {
|
/linux-master/drivers/net/ethernet/marvell/octeontx2/af/ |
H A D | rvu_nix.c | 3072 int mce, u8 op, u16 pcifunc, int next, 3088 aq_req.mce.next = next; 4177 /* Found a field marker, get the next field */ 3071 nix_blk_setup_mce(struct rvu *rvu, struct nix_hw *nix_hw, int mce, u8 op, u16 pcifunc, int next, int index, u8 mce_op, bool eol) argument
|
/linux-master/drivers/isdn/mISDN/ |
H A D | socket.c | 283 struct mISDNchannel *bchan, *next; local 297 list_for_each_entry_safe(bchan, next,
|
/linux-master/drivers/gpu/drm/qxl/ |
H A D | qxl_release.c | 11 * The above copyright notice and this permission notice (including the next 166 entry = container_of(release->bos.next,
|
/linux-master/drivers/gpu/drm/msm/adreno/ |
H A D | a6xx_gpu.c | 78 ring->cur = ring->next;
|
/linux-master/drivers/gpu/drm/i915/gt/uc/ |
H A D | intel_guc_submission.c | 3469 * get picked up on the next context deregistration event or 3495 * end of suspend.. OR.. this worker can be picked up later on the next 5326 goto next; 5330 goto next; 5333 goto next; 5356 next: 5385 goto next; 5389 goto next; 5392 goto next; 5400 next [all...] |
/linux-master/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_device_queue_manager.c | 1340 struct device_process_node *cur, *next; local 1348 list_for_each_entry_safe(cur, next, &dqm->queues, list) { 2829 * next queue anyway.
|
/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_device.c | 4899 struct dma_fence *fence = NULL, *next = NULL; local 4923 r = amdgpu_bo_restore_shadow(shadow, &next); 4930 fence = next; 4939 fence = next; 5388 * recovery is scheduled next. So add one check
|
/linux-master/drivers/cxl/core/ |
H A D | port.c | 155 struct cxl_dport *next = NULL; local 161 next = cxlsd->target[i + 1]; 163 next ? "," : ""); 905 * @parent_dport: next hop up in the CXL memory decode hierarchy
|
/linux-master/drivers/cxl/ |
H A D | cxl.h | 439 * @region_id: region id for next region provisioning event 685 * @next: cxl switch port across the link attached to @dport NULL if 691 struct cxl_port *next; member in struct:cxl_ep
|
H A D | acpi.c | 704 struct resource *res, *next, *cxl = data; local 706 for (res = cxl->child; res; res = next) { 709 next = res->sibling; 742 struct resource *res, *new, *next; local 744 for (res = cxl_res->child; res; res = next) { 762 next = res->sibling; 763 while (next && resource_overlaps(new, next)) { 764 if (resource_contains(new, next)) { 765 struct resource *_next = next [all...] |
/linux-master/arch/mips/kernel/ |
H A D | scall64-o32.S | 38 daddiu t1, 4 # skip to next instruction
|
H A D | scall64-n64.S | 43 daddiu t1, 4 # skip to next instruction
|
H A D | scall64-n32.S | 40 daddiu t1, 4 # skip to next instruction
|
H A D | scall32-o32.S | 34 addiu t1, 4 # skip to next instruction
|
H A D | asm-offsets.c | 335 OFFSET(PBE_NEXT, pbe, next);
|
/linux-master/fs/btrfs/ |
H A D | transaction.c | 43 * | To next stage: 51 * | To next stage: 76 * | To next stage: 91 * | To next stage: 105 * | To next stage: | 1330 struct list_head *next; local 1366 next = fs_info->dirty_cowonly_roots.next; 1367 list_del_init(next); 1368 root = list_entry(next, struc 1922 struct btrfs_pending_snapshot *pending, *next; local [all...] |
H A D | delayed-inode.c | 221 p = delayed_root->node_list.next; 235 struct btrfs_delayed_node *next = NULL; local 243 p = delayed_root->node_list.next; 247 p = node->n_list.next; 249 next = list_entry(p, struct btrfs_delayed_node, n_list); 250 refcount_inc(&next->refs); 254 return next; 305 p = delayed_root->prepare_list.next; 348 * the next item. 491 struct btrfs_delayed_item *next local 665 struct btrfs_delayed_item *next; local 852 struct btrfs_delayed_item *curr, *next; local 1739 struct btrfs_delayed_item *curr, *next; local 1784 struct btrfs_delayed_item *curr, *next; local 2177 struct btrfs_delayed_item *next; local [all...] |