/linux-master/fs/btrfs/ |
H A D | inode-item.c | 101 return btrfs_find_name_in_ext_backref(path->nodes[0], path->slots[0], 141 extref = btrfs_find_name_in_ext_backref(path->nodes[0], path->slots[0], 149 leaf = path->nodes[0]; 210 ref = btrfs_find_name_in_backref(path->nodes[0], path->slots[0], name); 216 leaf = path->nodes[0]; 278 if (btrfs_find_name_in_ext_backref(path->nodes[0], 290 leaf = path->nodes[0]; 295 btrfs_set_inode_extref_name_len(path->nodes[0], extref, name->len); 296 btrfs_set_inode_extref_index(path->nodes[0], extref, index); 297 btrfs_set_inode_extref_parent(path->nodes[ [all...] |
H A D | free-space-cache.c | 107 leaf = path->nodes[0]; 185 leaf = path->nodes[0]; 214 leaf = path->nodes[0]; 771 leaf = path->nodes[0]; 1169 leaf = path->nodes[0]; 2806 * Since the allocation info of tree-log nodes are not recorded 2813 * Advance the pointer not to overwrite the tree-log nodes.
|
H A D | file.c | 256 leaf = path->nodes[0]; 264 leaf = path->nodes[0]; 274 leaf = path->nodes[0]; 359 leaf = path->nodes[0]; 525 leaf = path->nodes[0]; 640 leaf = path->nodes[0]; 738 leaf = path->nodes[0]; 2286 leaf = path->nodes[0]; 2455 leaf = path->nodes[0]; 3722 btrfs_item_key_to_cpu(path->nodes[ [all...] |
H A D | file-item.c | 179 leaf = path->nodes[0]; 221 leaf = path->nodes[0]; 296 if (path->nodes[0]) { 297 item = btrfs_item_ptr(path->nodes[0], path->slots[0], 299 btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]); 300 itemsize = btrfs_item_size(path->nodes[0], path->slots[0]); 317 btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]); 318 itemsize = btrfs_item_size(path->nodes[0], path->slots[0]); 327 read_extent_buffer(path->nodes[0], dst, (unsigned long)item, 495 leaf = path->nodes[ [all...] |
H A D | extent_io.c | 1608 * btree nodes/leafs whose content on disk is invalid - either garbage 2772 struct extent_buffer *clone = path->nodes[0]; 2778 if (path->slots[0] < btrfs_header_nritems(path->nodes[0])) 2797 btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]); 2813 clone->start = path->nodes[0]->start; 2815 copy_extent_buffer_full(clone, path->nodes[0]); 2819 path->nodes[0] = clone; 2852 btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0] - 1); 2857 if (path->slots[0] >= btrfs_header_nritems(path->nodes[0])) { 2862 btrfs_item_key_to_cpu(path->nodes[ [all...] |
H A D | extent-tree.c | 150 btrfs_item_key_to_cpu(path->nodes[0], &key, 160 leaf = path->nodes[0]; 474 leaf = path->nodes[0]; 485 leaf = path->nodes[0]; 542 leaf = path->nodes[0]; 569 leaf = path->nodes[0]; 603 leaf = path->nodes[0]; 647 leaf = path->nodes[0]; 747 if (!path->nodes[level]) 750 btrfs_header_nritems(path->nodes[leve [all...] |
H A D | export.c | 191 leaf = path->nodes[0]; 268 leaf = path->nodes[0];
|
H A D | disk-io.c | 2158 if (path->slots[0] >= btrfs_header_nritems(path->nodes[0])) { 2168 btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]); 4316 * and then a transaction commit runs the same delayed nodes 4317 * before any delayed worker has done something with the nodes. 4932 l = path->nodes[0];
|
H A D | delayed-inode.c | 46 delayed_root->nodes = 0; 189 root->nodes++; 201 root->nodes--; 767 data_ptr = btrfs_item_ptr(path->nodes[0], path->slots[0], char); 768 write_extent_buffer(path->nodes[0], &curr->data, 853 struct extent_buffer *leaf = path->nodes[0]; 1038 leaf = path->nodes[0]; 1068 leaf = path->nodes[0]; 1188 ASSERT(path->nodes[0] == NULL); 2164 * nodes) an [all...] |
H A D | ctree.c | 217 if (!p->nodes[i]) 220 btrfs_tree_unlock_rw(p->nodes[i], p->locks[i]); 223 free_extent_buffer(p->nodes[i]); 224 p->nodes[i] = NULL; 780 * leaves and nodes. 907 * node level balancing, used to make sure nodes are in proper order for 928 mid = path->nodes[level]; 936 parent = path->nodes[level + 1]; 979 path->nodes[level] = NULL; 1136 path->nodes[leve [all...] |
H A D | defrag.c | 504 * leafs from path->nodes[1], so set lowest_level to 1 to avoid later 514 if (!path->nodes[1]) { 525 path->nodes[1], 0, 541 path->slots[1] = btrfs_header_nritems(path->nodes[1]); 617 * will not visit leaves and nodes with a generation smaller than given 658 if (path.slots[0] >= btrfs_header_nritems(path.nodes[0])) { 664 ASSERT(btrfs_header_nritems(path.nodes[0])); 665 path.slots[0] = btrfs_header_nritems(path.nodes[0]) - 1; 667 btrfs_item_key_to_cpu(path.nodes[0], &key, path.slots[0]); 675 btrfs_item_key_to_cpu(path.nodes[ [all...] |
H A D | backref.c | 488 eb = path->nodes[level]; 505 eb = path->nodes[0]; 516 eb = path->nodes[0]; 605 * other tree nodes. This happens when qgroups does backref walks when 677 eb = path->nodes[level]; 684 eb = path->nodes[level]; 728 * New backrefs (i.e., for parent nodes) are added to the appropriate 1026 leaf = path->nodes[0]; 1162 leaf = path->nodes[0]; 1293 * If the node at this level is shared, than all nodes belo [all...] |
/linux-master/drivers/net/wireless/ath/ath12k/ |
H A D | dp_rx.c | 256 size_t nodes = 0; local 271 nodes++; 276 return nodes;
|
/linux-master/drivers/net/ethernet/wangxun/txgbe/ |
H A D | txgbe_phy.c | 25 struct txgbe_nodes *nodes = &txgbe->nodes; local 32 snprintf(nodes->gpio_name, sizeof(nodes->gpio_name), "txgbe_gpio-%x", id); 33 snprintf(nodes->i2c_name, sizeof(nodes->i2c_name), "txgbe_i2c-%x", id); 34 snprintf(nodes->sfp_name, sizeof(nodes->sfp_name), "txgbe_sfp-%x", id); 35 snprintf(nodes->phylink_name, sizeof(nodes [all...] |
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en/ |
H A D | ptp.c | 32 struct mlx5e_ptp_port_ts_cqe_tracker *nodes; member in struct:mlx5e_ptp_port_ts_cqe_list 41 struct mlx5e_ptp_port_ts_cqe_tracker *tracker = &list->nodes[metadata]; 53 struct mlx5e_ptp_port_ts_cqe_tracker *tracker = &list->nodes[metadata]; 192 if (likely(pending_cqe_list->nodes[metadata_id].inuse)) { 390 cqe_list->nodes = kvzalloc_node(array_size(db_sz, sizeof(*cqe_list->nodes)), 392 if (!cqe_list->nodes) 405 cqe_list->nodes[md].metadata_id = md; 422 kvfree(cqe_list->nodes); 444 kvfree(ptpsq->ts_cqe_pending_list->nodes); [all...] |
/linux-master/drivers/net/ethernet/hisilicon/hns3/hns3pf/ |
H A D | hclge_debugfs.c | 1570 struct hclge_tm_nodes_cmd *nodes; local 1579 "failed to dump tm nodes, ret = %d\n", ret); 1583 nodes = (struct hclge_tm_nodes_cmd *)desc.data; 1587 nodes->pg_base_id, nodes->pg_num); 1589 nodes->pri_base_id, nodes->pri_num); 1591 le16_to_cpu(nodes->qset_base_id), 1592 le16_to_cpu(nodes->qset_num)); 1594 le16_to_cpu(nodes [all...] |
/linux-master/drivers/md/ |
H A D | md.h | 505 int nodes; /* Maximum number of nodes in the cluster */ member in struct:mddev::__anon363 778 extern int md_setup_cluster(struct mddev *mddev, int nodes); 858 return mddev->cluster_info && mddev->bitmap_info.nodes > 1;
|
H A D | md.c | 6410 mddev->bitmap_info.nodes = 0; 6982 * check whether the device shows up in other nodes 7566 if (mddev->bitmap_info.nodes) { 7575 mddev->bitmap_info.nodes = 0; 8548 int md_setup_cluster(struct mddev *mddev, int nodes) argument 8562 ret = md_cluster_ops->join(mddev, nodes); 9227 * so other nodes are informed. It should be harmless for normal
|
H A D | md-bitmap.c | 502 sb->nodes = cpu_to_le32(bitmap->mddev->bitmap_info.nodes); 606 bitmap->mddev->bitmap_info.nodes = 0; 620 int nodes = 0; local 698 * Setup nodes/clustername only if bitmap version is 702 nodes = le32_to_cpu(sb->nodes); 721 if (!nodes && (events < bitmap->mddev->events)) { 738 if (err == 0 && nodes && (bitmap->cluster_slot < 0)) { 741 err = md_setup_cluster(bitmap->mddev, nodes); [all...] |
/linux-master/drivers/md/bcache/ |
H A D | sysfs.c | 618 size_t nodes; member in struct:bset_stats_op 626 op->nodes++; 645 "btree nodes: %zu\n" 652 op.nodes, 715 (c->gc_stats.nodes ?: 1) * btree_bytes(c)); 746 sysfs_print(btree_nodes, c->gc_stats.nodes);
|
H A D | btree.c | 17 * Indexing is done via a btree; nodes are not necessarily fully sorted, rather 367 * flush, and writes appending to leaf nodes aren't blocking anything so 689 * It's _really_ critical that we don't free too many btree nodes - we 909 * We can only have one thread cannibalizing other cached btree nodes at a time, 937 * the list. Check if there's any freed nodes there: 1236 * ptr_invalid() can't return true for the keys that mark btree nodes as 1315 gc->nodes++; 1364 unsigned int i, nodes = 0, keys = 0, blocks; local 1378 while (nodes < GC_MERGE_NODES && !IS_ERR_OR_NULL(r[nodes] [all...] |
/linux-master/drivers/acpi/acpica/ |
H A D | aclocal.h | 1228 u32 nodes; member in struct:acpi_integrity_info
|
/linux-master/Documentation/sphinx/ |
H A D | kernel_include.py | 36 from docutils import io, nodes, statemachine namespace 156 literal_block = nodes.literal_block(rawtext, source=path, 172 literal_block += nodes.inline(value, value, 175 literal_block += nodes.Text(value, value) 177 literal_block += nodes.Text(text, text)
|
/linux-master/lib/ |
H A D | maple_tree.c | 15 * pivots. In the case of dense nodes, the pivots are implied by the position 25 * The following illustrates the layout of a range64 nodes slots and pivots. 41 * Internal (non-leaf) nodes contain pointers to other nodes. 42 * Leaf nodes contain entries. 73 * * MA_STATE_PREALLOC - Preallocated nodes, WARN_ON allocation 147 struct ma_topiary *free; /* nodes to be freed */ 165 static inline int mt_alloc_bulk(gfp_t gfp, size_t size, void **nodes) argument 167 return kmem_cache_alloc_bulk(maple_node_cache, gfp, size, nodes); 175 static inline void mt_free_bulk(size_t size, void __rcu **nodes) argument [all...] |
/linux-master/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_chardev.c | 770 /* Return number of nodes, so that user space can alloacate 779 * nodes, but not more than args->num_of_nodes as that is 1546 args->gpu_id = dmabuf_adev->kfd.dev->nodes[xcp_id]->id; 2248 /* create the vm using render nodes for kfd pdd */
|