/linux-master/fs/btrfs/ |
H A D | uuid-tree.c | 55 eb = path->nodes[0]; 118 eb = path->nodes[0]; 127 eb = path->nodes[0]; 188 eb = path->nodes[0]; 327 leaf = path->nodes[0];
|
H A D | verity.c | 149 btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]); 257 leaf = path->nodes[0]; 335 leaf = path->nodes[0]; 392 if (path->slots[0] >= btrfs_header_nritems(path->nodes[0])) {
|
H A D | volumes.c | 1662 l = path->nodes[0]; 1783 leaf = path->nodes[0]; 1793 leaf = path->nodes[0]; 1861 btrfs_item_key_to_cpu(path->nodes[0], &found_key, 1900 leaf = path->nodes[0]; 2588 leaf = path->nodes[0]; 2596 leaf = path->nodes[0]; 2908 leaf = path->nodes[0]; 3471 leaf = path->nodes[0]; 3616 leaf = path->nodes[ [all...] |
H A D | xattr.c | 52 leaf = path->nodes[0]; 145 btrfs_assert_tree_write_locked(path->nodes[0]); 173 struct extent_buffer *leaf = path->nodes[0]; 310 leaf = path->nodes[0];
|
H A D | zoned.c | 311 if (path->slots[0] >= btrfs_header_nritems(path->nodes[0])) { 322 leaf = path->nodes[0]; 1262 btrfs_item_key_to_cpu(path->nodes[0], &found_key, path->slots[0]);
|
/linux-master/fs/dlm/ |
H A D | config.c | 27 * /config/dlm/<cluster>/spaces/<space>/nodes/<node>/nodeid 28 * /config/dlm/<cluster>/spaces/<space>/nodes/<node>/weight 485 config_group_init_type_name(&nds->ns_group, "nodes", &nodes_type); 877 struct dlm_config_node *nodes, *node; local 893 nodes = kcalloc(count, sizeof(struct dlm_config_node), GFP_NOFS); 894 if (!nodes) { 899 node = nodes; 911 *nodes_out = nodes;
|
H A D | dlm_internal.h | 123 struct dlm_config_node *nodes; member in struct:dlm_recover 160 * Some internal flags are shared between the master and process nodes; 390 /* dlm_header is first element of all structs sent between nodes */ 612 struct list_head ls_nodes; /* current nodes in ls */ 614 int ls_num_nodes; /* number of nodes in ls */
|
H A D | member.c | 156 /* for any nodes that do not support slots, we will not have set memb->slot 420 /* all nodes revert to weight of 1 if all have weight 0 */ 542 if (rv->nodes[i].nodeid == nodeid) 543 return &rv->nodes[i]; 591 node = &rv->nodes[i]; 611 log_rinfo(ls, "dlm_recover_members %d nodes", ls->ls_num_nodes); 615 /* Userspace guarantees that dlm_ls_stop() has completed on all nodes before 713 struct dlm_config_node *nodes = NULL; local 720 error = dlm_config_nodes(ls->ls_name, &nodes, &count); 735 rv->nodes [all...] |
H A D | recoverd.c | 144 * Add or remove nodes from the lockspace's ls_nodes list. 161 * dlm_recover_directory() other nodes can dump their necessary 191 * nodes their master rsb names that hash to us. 227 * Clear lkb's for departed nodes. 234 * departed nodes. 291 * rebuild and will be resent by the requesting nodes. 385 kfree(rv->nodes);
|
/linux-master/fs/f2fs/ |
H A D | segment.c | 451 unsigned int nodes = get_pages(sbi, F2FS_DIRTY_NODES); local 459 nodes >= threshold || meta >= threshold || 462 return dents + qdata + nodes + meta + imeta > global_threshold;
|
/linux-master/fs/jffs2/ |
H A D | build.c | 95 - Scan physical nodes. Build map of inodes/dirents. Allocate inocaches as we go 109 lists of physical nodes */ 251 raw = ic->nodes; 307 The erase code will do that, when all the nodes are completely gone. 322 of the medium to be available, for overhead caused by nodes being 335 /* When do we allow garbage collection to merge nodes to make 345 can't mark nodes obsolete on the medium, the old dirty nodes cause
|
H A D | erase.c | 214 /* Walk the inode's list once, removing any nodes from this eraseblock */ 221 prev = &ic->nodes; 249 jffs2_dbg(1, "Removed nodes in range 0x%08x-0x%08x from ino #%u\n", 257 this = ic->nodes; 282 if (ic->nodes == (void *)ic && ic->pino_nlink == 0)
|
H A D | gc.c | 214 finished checking all its nodes */ 298 pr_warn("eep. End of raw list while still supposedly nodes to GC\n"); 368 We can just copy any pristine nodes, but have 401 the alloc_sem() (for marking nodes invalid) so we must 445 the only valid nodes in the block, followed by erasure, 447 held _all_ the nodes of that inode.... never been seen but 549 frag->node->raw = f->inocache->nodes; 901 /* On a medium where we can't actually mark nodes obsolete 924 for (raw = f->inocache->nodes; raw != (void *)f->inocache; raw = raw->next_in_ino) { 1189 circles dirtying the nodes i [all...] |
H A D | nodelist.c | 220 /* Skip all the nodes which are completed before this one starts */ 383 /* If we now share a page with other nodes, mark either previous 563 /* Pass 'c' argument to indicate that nodes should be marked obsolete as 627 ref->next_in_ino = ic->nodes; 628 ic->nodes = ref; 691 /* Calculate totlen from surrounding nodes or eraseblock */
|
H A D | nodelist.h | 78 scan the nodes which are listed for it and build up a proper map of 79 which nodes are currently valid. JFFSv1 always used to keep that whole 129 /* flash_offset & 3 always has to be zero, because nodes are 141 /* Dirent nodes should be REF_PRISTINE only if they are not a deletion 148 copied. If you need to do anything different to GC inode-less nodes, then 155 dirent nodes) would take more space than this does. We also keep 167 struct jffs2_raw_node_ref *nodes; member in struct:jffs2_inode_cache 225 we're actually building up the original map of which nodes go where,
|
H A D | readinode.c | 156 /* We only check the data CRC of unchecked nodes */ 206 * This function is used when we read an inode. Data nodes arrive in 207 * arbitrary order -- they may be older or newer than the nodes which 229 Note that such nodes won't be REF_UNCHECKED since there are no data to 301 /* ... and is good. Kill 'this' and any subsequent nodes which are also overlapped */ 313 dbg_readinode("Done killing overlapped nodes\n"); 445 we add nodes to the real fragtree, as long as they don't overlap. And 446 having thrown away the majority of overlapped nodes as we went, there 447 really shouldn't be many sets of nodes which do overlap. If we start at 448 the end, we can use the overlap markers -- we can just eat nodes whic [all...] |
H A D | scan.c | 266 pr_notice("Cowardly refusing to erase blocks on filesystem with no valid JFFS2 nodes\n"); 933 /* We can't summarise nodes we don't grok */ 984 ic->nodes = (void *)ic;
|
H A D | wbuf.c | 156 /* It has to have had some nodes or we couldn't be here */ 272 /* Recover from failure to write wbuf. Recover the nodes up to the 318 /* All nodes were obsolete. Nothing to recover. */ 319 jffs2_dbg(1, "No non-obsolete nodes to be recovered. Just filing block bad\n"); 332 dbg_noderef("wbuf recover %08x-%08x (%d bytes in %d nodes)\n", 375 /* It wasn't. Go on and try to recover nodes complete in the wbuf */ 377 dbg_noderef("wbuf now recover %08x-%08x (%d bytes in %d nodes)\n", 489 struct jffs2_raw_node_ref **p = &ic->nodes;
|
/linux-master/fs/ntfs3/ |
H A D | index.c | 661 struct indx_node *n = fnd->nodes[i]; 667 fnd->nodes[i] = NULL; 678 if (i < 0 || i >= ARRAY_SIZE(fnd->nodes)) 680 fnd->nodes[i] = n; 692 n = fnd->nodes[i]; 693 fnd->nodes[i] = NULL; 1230 n = fnd->nodes[level - 1]; 1292 n = fnd->nodes[level - 1]; 1368 n = fnd->nodes[fnd->level - 1]; 1790 struct indx_node *n1 = fnd->nodes[leve [all...] |
H A D | ntfs_fs.h | 410 struct indx_node *nodes[20]; member in struct:ntfs_fnd
|
/linux-master/fs/ocfs2/cluster/ |
H A D | heartbeat.c | 54 * - o2hb_quorum_region_bitmap tracks live regions that have seen all nodes 405 printk(KERN_NOTICE "o2hb: all nodes hb write hung, maybe region %s (%pg) is down.\n", 863 * live nodes heartbeat on it. In other words, the region has been 864 * added to all nodes. 973 /* dead nodes only come to life after some number of 984 mlog(ML_HEARTBEAT, "o2hb: Add node %d to live nodes " 1000 /* We want to be sure that all nodes agree on the 1023 /* live nodes only go dead after enough consequtive missed 1036 "nodes bitmap\n", slot->ds_node_num); 1068 static int o2hb_highest_node(unsigned long *nodes, in argument 1073 o2hb_lowest_node(unsigned long *nodes, int numbits) argument [all...] |
/linux-master/fs/ubifs/ |
H A D | auth.c | 55 * @inhash: input hash of previous nodes 210 snod = list_first_entry(&sleb->nodes, struct ubifs_scan_node, list);
|
H A D | debug.c | 697 pr_err("LEB %-7d free %-8d dirty %-8d used %-8d free + dirty %-8d dark %-4d dead %-4d nodes fit %-3d flags %#-4x (", 851 pr_err("LEB %d has %d nodes ending at %d\n", lnum, 854 list_for_each_entry(snod, &sleb->nodes, list) { 1261 ubifs_err(c, "2 xent/dent nodes with the same name"); 1305 /* Only dirty zbranch may have no on-flash nodes */ 1621 /* Iterate over all indexing nodes */ 2282 * o makes sure that all leaf nodes exist and their CRCs are correct; 2286 * The function reads whole indexing tree and all nodes, so it is pretty 2318 * dbg_check_data_nodes_order - check that list of data nodes is sorted. 2320 * @head: the list of nodes ('struc [all...] |
H A D | gc.c | 14 * nodes) or not. For non-index LEBs, garbage collection finds a LEB which 15 * contains a lot of dirty space (obsolete nodes), and copies the non-obsolete 16 * nodes to the journal, at which point the garbage-collected LEB is free to be 17 * reused. For index LEBs, garbage collection marks the non-obsolete index nodes 19 * to be reused. Garbage collection will cause the number of dirty index nodes 33 * the UBIFS nodes GC deals with. Large nodes make GC waste more space. Indeed, 34 * if GC move data from LEB A to LEB B and nodes in LEB A are large, GC would 35 * have to waste large pieces of free space at the end of LEB B, because nodes 36 * from LEB A would not fit. And the worst situation is when all nodes ar [all...] |
H A D | lprops.c | 1114 list_for_each_entry(snod, &sleb->nodes, list) { 1161 * nodes from an uncompleted commit due to an
|