/linux-master/fs/xfs/scrub/ |
H A D | trace.c | 41 cur->bc_ops->type == XFS_BTREE_TYPE_INODE)
|
H A D | btree.c | 50 if (cur->bc_ops->type == XFS_BTREE_TYPE_INODE) 94 if (cur->bc_ops->type == XFS_BTREE_TYPE_INODE) 156 !cur->bc_ops->recs_inorder(cur, &bs->lastrec, rec)) 158 memcpy(&bs->lastrec, rec, cur->bc_ops->rec_len); 165 cur->bc_ops->init_key_from_rec(&key, rec); 171 if (!(cur->bc_ops->geom_flags & XFS_BTGEO_OVERLAPPING)) 175 cur->bc_ops->init_high_key_from_rec(&hkey, rec); 204 !cur->bc_ops->keys_inorder(cur, &bs->lastkey[level - 1].key, key)) 206 memcpy(&bs->lastkey[level - 1].key, key, cur->bc_ops->key_len); 218 if (!(cur->bc_ops [all...] |
H A D | health.c | 257 unsigned int mask = (*curp)->bc_ops->sick_mask;
|
H A D | ialloc.c | 79 ASSERT(xfs_btree_is_fino(cur->bc_ops)); 182 ASSERT(xfs_btree_is_ino(cur->bc_ops)); 517 if (xfs_btree_is_fino(bs->cur->bc_ops)) {
|
H A D | ialloc_repair.c | 545 cur->bc_ops->init_rec_from_cur(cur, block_rec); 572 cur->bc_ops->init_rec_from_cur(cur, block_rec);
|
H A D | trace.h | 465 __string(name, cur->bc_ops->name) 478 __assign_str(name, cur->bc_ops->name); 507 __string(name, cur->bc_ops->name) 521 __assign_str(name, cur->bc_ops->name); 550 __string(name, cur->bc_ops->name) 561 __assign_str(name, cur->bc_ops->name); 588 __string(name, cur->bc_ops->name) 601 __assign_str(name, cur->bc_ops->name); 628 __string(name, cur->bc_ops->name) 640 __assign_str(name, cur->bc_ops [all...] |
H A D | refcount_repair.c | 538 cur->bc_ops->init_rec_from_cur(cur, block_rec);
|
H A D | newbt.c | 546 if (cur->bc_ops->ptr_len == XFS_BTREE_LONG_PTR_LEN)
|
H A D | alloc_repair.c | 625 cur->bc_ops->init_rec_from_cur(cur, block_rec);
|
H A D | bmap_repair.c | 455 cur->bc_ops->init_rec_from_cur(cur, block_rec);
|
H A D | rmap_repair.c | 1203 cur->bc_ops->init_rec_from_cur(cur, block_rec);
|
/linux-master/fs/xfs/libxfs/ |
H A D | xfs_btree.c | 137 if (be32_to_cpu(block->bb_magic) != xfs_btree_magic(mp, cur->bc_ops)) 142 cur->bc_ops->get_maxrecs(cur, level)) 239 if (be32_to_cpu(block->bb_magic) != xfs_btree_magic(mp, cur->bc_ops)) 244 cur->bc_ops->get_maxrecs(cur, level)) 268 switch (cur->bc_ops->type) { 283 if (cur->bc_ops->ptr_len == XFS_BTREE_SHORT_PTR_LEN) 322 switch (cur->bc_ops->type) { 358 switch (cur->bc_ops->type) { 362 cur->bc_ops->name, cur->bc_flags, level, index, 369 cur->bc_ino.whichfork, cur->bc_ops [all...] |
H A D | xfs_btree_staging.c | 64 ASSERT(cur->bc_ops->type != XFS_BTREE_TYPE_INODE); 121 * root. The btree cursor's bc_ops will be overridden as needed to make the 131 ASSERT(cur->bc_ops->type == XFS_BTREE_TYPE_INODE); 310 xfs_btree_init_block(cur->bc_mp, ifp->if_broot, cur->bc_ops, 442 if (level == cur->bc_nlevels - 1 && cur->bc_ops->get_dmaxrecs) 443 return cur->bc_ops->get_dmaxrecs(cur, level); 445 ret = cur->bc_ops->get_maxrecs(cur, level); 470 return max_t(unsigned int, cur->bc_ops->get_minrecs(cur, level), npb); 498 if (cur->bc_ops->get_dmaxrecs) 499 maxnr = cur->bc_ops [all...] |
H A D | xfs_btree_mem.c | 29 ASSERT(cur->bc_ops->type == XFS_BTREE_TYPE_MEM); 41 ASSERT(cur->bc_ops->type == XFS_BTREE_TYPE_MEM); 53 ASSERT(cur->bc_ops->type == XFS_BTREE_TYPE_MEM); 55 ncur = xfs_btree_alloc_cursor(cur->bc_mp, cur->bc_tp, cur->bc_ops, 172 ASSERT(cur->bc_ops->type == XFS_BTREE_TYPE_MEM); 198 ASSERT(cur->bc_ops->type == XFS_BTREE_TYPE_MEM);
|
H A D | xfs_btree.h | 83 (cur)->bc_ops->statoff + __XBTS_ ## stat) 86 (cur)->bc_ops->statoff + __XBTS_ ## stat, val) 265 const struct xfs_btree_ops *bc_ops; member in struct:xfs_btree_cur 548 return cur->bc_ops->diff_two_keys(cur, key1, key2, NULL) < 0; 557 return cur->bc_ops->diff_two_keys(cur, key1, key2, NULL) > 0; 566 return cur->bc_ops->diff_two_keys(cur, key1, key2, NULL) == 0; 604 return cur->bc_ops->diff_two_keys(cur, key1, key2, mask) < 0; 614 return cur->bc_ops->diff_two_keys(cur, key1, key2, mask) > 0; 638 if (cur->bc_ops->ptr_len == XFS_BTREE_LONG_PTR_LEN) 681 cur->bc_ops [all...] |
H A D | xfs_ag.c | 496 xfs_btree_init_buf(mp, bp, id->bc_ops, 0, 0, id->agno); 562 xfs_btree_init_buf(mp, bp, id->bc_ops, 0, 0, id->agno); 578 xfs_btree_init_buf(mp, bp, id->bc_ops, 0, 4, id->agno); 790 const struct xfs_btree_ops *bc_ops; member in struct:xfs_aghdr_grow_data 844 .bc_ops = &xfs_bnobt_ops, 852 .bc_ops = &xfs_cntbt_ops, 860 .bc_ops = &xfs_inobt_ops, 868 .bc_ops = &xfs_finobt_ops, 876 .bc_ops = &xfs_rmapbt_ops, 884 .bc_ops [all...] |
H A D | xfs_alloc_btree.c | 54 if (xfs_btree_is_bno(cur->bc_ops)) { 134 ASSERT(!xfs_btree_is_bno(cur->bc_ops)); 244 if (xfs_btree_is_bno(cur->bc_ops)) 610 if (xfs_btree_is_bno(cur->bc_ops)) {
|
H A D | xfs_ag.h | 338 const struct xfs_btree_ops *bc_ops; /* btree ops */ member in struct:aghdr_init_data
|
H A D | xfs_ialloc_btree.c | 93 if (xfs_btree_is_fino(cur->bc_ops)) 538 if (xfs_btree_is_ino(cur->bc_ops)) {
|
H A D | xfs_alloc.c | 277 cur->bc_ops->name, cur->bc_ag.pag->pag_agno, fa); 921 bool isbnobt = xfs_btree_is_bno(cur->bc_ops); 4029 ASSERT(xfs_btree_is_bno(cur->bc_ops)); 4043 ASSERT(xfs_btree_is_bno(cur->bc_ops));
|
H A D | xfs_ialloc.c | 145 cur->bc_ops->name, cur->bc_ag.pag->pag_agno, fa); 2851 ASSERT(xfs_btree_is_ino(cur->bc_ops));
|
H A D | xfs_rmap.c | 277 if (xfs_btree_is_mem_rmap(cur->bc_ops)) 290 if (xfs_btree_is_mem_rmap(cur->bc_ops))
|
/linux-master/fs/xfs/ |
H A D | xfs_health.c | 529 switch (cur->bc_ops->type) { 534 ASSERT(cur->bc_ops->sick_mask); 535 xfs_ag_mark_sick(cur->bc_ag.pag, cur->bc_ops->sick_mask); 538 if (xfs_btree_is_bmap(cur->bc_ops)) {
|
H A D | xfs_trace.h | 1900 __string(name, cur->bc_ops->name) 1908 __assign_str(name, cur->bc_ops->name); 2462 __string(name, cur->bc_ops->name) 2470 __assign_str(name, cur->bc_ops->name); 2500 __string(name, cur->bc_ops->name) 2506 switch (cur->bc_ops->type) { 2520 __assign_str(name, cur->bc_ops->name); 2523 if (cur->bc_ops->ptr_len == XFS_BTREE_LONG_PTR_LEN) { 2553 __string(name, cur->bc_ops->name) 2560 if (cur->bc_ops [all...] |