Lines Matching refs:cur

28 	struct xfs_btree_cur	*cur,
50 if (cur->bc_ops->type == XFS_BTREE_TYPE_INODE)
51 trace_xchk_ifork_btree_op_error(sc, cur, level,
54 trace_xchk_btree_op_error(sc, cur, level,
64 struct xfs_btree_cur *cur,
68 return __xchk_btree_process_error(sc, cur, level, error,
75 struct xfs_btree_cur *cur,
79 return __xchk_btree_process_error(sc, cur, level, error,
87 struct xfs_btree_cur *cur,
94 if (cur->bc_ops->type == XFS_BTREE_TYPE_INODE)
95 trace_xchk_ifork_btree_error(sc, cur, level,
98 trace_xchk_btree_error(sc, cur, level,
105 struct xfs_btree_cur *cur,
108 __xchk_btree_set_corrupt(sc, cur, level, XFS_SCRUB_OFLAG_CORRUPT,
115 struct xfs_btree_cur *cur,
118 __xchk_btree_set_corrupt(sc, cur, level, XFS_SCRUB_OFLAG_XCORRUPT,
125 struct xfs_btree_cur *cur,
128 __xchk_btree_set_corrupt(sc, cur, level, XFS_SCRUB_OFLAG_PREEN,
140 struct xfs_btree_cur *cur = bs->cur;
149 block = xfs_btree_get_block(cur, 0, &bp);
150 rec = xfs_btree_rec_addr(cur, cur->bc_levels[0].ptr, block);
152 trace_xchk_btree_rec(bs->sc, cur, 0);
156 !cur->bc_ops->recs_inorder(cur, &bs->lastrec, rec))
157 xchk_btree_set_corrupt(bs->sc, cur, 0);
158 memcpy(&bs->lastrec, rec, cur->bc_ops->rec_len);
161 if (cur->bc_nlevels == 1)
165 cur->bc_ops->init_key_from_rec(&key, rec);
166 keyblock = xfs_btree_get_block(cur, 1, &bp);
167 keyp = xfs_btree_key_addr(cur, cur->bc_levels[1].ptr, keyblock);
168 if (xfs_btree_keycmp_lt(cur, &key, keyp))
169 xchk_btree_set_corrupt(bs->sc, cur, 1);
171 if (!(cur->bc_ops->geom_flags & XFS_BTGEO_OVERLAPPING))
175 cur->bc_ops->init_high_key_from_rec(&hkey, rec);
176 keyp = xfs_btree_high_key_addr(cur, cur->bc_levels[1].ptr, keyblock);
177 if (xfs_btree_keycmp_lt(cur, keyp, &hkey))
178 xchk_btree_set_corrupt(bs->sc, cur, 1);
190 struct xfs_btree_cur *cur = bs->cur;
197 block = xfs_btree_get_block(cur, level, &bp);
198 key = xfs_btree_key_addr(cur, cur->bc_levels[level].ptr, block);
200 trace_xchk_btree_key(bs->sc, cur, level);
204 !cur->bc_ops->keys_inorder(cur, &bs->lastkey[level - 1].key, key))
205 xchk_btree_set_corrupt(bs->sc, cur, level);
206 memcpy(&bs->lastkey[level - 1].key, key, cur->bc_ops->key_len);
209 if (level + 1 >= cur->bc_nlevels)
213 keyblock = xfs_btree_get_block(cur, level + 1, &bp);
214 keyp = xfs_btree_key_addr(cur, cur->bc_levels[level + 1].ptr, keyblock);
215 if (xfs_btree_keycmp_lt(cur, key, keyp))
216 xchk_btree_set_corrupt(bs->sc, cur, level);
218 if (!(cur->bc_ops->geom_flags & XFS_BTGEO_OVERLAPPING))
222 key = xfs_btree_high_key_addr(cur, cur->bc_levels[level].ptr, block);
223 keyp = xfs_btree_high_key_addr(cur, cur->bc_levels[level + 1].ptr,
225 if (xfs_btree_keycmp_lt(cur, keyp, key))
226 xchk_btree_set_corrupt(bs->sc, cur, level);
240 if (bs->cur->bc_ops->type == XFS_BTREE_TYPE_INODE &&
241 level == bs->cur->bc_nlevels)
245 if (__xfs_btree_check_ptr(bs->cur, ptr, 0, level)) {
246 xchk_btree_set_corrupt(bs->sc, bs->cur, level);
261 struct xfs_btree_cur *cur = bs->cur;
269 error = xfs_btree_dup_cursor(cur, &ncur);
270 if (!xchk_btree_process_error(bs->sc, cur, level + 1, &error) ||
278 if (xfs_btree_ptr_is_null(cur, sibling)) {
284 xchk_btree_set_corrupt(bs->sc, cur, level);
294 if (!xchk_btree_process_error(bs->sc, cur, level + 1, &error))
297 xchk_btree_set_corrupt(bs->sc, cur, level + 1);
309 if (xfs_btree_diff_two_ptrs(cur, pp, sibling))
310 xchk_btree_set_corrupt(bs->sc, cur, level);
322 struct xfs_btree_cur *cur = bs->cur;
328 xfs_btree_get_sibling(cur, block, &leftsib, XFS_BB_LEFTSIB);
329 xfs_btree_get_sibling(cur, block, &rightsib, XFS_BB_RIGHTSIB);
333 if (level == cur->bc_nlevels - 1) {
334 if (!xfs_btree_ptr_is_null(cur, &leftsib) ||
335 !xfs_btree_ptr_is_null(cur, &rightsib))
336 xchk_btree_set_corrupt(bs->sc, cur, level);
376 if (!bs->cur)
379 agno = xfs_daddr_to_agno(bs->cur->bc_mp, daddr);
380 agbno = xfs_daddr_to_agbno(bs->cur->bc_mp, daddr);
387 init_sa = bs->cur->bc_ops->type != XFS_BTREE_TYPE_AG;
390 if (!xchk_btree_xref_process_error(bs->sc, bs->cur,
397 * The bnobt scrubber aliases bs->cur to bs->sc->sa.bno_cur, so we
401 if (!bs->sc->sa.bno_cur && xfs_btree_is_bno(bs->cur->bc_ops))
402 bs->cur = NULL;
405 if (!bs->sc->sa.rmap_cur && xfs_btree_is_rmap(bs->cur->bc_ops))
406 bs->cur = NULL;
422 struct xfs_btree_cur *cur = bs->cur;
431 if (cur->bc_ops->type != XFS_BTREE_TYPE_INODE)
432 xchk_btree_set_corrupt(bs->sc, bs->cur, level);
444 if (xfs_btree_is_bno(cur->bc_ops) || xfs_btree_is_rmap(cur->bc_ops)) {
477 if (xfs_btree_is_bmap(bs->cur->bc_ops) &&
478 bs->cur->bc_ino.whichfork == XFS_DATA_FORK &&
495 struct xfs_btree_cur *cur = bs->cur;
496 unsigned int root_level = cur->bc_nlevels - 1;
500 if (numrecs >= cur->bc_ops->get_minrecs(cur, level))
510 if (cur->bc_ops->type == XFS_BTREE_TYPE_INODE &&
511 level == cur->bc_nlevels - 2) {
516 root_block = xfs_btree_get_block(cur, root_level, &root_bp);
517 root_maxrecs = cur->bc_ops->get_dmaxrecs(cur, root_level);
521 xchk_btree_set_corrupt(bs->sc, cur, level);
530 xchk_btree_set_corrupt(bs->sc, cur, level);
546 struct xfs_btree_cur *cur = bs->cur;
550 if (level == cur->bc_nlevels - 1)
553 xfs_btree_get_keys(cur, block, &block_key);
556 parent_block = xfs_btree_get_block(cur, level + 1, &bp);
557 parent_low_key = xfs_btree_key_addr(cur, cur->bc_levels[level + 1].ptr,
559 if (xfs_btree_keycmp_ne(cur, &block_key, parent_low_key)) {
560 xchk_btree_set_corrupt(bs->sc, bs->cur, level);
564 if (!(cur->bc_ops->geom_flags & XFS_BTGEO_OVERLAPPING))
568 parent_high_key = xfs_btree_high_key_addr(cur,
569 cur->bc_levels[level + 1].ptr, parent_block);
570 block_high_key = xfs_btree_high_key_from_key(cur, &block_key);
571 if (xfs_btree_keycmp_ne(cur, block_high_key, parent_high_key))
572 xchk_btree_set_corrupt(bs->sc, bs->cur, level);
592 error = xfs_btree_lookup_get_block(bs->cur, level, pp, pblock);
593 if (!xchk_btree_process_error(bs->sc, bs->cur, level, &error) ||
597 xfs_btree_get_block(bs->cur, level, pbp);
598 if (__xfs_btree_check_block(bs->cur, *pblock, level, *pbp)) {
599 xchk_btree_set_corrupt(bs->sc, bs->cur, level);
638 struct xfs_btree_cur *cur = bs->cur;
645 if (level >= cur->bc_nlevels - 1)
649 xfs_btree_get_keys(cur, block, &block_keys);
652 parent_block = xfs_btree_get_block(cur, level + 1, &bp);
653 parent_keys = xfs_btree_key_addr(cur, cur->bc_levels[level + 1].ptr,
656 if (xfs_btree_keycmp_ne(cur, &block_keys, parent_keys))
657 xchk_btree_set_corrupt(bs->sc, cur, 1);
659 if (!(cur->bc_ops->geom_flags & XFS_BTGEO_OVERLAPPING))
663 high_bk = xfs_btree_high_key_from_key(cur, &block_keys);
664 high_pk = xfs_btree_high_key_addr(cur, cur->bc_levels[level + 1].ptr,
667 if (xfs_btree_keycmp_ne(cur, high_bk, high_pk))
668 xchk_btree_set_corrupt(bs->sc, cur, 1);
679 struct xfs_btree_cur *cur,
701 cur_sz = xchk_btree_sizeof(cur->bc_nlevels);
703 xchk_btree_set_corrupt(sc, cur, 0);
709 bs->cur = cur;
722 level = cur->bc_nlevels - 1;
723 xfs_btree_init_ptr_from_cur(cur, &ptr);
724 if (!xchk_btree_ptr_ok(bs, cur->bc_nlevels, &ptr))
730 cur->bc_levels[level].ptr = 1;
732 while (level < cur->bc_nlevels) {
733 block = xfs_btree_get_block(cur, level, &bp);
737 if (cur->bc_levels[level].ptr >
740 if (level < cur->bc_nlevels - 1)
741 cur->bc_levels[level + 1].ptr++;
750 recp = xfs_btree_rec_addr(cur, cur->bc_levels[0].ptr,
759 cur->bc_levels[level].ptr++;
764 if (cur->bc_levels[level].ptr >
767 if (level < cur->bc_nlevels - 1)
768 cur->bc_levels[level + 1].ptr++;
777 pp = xfs_btree_ptr_addr(cur, cur->bc_levels[level].ptr, block);
779 cur->bc_levels[level].ptr++;
787 cur->bc_levels[level].ptr = 1;
793 if (!error && bs->cur)