Lines Matching refs:sleb
501 struct ubifs_scan_leb *sleb;
507 sleb = ubifs_recover_leb(c, lnum, offs, c->sbuf, jhead != GCHD);
509 sleb = ubifs_scan(c, lnum, offs, c->sbuf, 0);
510 if (IS_ERR(sleb))
511 return PTR_ERR(sleb);
524 * sleb->endpt', and the space in the 'lnum' LEB between 'offs' and
525 * 'sleb->endpt' is used by bud data. We have to correctly calculate
535 list_for_each_entry(snod, &sleb->nodes, list) {
627 ubifs_assert(sleb->endpt - offs >= used);
628 ubifs_assert(sleb->endpt % c->min_io_size == 0);
630 if (sleb->endpt + c->min_io_size <= c->leb_size &&
633 sleb->endpt, UBI_SHORTTERM);
635 *dirty = sleb->endpt - offs - used;
636 *free = c->leb_size - sleb->endpt;
639 ubifs_scan_destroy(sleb);
645 ubifs_scan_destroy(sleb);
834 struct ubifs_scan_leb *sleb;
839 sleb = ubifs_scan(c, lnum, offs, sbuf, c->need_recovery);
840 if (IS_ERR(sleb)) {
841 if (PTR_ERR(sleb) != -EUCLEAN || !c->need_recovery)
842 return PTR_ERR(sleb);
843 sleb = ubifs_recover_log_leb(c, lnum, offs, sbuf);
844 if (IS_ERR(sleb))
845 return PTR_ERR(sleb);
848 if (sleb->nodes_cnt == 0) {
853 node = sleb->buf;
855 snod = list_entry(sleb->nodes.next, struct ubifs_scan_node, list);
900 list_for_each_entry(snod, &sleb->nodes, list) {
950 if (sleb->endpt || c->lhead_offs >= c->leb_size) {
952 c->lhead_offs = sleb->endpt;
955 err = !sleb->endpt;
957 ubifs_scan_destroy(sleb);
964 ubifs_scan_destroy(sleb);