/linux-master/fs/ubifs/ |
H A D | lpt_commit.c | 157 if (c->ltab[i].free == c->leb_size) { 167 if (c->ltab[i].free == c->leb_size) { 198 if (!done_lsave && offs + c->lsave_sz <= c->leb_size) { 206 if (offs + c->ltab_sz <= c->leb_size) { 222 while (offs + len > c->leb_size) { 224 upd_ltab(c, lnum, c->leb_size - alen, alen - offs); 225 dbg_chk_lpt_sz(c, 2, c->leb_size - offs); 265 if (offs + c->lsave_sz > c->leb_size) { 267 upd_ltab(c, lnum, c->leb_size - alen, alen - offs); 268 dbg_chk_lpt_sz(c, 2, c->leb_size [all...] |
H A D | lprops.c | 402 if (lprops->free == c->leb_size) { 407 if (lprops->free + lprops->dirty == c->leb_size) { 557 if ((lprops->flags & LPROPS_TAKEN) && lprops->free == c->leb_size) 569 c->lst.total_used -= c->leb_size - old_spc; 577 if (free == c->leb_size) { 578 if (lprops->free != c->leb_size) 580 } else if (lprops->free == c->leb_size) 610 c->lst.total_used += c->leb_size - new_spc; 613 if ((lprops->flags & LPROPS_TAKEN) && lprops->free == c->leb_size) 790 ubifs_assert(c, lprops->free == c->leb_size); [all...] |
H A D | io.c | 252 ubifs_assert(c, !(offs & 7) && offs < c->leb_size); 273 if (node_len + offs > c->leb_size) 571 ubifs_assert(c, wbuf->offs + wbuf->size <= c->leb_size); 576 if (c->leb_size - wbuf->offs >= c->max_write_size) 606 if (c->leb_size - wbuf->offs < c->max_write_size) 607 wbuf->size = c->leb_size - wbuf->offs; 619 c->leb_size - wbuf->offs, dirt); 639 ubifs_assert(c, offs >= 0 && offs <= c->leb_size); 647 if (c->leb_size - wbuf->offs < c->max_write_size) 648 wbuf->size = c->leb_size [all...] |
H A D | log.c | 94 h = (long long)c->lhead_lnum * c->leb_size + c->lhead_offs; 95 t = (long long)c->ltail_lnum * c->leb_size; 144 c->bud_bytes += c->leb_size - bud->start; 204 if (c->bud_bytes + c->leb_size - offs > c->max_bud_bytes) { 234 if (c->lhead_offs > c->leb_size - c->ref_node_alsz) { 325 c->cmt_bud_bytes += c->leb_size - bud->start; 328 c->leb_size - bud->start, c->cmt_bud_bytes); 397 if (lnum == -1 || offs == c->leb_size) 441 ubifs_assert(c, c->lhead_offs < c->leb_size); 485 c->min_log_bytes = c->leb_size; [all...] |
H A D | recovery.c | 104 sbuf = vmalloc(c->leb_size); 108 err = ubifs_leb_read(c, lnum, sbuf, 0, c->leb_size, 0); 115 len = c->leb_size; 116 while (offs + UBIFS_MST_NODE_SZ <= c->leb_size) { 160 if (offs < c->leb_size) { 170 if (offs < c->leb_size) 267 c->leb_size - offs2 - sz < sz) { 292 if (offs2 + sz + sz <= c->leb_size) 355 ubifs_dump_node(c, mst1, c->leb_size - ((void *)mst1 - buf1)); 359 ubifs_dump_node(c, mst2, c->leb_size [all...] |
H A D | scan.c | 89 offs + node_len + pad_len > c->leb_size) { 120 * @sbuf: scan buffer (must be c->leb_size) 141 err = ubifs_leb_read(c, lnum, sbuf + offs, offs, c->leb_size - offs, 0); 144 c->leb_size - offs, lnum, offs, err); 231 len = c->leb_size - offs; 243 * @sbuf: scan buffer (must be of @c->leb_size bytes in size) 258 int err, len = c->leb_size - offs;
|
H A D | sb.c | 109 if (jnl_lebs * c->leb_size > DEFAULT_MAX_JNL) 110 jnl_lebs = DEFAULT_MAX_JNL / c->leb_size; 118 tmp = 2 * (c->ref_node_alsz * jnl_lebs) + c->leb_size - 1; 119 log_lebs = tmp / c->leb_size; 177 tmp64 = (long long)max_buds * c->leb_size; 197 sup->leb_size = cpu_to_le32(c->leb_size); 216 main_bytes = (long long)main_lebs * c->leb_size; 400 if (le32_to_cpu(sup->leb_size) != c->leb_size) { [all...] |
H A D | lpt.c | 70 c->space_bits = fls(c->leb_size) - 3; 72 c->lpt_offs_bits = fls(c->leb_size - 1); 73 c->lpt_spc_bits = fls(c->leb_size); 110 while (sz > c->leb_size) { 112 sz -= c->leb_size; 134 lebs_needed = div_u64(sz + c->leb_size - 1, c->leb_size); 141 if (c->ltab_sz > c->leb_size) { 183 /* Small LPT model must have lpt_sz < leb_size */ 184 if (c->lpt_sz > c->leb_size) { [all...] |
H A D | tnc_commit.c | 141 if (gap_end == c->leb_size) { 287 c->ileb_len = c->leb_size; 288 gap_end = c->leb_size; 301 if (lp.free == c->leb_size) { 307 c->leb_size - c->ileb_len, 314 err = ubifs_change_one_lp(c, lnum, c->leb_size - c->ileb_len, dirt, 339 cnt -= (c->leb_size - c->ihead_offs) / c->max_idx_node_sz; 342 d = c->leb_size / c->max_idx_node_sz; 449 if (buf_offs + next_len > c->leb_size) 515 buf_offs + used + next_len <= c->leb_size [all...] |
H A D | master.c | 193 c->lhead_offs < 0 || c->lhead_offs >= c->leb_size || 200 c->zroot.offs >= c->leb_size || c->zroot.offs & 7) { 218 c->ihead_offs > c->leb_size || c->ihead_offs & 7) { 223 main_sz = (long long)c->main_lebs * c->leb_size; 230 c->lpt_offs < 0 || c->lpt_offs + c->nnode_sz > c->leb_size) { 237 c->nhead_offs > c->leb_size) { 244 c->ltab_offs + c->ltab_sz > c->leb_size) { 251 c->lsave_offs + c->lsave_sz > c->leb_size)) { 402 c->lst.total_free += growth * (long long)c->leb_size; 446 if (offs + UBIFS_MST_NODE_SZ > c->leb_size) { [all...] |
H A D | super.c | 527 c->leb_size = c->vi.usable_leb_size; 529 c->half_leb_size = c->leb_size / 2; 535 if (c->leb_size < UBIFS_MIN_LEB_SZ) { 537 c->leb_size, UBIFS_MIN_LEB_SZ); 595 c->ranges[UBIFS_SIG_NODE].max_len = c->leb_size - UBIFS_SB_NODE_SZ; 601 c->ranges[UBIFS_ORPH_NODE].max_len = c->leb_size; 631 c->leb_overhead = c->leb_size % UBIFS_MAX_DATA_NODE_SZ; 635 if (c->max_bu_buf_len > c->leb_size) 636 c->max_bu_buf_len = c->leb_size; 639 c->min_log_bytes = c->leb_size; [all...] |
H A D | gc.c | 73 c->leb_size - wbuf->offs - wbuf->used); 358 avail = c->leb_size - wbuf->offs - wbuf->used - 380 avail = c->leb_size - wbuf->offs - wbuf->used - 503 if (lp->free + lp->dirty == c->leb_size) { 508 if (lp->free != c->leb_size) { 517 err = ubifs_change_one_lp(c, lp->lnum, c->leb_size, 536 * (c->leb_size - lp->free). 578 err = ubifs_change_one_lp(c, lnum, c->leb_size, 0, 0, 595 err = ubifs_change_one_lp(c, lnum, c->leb_size, 0, 0, 0, 0); 745 space_before = c->leb_size [all...] |
H A D | find.c | 103 if (lprops->free + lprops->dirty == c->leb_size) { 183 lprops->free + lprops->dirty == c->leb_size)); 318 (pick_free && lp->free + lp->dirty == c->leb_size)); 370 if (!data->pick_free && lprops->free == c->leb_size) 378 if (lprops->free + lprops->dirty == c->leb_size && lprops->dirty > 0) 550 *offs = c->leb_size - lprops->free; 557 * LEBs that were freeable LEBs (free + dirty == leb_size) will 565 dbg_find("found LEB %d, free %d", lnum, c->leb_size - *offs); 566 ubifs_assert(c, *offs <= c->leb_size - min_space); 608 if (lprops->free + lprops->dirty != c->leb_size) [all...] |
H A D | orphan.c | 30 * (c->leb_size - UBIFS_ORPH_NODE_SZ) / sizeof(__le64) 278 ((c->leb_size - UBIFS_ORPH_NODE_SZ) / sizeof(__le64)); 279 gap = c->leb_size - c->ohead_offs; 298 ((c->leb_size - UBIFS_ORPH_NODE_SZ) / sizeof(__le64)); 350 gap = c->leb_size - c->ohead_offs; 354 gap = c->leb_size; 389 ubifs_assert(c, c->ohead_offs + len <= c->leb_size); 650 c->leb_size - snod->offs); 679 c->leb_size - snod->offs); 823 c->orph_buf = vmalloc(c->leb_size); [all...] |
H A D | replay.c | 101 if (b->bud->start == 0 && (lp->free != c->leb_size || lp->dirty != 0)) { 105 * property values should be @lp->free == @c->leb_size and 125 dirty -= c->leb_size - lp->free; 130 * that the dirty space may temporarily exceed c->leb_size 147 b->bud->lnum, c->leb_size - b->free); 706 * At this point lprops thinks that this LEB has 'c->leb_size - offs' 710 * But we know that real amount of free space is 'c->leb_size - 824 b->free = c->leb_size - sleb->endpt; 834 ubifs_dump_node(c, snod->node, c->leb_size - snod->offs); 953 * So this is why we require 'offs > c->leb_size' [all...] |
H A D | misc.h | 281 int max_xattrs = (c->leb_size / 2) / UBIFS_INO_NODE_SZ;
|
H A D | debug.c | 365 pr_err("\tleb_size %u\n", le32_to_cpu(sup->leb_size)); 694 lp->lnum, lp->free, lp->dirty, c->leb_size - spc, spc, 698 lp->lnum, lp->free, lp->dirty, c->leb_size - spc, spc, 839 buf = __vmalloc(c->leb_size, GFP_NOFS); 858 ubifs_dump_node(c, snod->node, c->leb_size - snod->offs); 2343 ubifs_dump_node(c, sa->node, c->leb_size - sa->offs); 2348 ubifs_dump_node(c, sb->node, c->leb_size - sb->offs); 2379 ubifs_dump_node(c, sa->node, c->leb_size - sa->offs); 2380 ubifs_dump_node(c, sb->node, c->leb_size - sb->offs); 2411 ubifs_dump_node(c, sa->node, c->leb_size [all...] |
/linux-master/drivers/mtd/ubi/ |
H A D | wl.h | 16 *count += (ubi->fm_size / ubi->leb_size) * 2 + ubi->fm_pool_rsv_cnt;
|
H A D | vtbl.c | 199 if (alignment > ubi->leb_size || alignment == 0) { 210 n = ubi->leb_size % alignment; 545 vol->usable_leb_size = ubi->leb_size - vol->data_pad; 635 vol->usable_leb_size = ubi->leb_size; 639 (long long)vol->used_ebs * (ubi->leb_size - vol->data_pad); 795 if (ubi->leb_size < UBI_VTBL_RECORD_SIZE) { 800 ubi->vtbl_slots = ubi->leb_size / UBI_VTBL_RECORD_SIZE;
|
H A D | block.c | 78 int leb_size; member in struct:ubiblock 188 int offset = do_div(pos, dev->leb_size); 209 if (offset + to_read > dev->leb_size) 210 to_read = dev->leb_size - offset; 381 dev->leb_size = vi->usable_leb_size;
|
H A D | vmt.c | 226 vol->usable_leb_size = ubi->leb_size - ubi->leb_size % req->alignment; 246 vol->data_pad = ubi->leb_size % vol->alignment; 705 if (vol->alignment > ubi->leb_size || vol->alignment == 0) { 716 n = ubi->leb_size % vol->alignment; 738 n = ubi->leb_size - vol->data_pad; 739 if (vol->usable_leb_size != ubi->leb_size - vol->data_pad) {
|
H A D | attach.c | 878 memset(ubi->peb_buf, 0x00, ubi->leb_size); 881 ubi->leb_size); 897 if (ubi_check_pattern(ubi->peb_buf, 0xFF, ubi->leb_size)) 905 pnum, ubi->leb_start, ubi->leb_size); 907 ubi->peb_buf, ubi->leb_size, 1); 1711 if (av->data_pad > ubi->leb_size / 2) {
|
H A D | fastmap-wl.c | 121 ubi->fm_size / ubi->leb_size - 1 + ubi->fm_pool_rsv_cnt; 155 fm_used = ubi->fm_size / ubi->leb_size - 1;
|
H A D | fastmap.c | 92 return roundup(size, ubi->leb_size); 933 fm_size = ubi->leb_size * used_blocks; 1025 ret = ubi_io_read_data(ubi, ubi->fm_buf + (ubi->leb_size * i), 1026 pnum, 0, ubi->leb_size); 1367 ret = ubi_io_write_data(ubi, fm_raw + (i * ubi->leb_size), 1368 new_fm->e[i]->pnum, 0, ubi->leb_size); 1512 new_fm->used_blocks = ubi->fm_size / ubi->leb_size;
|
/linux-master/include/linux/mtd/ |
H A D | ubi.h | 148 * @leb_size: logical eraseblock size on this UBI device 157 * Note, @leb_size is the logical eraseblock size offered by the UBI device. 179 int leb_size; member in struct:ubi_device_info
|