/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/fs/ubifs/ |
H A D | lprops.c | 410 if (lprops->free == c->leb_size) { 415 if (lprops->free + lprops->dirty == c->leb_size) { 565 if ((lprops->flags & LPROPS_TAKEN) && lprops->free == c->leb_size) 577 c->lst.total_used -= c->leb_size - old_spc; 585 if (free == c->leb_size) { 586 if (lprops->free != c->leb_size) 588 } else if (lprops->free == c->leb_size) 618 c->lst.total_used += c->leb_size - new_spc; 621 if ((lprops->flags & LPROPS_TAKEN) && lprops->free == c->leb_size) 798 ubifs_assert(lprops->free == c->leb_size); [all...] |
H A D | sb.c | 103 if (jnl_lebs * c->leb_size > DEFAULT_MAX_JNL) 104 jnl_lebs = DEFAULT_MAX_JNL / c->leb_size; 112 tmp = 2 * (c->ref_node_alsz * jnl_lebs) + c->leb_size - 1; 113 log_lebs = tmp / c->leb_size; 165 tmp64 = (long long)max_buds * c->leb_size; 173 sup->leb_size = cpu_to_le32(c->leb_size); 192 main_bytes = (long long)main_lebs * c->leb_size; 377 if (le32_to_cpu(sup->leb_size) != c->leb_size) { [all...] |
H A D | lpt_commit.c | 165 if (c->ltab[i].free == c->leb_size) { 175 if (c->ltab[i].free == c->leb_size) { 206 if (!done_lsave && offs + c->lsave_sz <= c->leb_size) { 214 if (offs + c->ltab_sz <= c->leb_size) { 230 while (offs + len > c->leb_size) { 232 upd_ltab(c, lnum, c->leb_size - alen, alen - offs); 233 dbg_chk_lpt_sz(c, 2, c->leb_size - offs); 273 if (offs + c->lsave_sz > c->leb_size) { 275 upd_ltab(c, lnum, c->leb_size - alen, alen - offs); 276 dbg_chk_lpt_sz(c, 2, c->leb_size [all...] |
H A D | master.c | 120 c->lhead_offs < 0 || c->lhead_offs >= c->leb_size || 127 c->zroot.offs >= c->leb_size || c->zroot.offs & 7) { 145 c->ihead_offs > c->leb_size || c->ihead_offs & 7) { 150 main_sz = (long long)c->main_lebs * c->leb_size; 157 c->lpt_offs < 0 || c->lpt_offs + c->nnode_sz > c->leb_size) { 164 c->nhead_offs > c->leb_size) { 171 c->ltab_offs + c->ltab_sz > c->leb_size) { 178 c->lsave_offs + c->lsave_sz > c->leb_size)) { 327 c->lst.total_free += growth * (long long)c->leb_size; 371 if (offs + UBIFS_MST_NODE_SZ > c->leb_size) { [all...] |
H A D | recovery.c | 99 sbuf = vmalloc(c->leb_size); 103 err = ubi_read(c->ubi, lnum, sbuf, 0, c->leb_size); 110 len = c->leb_size; 111 while (offs + UBIFS_MST_NODE_SZ <= c->leb_size) { 155 if (offs < c->leb_size) { 165 if (offs < c->leb_size) 260 } else if (offs1 == 0 && offs2 + sz >= c->leb_size) { 285 if (offs2 + sz + sz <= c->leb_size) 378 check_len = c->leb_size - empty_offs; 410 memset(*buf, 0xff, c->leb_size [all...] |
H A D | tnc_commit.c | 143 if (gap_end == c->leb_size) { 289 c->ileb_len = c->leb_size; 290 gap_end = c->leb_size; 303 if (lp.free == c->leb_size) { 309 c->leb_size - c->ileb_len, 316 err = ubifs_change_one_lp(c, lnum, c->leb_size - c->ileb_len, dirt, 342 cnt -= (c->leb_size - c->ihead_offs) / c->max_idx_node_sz; 345 d = c->leb_size / c->max_idx_node_sz; 434 if (buf_offs + next_len > c->leb_size) 499 if (buf_offs + next_len <= c->leb_size) [all...] |
H A D | gc.c | 89 c->leb_size - wbuf->offs - wbuf->used); 330 avail = c->leb_size - wbuf->offs - wbuf->used; 345 avail = c->leb_size - wbuf->offs - wbuf->used; 437 * (c->leb_size - lp->free). 479 err = ubifs_change_one_lp(c, lnum, c->leb_size, 0, 0, 496 err = ubifs_change_one_lp(c, lnum, c->leb_size, 0, 0, 0, 0); 593 int space_before = c->leb_size - wbuf->offs - wbuf->used; 643 if (lp.free + lp.dirty == c->leb_size) { 651 if (lp.free != c->leb_size) { 662 c->leb_size, [all...] |
H A D | log.c | 124 h = (long long)c->lhead_lnum * c->leb_size + c->lhead_offs; 125 t = (long long)c->ltail_lnum * c->leb_size; 170 c->bud_bytes += c->leb_size - bud->start; 250 if (c->bud_bytes + c->leb_size - offs > c->max_bud_bytes) { 279 if (c->lhead_offs > c->leb_size - c->ref_node_alsz) { 364 c->cmt_bud_bytes += c->leb_size - bud->start; 367 dbg_jhead(bud->jhead), c->leb_size - bud->start, 429 if (lnum == -1 || offs == c->leb_size) 468 if (c->lhead_offs == c->leb_size) { 515 c->min_log_bytes = c->leb_size; [all...] |
H A D | scan.c | 100 offs + node_len + pad_len > c->leb_size) { 131 * @sbuf: scan buffer (must be c->leb_size) 151 err = ubi_read(c->ubi, lnum, sbuf + offs, offs, c->leb_size - offs); 154 " error %d", c->leb_size - offs, lnum, offs, err); 243 len = c->leb_size - offs; 255 * @sbuf: scan buffer (must be of @c->leb_size bytes in size) 270 int err, len = c->leb_size - offs;
|
H A D | lpt.c | 82 c->space_bits = fls(c->leb_size) - 3; 84 c->lpt_offs_bits = fls(c->leb_size - 1); 85 c->lpt_spc_bits = fls(c->leb_size); 122 while (sz > c->leb_size) { 124 sz -= c->leb_size; 146 lebs_needed = div_u64(sz + c->leb_size - 1, c->leb_size); 153 if (c->ltab_sz > c->leb_size) { 195 /* Small LPT model must have lpt_sz < leb_size */ 196 if (c->lpt_sz > c->leb_size) { [all...] |
H A D | super.c | 506 c->leb_size = c->vi.usable_leb_size; 507 c->half_leb_size = c->leb_size / 2; 511 if (c->leb_size < UBIFS_MIN_LEB_SZ) { 513 c->leb_size, UBIFS_MIN_LEB_SZ); 556 c->ranges[UBIFS_ORPH_NODE].max_len = c->leb_size; 586 c->leb_overhead = c->leb_size % UBIFS_MAX_DATA_NODE_SZ; 590 if (c->max_bu_buf_len > c->leb_size) 591 c->max_bu_buf_len = c->leb_size; 629 c->main_bytes = (long long)c->main_lebs * c->leb_size; 644 if (tmp > c->leb_size) { [all...] |
H A D | io.c | 105 ubifs_assert(!(offs & 7) && offs < c->leb_size); 124 if (node_len + offs > c->leb_size) 361 ubifs_assert(wbuf->offs + c->min_io_size <= c->leb_size); 387 c->leb_size - wbuf->offs, dirt); 409 ubifs_assert(offs >= 0 && offs <= c->leb_size); 519 ubifs_assert(!(wbuf->offs & 7) && wbuf->offs <= c->leb_size); 523 if (c->leb_size - wbuf->offs - wbuf->used < aligned_len) { 620 int free = c->leb_size - wbuf->offs - wbuf->used; 665 ubifs_assert(offs % c->min_io_size == 0 && offs < c->leb_size); 707 ubifs_assert(!(offs & 7) && offs < c->leb_size); [all...] |
H A D | find.c | 114 if (lprops->free + lprops->dirty == c->leb_size) { 195 lprops->free + lprops->dirty == c->leb_size)); 330 (pick_free && lp->free + lp->dirty == c->leb_size)); 381 if (!data->pick_free && lprops->free == c->leb_size) 389 if (lprops->free + lprops->dirty == c->leb_size && lprops->dirty > 0) 561 *offs = c->leb_size - lprops->free; 568 * LEBs that were freeable LEBs (free + dirty == leb_size) will 576 dbg_find("found LEB %d, free %d", lnum, c->leb_size - *offs); 577 ubifs_assert(*offs <= c->leb_size - min_space); 618 if (lprops->free + lprops->dirty != c->leb_size) [all...] |
H A D | replay.c | 120 if (r->offs == 0 && (lp->free != c->leb_size || lp->dirty != 0)) { 124 * property values should be lp->free == c->leb_size and 143 dirty -= c->leb_size - lp->free; 148 * that the dirty space may temporarily exceed c->leb_size 519 * At this point lprops thinks that this LEB has 'c->leb_size - offs' 523 * But we know that real amount of free space is 'c->leb_size - 630 if (sleb->endpt + c->min_io_size <= c->leb_size && 636 *free = c->leb_size - sleb->endpt; 801 * So this is why we require 'offs > c->leb_size'. 804 lnum < c->main_first || offs > c->leb_size || [all...] |
H A D | orphan.c | 42 * (c->leb_size - UBIFS_ORPH_NODE_SZ) / sizeof(__le64) 209 ((c->leb_size - UBIFS_ORPH_NODE_SZ) / sizeof(__le64)); 210 gap = c->leb_size - c->ohead_offs; 229 ((c->leb_size - UBIFS_ORPH_NODE_SZ) / sizeof(__le64)); 282 gap = c->leb_size - c->ohead_offs; 286 gap = c->leb_size; 319 ubifs_assert(c->ohead_offs + len <= c->leb_size); 714 c->orph_buf = vmalloc(c->leb_size);
|
H A D | ubifs-media.h | 599 * @leb_size: logical eraseblock size in bytes 627 __le32 leb_size; member in struct:ubifs_sb_node
|
H A D | tnc_misc.c | 318 zbr->offs + zbr->len > c->leb_size || zbr->offs & 7) {
|
H A D | budget.c | 236 available -= (long long)subtract_lebs * c->leb_size;
|
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/include/linux/mtd/ |
H A D | ubi.h | 118 * @leb_size: logical eraseblock size on this UBI device 123 * Note, @leb_size is the logical eraseblock size offered by the UBI device. 129 int leb_size; member in struct:ubi_device_info
|
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/mtd/ubi/ |
H A D | vtbl.c | 219 if (alignment > ubi->leb_size || alignment == 0) { 230 n = ubi->leb_size % alignment; 574 vol->usable_leb_size = ubi->leb_size - vol->data_pad; 643 vol->usable_leb_size = ubi->leb_size; 647 (long long)vol->used_ebs * (ubi->leb_size - vol->data_pad); 794 ubi->vtbl_slots = ubi->leb_size / UBI_VTBL_RECORD_SIZE;
|
H A D | vmt.c | 257 vol->usable_leb_size = ubi->leb_size - ubi->leb_size % req->alignment; 273 vol->data_pad = ubi->leb_size % vol->alignment; 743 if (vol->alignment > ubi->leb_size || vol->alignment == 0) { 754 n = ubi->leb_size % vol->alignment; 776 n = ubi->leb_size - vol->data_pad; 777 if (vol->usable_leb_size != ubi->leb_size - vol->data_pad) {
|
H A D | kapi.c | 42 di->leb_size = ubi->leb_size;
|
H A D | ubi.h | 364 * @leb_size: logical eraseblock size 450 int leb_size; member in struct:ubi_device
|
H A D | upd.c | 161 vol->upd_buf = vmalloc(ubi->leb_size);
|
H A D | build.c | 337 ret = sprintf(buf, "%d\n", ubi->leb_size); 758 ubi->leb_size = ubi->peb_size - ubi->leb_start; 768 ubi_msg("logical eraseblock size: %d bytes", ubi->leb_size);
|