Lines Matching refs:reserved_pebs

85 		ret = sprintf(buf, "%d\n", vol->reserved_pebs);
211 vol->reserved_pebs = div_u64(req->bytes + vol->usable_leb_size - 1,
215 if (vol->reserved_pebs > ubi->avail_pebs) {
224 ubi->avail_pebs -= vol->reserved_pebs;
225 ubi->rsvd_pebs += vol->reserved_pebs;
244 vol->eba_tbl = kmalloc(vol->reserved_pebs * sizeof(int), GFP_KERNEL);
250 for (i = 0; i < vol->reserved_pebs; i++)
254 vol->used_ebs = vol->reserved_pebs;
295 vtbl_rec.reserved_pebs = cpu_to_be32(vol->reserved_pebs);
341 ubi->rsvd_pebs -= vol->reserved_pebs;
342 ubi->avail_pebs += vol->reserved_pebs;
367 int i, err, vol_id = vol->vol_id, reserved_pebs = vol->reserved_pebs;
394 for (i = 0; i < vol->reserved_pebs; i++) {
404 ubi->rsvd_pebs -= reserved_pebs;
405 ubi->avail_pebs += reserved_pebs;
428 * @reserved_pebs: new size in physical eraseblocks
434 int ubi_resize_volume(struct ubi_volume_desc *desc, int reserved_pebs)
446 ubi->ubi_num, vol_id, vol->reserved_pebs, reserved_pebs);
449 reserved_pebs < vol->used_ebs) {
451 reserved_pebs, vol->used_ebs);
456 if (reserved_pebs == vol->reserved_pebs)
459 new_mapping = kmalloc(reserved_pebs * sizeof(int), GFP_KERNEL);
463 for (i = 0; i < reserved_pebs; i++)
475 pebs = reserved_pebs - vol->reserved_pebs;
490 for (i = 0; i < vol->reserved_pebs; i++)
499 vtbl_rec.reserved_pebs = cpu_to_be32(reserved_pebs);
506 err = ubi_eba_unmap_leb(ubi, vol, reserved_pebs + i);
514 for (i = 0; i < reserved_pebs; i++)
521 vol->reserved_pebs = reserved_pebs;
523 vol->used_ebs = reserved_pebs;
658 int reserved_pebs, alignment, data_pad, vol_type, name_len, upd_marker;
664 reserved_pebs = be32_to_cpu(ubi->vtbl[vol_id].reserved_pebs);
668 if (reserved_pebs) {
676 if (vol->reserved_pebs < 0 || vol->alignment < 0 || vol->data_pad < 0 ||
709 if (vol->reserved_pebs > ubi->good_peb_count) {
710 ubi_err(ubi, "too large reserved_pebs");
738 if (vol->used_ebs != vol->reserved_pebs) {
756 if (vol->used_ebs < 0 || vol->used_ebs > vol->reserved_pebs) {