Lines Matching defs:lend
300 * @lend: offset to which to truncate (inclusive)
304 * if lstart or lend + 1 is not page aligned).
317 * truncate_inode_pages_range is able to handle cases where lend + 1 is not
321 loff_t lstart, loff_t lend)
339 * Note that 'end' is exclusive while 'lend' is inclusive.
342 if (lend == -1)
344 * lend == -1 indicates end-of-file so we have to set 'end'
350 end = (lend + 1) >> PAGE_SHIFT;
366 same_folio = (lstart >> PAGE_SHIFT) == (lend >> PAGE_SHIFT);
369 same_folio = lend < folio_pos(folio) + folio_size(folio);
370 if (!truncate_inode_partial_folio(folio, lstart, lend)) {
381 folio = __filemap_get_folio(mapping, lend >> PAGE_SHIFT,
384 if (!truncate_inode_partial_folio(folio, lstart, lend))
816 * @lend: offset of last byte of hole
825 void truncate_pagecache_range(struct inode *inode, loff_t lstart, loff_t lend)
829 loff_t unmap_end = round_down(1 + lend, PAGE_SIZE) - 1;
835 * allows holelen 0 for all, and we allow lend -1 for end of file.
846 truncate_inode_pages_range(mapping, lstart, lend);