/linux-master/fs/ext4/ |
H A D | symlink.c | 112 nd_terminate_link(bh->b_data, inode->i_size,
|
H A D | inode.c | 157 return S_ISLNK(inode->i_mode) && inode->i_size && 158 (inode->i_size < EXT4_N_BLOCKS * 4); 242 * Set inode->i_size to 0 before calling ext4_truncate(). We need 243 * special handling of symlinks here because i_size is used to 245 * block mappings. Setting i_size to 0 will remove its fast symlink 250 inode->i_size = 0; 1210 bool extended = (pos + len > inode->i_size) && 1216 * outside i_size. Trim these off again. Don't need 1277 loff_t old_size = inode->i_size; 1291 * it's important to update i_size whil 2267 loff_t i_size; local [all...] |
/linux-master/fs/ |
H A D | direct-io.c | 122 loff_t i_size; /* i_size when submitted */ member in struct:dio 266 ((offset + transferred) > dio->i_size)) 267 transferred = dio->i_size - offset; 614 loff_t i_size; local 632 * For writes that could fill holes inside i_size on a 644 i_size = i_size_read(dio->inode); 645 if (i_size && fs_startblk <= (i_size - 1) >> i_blkbits) 1150 /* Once we sampled i_size chec [all...] |
/linux-master/fs/erofs/ |
H A D | data.c | 103 map->m_plen = inode->i_size - offset; 115 vi->nid, inode->i_size, map->m_la); 136 if (map->m_la >= inode->i_size) { 164 round_up(inode->i_size - map->m_la, sb->s_blocksize));
|
H A D | erofs_fs.h | 163 __le32 i_size; member in struct:erofs_inode_compact 181 __le64 i_size; member in struct:erofs_inode_extended
|
/linux-master/fs/qnx6/ |
H A D | dir.c | 38 unsigned long last_byte = inode->i_size; 124 if (ctx->pos >= inode->i_size)
|
/linux-master/mm/ |
H A D | secretmem.c | 177 if ((ia_valid & ATTR_SIZE) && inode->i_size) 224 inode->i_size = 0;
|
H A D | shmem.c | 91 /* Pretend that each entry is of this size in directory's i_size */ 515 * only allocate huge pages if the page will be fully within i_size, 547 loff_t i_size; local 563 i_size = round_up(i_size_read(inode), PAGE_SIZE); 564 if (i_size >> PAGE_SHIFT >= index) 646 if (round_up(inode->i_size, PAGE_SIZE) == 647 round_up(inode->i_size, HPAGE_PMD_SIZE)) { 677 index = (inode->i_size & HPAGE_PMD_MASK) >> PAGE_SHIFT; 904 if (!vma->vm_pgoff && vma->vm_end - vma->vm_start >= inode->i_size) 938 * beyond i_size, an 2804 loff_t i_size = i_size_read(inode); local [all...] |
/linux-master/fs/hfs/ |
H A D | inode.c | 41 if (to > inode->i_size) { 42 truncate_pagecache(inode, inode->i_size); 137 * blocks outside i_size. Trim these off again. 204 inode->i_size = 2; 258 inode->i_size = 0; 279 inode->i_size = HFS_I(inode)->phys_size = log_size; 362 inode->i_size = be16_to_cpu(rec->dir.Val) + 2; 413 *log_size = cpu_to_be32(inode->i_size); 474 rec.dir.Val = cpu_to_be16(inode->i_size - 2);
|
/linux-master/fs/freevxfs/ |
H A D | vxfs_lookup.c | 62 loff_t limit = VXFS_DIRROUND(ip->i_size); 209 limit = VXFS_DIRROUND(ip->i_size);
|
/linux-master/fs/afs/ |
H A D | inode.c | 684 /* inode->i_size has now been changed. */ 700 loff_t i_size = op->setattr.old_i_size; local 702 if (size != i_size) { 729 loff_t i_size; local 741 i_size = i_size_read(inode); 750 if (attr->ia_size == i_size) 763 loff_t from = min(size, i_size); 764 loff_t to = max(size, i_size); 773 attr->ia_size < i_size &&
|
/linux-master/fs/nfs/ |
H A D | internal.h | 799 loff_t i_size = i_size_read(page_file_mapping(page)->host); local 801 if (i_size > 0) { 803 pgoff_t end_index = (i_size - 1) >> PAGE_SHIFT; 807 return ((i_size - 1) & ~PAGE_MASK) + 1; 817 loff_t i_size = i_size_read(folio_file_mapping(folio)->host); local 819 if (i_size > 0) { 821 pgoff_t end_index = (i_size - 1) >> folio_shift(folio); 825 return offset_in_folio(folio, i_size - 1) + 1;
|
/linux-master/fs/f2fs/ |
H A D | verity.c | 13 * boundary beyond i_size. This approach works because (a) verity files are 14 * readonly, and (b) pages fully beyond i_size aren't visible to userspace but 24 * since there can be a hole between i_size and the start of the Merkle tree. 36 return round_up(inode->i_size, 65536); 41 * we need to read beyond i_size. 174 * i_size won't be written properly. For crash consistency, this also 199 * metadata that was written beyond i_size (both from cache and from 204 * normal file accesses, garbage collection isn't limited by i_size). 207 truncate_inode_pages(inode->i_mapping, inode->i_size);
|
/linux-master/fs/netfs/ |
H A D | fscache_io.c | 212 loff_t start, size_t len, loff_t i_size, 242 ret = cres->ops->prepare_write(cres, &start, &len, len, i_size, false); 210 __fscache_write_to_cache(struct fscache_cookie *cookie, struct address_space *mapping, loff_t start, size_t len, loff_t i_size, netfs_io_terminated_t term_func, void *term_func_priv, bool cond) argument
|
/linux-master/fs/smb/client/ |
H A D | fscache.c | 153 loff_t i_size = i_size_read(inode); local 156 fscache_unuse_cookie(cifs_inode_cookie(inode), &cd, &i_size); local
|
/linux-master/fs/fat/ |
H A D | inode.c | 217 if (to > inode->i_size) { 218 truncate_pagecache(inode, inode->i_size); 219 fat_truncate_blocks(inode, inode->i_size); 481 inode->i_size = 0; 488 inode->i_size = (fclus + 1) << sbi->cluster_bits; 535 MSDOS_I(inode)->mmu_private = inode->i_size; 550 inode->i_size = le32_to_cpu(de->size); 554 MSDOS_I(inode)->mmu_private = inode->i_size; 562 inode->i_blocks = ((inode->i_size + (sbi->cluster_size - 1)) 657 inode->i_size [all...] |
/linux-master/fs/ubifs/ |
H A D | xattr.c | 133 inode->i_size = ui->ui_size = size; 201 ubifs_assert(c, ui->data_len == inode->i_size); 213 inode->i_size = ui->ui_size = size; 360 ubifs_assert(c, inode->i_size == ui->data_len); 470 ubifs_assert(c, ui->data_len == inode->i_size);
|
/linux-master/fs/ufs/ |
H A D | dir.c | 52 if (pos+len > dir->i_size) { 127 if ((dir->i_size >> PAGE_SHIFT) == page->index) { 128 limit = dir->i_size & ~PAGE_MASK; 220 unsigned last_byte = inode->i_size; 332 * This code plays outside i_size, so it locks the page 349 /* We hit i_size */ 444 if (pos > inode->i_size - UFS_DIR_REC_LEN(1))
|
/linux-master/fs/jfs/ |
H A D | namei.c | 429 if (dip->i_size > 1) 586 if (dip->i_size > 1) 636 if (ip->i_size < IDATASIZE) { 637 ip->i_size = 0; 672 if (ip->i_size) 705 if (ip->i_size < IDATASIZE) 758 if (ip->i_size) 957 ip->i_size = ssize - 1; 993 ip->i_size = ssize - 1; 1332 if (old_dir->i_size > [all...] |
/linux-master/fs/xfs/ |
H A D | xfs_iomap.c | 567 xfs_fsize_t i_size; local 632 i_size = XFS_FSB_TO_B(mp, offset_fsb + count_fsb); 633 if (i_size > offset + count) 634 i_size = offset + count; 635 if (update_isize && i_size > i_size_read(inode)) 636 i_size_write(inode, i_size); 637 i_size = xfs_new_eof(ip, i_size); 638 if (i_size) { 639 ip->i_disk_size = i_size; [all...] |
/linux-master/fs/hostfs/ |
H A D | hostfs_kern.c | 405 int end_index = inode->i_size >> PAGE_SHIFT; 409 count = inode->i_size & (PAGE_SIZE-1); 421 if (base > inode->i_size) 422 inode->i_size = base; 493 * i_size against the last byte written. 495 if (err > 0 && (pos > inode->i_size)) 496 inode->i_size = pos; 525 ino->i_size = st->size;
|
/linux-master/fs/btrfs/ |
H A D | reflink.c | 40 if (endoff > inode->i_size) { 517 * the i_size (which implies the whole inlined data). 588 * i_size, punching a hole in the target range will not create 590 * beyond current i_size. If the file previously had an i_size 591 * greater than the new i_size set by this clone operation, we 594 * i_size to the new i_size. If the clone range covers extents, 741 if (off + len == src->i_size) 742 len = ALIGN(src->i_size, b [all...] |
/linux-master/fs/hfsplus/ |
H A D | btree.c | 361 hip->phys_size = inode->i_size = 366 inode_set_bytes(inode, inode->i_size); 367 count = inode->i_size >> tree->node_size_shift;
|
/linux-master/fs/omfs/ |
H A D | dir.c | 27 int nbuckets = (dir->i_size - OMFS_DIR_START)/8; 218 int nbuckets = (inode->i_size - OMFS_DIR_START) / 8; 425 nbuckets = (dir->i_size - OMFS_DIR_START) / 8;
|
/linux-master/fs/cachefiles/ |
H A D | interface.c | 237 loff_t i_size, dio_size; local 241 i_size = i_size_read(inode); 243 trace_cachefiles_trunc(object, inode, i_size, dio_size,
|