/linux-master/fs/ext4/ |
H A D | truncate.h | 34 needed = inode->i_blocks >> (inode->i_sb->s_blocksize_bits - 9); 50 return EXT4_DATA_TRANS_BLOCKS(inode->i_sb) + needed;
|
H A D | fsync.c | 91 if (test_opt(inode->i_sb, BARRIER)) 101 journal_t *journal = EXT4_SB(inode->i_sb)->s_journal; 109 return ext4_force_commit(inode->i_sb); 135 if (unlikely(ext4_forced_shutdown(inode->i_sb))) 142 if (sb_rdonly(inode->i_sb)) { 145 if (ext4_forced_shutdown(inode->i_sb)) 150 if (!EXT4_SB(inode->i_sb)->s_journal) { 171 err = blkdev_issue_flush(inode->i_sb->s_bdev);
|
H A D | xattr_user.c | 26 if (!test_opt(inode->i_sb, XATTR_USER)) 39 if (!test_opt(inode->i_sb, XATTR_USER))
|
H A D | xattr_hurd.c | 26 if (!test_opt(inode->i_sb, XATTR_USER)) 40 if (!test_opt(inode->i_sb, XATTR_USER))
|
H A D | migrate.c | 96 unsigned long max_entries = inode->i_sb->s_blocksize >> 2; 98 bh = ext4_sb_bread(inode->i_sb, pblock, 0); 125 unsigned long max_entries = inode->i_sb->s_blocksize >> 2; 127 bh = ext4_sb_bread(inode->i_sb, pblock, 0); 155 unsigned long max_entries = inode->i_sb->s_blocksize >> 2; 157 bh = ext4_sb_bread(inode->i_sb, pblock, 0); 184 struct super_block *sb = inode->i_sb; 185 unsigned long max_entries = inode->i_sb->s_blocksize >> 2; 225 unsigned long max_entries = inode->i_sb->s_blocksize >> 2; 227 bh = ext4_sb_bread(inode->i_sb, le32_to_cp [all...] |
/linux-master/fs/jfs/ |
H A D | jfs_discard.c | 34 struct super_block *sb = ip->i_sb; 67 struct inode *ipbmap = JFS_SBI(ip->i_sb)->ipbmap; 68 struct bmap *bmp = JFS_SBI(ip->i_sb)->bmap; 69 struct super_block *sb = ipbmap->i_sb; 97 agno = BLKTOAG(start, JFS_SBI(ip->i_sb)); 98 agno_end = BLKTOAG(end, JFS_SBI(ip->i_sb));
|
/linux-master/fs/hpfs/ |
H A D | dir.c | 15 hpfs_lock(inode->i_sb); 18 hpfs_unlock(inode->i_sb); 31 struct super_block *s = i->i_sb; 77 hpfs_lock(inode->i_sb); 79 if (hpfs_sb(inode->i_sb)->sb_chk) { 80 if (hpfs_chk_sectors(inode->i_sb, inode->i_ino, 1, "dir_fnode")) { 84 if (hpfs_chk_sectors(inode->i_sb, hpfs_inode->i_dno, 4, "dir_dnode")) { 89 if (hpfs_sb(inode->i_sb)->sb_chk >= 2) { 93 if (!(fno = hpfs_map_fnode(inode->i_sb, inode->i_ino, &bh))) { 99 hpfs_error(inode->i_sb, "no [all...] |
H A D | inode.c | 16 struct super_block *sb = i->i_sb; 48 struct super_block *sb = i->i_sb; 62 if (hpfs_sb(i->i_sb)->sb_eas) { 63 if ((ea = hpfs_get_ea(i->i_sb, fnode, "UID", &ea_size))) { 70 if ((ea = hpfs_get_ea(i->i_sb, fnode, "GID", &ea_size))) { 77 if ((ea = hpfs_get_ea(i->i_sb, fnode, "SYMLINK", &ea_size))) { 89 if ((ea = hpfs_get_ea(i->i_sb, fnode, "MODE", &ea_size))) { 99 if ((ea = hpfs_get_ea(i->i_sb, fnode, "DEV", &ea_size))) { 128 hpfs_count_dnodes(i->i_sb, hpfs_inode->i_dno, &n_dnodes, &n_subdirs, NULL); 152 hpfs_error(i->i_sb, "fnod [all...] |
H A D | file.c | 19 hpfs_lock(inode->i_sb); 21 hpfs_unlock(inode->i_sb); 33 return sync_blockdev(inode->i_sb->s_bdev); 53 if (!(fnode = hpfs_map_fnode(inode->i_sb, inode->i_ino, &bh))) return 0; 54 disk_secno = hpfs_bplus_lookup(inode->i_sb, inode, &fnode->btree, file_secno, bh); 56 if (hpfs_chk_sectors(inode->i_sb, disk_secno, 1, "bmap")) return 0; 69 hpfs_lock_assert(i->i_sb); 74 hpfs_truncate_btree(i->i_sb, i->i_ino, 1, ((i->i_size + 511) >> 9)); 84 hpfs_lock(inode->i_sb); 89 n_secs = hpfs_search_hotfix_map_for_range(inode->i_sb, [all...] |
H A D | namei.c | 14 time64_t t = local_to_gmt(dir->i_sb, local_get_seconds(dir->i_sb)); 39 hpfs_lock(dir->i_sb); 41 fnode = hpfs_alloc_fnode(dir->i_sb, hpfs_i(dir)->i_dno, &fno, &bh); 44 dnode = hpfs_alloc_dnode(dir->i_sb, fno, &dno, &qbh0); 53 dee.creation_date = dee.write_date = dee.read_date = cpu_to_le32(local_get_seconds(dir->i_sb)); 54 result = new_inode(dir->i_sb); 62 inode_set_atime_to_ts(result, inode_set_ctime(result, local_to_gmt(dir->i_sb, le32_to_cpu(dee.creation_date)), 0))); 91 de = hpfs_add_de(dir->i_sb, dnode, "\001\001", 2, 0); 92 de->creation_date = de->write_date = de->read_date = cpu_to_le32(local_get_seconds(dir->i_sb)); [all...] |
H A D | dnode.c | 258 hpfs_error(i->i_sb, "%s(): namelen == %d", __func__, namelen); 263 if (!(d = hpfs_map_dnode(i->i_sb, dno, &qbh))) { 269 if (hpfs_sb(i->i_sb)->sb_chk) 270 if (hpfs_stop_cycles(i->i_sb, dno, &c1, &c2, "hpfs_add_to_dnode")) { 278 copy_de(de=hpfs_add_de(i->i_sb, d, name, namelen, down_ptr), new_de); 301 copy_de(de = hpfs_add_de(i->i_sb, nd, name, namelen, down_ptr), new_de); 304 if (!(ad = hpfs_alloc_dnode(i->i_sb, le32_to_cpu(d->up), &adno, &qbh1))) { 305 hpfs_error(i->i_sb, "unable to alloc dnode - dnode tree will be corrupted"); 315 copy_de(hpfs_add_de(i->i_sb, ad, de->name, de->namelen, de->down ? de_down_pointer(de) : 0), de); 325 set_last_pointer(i->i_sb, a [all...] |
/linux-master/fs/nilfs2/ |
H A D | file.c | 32 err = nilfs_construct_dsync_segment(inode->i_sb, inode, 35 err = nilfs_construct_segment(inode->i_sb); 38 nilfs = inode->i_sb->s_fs_info; 54 if (unlikely(nilfs_near_disk_full(inode->i_sb->s_fs_info))) 57 sb_start_pagefault(inode->i_sb); 95 ret = nilfs_transaction_begin(inode->i_sb, &ti, 1); 103 nilfs_transaction_abort(inode->i_sb); 107 nilfs_transaction_commit(inode->i_sb); 118 sb_end_pagefault(inode->i_sb);
|
H A D | namei.c | 63 inode = ino ? nilfs_iget(dir->i_sb, NILFS_I(dir)->i_root, ino) : NULL; 82 err = nilfs_transaction_begin(dir->i_sb, &ti, 1); 95 err = nilfs_transaction_commit(dir->i_sb); 97 nilfs_transaction_abort(dir->i_sb); 110 err = nilfs_transaction_begin(dir->i_sb, &ti, 1); 121 err = nilfs_transaction_commit(dir->i_sb); 123 nilfs_transaction_abort(dir->i_sb); 132 struct super_block *sb = dir->i_sb; 140 err = nilfs_transaction_begin(dir->i_sb, &ti, 1); 163 err = nilfs_transaction_commit(dir->i_sb); [all...] |
/linux-master/fs/udf/ |
H A D | ialloc.c | 27 udf_free_blocks(inode->i_sb, NULL, &UDF_I(inode)->i_location, 0, 1); 32 struct super_block *sb = dir->i_sb; 47 if (UDF_QUERY_FLAG(inode->i_sb, UDF_FLAG_USE_EXTENDED_FE)) { 51 iinfo->i_data = kzalloc(inode->i_sb->s_blocksize - 56 iinfo->i_data = kzalloc(inode->i_sb->s_blocksize - 67 block = udf_new_block(dir->i_sb, NULL, 97 if (UDF_QUERY_FLAG(inode->i_sb, UDF_FLAG_USE_AD_IN_ICB)) 99 else if (UDF_QUERY_FLAG(inode->i_sb, UDF_FLAG_USE_SHORT_AD))
|
H A D | truncate.c | 30 int last_block = (elen + inode->i_sb->s_blocksize - 1) >> 31 inode->i_sb->s_blocksize_bits; 32 int first_block = (nelen + inode->i_sb->s_blocksize - 1) >> 33 inode->i_sb->s_blocksize_bits; 37 udf_free_blocks(inode->i_sb, inode, eloc, 0, 52 udf_free_blocks(inode->i_sb, inode, eloc, 92 if (lbcount - inode->i_size >= inode->i_sb->s_blocksize) 93 udf_warn(inode->i_sb, 105 udf_err(inode->i_sb, 147 udf_free_blocks(inode->i_sb, inod [all...] |
/linux-master/fs/reiserfs/ |
H A D | namei.c | 312 if (namelen > REISERFS_MAX_NAME(dir->i_sb->s_blocksize)) 317 get_third_component(dir->i_sb, name, namelen), 322 search_by_entry_key(dir->i_sb, &key_to_search, 325 reiserfs_error(dir->i_sb, "zam-7001", "io error"); 361 if (REISERFS_MAX_NAME(dir->i_sb->s_blocksize) < dentry->d_name.len) 364 reiserfs_write_lock(dir->i_sb); 372 inode = reiserfs_iget(dir->i_sb, 375 reiserfs_write_unlock(dir->i_sb); 387 reiserfs_write_unlock(dir->i_sb); 412 reiserfs_write_lock(dir->i_sb); [all...] |
H A D | ioctl.c | 19 if (!reiserfs_attrs(inode->i_sb)) 34 reiserfs_write_lock(inode->i_sb); 37 if (!reiserfs_attrs(inode->i_sb)) 62 reiserfs_write_unlock(inode->i_sb); 81 reiserfs_write_lock(inode->i_sb); 119 reiserfs_write_unlock(inode->i_sb); 164 unsigned long blocksize = inode->i_sb->s_blocksize; 177 int depth = reiserfs_write_unlock_nested(inode->i_sb); 180 reiserfs_write_lock_nested(inode->i_sb, depth); 183 reiserfs_write_lock(inode->i_sb); [all...] |
H A D | inode.c | 35 2 * REISERFS_QUOTA_INIT_BLOCKS(inode->i_sb); 55 reiserfs_write_lock(inode->i_sb); 57 if (journal_begin(&th, inode->i_sb, jbegin_count)) 71 int depth = reiserfs_write_unlock_nested(inode->i_sb); 73 reiserfs_write_lock_nested(inode->i_sb, depth); 93 reiserfs_write_unlock(inode->i_sb); 234 map_bh(bh, inode->i_sb, block); 246 block < (1 << (31 - inode->i_sb->s_blocksize_bits))) 302 (loff_t) block * inode->i_sb->s_blocksize + 1, TYPE_ANY, 305 result = search_for_position_by_key(inode->i_sb, [all...] |
H A D | xattr.h | 55 ret = _ROUND_UP(xattr_size(size), inode->i_sb->s_blocksize); 56 ret >>= inode->i_sb->s_blocksize_bits; 75 size_t nblocks = JOURNAL_BLOCKS_PER_OBJECT(inode->i_sb); 78 nblocks += JOURNAL_BLOCKS_PER_OBJECT(inode->i_sb); 79 if (d_really_is_negative(REISERFS_SB(inode->i_sb)->xattr_root)) 80 nblocks += JOURNAL_BLOCKS_PER_OBJECT(inode->i_sb);
|
H A D | xattr_user.c | 14 if (!reiserfs_xattrs_user(inode->i_sb)) 26 if (!reiserfs_xattrs_user(inode->i_sb))
|
H A D | dir.c | 42 reiserfs_write_lock(inode->i_sb); 44 reiserfs_write_unlock(inode->i_sb); 55 struct dentry *privroot = REISERFS_SB(dir->i_sb)->priv_root; 79 reiserfs_write_lock(inode->i_sb); 81 reiserfs_check_lock_depth(inode->i_sb, "readdir"); 98 search_by_entry_key(inode->i_sb, &pos_key, &path_to_entry, 168 REISERFS_MAX_NAME(inode->i_sb-> 208 depth = reiserfs_write_unlock_nested(inode->i_sb); 212 reiserfs_write_lock_nested(inode->i_sb, depth); 218 reiserfs_write_lock_nested(inode->i_sb, dept [all...] |
H A D | file.c | 53 reiserfs_write_lock(inode->i_sb); 60 err = journal_begin(&th, inode->i_sb, 1); 68 err = journal_join_abort(&th, inode->i_sb); 81 reiserfs_warning(inode->i_sb, "clm-9001", 112 reiserfs_write_unlock(inode->i_sb); 158 reiserfs_write_lock(inode->i_sb); 160 reiserfs_write_unlock(inode->i_sb); 161 if (barrier_done != 1 && reiserfs_barrier_flush(inode->i_sb)) 162 blkdev_issue_flush(inode->i_sb->s_bdev); 180 struct super_block *s = inode->i_sb; [all...] |
/linux-master/fs/ext2/ |
H A D | xattr_user.c | 25 if (!test_opt(inode->i_sb, XATTR_USER)) 38 if (!test_opt(inode->i_sb, XATTR_USER))
|
/linux-master/fs/coda/ |
H A D | pioctl.c | 76 if (target_inode->i_sb != inode->i_sb) { 84 error = venus_pioctl(inode->i_sb, &(cnp->c_fid), cmd, &data);
|
/linux-master/fs/omfs/ |
H A D | file.c | 33 struct omfs_sb_info *sbi = OMFS_SB(inode->i_sb); 52 bh = omfs_bread(inode->i_sb, next); 79 omfs_clear_range(inode->i_sb, start, (int) count); 87 omfs_clear_range(inode->i_sb, last, sbi->s_mirrors); 92 bh = omfs_bread(inode->i_sb, next); 121 struct omfs_sb_info *sbi = OMFS_SB(inode->i_sb); 148 if (omfs_allocate_block(inode->i_sb, new_block)) { 162 ret = omfs_allocate_range(inode->i_sb, 1, sbi->s_clustersize, 197 int numblocks = clus_to_blk(OMFS_SB(inode->i_sb), 207 return clus_to_blk(OMFS_SB(inode->i_sb), [all...] |