Lines Matching defs:csum_size

131 	return (bytes >> fs_info->sectorsize_bits) * fs_info->csum_size;
134 static size_t csum_size_to_bytes(const struct btrfs_fs_info *fs_info, u32 csum_size)
136 ASSERT(IS_ALIGNED(csum_size, fs_info->csum_size));
138 return (csum_size / fs_info->csum_size) << fs_info->sectorsize_bits;
144 fs_info->csum_size);
212 const u32 csum_size = fs_info->csum_size;
234 csums_in_item /= csum_size;
245 csum_offset * csum_size);
286 const u32 csum_size = fs_info->csum_size;
303 csum_len = (itemsize / csum_size) * sectorsize;
321 csum_len = (itemsize / csum_size) * sectorsize;
328 ret * csum_size);
347 const u32 csum_size = fs_info->csum_size;
375 if (nblocks * csum_size > BTRFS_BIO_INLINE_CSUM_SIZE) {
376 bbio->csum = kmalloc_array(nblocks, csum_size, GFP_NOFS);
407 (bio_offset >> fs_info->sectorsize_bits) * csum_size;
430 memset(csum_dst, 0, csum_size);
772 index += fs_info->csum_size;
817 const u32 csum_size = fs_info->csum_size;
823 csum_end = btrfs_item_size(leaf, path->slots[0]) / csum_size;
835 new_size *= csum_size;
846 new_size *= csum_size;
870 const u32 csum_size = fs_info->csum_size;
906 csum_end = btrfs_item_size(leaf, path->slots[0]) / csum_size;
970 offset *= csum_size;
972 shift_len = (len >> blocksize_bits) * csum_size;
1054 const u32 csum_size = fs_info->csum_size;
1086 if ((item_size / csum_size) >=
1087 MAX_CSUM_ITEMS(fs_info, csum_size)) {
1110 if (btrfs_leaf_free_space(leaf) >= csum_size) {
1120 csum_size, 1);
1137 csum_offset >= MAX_CSUM_ITEMS(fs_info, csum_size)) {
1143 csum_size) {
1193 diff = (csum_offset + extend_nr) * csum_size;
1195 MAX_CSUM_ITEMS(fs_info, csum_size) * csum_size);
1199 diff /= csum_size;
1200 diff *= csum_size;
1219 tmp = min_t(u64, tmp, MAX_CSUM_ITEMS(fs_info, csum_size));
1220 ins_size = csum_size * tmp;
1222 ins_size = csum_size;
1234 csum_offset * csum_size);
1237 ins_size *= csum_size;
1244 ins_size /= csum_size;