Lines Matching refs:csum_size

134 	return (bytes >> fs_info->sectorsize_bits) * fs_info->csum_size;
137 static size_t csum_size_to_bytes(const struct btrfs_fs_info *fs_info, u32 csum_size)
139 ASSERT(IS_ALIGNED(csum_size, fs_info->csum_size));
141 return (csum_size / fs_info->csum_size) << fs_info->sectorsize_bits;
147 fs_info->csum_size);
216 const u32 csum_size = fs_info->csum_size;
238 csums_in_item /= csum_size;
249 csum_offset * csum_size);
290 const u32 csum_size = fs_info->csum_size;
307 csum_len = (itemsize / csum_size) * sectorsize;
325 csum_len = (itemsize / csum_size) * sectorsize;
332 ret * csum_size);
351 const u32 csum_size = fs_info->csum_size;
379 if (nblocks * csum_size > BTRFS_BIO_INLINE_CSUM_SIZE) {
380 bbio->csum = kmalloc_array(nblocks, csum_size, GFP_NOFS);
411 (bio_offset >> fs_info->sectorsize_bits) * csum_size;
434 memset(csum_dst, 0, csum_size);
776 index += fs_info->csum_size;
821 const u32 csum_size = fs_info->csum_size;
827 csum_end = btrfs_item_size(leaf, path->slots[0]) / csum_size;
839 new_size *= csum_size;
850 new_size *= csum_size;
874 const u32 csum_size = fs_info->csum_size;
910 csum_end = btrfs_item_size(leaf, path->slots[0]) / csum_size;
974 offset *= csum_size;
976 shift_len = (len >> blocksize_bits) * csum_size;
1058 const u32 csum_size = fs_info->csum_size;
1090 if ((item_size / csum_size) >=
1091 MAX_CSUM_ITEMS(fs_info, csum_size)) {
1114 if (btrfs_leaf_free_space(leaf) >= csum_size) {
1124 csum_size, 1);
1141 csum_offset >= MAX_CSUM_ITEMS(fs_info, csum_size)) {
1147 csum_size) {
1197 diff = (csum_offset + extend_nr) * csum_size;
1199 MAX_CSUM_ITEMS(fs_info, csum_size) * csum_size);
1203 diff /= csum_size;
1204 diff *= csum_size;
1223 tmp = min_t(u64, tmp, MAX_CSUM_ITEMS(fs_info, csum_size));
1224 ins_size = csum_size * tmp;
1226 ins_size = csum_size;
1240 csum_offset * csum_size);
1243 ins_size *= csum_size;
1250 ins_size /= csum_size;