Lines Matching refs:SECTOR_SHIFT
121 #define JOURNAL_SECTOR_DATA ((1 << SECTOR_SHIFT) - sizeof(commit_id_t))
445 ms += offset >> (SECTOR_SHIFT + ic->log2_buffer_sectors - ic->log2_tag_size);
446 mo = (offset << ic->log2_tag_size) & ((1U << SECTOR_SHIFT << ic->log2_buffer_sectors) - 1);
448 ms += (__u64)offset * ic->tag_size >> (SECTOR_SHIFT + ic->log2_buffer_sectors);
449 mo = (offset * ic->tag_size) & ((1U << SECTOR_SHIFT << ic->log2_buffer_sectors) - 1);
500 if (sizeof(struct superblock) + size > 1 << SECTOR_SHIFT) {
513 r = crypto_shash_update(desc, (__u8 *)ic->sb, (1 << SECTOR_SHIFT) - size);
520 r = crypto_shash_final(desc, (__u8 *)ic->sb + (1 << SECTOR_SHIFT) - size);
533 if (memcmp((__u8 *)ic->sb + (1 << SECTOR_SHIFT) - size, result, size)) {
740 *pl_index = sector >> (PAGE_SHIFT - SECTOR_SHIFT);
741 *pl_offset = (sector << SECTOR_SHIFT) & (PAGE_SIZE - 1);
753 *n_sectors = (PAGE_SIZE - pl_offset) >> SECTOR_SHIFT;
897 size_t n_bytes = (size_t)(n_sections * ic->journal_section_sectors) << SECTOR_SHIFT;
1065 pl_index = sector >> (PAGE_SHIFT - SECTOR_SHIFT);
1066 pl_offset = (sector << SECTOR_SHIFT) & (PAGE_SIZE - 1);
1189 pl_index = sector >> (PAGE_SHIFT - SECTOR_SHIFT);
1190 pl_offset = (sector << SECTOR_SHIFT) & (PAGE_SIZE - 1);
1437 to_copy = min((1U << SECTOR_SHIFT << ic->log2_buffer_sectors) - *metadata_offset, total_size);
1481 if (unlikely(*metadata_offset == 1U << SECTOR_SHIFT << ic->log2_buffer_sectors)) {
1628 bio_advance(bio, dio->range.n_sectors << SECTOR_SHIFT);
1681 r = crypto_shash_update(req, data, ic->sectors_per_block << SECTOR_SHIFT);
1741 alignment = dio->range.logical_sector | bio_sectors(bio) | (PAGE_SIZE >> SECTOR_SHIFT);
1745 buffer += (sector - io_loc.sector) << SECTOR_SHIFT;
1771 memcpy(mem + pos, buffer, ic->sectors_per_block << SECTOR_SHIFT);
1774 pos += ic->sectors_per_block << SECTOR_SHIFT;
1805 checksums = kmalloc((PAGE_SIZE >> SECTOR_SHIFT >> ic->sb->log2_sectors_per_block) * ic->tag_size + extra_space,
1826 unsigned int this_step_blocks = bi_size >> (SECTOR_SHIFT + ic->sb->log2_sectors_per_block);
1837 bi_size -= this_step_blocks << (SECTOR_SHIFT + ic->sb->log2_sectors_per_block);
1861 pos += ic->sectors_per_block << SECTOR_SHIFT;
1985 if (unlikely(bv.bv_len & ((ic->sectors_per_block << SECTOR_SHIFT) - 1))) {
2039 if (unlikely(bv.bv_len >> SECTOR_SHIFT > n_sectors))
2040 bv.bv_len = n_sectors << SECTOR_SHIFT;
2041 n_sectors -= bv.bv_len >> SECTOR_SHIFT;
2072 mem_ptr += 1 << SECTOR_SHIFT;
2119 memcpy(js, mem + bv.bv_offset, ic->sectors_per_block << SECTOR_SHIFT);
2149 bv.bv_offset += ic->sectors_per_block << SECTOR_SHIFT;
2150 } while (bv.bv_len -= ic->sectors_per_block << SECTOR_SHIFT);
2361 bio->bi_iter.bi_size = dio->range.n_sectors << SECTOR_SHIFT;
2739 recalc_buffer = __vmalloc(recalc_sectors << SECTOR_SHIFT, GFP_NOIO);
2835 integrity_sector_checksum(ic, logical_sector + i, recalc_buffer + (i << SECTOR_SHIFT), t);
2917 bbs->idx * (BITMAP_BLOCK_SIZE >> SECTOR_SHIFT),
2918 BITMAP_BLOCK_SIZE >> SECTOR_SHIFT, NULL);
2963 ic->n_bitmap_blocks * (BITMAP_BLOCK_SIZE >> SECTOR_SHIFT), NULL);
3269 ic->n_bitmap_blocks * (BITMAP_BLOCK_SIZE >> SECTOR_SHIFT), NULL);
3273 ic->n_bitmap_blocks * (BITMAP_BLOCK_SIZE >> SECTOR_SHIFT), NULL);
3285 ic->n_bitmap_blocks * (BITMAP_BLOCK_SIZE >> SECTOR_SHIFT), NULL);
3304 ic->n_bitmap_blocks * (BITMAP_BLOCK_SIZE >> SECTOR_SHIFT), NULL);
3348 ic->n_bitmap_blocks * (BITMAP_BLOCK_SIZE >> SECTOR_SHIFT), NULL);
3420 DMEMIT(" block_size:%u", ic->sectors_per_block << SECTOR_SHIFT);
3466 DMEMIT(",block_size=%u", ic->sectors_per_block << SECTOR_SHIFT);
3501 limits->logical_block_size = ic->sectors_per_block << SECTOR_SHIFT;
3502 limits->physical_block_size = ic->sectors_per_block << SECTOR_SHIFT;
3503 blk_limits_io_min(limits, ic->sectors_per_block << SECTOR_SHIFT);
3540 (__u64)(METADATA_PADDING_SECTORS << SECTOR_SHIFT) :
3541 (__u64)(1 << SECTOR_SHIFT << METADATA_PADDING_SECTORS);
3544 metadata_run_padding) >> SECTOR_SHIFT;
3557 meta_size = (meta_size + ((1U << (ic->log2_buffer_sectors + SECTOR_SHIFT)) - 1))
3558 >> (ic->log2_buffer_sectors + SECTOR_SHIFT);
3595 memset(ic->sb, 0, SB_SECTORS << SECTOR_SHIFT);
3670 bi.interval_exp = ic->sb->log2_sectors_per_block + SECTOR_SHIFT;
3760 end = end_offset + (1 << SECTOR_SHIFT);
3854 PAGE_SIZE >> SECTOR_SHIFT) >> (PAGE_SHIFT - SECTOR_SHIFT);
4040 section_req->cryptlen = (size_t)ic->journal_section_sectors << SECTOR_SHIFT;
4223 if (val < 1 << SECTOR_SHIFT ||
4224 val > MAX_SECTORS_PER_BLOCK << SECTOR_SHIFT ||
4230 ic->sectors_per_block = val >> SECTOR_SHIFT;
4288 ic->log2_buffer_sectors = min((int)__fls(buffer_sectors), 31 - SECTOR_SHIFT);
4399 ic->sb = alloc_pages_exact(SB_SECTORS << SECTOR_SHIFT, GFP_KERNEL);
4414 if (memchr_inv(ic->sb, 0, SB_SECTORS << SECTOR_SHIFT)) {
4496 bits_in_journal = ((__u64)ic->journal_section_sectors * ic->journal_sections) << (SECTOR_SHIFT + 3);
4573 1U << (SECTOR_SHIFT + ic->log2_buffer_sectors), 1, 0, NULL, NULL, 0);
4619 sector = i * (BITMAP_BLOCK_SIZE >> SECTOR_SHIFT);
4620 pl_index = sector >> (PAGE_SHIFT - SECTOR_SHIFT);
4621 pl_offset = (sector << SECTOR_SHIFT) & (PAGE_SIZE - 1);
4735 free_pages_exact(ic->sb, SB_SECTORS << SECTOR_SHIFT);