Lines Matching refs:sectors

346 		r10_bio->devs[slot].addr + (r10_bio->sectors);
432 r10_bio->sectors,
538 r10_bio->sectors,
735 int sectors = r10_bio->sectors;
762 && (this_sector + sectors >= conf->next_resync)) ||
765 this_sector + sectors)))
780 r10_bio->devs[slot].addr + sectors >
794 r10_bio->devs[slot].addr + sectors > rdev->recovery_offset)
798 if (is_badblock(rdev, dev_sector, sectors,
809 if (!do_balance && sectors > bad_sectors)
810 sectors = bad_sectors;
811 if (best_good_sectors > sectors)
812 best_good_sectors = sectors;
827 best_good_sectors = sectors;
1147 struct bio *bio, sector_t sectors)
1156 bio->bi_iter.bi_sector + sectors > conf->reshape_progress) {
1166 sectors);
1214 if (!regular_request_wait(mddev, conf, bio, r10_bio->sectors))
1240 r10_bio->sectors = max_sectors;
1377 if (!r10_bio->sectors)
1380 is_bad = is_badblock(rdev, dev_sector, r10_bio->sectors,
1412 sector_t sectors;
1436 sectors = r10_bio->sectors;
1437 if (!regular_request_wait(mddev, conf, bio, sectors))
1442 bio->bi_iter.bi_sector + sectors > conf->reshape_progress)
1443 : (bio->bi_iter.bi_sector + sectors > conf->reshape_safe &&
1478 max_sectors = r10_bio->sectors;
1540 if (max_sectors < r10_bio->sectors)
1541 r10_bio->sectors = max_sectors;
1543 if (r10_bio->sectors < bio_sectors(bio)) {
1544 struct bio *split = bio_split(bio, r10_bio->sectors,
1557 md_bitmap_startwrite(mddev->bitmap, r10_bio->sector, r10_bio->sectors, 0);
1568 static void __make_request(struct mddev *mddev, struct bio *bio, int sectors)
1576 r10_bio->sectors = sectors;
1764 r10_bio->sectors = 0;
1905 int sectors = bio_sectors(bio);
1923 sectors > chunk_sects
1927 sectors = chunk_sects -
1930 __make_request(mddev, bio, sectors);
2288 atomic_add(r10_bio->sectors,
2328 sector_t s = r10_bio->sectors;
2378 r10_bio->sectors,
2423 fbio->bi_iter.bi_size = r10_bio->sectors << 9;
2427 vcnt = (r10_bio->sectors + (PAGE_SIZE >> 9) - 1) >> (PAGE_SHIFT - 9);
2449 int sectors = r10_bio->sectors;
2452 if (sectors < (len / 512))
2453 len = sectors * 512;
2458 sectors -= len/512;
2462 atomic64_add(r10_bio->sectors, &mddev->resync_mismatches);
2519 md_done_sync(mddev, r10_bio->sectors, 1);
2526 * Recovery happens across physical sectors.
2547 int sectors = r10_bio->sectors;
2553 while (sectors) {
2554 int s = sectors;
2611 sectors -= s;
2695 int sectors, struct page *page, enum req_op op)
2700 if (is_badblock(rdev, sector, sectors, &first_bad, &bad_sectors)
2703 if (sync_page_io(rdev, sector, sectors << 9, page, op, false))
2713 if (!rdev_set_badblocks(rdev, sector, sectors, 0))
2729 int sectors = r10_bio->sectors, slot = r10_bio->read_slot;
2757 while(sectors) {
2758 int s = sectors;
2839 pr_notice("md/raid10:%s: read correction write failed (%d sectors at %llu on %pg)\n",
2873 pr_notice("md/raid10:%s: unable to read back corrected sectors (%d sectors at %llu on %pg)\n",
2884 pr_info("md/raid10:%s: read error corrected (%d sectors at %llu on %pg)\n",
2898 sectors -= s;
2922 int sectors;
2923 int sect_to_write = r10_bio->sectors;
2932 sectors = ((r10_bio->sector + block_sectors)
2939 if (sectors > sect_to_write)
2940 sectors = sect_to_write;
2941 /* Write at 'sector' for 'sectors' */
2944 bio_trim(wbio, sector - bio->bi_iter.bi_sector, sectors);
2953 sectors, 0)
2957 sect_to_write -= sectors;
2958 sector += sectors;
2959 sectors = block_sectors;
3025 r10_bio->sectors, 0);
3030 r10_bio->sectors, 0))
3042 r10_bio->sectors, 0);
3047 r10_bio->sectors, 0))
3062 r10_bio->sectors, 0);
3079 r10_bio->sectors, 0);
3737 r10_bio->sectors = (sector_nr | chunk_mask) - sector_nr + 1;
3845 r10_bio->sectors = nr_sectors;
3899 r10_bio->sectors = nr_sectors;
3930 raid10_size(struct mddev *mddev, sector_t sectors, int raid_disks)
3938 if (!sectors)
3939 sectors = conf->dev_sectors;
3941 size = sectors >> conf->geo.chunk_shift;
3951 /* Calculate the number of sectors-per-device that will
4345 static int raid10_resize(struct mddev *mddev, sector_t sectors)
4349 * number of sectors used to be an appropriate multiple
4369 size = raid10_size(mddev, sectors, 0);
4379 if (sectors > mddev->dev_sectors &&
4384 calc_sectors(conf, sectors);
4418 rdev->sectors = size;
4826 /* If restarting in the middle, skip the initial sectors */
4925 r10_bio->sectors = last - sector_nr + 1;
5024 r10_bio->sectors = nr_sectors;
5027 md_sync_acct_bio(read_bio, r10_bio->sectors);
5064 md_done_sync(mddev, r10_bio->sectors, 0);
5090 md_sync_acct_bio(b, r10_bio->sectors);
5133 int sectors = r10_bio->sectors;
5152 while (sectors) {
5153 int s = sectors;
5197 sectors -= s;
5235 md_done_sync(r10_bio->mddev, r10_bio->sectors, 1);