Lines Matching refs:sector_nr
71 static sector_t reshape_request(struct mddev *mddev, sector_t sector_nr,
3293 static sector_t raid10_sync_request(struct mddev *mddev, sector_t sector_nr,
3321 return mddev->dev_sectors - sector_nr;
3333 if (sector_nr >= max_sector) {
3388 return reshape_request(mddev, sector_nr, skipped);
3408 return (max_sector - sector_nr) + sectors_skipped;
3418 max_sector > (sector_nr | chunk_mask))
3419 max_sector = (sector_nr | chunk_mask) + 1;
3476 sect = raid10_find_virt(conf, sector_nr, i);
3707 * Let's check against "sector_nr + 2 * RESYNC_SECTORS" for
3711 md_bitmap_cond_end_sync(mddev->bitmap, sector_nr,
3713 (sector_nr + 2 * RESYNC_SECTORS > conf->cluster_sync_high));
3715 if (!md_bitmap_start_sync(mddev->bitmap, sector_nr,
3731 conf->next_resync = sector_nr;
3734 r10_bio->sector = sector_nr;
3737 r10_bio->sectors = (sector_nr | chunk_mask) - sector_nr + 1;
3824 if (sector_nr + max_sync < max_sector)
3825 max_sector = sector_nr + max_sync;
3829 if (sector_nr + (len>>9) > max_sector)
3830 len = (max_sector - sector_nr) << 9;
3843 sector_nr += len>>9;
3850 if (conf->cluster_sync_high < sector_nr + nr_sectors) {
3865 * sector_nr is a device address for recovery, so we
3869 sect_va1 = raid10_find_virt(conf, sector_nr, i);
3875 * sector_nr, so make the translation too.
3920 if (sector_nr + max_sync < max_sector)
3921 max_sector = sector_nr + max_sync;
3923 sectors_skipped += (max_sector - sector_nr);
3925 sector_nr = max_sector;
4772 static sector_t reshape_request(struct mddev *mddev, sector_t sector_nr,
4791 * We interpret 'sector_nr' as an address that we want to write to.
4825 if (sector_nr == 0) {
4829 sector_nr = (raid10_size(mddev, 0, 0)
4833 sector_nr = conf->reshape_progress;
4834 if (sector_nr) {
4835 mddev->curr_resync_completed = sector_nr;
4838 return sector_nr;
4842 /* We don't use sector_nr to track where we are up to
4863 sector_nr = last & ~(sector_t)(conf->geo.chunk_mask
4865 if (sector_nr + RESYNC_SECTORS < last)
4866 sector_nr = last + 1 - RESYNC_SECTORS;
4884 sector_nr = conf->reshape_progress;
4885 last = sector_nr | (conf->geo.chunk_mask
4888 if (sector_nr + RESYNC_SECTORS <= last)
4889 last = sector_nr + RESYNC_SECTORS - 1;
4917 /* Now schedule reads for blocks from sector_nr to last */
4923 r10_bio->sector = sector_nr;
4925 r10_bio->sectors = last - sector_nr + 1;
4952 if (mddev_is_clustered(mddev) && conf->cluster_sync_high <= sector_nr) {
4956 conf->cluster_sync_low = sector_nr;
4957 conf->cluster_sync_high = sector_nr + CLUSTER_RESYNC_WINDOW_SECTORS;
5020 sector_nr += len >> 9;
5032 if (sector_nr <= last)