Home
last modified time | relevance | path

Searched refs:dev_sectors (Results 1 - 25 of 26) sorted by relevance

12

/kernel/linux/linux-5.10/drivers/md/
H A Ddm-raid.c247 sector_t dev_sectors; member
439 return rs->md.recovery_cp < rs->md.dev_sectors; in rs_is_recovering()
693 rdev->sectors = mddev->dev_sectors; in rs_set_rdev_sectors()
777 * rs->md.dev_sectors in raid_set_alloc()
1605 /* Check that calculated dev_sectors fits all component devices. */
1614 if (ds < rs->md.dev_sectors) { in _check_data_dev_sectors()
1628 sector_t array_sectors = sectors, dev_sectors = sectors; in rs_set_dev_and_array_sectors() local
1649 dev_sectors *= rs->raid10_copies; in rs_set_dev_and_array_sectors()
1650 if (sector_div(dev_sectors, data_stripes)) in rs_set_dev_and_array_sectors()
1653 array_sectors = (data_stripes + delta_disks) * dev_sectors; in rs_set_dev_and_array_sectors()
1675 rs_setup_recovery(struct raid_set *rs, sector_t dev_sectors) rs_setup_recovery() argument
[all...]
H A Draid10.h62 sector_t dev_sectors; /* temp copy of member
63 * mddev->dev_sectors */
H A Dmd.c1335 mddev->dev_sectors = ((sector_t)sb->size) * 2; in super_90_validate()
1476 sb->size = mddev->dev_sectors / 2; in super_90_sync()
1587 if (num_sectors && num_sectors < rdev->mddev->dev_sectors) in super_90_rdev_size_change()
1848 mddev->dev_sectors = le64_to_cpu(sb->size); in super_1_validate()
2057 sb->size = cpu_to_le64(mddev->dev_sectors); in super_1_sync()
2222 if (num_sectors && num_sectors < rdev->mddev->dev_sectors) in super_1_rdev_size_change()
2451 /* make sure rdev->sectors exceeds mddev->dev_sectors */ in bind_rdev_to_array()
2454 (mddev->dev_sectors == 0 || rdev->sectors < mddev->dev_sectors)) { in bind_rdev_to_array()
2463 mddev->dev_sectors in bind_rdev_to_array()
[all...]
H A Draid10.c2934 return mddev->dev_sectors - sector_nr; in raid10_sync_request()
2942 max_sector = mddev->dev_sectors; in raid10_sync_request()
3536 sectors = conf->dev_sectors; in raid10_size()
3549 * actually be used, and set conf->dev_sectors and in calc_sectors()
3566 conf->dev_sectors = size << conf->geo.chunk_shift; in calc_sectors()
3695 calc_sectors(conf, mddev->dev_sectors); in setup_conf()
3709 conf->prev.stride = conf->dev_sectors; in setup_conf()
3897 mddev->dev_sectors = conf->dev_sectors; in raid10_run()
3990 if (sectors > mddev->dev_sectors in raid10_resize()
[all...]
H A Dmd-faulty.c289 return mddev->dev_sectors; in faulty_size()
H A Draid1.c2639 max_sector = mddev->dev_sectors; in raid1_sync_request()
2921 return mddev->dev_sectors; in raid1_size()
3202 if (sectors > mddev->dev_sectors && in raid1_resize()
3203 mddev->recovery_cp > mddev->dev_sectors) { in raid1_resize()
3204 mddev->recovery_cp = mddev->dev_sectors; in raid1_resize()
3207 mddev->dev_sectors = sectors; in raid1_resize()
H A Dmd.h320 sector_t dev_sectors; /* used size of member
H A Dmd-multipath.c340 return mddev->dev_sectors; in multipath_size()
H A Draid5.c6028 BUG_ON((mddev->dev_sectors & in reshape_request()
6148 if (last_sector >= mddev->dev_sectors) in reshape_request()
6149 last_sector = mddev->dev_sectors - 1; in reshape_request()
6213 sector_t max_sector = mddev->dev_sectors; in raid5_sync_request()
6254 sector_t rv = mddev->dev_sectors - sector_nr; in raid5_sync_request()
7013 sectors = mddev->dev_sectors; in raid5_size()
7663 mddev->dev_sectors &= ~(mddev->chunk_sectors - 1); in raid5_run()
7664 mddev->resync_max_sectors = mddev->dev_sectors; in raid5_run()
8068 if (sectors > mddev->dev_sectors && in raid5_resize()
8069 mddev->recovery_cp > mddev->dev_sectors) { in raid5_resize()
[all...]
H A Draid0.c664 rdev->sectors = mddev->dev_sectors; in raid0_takeover_raid45()
H A Dmd-bitmap.c237 < (rdev->data_offset + mddev->dev_sectors in write_sb_page()
247 if (rdev->data_offset + mddev->dev_sectors in write_sb_page()
/kernel/linux/linux-6.6/drivers/md/
H A Ddm-raid.c248 sector_t dev_sectors; member
440 return rs->md.recovery_cp < rs->md.dev_sectors; in rs_is_recovering()
694 rdev->sectors = mddev->dev_sectors; in rs_set_rdev_sectors()
777 * rs->md.dev_sectors in raid_set_alloc()
1605 /* Check that calculated dev_sectors fits all component devices. */
1614 if (ds < rs->md.dev_sectors) { in _check_data_dev_sectors()
1628 sector_t array_sectors = sectors, dev_sectors = sectors; in rs_set_dev_and_array_sectors() local
1649 dev_sectors *= rs->raid10_copies; in rs_set_dev_and_array_sectors()
1650 if (sector_div(dev_sectors, data_stripes)) in rs_set_dev_and_array_sectors()
1653 array_sectors = (data_stripes + delta_disks) * dev_sectors; in rs_set_dev_and_array_sectors()
1675 rs_setup_recovery(struct raid_set *rs, sector_t dev_sectors) rs_setup_recovery() argument
[all...]
H A Draid10.h62 sector_t dev_sectors; /* temp copy of member
63 * mddev->dev_sectors */
H A Dmd.c1294 mddev->dev_sectors = ((sector_t)sb->size) * 2; in super_90_validate()
1435 sb->size = mddev->dev_sectors / 2; in super_90_sync()
1546 if (num_sectors && num_sectors < rdev->mddev->dev_sectors) in super_90_rdev_size_change()
1804 mddev->dev_sectors = le64_to_cpu(sb->size); in super_1_validate()
2013 sb->size = cpu_to_le64(mddev->dev_sectors); in super_1_sync()
2178 if (num_sectors && num_sectors < rdev->mddev->dev_sectors) in super_1_rdev_size_change()
2417 /* make sure rdev->sectors exceeds mddev->dev_sectors */ in bind_rdev_to_array()
2420 (mddev->dev_sectors == 0 || rdev->sectors < mddev->dev_sectors)) { in bind_rdev_to_array()
2429 mddev->dev_sectors in bind_rdev_to_array()
[all...]
H A Draid10.c3321 return mddev->dev_sectors - sector_nr; in raid10_sync_request()
3329 max_sector = mddev->dev_sectors; in raid10_sync_request()
3939 sectors = conf->dev_sectors; in raid10_size()
3952 * actually be used, and set conf->dev_sectors and in calc_sectors()
3969 conf->dev_sectors = size << conf->geo.chunk_shift; in calc_sectors()
4098 calc_sectors(conf, mddev->dev_sectors); in setup_conf()
4112 conf->prev.stride = conf->dev_sectors; in setup_conf()
4286 mddev->dev_sectors = conf->dev_sectors; in raid10_run()
4379 if (sectors > mddev->dev_sectors in raid10_resize()
[all...]
H A Dmd-faulty.c291 return mddev->dev_sectors; in faulty_size()
H A Draid1.c2683 max_sector = mddev->dev_sectors; in raid1_sync_request()
2965 return mddev->dev_sectors; in raid1_size()
3232 if (sectors > mddev->dev_sectors && in raid1_resize()
3233 mddev->recovery_cp > mddev->dev_sectors) { in raid1_resize()
3234 mddev->recovery_cp = mddev->dev_sectors; in raid1_resize()
3237 mddev->dev_sectors = sectors; in raid1_resize()
H A Dmd.h346 sector_t dev_sectors; /* used size of member
H A Dmd-multipath.c334 return mddev->dev_sectors; in multipath_size()
H A Draid5.c6350 BUG_ON((mddev->dev_sectors & in reshape_request()
6471 if (last_sector >= mddev->dev_sectors) in reshape_request()
6472 last_sector = mddev->dev_sectors - 1; in reshape_request()
6537 sector_t max_sector = mddev->dev_sectors; in raid5_sync_request()
6578 sector_t rv = mddev->dev_sectors - sector_nr; in raid5_sync_request()
7348 sectors = mddev->dev_sectors; in raid5_size()
8005 mddev->dev_sectors &= ~((sector_t)mddev->chunk_sectors - 1); in raid5_run()
8006 mddev->resync_max_sectors = mddev->dev_sectors; in raid5_run()
8422 if (sectors > mddev->dev_sectors && in raid5_resize()
8423 mddev->recovery_cp > mddev->dev_sectors) { in raid5_resize()
[all...]
H A Draid0.c659 rdev->sectors = mddev->dev_sectors; in raid0_takeover_raid45()
/kernel/linux/linux-6.6/include/uapi/linux/
H A Dublk_cmd.h345 __u64 dev_sectors; member
/kernel/linux/linux-5.10/drivers/ata/
H A Dlibata-scsi.c1388 u64 dev_sectors = qc->dev->n_sectors; in ata_scsi_verify_xlat() local
1416 if (block >= dev_sectors) in ata_scsi_verify_xlat()
1418 if ((block + n_block) > dev_sectors) in ata_scsi_verify_xlat()
/kernel/linux/linux-6.6/drivers/ata/
H A Dlibata-scsi.c1406 u64 dev_sectors = qc->dev->n_sectors; in ata_scsi_verify_xlat() local
1437 if (block >= dev_sectors) in ata_scsi_verify_xlat()
1439 if ((block + n_block) > dev_sectors) in ata_scsi_verify_xlat()
/kernel/linux/linux-6.6/drivers/block/
H A Dublk_drv.c223 return p->dev_sectors >> ilog2(p->chunk_sectors); in ublk_get_nr_zones()
523 set_capacity(ub->ub_disk, p->dev_sectors); in ublk_dev_param_basic_apply()

Completed in 110 milliseconds

12