Lines Matching defs:bdev

200 		rdev->bdev->bd_disk->queue->nr_hw_queues != 1 &&
581 bio_set_dev(bi, rdev->bdev);
879 if (rdev->bdev->bd_dev == dev)
890 if (rdev->bdev->bd_dev == dev)
912 sector_t num_sectors = i_size_read(rdev->bdev->bd_inode) / 512;
988 bio_set_dev(bio, rdev->meta_bdev ? rdev->meta_bdev : rdev->bdev);
1022 bio_set_dev(bio, rdev->bdev);
1056 bdevname(rdev->bdev,b));
1222 bdevname(rdev->bdev, b);
1275 b, bdevname(refdev->bdev,b2));
1280 b, bdevname(refdev->bdev, b2));
1536 d->major = MAJOR(rdev2->bdev->bd_dev);
1537 d->minor = MINOR(rdev2->bdev->bd_dev);
1659 sb_start = i_size_read(rdev->bdev->bd_inode) >> 9;
1691 bdevname(rdev->bdev,b));
1696 bdevname(rdev->bdev,b));
1714 bmask = queue_logical_block_size(rdev->bdev->bd_disk->queue)-1;
1803 bdevname(rdev->bdev,b),
1804 bdevname(refdev->bdev,b2));
1816 sectors = (i_size_read(rdev->bdev->bd_inode) >> 9);
1972 mdname(mddev), rdev->bdev, rdev->desc_nr,
1973 freshest->bdev, freshest_max_dev);
1979 mdname(mddev), rdev->bdev, role, role, freshest->bdev);
2160 bmask = queue_logical_block_size(rdev->bdev->bd_disk->queue)-1;
2228 max_sectors = i_size_read(rdev->bdev->bd_inode) >> 9;
2238 sector_t dev_size = i_size_read(rdev->bdev->bd_inode) >> 9;
2348 if (rdev->bdev->bd_disk == rdev2->bdev->bd_disk) {
2387 if (blk_integrity_compare(reference->bdev->bd_disk,
2388 rdev->bdev->bd_disk) < 0)
2391 if (!reference || !bdev_get_integrity(reference->bdev))
2398 bdev_get_integrity(reference->bdev));
2427 if (blk_integrity_compare(mddev->gendisk, rdev->bdev->bd_disk) != 0) {
2429 mdname(mddev), bdevname(rdev->bdev, name));
2444 if (find_rdev(mddev, rdev->bdev->bd_dev))
2447 if ((bdev_read_only(rdev->bdev) || bdev_read_only(rdev->meta_bdev)) &&
2491 bdevname(rdev->bdev,b);
2503 ko = &part_to_dev(rdev->bdev->bd_part)->kobj;
2513 bd_link_disk_holder(rdev->bdev, mddev->gendisk);
2538 bd_unlink_disk_holder(rdev->bdev, rdev->mddev->gendisk);
2540 pr_debug("md: unbind<%s>\n", bdevname(rdev->bdev,b));
2569 struct block_device *bdev;
2571 bdev = blkdev_get_by_dev(dev, FMODE_READ|FMODE_WRITE|FMODE_EXCL,
2573 if (IS_ERR(bdev)) {
2576 return PTR_ERR(bdev);
2578 rdev->bdev = bdev;
2584 struct block_device *bdev = rdev->bdev;
2585 rdev->bdev = NULL;
2586 blkdev_put(bdev, FMODE_READ|FMODE_WRITE|FMODE_EXCL);
2595 pr_debug("md: export_rdev(%s)\n", bdevname(rdev->bdev,b));
2599 md_autodetect_dev(rdev->bdev->bd_dev);
2863 bdevname(rdev->bdev, b),
2876 bdevname(rdev->bdev, b));
3447 sectors = (i_size_read(rdev->bdev->bd_inode) >> 9) -
3459 * ->bdev do not overlap. 'rcu' is sufficient to walk
3473 if (rdev->bdev == rdev2->bdev &&
3774 size = i_size_read(rdev->bdev->bd_inode) >> BLOCK_SIZE_BITS;
3777 bdevname(rdev->bdev,b));
3787 bdevname(rdev->bdev,b),
3793 bdevname(rdev->bdev,b));
3801 if (rdev->bdev)
3829 bdevname(rdev->bdev,b));
3848 mdname(mddev), bdevname(rdev->bdev, b),
3857 bdevname(rdev->bdev,b));
4467 static int do_md_stop(struct mddev *mddev, int ro, struct block_device *bdev);
4468 static int md_set_readonly(struct mddev *mddev, struct block_device *bdev);
5899 sync_blockdev(rdev->bdev);
5900 invalidate_bdev(rdev->bdev);
5902 (bdev_read_only(rdev->bdev) ||
5987 rdev->bdev->bd_disk ==
5988 rdev2->bdev->bd_disk) {
5991 bdevname(rdev->bdev,b),
5992 bdevname(rdev2->bdev,b2));
6062 !blk_queue_nonrot(bdev_get_queue(rdev->bdev))) {
6199 if (bdev_read_only(rdev->bdev))
6355 static int md_set_readonly(struct mddev *mddev, struct block_device *bdev)
6382 if ((mddev->pers && atomic_read(&mddev->openers) > !!bdev) ||
6418 struct block_device *bdev)
6443 if ((mddev->pers && atomic_read(&mddev->openers) > !!bdev) ||
6518 pr_cont("<%s>", bdevname(rdev->bdev,b));
6555 pr_debug("md: considering %s ...\n", bdevname(rdev0->bdev,b));
6560 bdevname(rdev->bdev,b));
6578 bdevname(rdev0->bdev, b), rdev0->preferred_minor);
6592 mdname(mddev), bdevname(rdev0->bdev,b));
6735 info.major = MAJOR(rdev->bdev->bd_dev);
6736 info.minor = MINOR(rdev->bdev->bd_dev);
6797 bdevname(rdev->bdev,b),
6798 bdevname(rdev0->bdev,b2));
6956 rdev->sb_start = i_size_read(rdev->bdev->bd_inode) / 512;
7009 bdevname(rdev->bdev,b), mdname(mddev));
7043 rdev->sb_start = i_size_read(rdev->bdev->bd_inode) / 512;
7049 bdevname(rdev->bdev,b), mdname(mddev));
7514 static int md_getgeo(struct block_device *bdev, struct hd_geometry *geo)
7516 struct mddev *mddev = bdev->bd_disk->private_data;
7549 static int md_ioctl(struct block_device *bdev, fmode_t mode,
7586 mddev = bdev->bd_disk->private_data;
7645 sync_blockdev(bdev);
7710 err = do_md_stop(mddev, 0, bdev);
7714 err = md_set_readonly(mddev, bdev);
7746 /* if the bdev is going readonly the value of mddev->ro
7841 static int md_compat_ioctl(struct block_device *bdev, fmode_t mode,
7856 return md_ioctl(bdev, mode, cmd, arg);
7860 static int md_open(struct block_device *bdev, fmode_t mode)
7866 struct mddev *mddev = mddev_find(bdev->bd_dev);
7872 if (mddev->gendisk != bdev->bd_disk) {
7877 /* Wait until bdev->bd_disk is definitely gone */
7882 BUG_ON(mddev != bdev->bd_disk->private_data);
7897 bdev_check_media_change(bdev);
8073 bdevname(rdev->bdev,b));
8315 bdevname(rdev->bdev,b), rdev->desc_nr);
8503 struct gendisk *disk = rdev->bdev->bd_disk;
9666 pr_info("md: Removing Candidate device %s because add failed\n", bdevname(rdev2->bdev,b));
9684 bdevname(rdev2->bdev,b));