Lines Matching defs:bdev
521 struct block_device *bdev)
523 int ret = blkdev_issue_flush(bdev);
525 trace_f2fs_issue_flush(bdev, test_opt(sbi, NOBARRIER),
543 ret = __submit_flush_wait(sbi, FDEV(i).bdev);
724 ret = __submit_flush_wait(sbi, FDEV(i).bdev);
932 struct block_device *bdev, block_t lstart,
945 dc->bdev = bdev;
1089 trace_f2fs_remove_discard(dc->bdev, dc->di.start, dc->di.len);
1200 struct block_device *bdev, block_t lstart,
1210 struct block_device *bdev = dc->bdev;
1211 struct bio *bio = bio_alloc(bdev, 0, REQ_OP_ZONE_RESET | flag, GFP_NOFS);
1214 trace_f2fs_issue_reset_zone(bdev, dc->di.start);
1246 struct block_device *bdev = dc->bdev;
1248 SECTOR_TO_BLOCK(bdev_max_discard_sectors(bdev));
1263 if (f2fs_sb_has_blkzoned(sbi) && bdev_is_zoned(bdev)) {
1264 int devi = f2fs_bdev_index(sbi, bdev);
1277 trace_f2fs_issue_discard(bdev, dc->di.start, dc->di.len);
1305 err = __blkdev_issue_discard(bdev,
1357 __update_discard_tree_range(sbi, bdev, lstart, start, len);
1363 struct block_device *bdev, block_t lstart,
1387 dc = __create_discard_cmd(sbi, bdev, lstart, start, len);
1422 __insert_discard_cmd(sbi, dc->bdev, blkaddr + 1,
1436 struct block_device *bdev, block_t lstart,
1445 SECTOR_TO_BLOCK(bdev_max_discard_sectors(bdev));
1483 prev_dc->bdev == bdev &&
1495 next_dc->bdev == bdev &&
1509 __insert_discard_cmd(sbi, bdev,
1523 struct block_device *bdev, block_t blkstart, block_t lblkstart,
1526 trace_f2fs_queue_reset_zone(bdev, blkstart);
1529 __insert_discard_cmd(sbi, bdev, lblkstart, blkstart, blklen);
1535 struct block_device *bdev, block_t blkstart, block_t blklen)
1539 if (!f2fs_bdev_support_discard(bdev))
1542 trace_f2fs_queue_discard(bdev, blkstart, blklen);
1550 __update_discard_tree_range(sbi, bdev, lblkstart, blkstart, blklen);
1798 if (dc && f2fs_sb_has_blkzoned(sbi) && bdev_is_zoned(dc->bdev)) {
1799 int devi = f2fs_bdev_index(sbi, dc->bdev);
1942 struct block_device *bdev, block_t blkstart, block_t blklen)
1963 div64_u64_rem(sector, bdev_zone_sectors(bdev), &remainder);
1965 if (remainder || nr_sects != bdev_zone_sectors(bdev)) {
1973 trace_f2fs_issue_reset_zone(bdev, blkstart);
1974 return blkdev_zone_mgmt(bdev, REQ_OP_ZONE_RESET,
1978 __queue_zone_reset_cmd(sbi, bdev, blkstart, lblkstart, blklen);
1983 __queue_discard_cmd(sbi, bdev, lblkstart, blklen);
1989 struct block_device *bdev, block_t blkstart, block_t blklen)
1992 if (f2fs_sb_has_blkzoned(sbi) && bdev_is_zoned(bdev))
1993 return __f2fs_issue_discard_zone(sbi, bdev, blkstart, blklen);
1995 __queue_discard_cmd(sbi, bdev, blkstart, blklen);
2003 struct block_device *bdev;
2009 bdev = f2fs_target_device(sbi, blkstart, NULL);
2016 if (bdev2 != bdev) {
2017 err = __issue_discard_async(sbi, bdev,
2021 bdev = bdev2;
2036 err = __issue_discard_async(sbi, bdev, start, len);
3562 /* writeout dirty page into bdev */
4922 ret = __f2fs_issue_discard_zone(sbi, fdev->bdev, zone_block,
4944 ret = blkdev_zone_mgmt(fdev->bdev, REQ_OP_ZONE_FINISH,
4947 ret = blkdev_issue_zeroout(fdev->bdev, zone->wp,
4967 if (!bdev_is_zoned(FDEV(i).bdev))
5005 err = blkdev_report_zones(zbd->bdev, zone_sector, 1,
5048 err = blkdev_report_zones(zbd->bdev, zone_sector, 1,
5064 err = __f2fs_issue_discard_zone(sbi, zbd->bdev, cs_zone_block,
5110 if (!bdev_is_zoned(FDEV(i).bdev))
5115 ret = blkdev_report_zones(FDEV(i).bdev, 0, BLK_ALL_ZONES,