Lines Matching refs:bdev
26 int __blkdev_issue_discard(struct block_device *bdev, sector_t sector, in __blkdev_issue_discard() argument
30 struct request_queue *q = bdev_get_queue(bdev); in __blkdev_issue_discard()
38 if (bdev_read_only(bdev)) in __blkdev_issue_discard()
55 bdevname(bdev, dev_name); in __blkdev_issue_discard()
60 bs_mask = (bdev_logical_block_size(bdev) >> 9) - 1; in __blkdev_issue_discard()
68 if (bdev_is_partition(bdev)) in __blkdev_issue_discard()
69 part_offset = bdev->bd_start_sect; in __blkdev_issue_discard()
100 bio_set_dev(bio, bdev); in __blkdev_issue_discard()
132 int blkdev_issue_discard(struct block_device *bdev, sector_t sector, in blkdev_issue_discard() argument
140 ret = __blkdev_issue_discard(bdev, sector, nr_sects, gfp_mask, flags, in blkdev_issue_discard()
166 static int __blkdev_issue_write_same(struct block_device *bdev, sector_t sector, in __blkdev_issue_write_same() argument
170 struct request_queue *q = bdev_get_queue(bdev); in __blkdev_issue_write_same()
178 if (bdev_read_only(bdev)) in __blkdev_issue_write_same()
181 bs_mask = (bdev_logical_block_size(bdev) >> 9) - 1; in __blkdev_issue_write_same()
185 if (!bdev_write_same(bdev)) in __blkdev_issue_write_same()
194 bio_set_dev(bio, bdev); in __blkdev_issue_write_same()
198 bio->bi_io_vec->bv_len = bdev_logical_block_size(bdev); in __blkdev_issue_write_same()
227 int blkdev_issue_write_same(struct block_device *bdev, sector_t sector, in blkdev_issue_write_same() argument
236 ret = __blkdev_issue_write_same(bdev, sector, nr_sects, gfp_mask, page, in blkdev_issue_write_same()
247 static int __blkdev_issue_write_zeroes(struct block_device *bdev, in __blkdev_issue_write_zeroes() argument
253 struct request_queue *q = bdev_get_queue(bdev); in __blkdev_issue_write_zeroes()
258 if (bdev_read_only(bdev)) in __blkdev_issue_write_zeroes()
262 max_write_zeroes_sectors = bdev_write_zeroes_sectors(bdev); in __blkdev_issue_write_zeroes()
270 bio_set_dev(bio, bdev); in __blkdev_issue_write_zeroes()
303 static int __blkdev_issue_zero_pages(struct block_device *bdev, in __blkdev_issue_zero_pages() argument
307 struct request_queue *q = bdev_get_queue(bdev); in __blkdev_issue_zero_pages()
315 if (bdev_read_only(bdev)) in __blkdev_issue_zero_pages()
322 bio_set_dev(bio, bdev); in __blkdev_issue_zero_pages()
359 int __blkdev_issue_zeroout(struct block_device *bdev, sector_t sector, in __blkdev_issue_zeroout() argument
366 bs_mask = (bdev_logical_block_size(bdev) >> 9) - 1; in __blkdev_issue_zeroout()
370 ret = __blkdev_issue_write_zeroes(bdev, sector, nr_sects, gfp_mask, in __blkdev_issue_zeroout()
375 return __blkdev_issue_zero_pages(bdev, sector, nr_sects, gfp_mask, in __blkdev_issue_zeroout()
393 int blkdev_issue_zeroout(struct block_device *bdev, sector_t sector, in blkdev_issue_zeroout() argument
400 bool try_write_zeroes = !!bdev_write_zeroes_sectors(bdev); in blkdev_issue_zeroout()
402 bs_mask = (bdev_logical_block_size(bdev) >> 9) - 1; in blkdev_issue_zeroout()
410 ret = __blkdev_issue_write_zeroes(bdev, sector, nr_sects, in blkdev_issue_zeroout()
413 ret = __blkdev_issue_zero_pages(bdev, sector, nr_sects, in blkdev_issue_zeroout()
429 if (!bdev_write_zeroes_sectors(bdev)) { in blkdev_issue_zeroout()