1 /* SPDX-License-Identifier: GPL-2.0-or-later */ 2 /* 3 * RDMA Network Block Driver 4 * 5 * Copyright (c) 2014 - 2018 ProfitBricks GmbH. All rights reserved. 6 * Copyright (c) 2018 - 2019 1&1 IONOS Cloud GmbH. All rights reserved. 7 * Copyright (c) 2019 - 2020 1&1 IONOS SE. All rights reserved. 8 */ 9 #ifndef RNBD_SRV_DEV_H 10 #define RNBD_SRV_DEV_H 11 12 #include <linux/fs.h> 13 #include "rnbd-proto.h" 14 15 struct rnbd_dev { 16 struct block_device *bdev; 17 struct bio_set *ibd_bio_set; 18 fmode_t blk_open_flags; 19 char name[BDEVNAME_SIZE]; 20 }; 21 22 struct rnbd_dev_blk_io { 23 struct rnbd_dev *dev; 24 void *priv; 25 /* have to be last member for front_pad usage of bioset_init */ 26 struct bio bio; 27 }; 28 29 /** 30 * rnbd_dev_open() - Open a device 31 * @flags: open flags 32 * @bs: bio_set to use during block io, 33 */ 34 struct rnbd_dev *rnbd_dev_open(const char *path, fmode_t flags, 35 struct bio_set *bs); 36 37 /** 38 * rnbd_dev_close() - Close a device 39 */ 40 void rnbd_dev_close(struct rnbd_dev *dev); 41 42 void rnbd_endio(void *priv, int error); 43 44 void rnbd_dev_bi_end_io(struct bio *bio); 45 46 struct bio *rnbd_bio_map_kern(void *data, struct bio_set *bs, 47 unsigned int len, gfp_t gfp_mask); 48 rnbd_dev_get_max_segs(const struct rnbd_dev * dev)49static inline int rnbd_dev_get_max_segs(const struct rnbd_dev *dev) 50 { 51 return queue_max_segments(bdev_get_queue(dev->bdev)); 52 } 53 rnbd_dev_get_max_hw_sects(const struct rnbd_dev * dev)54static inline int rnbd_dev_get_max_hw_sects(const struct rnbd_dev *dev) 55 { 56 return queue_max_hw_sectors(bdev_get_queue(dev->bdev)); 57 } 58 rnbd_dev_get_secure_discard(const struct rnbd_dev * dev)59static inline int rnbd_dev_get_secure_discard(const struct rnbd_dev *dev) 60 { 61 return blk_queue_secure_erase(bdev_get_queue(dev->bdev)); 62 } 63 rnbd_dev_get_max_discard_sects(const struct rnbd_dev * dev)64static inline int rnbd_dev_get_max_discard_sects(const struct rnbd_dev *dev) 65 { 66 if (!blk_queue_discard(bdev_get_queue(dev->bdev))) 67 return 0; 68 69 return blk_queue_get_max_sectors(bdev_get_queue(dev->bdev), 70 REQ_OP_DISCARD); 71 } 72 rnbd_dev_get_discard_granularity(const struct rnbd_dev * dev)73static inline int rnbd_dev_get_discard_granularity(const struct rnbd_dev *dev) 74 { 75 return bdev_get_queue(dev->bdev)->limits.discard_granularity; 76 } 77 rnbd_dev_get_discard_alignment(const struct rnbd_dev * dev)78static inline int rnbd_dev_get_discard_alignment(const struct rnbd_dev *dev) 79 { 80 return bdev_get_queue(dev->bdev)->limits.discard_alignment; 81 } 82 83 #endif /* RNBD_SRV_DEV_H */ 84