Home
last modified time | relevance | path

Searched refs:ibmr (Results 1 – 25 of 54) sorted by relevance

123

/linux/net/rds/
A Dib_frmr.c67 if (ibmr) in rds_ib_alloc_frmr()
68 return ibmr; in rds_ib_alloc_frmr()
70 ibmr = kzalloc_node(sizeof(*ibmr), GFP_KERNEL, in rds_ib_alloc_frmr()
72 if (!ibmr) { in rds_ib_alloc_frmr()
98 return ibmr; in rds_ib_alloc_frmr()
101 kfree(ibmr); in rds_ib_alloc_frmr()
203 ibmr->sg = sg; in rds_ib_map_frmr()
208 ibmr->sg_dma_len = ib_dma_map_sg(dev, ibmr->sg, ibmr->sg_len, in rds_ib_map_frmr()
260 ib_dma_unmap_sg(rds_ibdev->dev, ibmr->sg, ibmr->sg_len, in rds_ib_map_frmr()
416 if (ibmr) in rds_ib_reg_frmr()
[all …]
A Dib_rdma.c215 return ibmr; in rds_ib_reuse_mr()
223 if (ibmr->odp) in rds_ib_sync_mr()
244 ibmr->sg, ibmr->sg_len, in __rds_ib_teardown_mr()
361 if (ibmr) { in rds_ib_flush_mr_pool()
373 if (ibmr) { in rds_ib_flush_mr_pool()
385 if (ibmr) { in rds_ib_flush_mr_pool()
447 if (ibmr) in rds_ib_try_reuse_ibmr()
470 if (ibmr) in rds_ib_try_reuse_ibmr()
592 ibmr = kzalloc(sizeof(*ibmr), GFP_KERNEL); in rds_ib_get_mr()
593 if (!ibmr) { in rds_ib_get_mr()
[all …]
/linux/drivers/infiniband/hw/vmw_pvrdma/
A Dpvrdma_mr.c94 mr->ibmr.lkey = resp->lkey; in pvrdma_get_dma_mr()
95 mr->ibmr.rkey = resp->rkey; in pvrdma_get_dma_mr()
97 return &mr->ibmr; in pvrdma_get_dma_mr()
182 mr->ibmr.lkey = resp->lkey; in pvrdma_reg_user_mr()
183 mr->ibmr.rkey = resp->rkey; in pvrdma_reg_user_mr()
185 return &mr->ibmr; in pvrdma_reg_user_mr()
254 mr->ibmr.lkey = resp->lkey; in pvrdma_alloc_mr()
255 mr->ibmr.rkey = resp->rkey; in pvrdma_alloc_mr()
259 return &mr->ibmr; in pvrdma_alloc_mr()
279 struct pvrdma_user_mr *mr = to_vmr(ibmr); in pvrdma_dereg_mr()
[all …]
A Dpvrdma.h141 struct ib_mr ibmr; member
284 static inline struct pvrdma_user_mr *to_vmr(struct ib_mr *ibmr) in to_vmr() argument
286 return container_of(ibmr, struct pvrdma_user_mr, ibmr); in to_vmr()
/linux/drivers/infiniband/hw/mlx5/
A Dmr.c825 return &mr->ibmr; in mlx5_ib_get_dma_mr()
973 mr->ibmr.pd = pd; in alloc_cacheable_mr()
1287 mr->ibmr.pd = pd; in reg_create()
1387 return &mr->ibmr; in mlx5_ib_get_dm_mr()
1486 return &mr->ibmr; in create_real_mr()
1540 return &mr->ibmr; in create_user_odp_mr()
1637 return &mr->ibmr; in mlx5_ib_reg_user_mr_dmabuf()
2065 mr->ibmr.pd = pd; in mlx5_ib_alloc_pi_mr()
2530 pi_mr->ibmr.page_size = ibmr->page_size; in mlx5_ib_map_mtt_mr_sg_pi()
2606 ibmr->length = pi_mr->ibmr.length; in mlx5_ib_map_klm_mr_sg_pi()
[all …]
A Drestrack.c81 static int fill_stat_mr_entry(struct sk_buff *msg, struct ib_mr *ibmr) in fill_stat_mr_entry() argument
83 struct mlx5_ib_mr *mr = to_mmr(ibmr); in fill_stat_mr_entry()
115 static int fill_res_mr_entry_raw(struct sk_buff *msg, struct ib_mr *ibmr) in fill_res_mr_entry_raw() argument
117 struct mlx5_ib_mr *mr = to_mmr(ibmr); in fill_res_mr_entry_raw()
123 static int fill_res_mr_entry(struct sk_buff *msg, struct ib_mr *ibmr) in fill_res_mr_entry() argument
125 struct mlx5_ib_mr *mr = to_mmr(ibmr); in fill_res_mr_entry()
A Dodp.c205 mlx5_ib_dereg_mr(&mr->ibmr, NULL); in free_implicit_child_mr_work()
428 mr->ibmr.pd = imr->ibmr.pd; in implicit_get_child_mr()
431 mr->ibmr.lkey = mr->mmkey.key; in implicit_get_child_mr()
432 mr->ibmr.rkey = mr->mmkey.key; in implicit_get_child_mr()
476 mlx5_ib_dereg_mr(&mr->ibmr, NULL); in implicit_get_child_mr()
502 imr->ibmr.pd = &pd->ibpd; in mlx5_ib_alloc_implicit_mr()
503 imr->ibmr.iova = 0; in mlx5_ib_alloc_implicit_mr()
505 imr->ibmr.lkey = imr->mmkey.key; in mlx5_ib_alloc_implicit_mr()
506 imr->ibmr.rkey = imr->mmkey.key; in mlx5_ib_alloc_implicit_mr()
507 imr->ibmr.device = &dev->ib_dev; in mlx5_ib_alloc_implicit_mr()
[all …]
A Dwr.c390 seg->start_addr = cpu_to_be64(mr->ibmr.iova); in set_reg_mkey_seg()
391 seg->len = cpu_to_be64(mr->ibmr.length); in set_reg_mkey_seg()
657 data_key = pi_mr->ibmr.lkey; in set_sig_data_segment()
661 prot_key = pi_mr->ibmr.lkey; in set_sig_data_segment()
797 region_len = pi_mr->ibmr.length; in set_pi_umr_wr()
1091 reg_pi_wr.mr = &pi_mr->ibmr; in handle_reg_mr_integrity()
1093 reg_pi_wr.key = pi_mr->ibmr.rkey; in handle_reg_mr_integrity()
1113 pa_pi_mr.ibmr.lkey = mr->ibmr.pd->local_dma_lkey; in handle_reg_mr_integrity()
1123 pa_pi_mr.ibmr.length = mr->ibmr.length; in handle_reg_mr_integrity()
1126 (*ctrl)->imm = cpu_to_be32(mr->ibmr.rkey); in handle_reg_mr_integrity()
[all …]
A Dmlx5_ib.h662 struct ib_mr ibmr; member
1153 return to_mdev(mr->ibmr.device); in mr_to_mdev()
1209 static inline struct mlx5_ib_mr *to_mmr(struct ib_mr *ibmr) in to_mmr() argument
1211 return container_of(ibmr, struct mlx5_ib_mr, ibmr); in to_mmr()
1303 int mlx5_ib_dereg_mr(struct ib_mr *ibmr, struct ib_udata *udata);
1309 int mlx5_ib_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, int sg_nents,
1311 int mlx5_ib_map_mr_sg_pi(struct ib_mr *ibmr, struct scatterlist *data_sg,
1348 int mlx5_ib_check_mr_status(struct ib_mr *ibmr, u32 check_mask,
/linux/drivers/infiniband/hw/mlx4/
A Dmr.c76 mr->ibmr.rkey = mr->ibmr.lkey = mr->mmr.key; in mlx4_ib_get_dma_mr()
79 return &mr->ibmr; in mlx4_ib_get_dma_mr()
441 mr->ibmr.rkey = mr->ibmr.lkey = mr->mmr.key; in mlx4_ib_reg_user_mr()
442 mr->ibmr.length = length; in mlx4_ib_reg_user_mr()
443 mr->ibmr.page_size = 1U << shift; in mlx4_ib_reg_user_mr()
445 return &mr->ibmr; in mlx4_ib_reg_user_mr()
599 struct mlx4_ib_mr *mr = to_mmr(ibmr); in mlx4_ib_dereg_mr()
674 mr->ibmr.rkey = mr->ibmr.lkey = mr->mmr.key; in mlx4_ib_alloc_mr()
677 return &mr->ibmr; in mlx4_ib_alloc_mr()
680 mr->ibmr.device = pd->device; in mlx4_ib_alloc_mr()
[all …]
/linux/drivers/infiniband/sw/rdmavt/
A Dmr.h12 struct ib_mr ibmr; member
17 static inline struct rvt_mr *to_imr(struct ib_mr *ibmr) in to_imr() argument
19 return container_of(ibmr, struct rvt_mr, ibmr); in to_imr()
30 int rvt_dereg_mr(struct ib_mr *ibmr, struct ib_udata *udata);
33 int rvt_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg,
A Dmr.c259 mr->ibmr.lkey = mr->mr.lkey; in __rvt_alloc_mr()
260 mr->ibmr.rkey = mr->mr.lkey; in __rvt_alloc_mr()
314 ret = &mr->ibmr; in rvt_get_dma_mr()
387 return &mr->ibmr; in rvt_reg_user_mr()
511 struct rvt_mr *mr = to_imr(ibmr); in rvt_dereg_mr()
547 return &mr->ibmr; in rvt_alloc_mr()
597 mr->mr.user_base = ibmr->iova; in rvt_map_mr_sg()
598 mr->mr.iova = ibmr->iova; in rvt_map_mr_sg()
629 ibmr->lkey = key; in rvt_fast_reg_mr()
630 ibmr->rkey = key; in rvt_fast_reg_mr()
[all …]
A Dtrace_mr.h144 TP_PROTO(struct ib_mr *ibmr, int sg_nents, unsigned int *sg_offset),
145 TP_ARGS(ibmr, sg_nents, sg_offset),
147 RDI_DEV_ENTRY(ib_to_rvt(to_imr(ibmr)->mr.pd->device))
156 RDI_DEV_ASSIGN(ib_to_rvt(to_imr(ibmr)->mr.pd->device));
157 __entry->ibmr_iova = ibmr->iova;
158 __entry->iova = to_imr(ibmr)->mr.iova;
159 __entry->user_base = to_imr(ibmr)->mr.user_base;
160 __entry->ibmr_length = to_imr(ibmr)->mr.length;
/linux/drivers/infiniband/hw/hns/
A Dhns_roce_mr.c233 mr->ibmr.rkey = mr->ibmr.lkey = mr->key; in hns_roce_get_dma_mr()
235 return &mr->ibmr; in hns_roce_get_dma_mr()
274 mr->ibmr.rkey = mr->ibmr.lkey = mr->key; in hns_roce_reg_user_mr()
275 mr->ibmr.length = length; in hns_roce_reg_user_mr()
277 return &mr->ibmr; in hns_roce_reg_user_mr()
295 struct hns_roce_mr *mr = to_hr_mr(ibmr); in hns_roce_rereg_user_mr()
363 struct hns_roce_mr *mr = to_hr_mr(ibmr); in hns_roce_dereg_mr()
414 mr->ibmr.rkey = mr->ibmr.lkey = mr->key; in hns_roce_alloc_mr()
415 mr->ibmr.length = mr->size; in hns_roce_alloc_mr()
417 return &mr->ibmr; in hns_roce_alloc_mr()
[all …]
/linux/drivers/infiniband/hw/usnic/
A Dusnic_ib.h69 struct ib_mr ibmr; member
125 struct usnic_ib_mr *to_umr(struct ib_mr *ibmr) in to_umr() argument
127 return container_of(ibmr, struct usnic_ib_mr, ibmr); in to_umr()
/linux/net/sunrpc/xprtrdma/
A Dfrwr_ops.c300 struct ib_mr *ibmr; in frwr_map() local
326 ibmr = mr->mr_ibmr; in frwr_map()
331 ibmr->iova &= 0x00000000ffffffff; in frwr_map()
332 ibmr->iova |= ((u64)be32_to_cpu(xid)) << 32; in frwr_map()
333 key = (u8)(ibmr->rkey & 0x000000FF); in frwr_map()
334 ib_update_fast_reg_key(ibmr, ++key); in frwr_map()
337 reg_wr->mr = ibmr; in frwr_map()
338 reg_wr->key = ibmr->rkey; in frwr_map()
343 mr->mr_handle = ibmr->rkey; in frwr_map()
344 mr->mr_length = ibmr->length; in frwr_map()
[all …]
/linux/drivers/infiniband/sw/rxe/
A Drxe_mr.c61 mr->lkey = mr->ibmr.lkey = lkey; in rxe_mr_init()
62 mr->rkey = mr->ibmr.rkey = rkey; in rxe_mr_init()
158 mr->ibmr.pd = &pd->ibpd; in rxe_mr_init_dma()
227 mr->ibmr.pd = &pd->ibpd; in rxe_mr_init_user()
257 mr->ibmr.pd = &pd->ibpd; in rxe_mr_init_fast()
643 if (unlikely(qp->ibqp.pd != mr->ibmr.pd)) { in rxe_reg_fast_mr()
662 int rxe_mr_set_page(struct ib_mr *ibmr, u64 addr) in rxe_mr_set_page() argument
664 struct rxe_mr *mr = to_rmr(ibmr); in rxe_mr_set_page()
676 buf->size = ibmr->page_size; in rxe_mr_set_page()
682 int rxe_dereg_mr(struct ib_mr *ibmr, struct ib_udata *udata) in rxe_dereg_mr() argument
[all …]
A Drxe_verbs.c905 return &mr->ibmr; in rxe_get_dma_mr()
933 return &mr->ibmr; in rxe_reg_user_mr()
968 return &mr->ibmr; in rxe_alloc_mr()
981 static int rxe_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, in rxe_map_mr_sg() argument
984 struct rxe_mr *mr = to_rmr(ibmr); in rxe_map_mr_sg()
990 n = ib_sg_to_pages(ibmr, sg, sg_nents, sg_offset, rxe_mr_set_page); in rxe_map_mr_sg()
992 set->va = ibmr->iova; in rxe_map_mr_sg()
993 set->iova = ibmr->iova; in rxe_map_mr_sg()
994 set->length = ibmr->length; in rxe_map_mr_sg()
995 set->page_shift = ilog2(ibmr->page_size); in rxe_map_mr_sg()
[all …]
A Drxe_verbs.h313 struct ib_mr ibmr; member
462 return mr ? container_of(mr, struct rxe_mr, ibmr) : NULL; in to_rmr()
477 return to_rpd(mr->ibmr.pd); in mr_pd()
/linux/drivers/infiniband/hw/cxgb4/
A Dmem.c384 mhp->ibmr.rkey = mhp->ibmr.lkey = stag; in finish_mem_reg()
385 mhp->ibmr.length = mhp->attr.len; in finish_mem_reg()
386 mhp->ibmr.page_size = 1U << (mhp->attr.page_size + 12); in finish_mem_reg()
478 return &mhp->ibmr; in c4iw_get_dma_mr()
582 return &mhp->ibmr; in c4iw_reg_user_mr()
653 mhp->ibmr.rkey = mhp->ibmr.lkey = stag; in c4iw_alloc_mr()
660 return &(mhp->ibmr); in c4iw_alloc_mr()
678 static int c4iw_set_page(struct ib_mr *ibmr, u64 addr) in c4iw_set_page() argument
680 struct c4iw_mr *mhp = to_c4iw_mr(ibmr); in c4iw_set_page()
693 struct c4iw_mr *mhp = to_c4iw_mr(ibmr); in c4iw_map_mr_sg()
[all …]
/linux/drivers/infiniband/hw/mthca/
A Dmthca_provider.h74 struct ib_mr ibmr; member
286 static inline struct mthca_mr *to_mmr(struct ib_mr *ibmr) in to_mmr() argument
288 return container_of(ibmr, struct mthca_mr, ibmr); in to_mmr()
A Dmthca_mr.c444 mr->ibmr.rkey = mr->ibmr.lkey = hw_index_to_key(dev, key); in mthca_mr_alloc()
481 mthca_dbg(dev, "Dumping MPT entry %08x:\n", mr->ibmr.lkey); in mthca_mr_alloc()
558 key_to_hw_index(dev, mr->ibmr.lkey) & in mthca_free_mr()
563 mthca_free_region(dev, mr->ibmr.lkey); in mthca_free_mr()
/linux/drivers/infiniband/hw/ocrdma/
A Docrdma.h192 struct ib_mr ibmr; member
481 static inline struct ocrdma_mr *get_ocrdma_mr(struct ib_mr *ibmr) in get_ocrdma_mr() argument
483 return container_of(ibmr, struct ocrdma_mr, ibmr); in get_ocrdma_mr()
/linux/drivers/i2c/busses/
A Di2c-pxa.c135 u32 ibmr; member
157 .ibmr = 0x00,
166 .ibmr = 0x00,
175 .ibmr = 0x14,
184 .ibmr = 0x00,
195 .ibmr = 0x00,
1297 u32 ibmr = readl(_IBMR(i2c)); in i2c_pxa_prepare_recovery() local
1303 gpiod_set_value(i2c->recovery.scl_gpiod, ibmr & IBMR_SCLS); in i2c_pxa_prepare_recovery()
1304 gpiod_set_value(i2c->recovery.sda_gpiod, ibmr & IBMR_SDAS); in i2c_pxa_prepare_recovery()
1414 i2c->reg_ibmr = i2c->reg_base + pxa_reg_layout[i2c_type].ibmr; in i2c_pxa_probe()
/linux/drivers/infiniband/hw/qedr/
A Dqedr.h484 struct ib_mr ibmr; member
603 static inline struct qedr_mr *get_qedr_mr(struct ib_mr *ibmr) in get_qedr_mr() argument
605 return container_of(ibmr, struct qedr_mr, ibmr); in get_qedr_mr()

Completed in 70 milliseconds

123