/linux/net/rds/ |
A D | ib_frmr.c | 67 if (ibmr) in rds_ib_alloc_frmr() 68 return ibmr; in rds_ib_alloc_frmr() 70 ibmr = kzalloc_node(sizeof(*ibmr), GFP_KERNEL, in rds_ib_alloc_frmr() 72 if (!ibmr) { in rds_ib_alloc_frmr() 98 return ibmr; in rds_ib_alloc_frmr() 101 kfree(ibmr); in rds_ib_alloc_frmr() 203 ibmr->sg = sg; in rds_ib_map_frmr() 208 ibmr->sg_dma_len = ib_dma_map_sg(dev, ibmr->sg, ibmr->sg_len, in rds_ib_map_frmr() 260 ib_dma_unmap_sg(rds_ibdev->dev, ibmr->sg, ibmr->sg_len, in rds_ib_map_frmr() 416 if (ibmr) in rds_ib_reg_frmr() [all …]
|
A D | ib_rdma.c | 215 return ibmr; in rds_ib_reuse_mr() 223 if (ibmr->odp) in rds_ib_sync_mr() 244 ibmr->sg, ibmr->sg_len, in __rds_ib_teardown_mr() 361 if (ibmr) { in rds_ib_flush_mr_pool() 373 if (ibmr) { in rds_ib_flush_mr_pool() 385 if (ibmr) { in rds_ib_flush_mr_pool() 447 if (ibmr) in rds_ib_try_reuse_ibmr() 470 if (ibmr) in rds_ib_try_reuse_ibmr() 592 ibmr = kzalloc(sizeof(*ibmr), GFP_KERNEL); in rds_ib_get_mr() 593 if (!ibmr) { in rds_ib_get_mr() [all …]
|
/linux/drivers/infiniband/hw/vmw_pvrdma/ |
A D | pvrdma_mr.c | 94 mr->ibmr.lkey = resp->lkey; in pvrdma_get_dma_mr() 95 mr->ibmr.rkey = resp->rkey; in pvrdma_get_dma_mr() 97 return &mr->ibmr; in pvrdma_get_dma_mr() 182 mr->ibmr.lkey = resp->lkey; in pvrdma_reg_user_mr() 183 mr->ibmr.rkey = resp->rkey; in pvrdma_reg_user_mr() 185 return &mr->ibmr; in pvrdma_reg_user_mr() 254 mr->ibmr.lkey = resp->lkey; in pvrdma_alloc_mr() 255 mr->ibmr.rkey = resp->rkey; in pvrdma_alloc_mr() 259 return &mr->ibmr; in pvrdma_alloc_mr() 279 struct pvrdma_user_mr *mr = to_vmr(ibmr); in pvrdma_dereg_mr() [all …]
|
A D | pvrdma.h | 141 struct ib_mr ibmr; member 284 static inline struct pvrdma_user_mr *to_vmr(struct ib_mr *ibmr) in to_vmr() argument 286 return container_of(ibmr, struct pvrdma_user_mr, ibmr); in to_vmr()
|
/linux/drivers/infiniband/hw/mlx5/ |
A D | mr.c | 825 return &mr->ibmr; in mlx5_ib_get_dma_mr() 973 mr->ibmr.pd = pd; in alloc_cacheable_mr() 1287 mr->ibmr.pd = pd; in reg_create() 1387 return &mr->ibmr; in mlx5_ib_get_dm_mr() 1486 return &mr->ibmr; in create_real_mr() 1540 return &mr->ibmr; in create_user_odp_mr() 1637 return &mr->ibmr; in mlx5_ib_reg_user_mr_dmabuf() 2065 mr->ibmr.pd = pd; in mlx5_ib_alloc_pi_mr() 2530 pi_mr->ibmr.page_size = ibmr->page_size; in mlx5_ib_map_mtt_mr_sg_pi() 2606 ibmr->length = pi_mr->ibmr.length; in mlx5_ib_map_klm_mr_sg_pi() [all …]
|
A D | restrack.c | 81 static int fill_stat_mr_entry(struct sk_buff *msg, struct ib_mr *ibmr) in fill_stat_mr_entry() argument 83 struct mlx5_ib_mr *mr = to_mmr(ibmr); in fill_stat_mr_entry() 115 static int fill_res_mr_entry_raw(struct sk_buff *msg, struct ib_mr *ibmr) in fill_res_mr_entry_raw() argument 117 struct mlx5_ib_mr *mr = to_mmr(ibmr); in fill_res_mr_entry_raw() 123 static int fill_res_mr_entry(struct sk_buff *msg, struct ib_mr *ibmr) in fill_res_mr_entry() argument 125 struct mlx5_ib_mr *mr = to_mmr(ibmr); in fill_res_mr_entry()
|
A D | odp.c | 205 mlx5_ib_dereg_mr(&mr->ibmr, NULL); in free_implicit_child_mr_work() 428 mr->ibmr.pd = imr->ibmr.pd; in implicit_get_child_mr() 431 mr->ibmr.lkey = mr->mmkey.key; in implicit_get_child_mr() 432 mr->ibmr.rkey = mr->mmkey.key; in implicit_get_child_mr() 476 mlx5_ib_dereg_mr(&mr->ibmr, NULL); in implicit_get_child_mr() 502 imr->ibmr.pd = &pd->ibpd; in mlx5_ib_alloc_implicit_mr() 503 imr->ibmr.iova = 0; in mlx5_ib_alloc_implicit_mr() 505 imr->ibmr.lkey = imr->mmkey.key; in mlx5_ib_alloc_implicit_mr() 506 imr->ibmr.rkey = imr->mmkey.key; in mlx5_ib_alloc_implicit_mr() 507 imr->ibmr.device = &dev->ib_dev; in mlx5_ib_alloc_implicit_mr() [all …]
|
A D | wr.c | 390 seg->start_addr = cpu_to_be64(mr->ibmr.iova); in set_reg_mkey_seg() 391 seg->len = cpu_to_be64(mr->ibmr.length); in set_reg_mkey_seg() 657 data_key = pi_mr->ibmr.lkey; in set_sig_data_segment() 661 prot_key = pi_mr->ibmr.lkey; in set_sig_data_segment() 797 region_len = pi_mr->ibmr.length; in set_pi_umr_wr() 1091 reg_pi_wr.mr = &pi_mr->ibmr; in handle_reg_mr_integrity() 1093 reg_pi_wr.key = pi_mr->ibmr.rkey; in handle_reg_mr_integrity() 1113 pa_pi_mr.ibmr.lkey = mr->ibmr.pd->local_dma_lkey; in handle_reg_mr_integrity() 1123 pa_pi_mr.ibmr.length = mr->ibmr.length; in handle_reg_mr_integrity() 1126 (*ctrl)->imm = cpu_to_be32(mr->ibmr.rkey); in handle_reg_mr_integrity() [all …]
|
A D | mlx5_ib.h | 662 struct ib_mr ibmr; member 1153 return to_mdev(mr->ibmr.device); in mr_to_mdev() 1209 static inline struct mlx5_ib_mr *to_mmr(struct ib_mr *ibmr) in to_mmr() argument 1211 return container_of(ibmr, struct mlx5_ib_mr, ibmr); in to_mmr() 1303 int mlx5_ib_dereg_mr(struct ib_mr *ibmr, struct ib_udata *udata); 1309 int mlx5_ib_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, int sg_nents, 1311 int mlx5_ib_map_mr_sg_pi(struct ib_mr *ibmr, struct scatterlist *data_sg, 1348 int mlx5_ib_check_mr_status(struct ib_mr *ibmr, u32 check_mask,
|
/linux/drivers/infiniband/hw/mlx4/ |
A D | mr.c | 76 mr->ibmr.rkey = mr->ibmr.lkey = mr->mmr.key; in mlx4_ib_get_dma_mr() 79 return &mr->ibmr; in mlx4_ib_get_dma_mr() 441 mr->ibmr.rkey = mr->ibmr.lkey = mr->mmr.key; in mlx4_ib_reg_user_mr() 442 mr->ibmr.length = length; in mlx4_ib_reg_user_mr() 443 mr->ibmr.page_size = 1U << shift; in mlx4_ib_reg_user_mr() 445 return &mr->ibmr; in mlx4_ib_reg_user_mr() 599 struct mlx4_ib_mr *mr = to_mmr(ibmr); in mlx4_ib_dereg_mr() 674 mr->ibmr.rkey = mr->ibmr.lkey = mr->mmr.key; in mlx4_ib_alloc_mr() 677 return &mr->ibmr; in mlx4_ib_alloc_mr() 680 mr->ibmr.device = pd->device; in mlx4_ib_alloc_mr() [all …]
|
/linux/drivers/infiniband/sw/rdmavt/ |
A D | mr.h | 12 struct ib_mr ibmr; member 17 static inline struct rvt_mr *to_imr(struct ib_mr *ibmr) in to_imr() argument 19 return container_of(ibmr, struct rvt_mr, ibmr); in to_imr() 30 int rvt_dereg_mr(struct ib_mr *ibmr, struct ib_udata *udata); 33 int rvt_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg,
|
A D | mr.c | 259 mr->ibmr.lkey = mr->mr.lkey; in __rvt_alloc_mr() 260 mr->ibmr.rkey = mr->mr.lkey; in __rvt_alloc_mr() 314 ret = &mr->ibmr; in rvt_get_dma_mr() 387 return &mr->ibmr; in rvt_reg_user_mr() 511 struct rvt_mr *mr = to_imr(ibmr); in rvt_dereg_mr() 547 return &mr->ibmr; in rvt_alloc_mr() 597 mr->mr.user_base = ibmr->iova; in rvt_map_mr_sg() 598 mr->mr.iova = ibmr->iova; in rvt_map_mr_sg() 629 ibmr->lkey = key; in rvt_fast_reg_mr() 630 ibmr->rkey = key; in rvt_fast_reg_mr() [all …]
|
A D | trace_mr.h | 144 TP_PROTO(struct ib_mr *ibmr, int sg_nents, unsigned int *sg_offset), 145 TP_ARGS(ibmr, sg_nents, sg_offset), 147 RDI_DEV_ENTRY(ib_to_rvt(to_imr(ibmr)->mr.pd->device)) 156 RDI_DEV_ASSIGN(ib_to_rvt(to_imr(ibmr)->mr.pd->device)); 157 __entry->ibmr_iova = ibmr->iova; 158 __entry->iova = to_imr(ibmr)->mr.iova; 159 __entry->user_base = to_imr(ibmr)->mr.user_base; 160 __entry->ibmr_length = to_imr(ibmr)->mr.length;
|
/linux/drivers/infiniband/hw/hns/ |
A D | hns_roce_mr.c | 233 mr->ibmr.rkey = mr->ibmr.lkey = mr->key; in hns_roce_get_dma_mr() 235 return &mr->ibmr; in hns_roce_get_dma_mr() 274 mr->ibmr.rkey = mr->ibmr.lkey = mr->key; in hns_roce_reg_user_mr() 275 mr->ibmr.length = length; in hns_roce_reg_user_mr() 277 return &mr->ibmr; in hns_roce_reg_user_mr() 295 struct hns_roce_mr *mr = to_hr_mr(ibmr); in hns_roce_rereg_user_mr() 363 struct hns_roce_mr *mr = to_hr_mr(ibmr); in hns_roce_dereg_mr() 414 mr->ibmr.rkey = mr->ibmr.lkey = mr->key; in hns_roce_alloc_mr() 415 mr->ibmr.length = mr->size; in hns_roce_alloc_mr() 417 return &mr->ibmr; in hns_roce_alloc_mr() [all …]
|
/linux/drivers/infiniband/hw/usnic/ |
A D | usnic_ib.h | 69 struct ib_mr ibmr; member 125 struct usnic_ib_mr *to_umr(struct ib_mr *ibmr) in to_umr() argument 127 return container_of(ibmr, struct usnic_ib_mr, ibmr); in to_umr()
|
/linux/net/sunrpc/xprtrdma/ |
A D | frwr_ops.c | 300 struct ib_mr *ibmr; in frwr_map() local 326 ibmr = mr->mr_ibmr; in frwr_map() 331 ibmr->iova &= 0x00000000ffffffff; in frwr_map() 332 ibmr->iova |= ((u64)be32_to_cpu(xid)) << 32; in frwr_map() 333 key = (u8)(ibmr->rkey & 0x000000FF); in frwr_map() 334 ib_update_fast_reg_key(ibmr, ++key); in frwr_map() 337 reg_wr->mr = ibmr; in frwr_map() 338 reg_wr->key = ibmr->rkey; in frwr_map() 343 mr->mr_handle = ibmr->rkey; in frwr_map() 344 mr->mr_length = ibmr->length; in frwr_map() [all …]
|
/linux/drivers/infiniband/sw/rxe/ |
A D | rxe_mr.c | 61 mr->lkey = mr->ibmr.lkey = lkey; in rxe_mr_init() 62 mr->rkey = mr->ibmr.rkey = rkey; in rxe_mr_init() 158 mr->ibmr.pd = &pd->ibpd; in rxe_mr_init_dma() 227 mr->ibmr.pd = &pd->ibpd; in rxe_mr_init_user() 257 mr->ibmr.pd = &pd->ibpd; in rxe_mr_init_fast() 643 if (unlikely(qp->ibqp.pd != mr->ibmr.pd)) { in rxe_reg_fast_mr() 662 int rxe_mr_set_page(struct ib_mr *ibmr, u64 addr) in rxe_mr_set_page() argument 664 struct rxe_mr *mr = to_rmr(ibmr); in rxe_mr_set_page() 676 buf->size = ibmr->page_size; in rxe_mr_set_page() 682 int rxe_dereg_mr(struct ib_mr *ibmr, struct ib_udata *udata) in rxe_dereg_mr() argument [all …]
|
A D | rxe_verbs.c | 905 return &mr->ibmr; in rxe_get_dma_mr() 933 return &mr->ibmr; in rxe_reg_user_mr() 968 return &mr->ibmr; in rxe_alloc_mr() 981 static int rxe_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, in rxe_map_mr_sg() argument 984 struct rxe_mr *mr = to_rmr(ibmr); in rxe_map_mr_sg() 990 n = ib_sg_to_pages(ibmr, sg, sg_nents, sg_offset, rxe_mr_set_page); in rxe_map_mr_sg() 992 set->va = ibmr->iova; in rxe_map_mr_sg() 993 set->iova = ibmr->iova; in rxe_map_mr_sg() 994 set->length = ibmr->length; in rxe_map_mr_sg() 995 set->page_shift = ilog2(ibmr->page_size); in rxe_map_mr_sg() [all …]
|
A D | rxe_verbs.h | 313 struct ib_mr ibmr; member 462 return mr ? container_of(mr, struct rxe_mr, ibmr) : NULL; in to_rmr() 477 return to_rpd(mr->ibmr.pd); in mr_pd()
|
/linux/drivers/infiniband/hw/cxgb4/ |
A D | mem.c | 384 mhp->ibmr.rkey = mhp->ibmr.lkey = stag; in finish_mem_reg() 385 mhp->ibmr.length = mhp->attr.len; in finish_mem_reg() 386 mhp->ibmr.page_size = 1U << (mhp->attr.page_size + 12); in finish_mem_reg() 478 return &mhp->ibmr; in c4iw_get_dma_mr() 582 return &mhp->ibmr; in c4iw_reg_user_mr() 653 mhp->ibmr.rkey = mhp->ibmr.lkey = stag; in c4iw_alloc_mr() 660 return &(mhp->ibmr); in c4iw_alloc_mr() 678 static int c4iw_set_page(struct ib_mr *ibmr, u64 addr) in c4iw_set_page() argument 680 struct c4iw_mr *mhp = to_c4iw_mr(ibmr); in c4iw_set_page() 693 struct c4iw_mr *mhp = to_c4iw_mr(ibmr); in c4iw_map_mr_sg() [all …]
|
/linux/drivers/infiniband/hw/mthca/ |
A D | mthca_provider.h | 74 struct ib_mr ibmr; member 286 static inline struct mthca_mr *to_mmr(struct ib_mr *ibmr) in to_mmr() argument 288 return container_of(ibmr, struct mthca_mr, ibmr); in to_mmr()
|
A D | mthca_mr.c | 444 mr->ibmr.rkey = mr->ibmr.lkey = hw_index_to_key(dev, key); in mthca_mr_alloc() 481 mthca_dbg(dev, "Dumping MPT entry %08x:\n", mr->ibmr.lkey); in mthca_mr_alloc() 558 key_to_hw_index(dev, mr->ibmr.lkey) & in mthca_free_mr() 563 mthca_free_region(dev, mr->ibmr.lkey); in mthca_free_mr()
|
/linux/drivers/infiniband/hw/ocrdma/ |
A D | ocrdma.h | 192 struct ib_mr ibmr; member 481 static inline struct ocrdma_mr *get_ocrdma_mr(struct ib_mr *ibmr) in get_ocrdma_mr() argument 483 return container_of(ibmr, struct ocrdma_mr, ibmr); in get_ocrdma_mr()
|
/linux/drivers/i2c/busses/ |
A D | i2c-pxa.c | 135 u32 ibmr; member 157 .ibmr = 0x00, 166 .ibmr = 0x00, 175 .ibmr = 0x14, 184 .ibmr = 0x00, 195 .ibmr = 0x00, 1297 u32 ibmr = readl(_IBMR(i2c)); in i2c_pxa_prepare_recovery() local 1303 gpiod_set_value(i2c->recovery.scl_gpiod, ibmr & IBMR_SCLS); in i2c_pxa_prepare_recovery() 1304 gpiod_set_value(i2c->recovery.sda_gpiod, ibmr & IBMR_SDAS); in i2c_pxa_prepare_recovery() 1414 i2c->reg_ibmr = i2c->reg_base + pxa_reg_layout[i2c_type].ibmr; in i2c_pxa_probe()
|
/linux/drivers/infiniband/hw/qedr/ |
A D | qedr.h | 484 struct ib_mr ibmr; member 603 static inline struct qedr_mr *get_qedr_mr(struct ib_mr *ibmr) in get_qedr_mr() argument 605 return container_of(ibmr, struct qedr_mr, ibmr); in get_qedr_mr()
|