/linux/net/9p/ |
A D | trans_rdma.c | 261 if (rdma) in p9_cm_event_handler() 365 if (!rdma) in rdma_destroy_trans() 368 if (rdma->qp && !IS_ERR(rdma->qp)) in rdma_destroy_trans() 371 if (rdma->pd && !IS_ERR(rdma->pd)) in rdma_destroy_trans() 374 if (rdma->cq && !IS_ERR(rdma->cq)) in rdma_destroy_trans() 377 if (rdma->cm_id && !IS_ERR(rdma->cm_id)) in rdma_destroy_trans() 551 if (!rdma) in rdma_close() 578 sema_init(&rdma->sq_sem, rdma->sq_depth); in alloc_rdma() 579 sema_init(&rdma->rq_sem, rdma->rq_depth); in alloc_rdma() 696 rdma->pd = ib_alloc_pd(rdma->cm_id->device, 0); in rdma_create_trans() [all …]
|
/linux/drivers/gpu/drm/mediatek/ |
A D | mtk_disp_rdma.c | 49 #define RDMA_FIFO_SIZE(rdma) ((rdma)->data->fifo_size) argument 113 rdma->vblank_cb = NULL; in mtk_rdma_disable_vblank() 152 mtk_ddp_write_mask(cmdq_pkt, width, &rdma->cmdq_reg, rdma->regs, in mtk_rdma_config() 154 mtk_ddp_write_mask(cmdq_pkt, height, &rdma->cmdq_reg, rdma->regs, in mtk_rdma_config() 157 if (rdma->fifo_size) in mtk_rdma_config() 236 &rdma->cmdq_reg, rdma->regs, DISP_REG_RDMA_SIZE_CON_0, in mtk_rdma_layer_config() 239 mtk_ddp_write_mask(cmdq_pkt, 0, &rdma->cmdq_reg, rdma->regs, in mtk_rdma_layer_config() 243 mtk_ddp_write_relaxed(cmdq_pkt, addr, &rdma->cmdq_reg, rdma->regs, in mtk_rdma_layer_config() 245 mtk_ddp_write_relaxed(cmdq_pkt, pitch, &rdma->cmdq_reg, rdma->regs, in mtk_rdma_layer_config() 247 mtk_ddp_write(cmdq_pkt, RDMA_MEM_GMC, &rdma->cmdq_reg, rdma->regs, in mtk_rdma_layer_config() [all …]
|
/linux/net/sunrpc/xprtrdma/ |
A D | svc_rdma_sendto.c | 155 rdma->sc_max_req_size); in svc_rdma_send_ctxt_alloc() 184 rdma->sc_max_req_size, in svc_rdma_send_ctxts_destroy() 203 spin_lock(&rdma->sc_send_lock); in svc_rdma_send_ctxt_get() 265 wake_up(&rdma->sc_send_wait); in svc_rdma_wake_send_waiters() 325 trace_svcrdma_sq_full(rdma); in svc_rdma_send() 344 wake_up(&rdma->sc_send_wait); in svc_rdma_send() 778 .md_rdma = rdma, in svc_rdma_map_reply_msg() 881 *p++ = rdma->sc_fc_credits; in svc_rdma_send_error_msg() 932 struct svcxprt_rdma *rdma = in svc_rdma_sendto() local 963 *p++ = rdma->sc_fc_credits; in svc_rdma_sendto() [all …]
|
A D | svc_rdma_recvfrom.c | 253 struct svcxprt_rdma *rdma = in svc_rdma_release_rqst() local 282 rdma->sc_pending_recvs++; in svc_rdma_refresh_recvs() 313 return svc_rdma_refresh_recvs(rdma, rdma->sc_max_requests, true); in svc_rdma_post_recvs() 328 rdma->sc_pending_recvs--; in svc_rdma_wc_receive() 346 if (rdma->sc_pending_recvs < rdma->sc_max_requests) in svc_rdma_wc_receive() 347 if (!svc_rdma_refresh_recvs(rdma, rdma->sc_recv_batch, false)) in svc_rdma_wc_receive() 353 spin_lock(&rdma->sc_rq_dto_lock); in svc_rdma_wc_receive() 357 spin_unlock(&rdma->sc_rq_dto_lock); in svc_rdma_wc_receive() 359 svc_xprt_enqueue(&rdma->sc_xprt); in svc_rdma_wc_receive() 368 svc_rdma_recv_ctxt_put(rdma, ctxt); in svc_rdma_wc_receive() [all …]
|
A D | svc_rdma_rw.c | 59 spin_lock(&rdma->sc_rw_ctxt_lock); in svc_rdma_get_rw_ctxt() 61 spin_unlock(&rdma->sc_rw_ctxt_lock); in svc_rdma_get_rw_ctxt() 98 __svc_rdma_put_rw_ctxt(rdma, ctxt, &rdma->sc_rw_ctxts); in svc_rdma_put_rw_ctxt() 135 ret = rdma_rw_ctx_init(&ctxt->rw_ctx, rdma->sc_qp, rdma->sc_port_num, in svc_rdma_rw_ctx_init() 139 svc_rdma_put_rw_ctxt(rdma, ctxt); in svc_rdma_rw_ctx_init() 175 cc->cc_rdma = rdma; in svc_rdma_cc_init() 395 trace_svcrdma_sq_full(rdma); in svc_rdma_post_chunk_ctxt() 397 wait_event(rdma->sc_send_wait, in svc_rdma_post_chunk_ctxt() 399 trace_svcrdma_sq_retry(rdma); in svc_rdma_post_chunk_ctxt() 410 wake_up(&rdma->sc_send_wait); in svc_rdma_post_chunk_ctxt() [all …]
|
A D | svc_rdma_transport.c | 538 struct svcxprt_rdma *rdma = in svc_rdma_detach() local 546 struct svcxprt_rdma *rdma = in __svc_rdma_free() local 550 if (rdma->sc_qp && !IS_ERR(rdma->sc_qp)) in __svc_rdma_free() 551 ib_drain_qp(rdma->sc_qp); in __svc_rdma_free() 560 if (rdma->sc_qp && !IS_ERR(rdma->sc_qp)) in __svc_rdma_free() 563 if (rdma->sc_sq_cq && !IS_ERR(rdma->sc_sq_cq)) in __svc_rdma_free() 566 if (rdma->sc_rq_cq && !IS_ERR(rdma->sc_rq_cq)) in __svc_rdma_free() 569 if (rdma->sc_pd && !IS_ERR(rdma->sc_pd)) in __svc_rdma_free() 575 kfree(rdma); in __svc_rdma_free() 580 struct svcxprt_rdma *rdma = in svc_rdma_free() local [all …]
|
A D | svc_rdma_backchannel.c | 75 static int svc_rdma_bc_sendto(struct svcxprt_rdma *rdma, in svc_rdma_bc_sendto() argument 82 rctxt = svc_rdma_recv_ctxt_get(rdma); in svc_rdma_bc_sendto() 86 ret = svc_rdma_map_reply_msg(rdma, sctxt, rctxt, &rqst->rq_snd_buf); in svc_rdma_bc_sendto() 87 svc_rdma_recv_ctxt_put(rdma, rctxt); in svc_rdma_bc_sendto() 96 ret = svc_rdma_send(rdma, sctxt); in svc_rdma_bc_sendto() 101 svc_rdma_send_ctxt_put(rdma, sctxt); in svc_rdma_bc_sendto() 153 ctxt = svc_rdma_send_ctxt_get(rdma); in rpcrdma_bc_send_request() 169 rc = svc_rdma_bc_sendto(rdma, rqst, ctxt); in rpcrdma_bc_send_request() 175 svc_rdma_send_ctxt_put(rdma, ctxt); in rpcrdma_bc_send_request() 192 struct svcxprt_rdma *rdma = in xprt_rdma_bc_send_request() local [all …]
|
/linux/drivers/gpu/drm/meson/ |
A D | meson_rdma.c | 26 if (!priv->rdma.addr) { in meson_rdma_init() 28 priv->rdma.addr = in meson_rdma_init() 32 if (!priv->rdma.addr) in meson_rdma_init() 36 priv->rdma.offset = 0; in meson_rdma_init() 50 if (!priv->rdma.addr && !priv->rdma.addr_dma) in meson_rdma_free() 56 priv->rdma.addr, priv->rdma.addr_dma); in meson_rdma_free() 58 priv->rdma.addr = NULL; in meson_rdma_free() 88 priv->rdma.offset = 0; in meson_rdma_reset() 99 priv->rdma.addr[priv->rdma.offset++] = reg; in meson_rdma_writel() 100 priv->rdma.addr[priv->rdma.offset++] = val; in meson_rdma_writel() [all …]
|
/linux/include/linux/sunrpc/ |
A D | svc_rdma.h | 173 extern bool svc_rdma_post_recvs(struct svcxprt_rdma *rdma); 175 svc_rdma_recv_ctxt_get(struct svcxprt_rdma *rdma); 176 extern void svc_rdma_recv_ctxt_put(struct svcxprt_rdma *rdma, 184 extern int svc_rdma_send_write_chunk(struct svcxprt_rdma *rdma, 187 extern int svc_rdma_send_reply_chunk(struct svcxprt_rdma *rdma, 190 extern int svc_rdma_process_read_list(struct svcxprt_rdma *rdma, 197 svc_rdma_send_ctxt_get(struct svcxprt_rdma *rdma); 198 extern void svc_rdma_send_ctxt_put(struct svcxprt_rdma *rdma, 200 extern int svc_rdma_send(struct svcxprt_rdma *rdma, 202 extern int svc_rdma_map_reply_msg(struct svcxprt_rdma *rdma, [all …]
|
/linux/Documentation/admin-guide/cgroup-v1/ |
A D | rdma.rst | 28 Currently user space applications can easily take away all the rdma verb 31 rdma resources. This can lead to service unavailability. 34 of processes can be limited. Through this controller different rdma 43 by rdma cgroup, which can be extended later if required. 60 Whenever RDMA resource charging occurs, owner rdma cgroup is returned to 86 Following resources can be accounted by rdma controller. 98 echo mlx4_0 hca_handle=2 hca_object=2000 > /sys/fs/cgroup/rdma/1/rdma.max 99 echo ocrdma1 hca_handle=3 > /sys/fs/cgroup/rdma/2/rdma.max 103 cat /sys/fs/cgroup/rdma/2/rdma.max 110 cat /sys/fs/cgroup/rdma/2/rdma.current [all …]
|
/linux/include/trace/events/ |
A D | rpcrdma.h | 1491 TP_ARGS(rdma, status), 1514 TP_ARGS(rdma, status)) 1791 TP_ARGS(rdma, dma_addr, length), 1833 TP_ARGS(rdma, nents, status), 1861 TP_ARGS(rdma, num_sges), 1887 TP_ARGS(rdma, rqst, pageno), 2059 TP_ARGS(rdma, status), 2181 const struct svcxprt_rdma *rdma 2184 TP_ARGS(rdma), 2208 TP_ARGS(rdma)) [all …]
|
/linux/Documentation/ABI/stable/ |
A D | sysfs-driver-ib_srp | 4 Contact: linux-rdma@vger.kernel.org 77 Contact: linux-rdma@vger.kernel.org 83 Contact: linux-rdma@vger.kernel.org 89 Contact: linux-rdma@vger.kernel.org 97 Contact: linux-rdma@vger.kernel.org 104 Contact: linux-rdma@vger.kernel.org 111 Contact: linux-rdma@vger.kernel.org 117 Contact: linux-rdma@vger.kernel.org 124 Contact: linux-rdma@vger.kernel.org 131 Contact: linux-rdma@vger.kernel.org [all …]
|
A D | sysfs-class-infiniband | 9 Contact: linux-rdma@vger.kernel.org 24 Contact: linux-rdma@vger.kernel.org 34 Contact: linux-rdma@vger.kernel.org 49 Contact: linux-rdma@vger.kernel.org 80 Contact: linux-rdma@vger.kernel.org 108 Contact: linux-rdma@vger.kernel.org 203 Contact: linux-rdma@vger.kernel.org 222 Contact: linux-rdma@vger.kernel.org 229 Contact: linux-rdma@vger.kernel.org 570 linux-rdma@vger.kernel.org [all …]
|
A D | sysfs-transport-srp | 4 Contact: linux-scsi@vger.kernel.org, linux-rdma@vger.kernel.org 11 Contact: linux-scsi@vger.kernel.org, linux-rdma@vger.kernel.org 20 Contact: linux-scsi@vger.kernel.org, linux-rdma@vger.kernel.org 36 Contact: linux-scsi@vger.kernel.org, linux-rdma@vger.kernel.org 50 Contact: linux-scsi@vger.kernel.org, linux-rdma@vger.kernel.org
|
/linux/Documentation/devicetree/bindings/display/mediatek/ |
A D | mediatek,disp.txt | 61 "mediatek,<chip>-disp-rdma" 71 - mediatek,rdma-fifo-size: rdma fifo size may be different even in same SOC, add this 72 property to the corresponding rdma 74 mediatek,rdma-fifo-size of mt8173-rdma0 is 8K 107 rdma0: rdma@1400e000 { 108 compatible = "mediatek,mt8173-disp-rdma"; 115 mediatek,rdma-fifosize = <8192>; 118 rdma1: rdma@1400f000 { 119 compatible = "mediatek,mt8173-disp-rdma"; 128 rdma2: rdma@14010000 { [all …]
|
/linux/drivers/net/ethernet/seeq/ |
A D | sgiseeq.c | 69 volatile struct hpc_dma_desc rdma; member 213 sp->rx_desc[i].rdma.pbuf = dma_addr; in seeq_init_ring() 263 i, (&r[i]), r[i].rdma.pbuf, r[i].rdma.cntinfo, in sgiseeq_dump_rings() 264 r[i].rdma.pnext); in sgiseeq_dump_rings() 267 i, (&r[i]), r[i].rdma.pbuf, r[i].rdma.cntinfo, in sgiseeq_dump_rings() 268 r[i].rdma.pnext); in sgiseeq_dump_rings() 355 while (!(rd->rdma.cntinfo & HPCDMA_OWN)) { in sgiseeq_rx() 404 rd->rdma.cntinfo = RCNTINFO_INIT; in sgiseeq_rx() 711 buf[i].rdma.pbuf = 0; in setup_rx_ring() 715 buf[i].rdma.pbuf = 0; in setup_rx_ring() [all …]
|
/linux/drivers/net/ethernet/chelsio/cxgb3/ |
A D | cxgb3_offload.c | 281 struct rdma_info *rdma = data; in cxgb_rdma_ctl() local 286 rdma->tpt_base = in cxgb_rdma_ctl() 289 rdma->pbl_base = in cxgb_rdma_ctl() 295 rdma->pdev = pdev; in cxgb_rdma_ctl() 300 struct rdma_cq_op *rdma = data; in cxgb_rdma_ctl() local 304 ret = t3_sge_cqcntxt_op(adapter, rdma->id, rdma->op, in cxgb_rdma_ctl() 305 rdma->credits); in cxgb_rdma_ctl() 337 rdma->base_addr, rdma->size, in cxgb_rdma_ctl() 339 rdma->ovfl_mode, rdma->credits, in cxgb_rdma_ctl() 340 rdma->credit_thres); in cxgb_rdma_ctl() [all …]
|
/linux/drivers/infiniband/ulp/rtrs/ |
A D | rtrs-clt-stats.c | 38 s->rdma.failover_cnt++; in rtrs_clt_inc_failover_cnt() 94 r = &per_cpu_ptr(stats->pcpu_stats, cpu)->rdma; in rtrs_clt_stats_rdma_to_str() 124 memset(&s->rdma, 0, sizeof(s->rdma)); in rtrs_clt_reset_rdma_stats() 175 s->rdma.dir[d].cnt++; in rtrs_clt_update_rdma_stats() 176 s->rdma.dir[d].size_total += size; in rtrs_clt_update_rdma_stats()
|
A D | README | 29 server side for a given client for rdma transfer. A session 50 invalidate each rdma buffer before we hand it over to RNBD server and 95 6. Server and client exchange periodically heartbeat messages (empty rdma 123 on the server side and rdma writes there the user data, user header and the 129 2. When confirming a write request server sends an "empty" rdma message with 140 on the server side and rdma writes there the user data, user header and the 147 2. When confirming a write request server sends an "empty" rdma message with 163 on the server side and rdma writes there the user header and the 169 attaches an invalidation message if requested and finally an "empty" rdma 182 on the server side and rdma writes there the user header and the [all …]
|
/linux/Documentation/devicetree/bindings/media/ |
A D | mediatek-mdp.txt | 12 "mediatek,mt8173-mdp-rdma" - read DMA 24 "mediatek,mt8173-mdp-rdma" 35 mdp_rdma0: rdma@14001000 { 36 compatible = "mediatek,mt8173-mdp-rdma"; 47 mdp_rdma1: rdma@14002000 { 48 compatible = "mediatek,mt8173-mdp-rdma";
|
/linux/drivers/infiniband/ |
A D | Kconfig | 25 need libibumad from rdma-core 26 <https://github.com/linux-rdma/rdma-core>. 38 rdma-core <https://github.com/linux-rdma/rdma-core>.
|
/linux/drivers/nvme/host/ |
A D | Makefile | 8 obj-$(CONFIG_NVME_RDMA) += nvme-rdma.o 23 nvme-rdma-y += rdma.o
|
/linux/drivers/nvme/target/ |
A D | Makefile | 7 obj-$(CONFIG_NVME_TARGET_RDMA) += nvmet-rdma.o 17 nvmet-rdma-y += rdma.o
|
/linux/drivers/macintosh/ |
A D | rack-meter.c | 147 struct rackmeter_dma *rdma = rm->dma_buf_v; in rackmeter_do_pause() local 156 memset(rdma->buf1, 0, sizeof(rdma->buf1)); in rackmeter_do_pause() 157 memset(rdma->buf2, 0, sizeof(rdma->buf2)); in rackmeter_do_pause() 372 struct resource ri2s, rdma; in rackmeter_probe() local 432 of_address_to_resource(i2s, 1, &rdma)) { in rackmeter_probe() 442 pr_debug(" dma @0x%08x\n", (unsigned int)rdma.start); in rackmeter_probe() 476 rm->dma_regs = ioremap(rdma.start, 0x100); in rackmeter_probe()
|
/linux/drivers/infiniband/hw/mlx5/ |
A D | odp.c | 76 } rdma; member 1293 u32 rkey = pfault->rdma.r_key; in mlx5_ib_mr_rdma_pfault_handler() 1304 pfault->rdma.rdma_op_len); in mlx5_ib_mr_rdma_pfault_handler() 1307 address = pfault->rdma.rdma_va; in mlx5_ib_mr_rdma_pfault_handler() 1414 pfault->rdma.r_key = in mlx5_ib_eq_pf_process() 1416 pfault->rdma.packet_size = in mlx5_ib_eq_pf_process() 1418 pfault->rdma.rdma_op_len = in mlx5_ib_eq_pf_process() 1420 pfault->rdma.rdma_va = in mlx5_ib_eq_pf_process() 1425 pfault->rdma.r_key); in mlx5_ib_eq_pf_process() 1428 pfault->rdma.rdma_op_len, in mlx5_ib_eq_pf_process() [all …]
|