Home
last modified time | relevance | path

Searched refs:mqp (Results 1 – 21 of 21) sorted by relevance

/linux/drivers/infiniband/hw/mlx5/
A Dgsi.c49 struct mlx5_ib_gsi_qp *gsi = &mqp->gsi; in generate_completions()
82 wr->wc.qp = &mqp->ibqp; in handle_single_completion()
84 generate_completions(mqp); in handle_single_completion()
106 gsi = &mqp->gsi; in mlx5_ib_create_gsi()
264 struct mlx5_ib_qp *mqp; in setup_qp() local
302 mqp = to_mqp(qp); in setup_qp()
304 mqp->gsi_lag_port = qp_index + 1; in setup_qp()
324 struct mlx5_ib_qp *mqp = to_mqp(qp); in mlx5_ib_gsi_modify_qp() local
349 struct mlx5_ib_qp *mqp = to_mqp(qp); in mlx5_ib_gsi_query_qp() local
397 .qp = &mqp->ibqp, in mlx5_ib_gsi_silent_drop()
[all …]
A Dqp.c1562 qp->trans_qp.base.mqp.qpn = qp->sq.wqe_cnt ? sq->base.mqp.qpn : in create_raw_packet_qp()
1563 rq->base.mqp.qpn; in create_raw_packet_qp()
1826 qp->trans_qp.base.mqp.qpn = 0; in create_rss_raw_qp_tir()
2600 base->mqp.qpn); in destroy_qp_common()
2633 base->mqp.qpn); in destroy_qp_common()
3112 kfree(mqp->dct.in); in mlx5_ib_destroy_dct()
3251 if (mqp->type == IB_QPT_GSI) in mlx5_ib_destroy_qp()
3973 base = &mqp->trans_qp.base; in __mlx5_ib_qp_set_counter()
5673 mutex_lock(&mqp->mutex); in mlx5_ib_qp_set_counter()
5692 mqp->counter_pending = 1; in mlx5_ib_qp_set_counter()
[all …]
A Dqpc.c193 struct mlx5_core_qp *qp = &dct->mqp; in _mlx5_core_destroy_dct()
206 destroy_resource_common(dev, &dct->mqp); in _mlx5_core_destroy_dct()
217 struct mlx5_core_qp *qp = &dct->mqp; in mlx5_core_create_dct()
275 struct mlx5_core_qp *qp = &dct->mqp; in mlx5_core_drain_dct()
520 struct mlx5_core_qp *qp = &dct->mqp; in mlx5_core_dct_query()
A Dmlx5_ib.h412 struct mlx5_core_qp mqp; member
1169 static inline struct mlx5_ib_qp *to_mibqp(struct mlx5_core_qp *mqp) in to_mibqp() argument
1171 return container_of(mqp, struct mlx5_ib_qp_base, mqp)->container_mibqp; in to_mibqp()
1438 int mlx5_ib_create_gsi(struct ib_pd *pd, struct mlx5_ib_qp *mqp,
1440 int mlx5_ib_destroy_gsi(struct mlx5_ib_qp *mqp);
A Dfs.c1338 struct mlx5_ib_qp *mqp = to_mqp(qp); in mlx5_ib_create_flow() local
1424 if (mqp->is_rss) in mlx5_ib_create_flow()
1425 dst->tir_num = mqp->rss_qp.tirn; in mlx5_ib_create_flow()
1427 dst->tir_num = mqp->raw_packet_qp.rq.tirn; in mlx5_ib_create_flow()
1432 underlay_qpn = (mqp->flags & IB_QP_CREATE_SOURCE_QPN) ? in mlx5_ib_create_flow()
1433 mqp->underlay_qpn : in mlx5_ib_create_flow()
2058 struct mlx5_ib_qp *mqp; in get_dests() local
2068 mqp = to_mqp(*qp); in get_dests()
2069 if (mqp->is_rss) in get_dests()
2070 *dest_id = mqp->rss_qp.tirn; in get_dests()
[all …]
A Dmain.c2402 struct mlx5_ib_qp *mqp = to_mqp(ibqp); in mlx5_ib_mcg_attach() local
2409 if (mqp->flags & IB_QP_CREATE_SOURCE_QPN) { in mlx5_ib_mcg_attach()
2533 struct mlx5_ib_qp *mqp; in mlx5_ib_handle_internal_error() local
2546 spin_lock_irqsave(&mqp->sq.lock, flags_qp); in mlx5_ib_handle_internal_error()
2547 if (mqp->sq.tail != mqp->sq.head) { in mlx5_ib_handle_internal_error()
2548 send_mcq = to_mcq(mqp->ibqp.send_cq); in mlx5_ib_handle_internal_error()
2551 mqp->ibqp.send_cq->comp_handler) { in mlx5_ib_handle_internal_error()
2563 if (!mqp->ibqp.srq) { in mlx5_ib_handle_internal_error()
2564 if (mqp->rq.tail != mqp->rq.head) { in mlx5_ib_handle_internal_error()
2565 recv_mcq = to_mcq(mqp->ibqp.recv_cq); in mlx5_ib_handle_internal_error()
[all …]
A Dib_rep.c189 return mlx5_eswitch_add_send_to_vport_rule(esw, esw, rep, sq->base.mqp.qpn); in create_flow_rule_vport_sq()
A Dmem.c128 (qp->trans_qp.base.mqp.qpn << 8)); in post_send_nop()
A Dcq.c448 struct mlx5_core_qp *mqp; in mlx5_poll_one() local
489 mqp = radix_tree_lookup(&dev->qp_table.tree, qpn); in mlx5_poll_one()
490 *cur_qp = to_mibqp(mqp); in mlx5_poll_one()
A Dodp.c1059 u32 qpn = qp->trans_qp.base.mqp.qpn; in mlx5_ib_mr_initiator_pfault_handler()
1177 struct mlx5_core_qp *mqp = (struct mlx5_core_qp *)res; in res_to_qp() local
1179 return to_mibqp(mqp); in res_to_qp()
A Ddevx.c642 rq->base.mqp.qpn) == obj_id || in devx_is_valid_obj_id()
644 sq->base.mqp.qpn) == obj_id || in devx_is_valid_obj_id()
653 qp->dct.mdct.mqp.qpn) == obj_id; in devx_is_valid_obj_id()
A Dwr.c990 ctrl->qpn_ds = cpu_to_be32(size | (qp->trans_qp.base.mqp.qpn << 8)); in finish_wqe()
/linux/drivers/infiniband/sw/rdmavt/
A Dmcast.c35 struct rvt_mcast_qp *mqp; in rvt_mcast_qp_alloc() local
37 mqp = kmalloc(sizeof(*mqp), GFP_KERNEL); in rvt_mcast_qp_alloc()
38 if (!mqp) in rvt_mcast_qp_alloc()
41 mqp->qp = qp; in rvt_mcast_qp_alloc()
45 return mqp; in rvt_mcast_qp_alloc()
50 struct rvt_qp *qp = mqp->qp; in rvt_mcast_qp_free()
55 kfree(mqp); in rvt_mcast_qp_free()
186 if (p->qp == mqp->qp) { in rvt_mcast_add()
244 struct rvt_mcast_qp *mqp; in rvt_attach_mcast() local
259 if (!mqp) in rvt_attach_mcast()
[all …]
/linux/drivers/scsi/arm/
A Dmsgqueue.c121 struct msgqueue_entry **mqp; in msgqueue_addmsg() local
133 mqp = &msgq->qe; in msgqueue_addmsg()
134 while (*mqp) in msgqueue_addmsg()
135 mqp = &(*mqp)->next; in msgqueue_addmsg()
137 *mqp = mq; in msgqueue_addmsg()
/linux/drivers/infiniband/hw/mlx4/
A Dmain.c1280 ge->port = mqp->port; in add_gid_entry()
1313 if (!mqp->port) in mlx4_ib_add_mc()
1857 err = mlx4_multicast_attach(mdev->dev, &mqp->mqp, gid->raw, mqp->port, in mlx4_ib_mcg_attach()
1858 !!(mqp->flags & in mlx4_ib_mcg_attach()
1868 err = mlx4_multicast_attach(mdev->dev, &mqp->mqp, gid->raw, in mlx4_ib_mcg_attach()
1891 mlx4_multicast_detach(mdev->dev, &mqp->mqp, gid->raw, in mlx4_ib_mcg_attach()
1894 mlx4_multicast_detach(mdev->dev, &mqp->mqp, gid->raw, in mlx4_ib_mcg_attach()
1949 err = mlx4_multicast_detach(mdev->dev, &mqp->mqp, gid->raw, in mlx4_ib_mcg_detach()
1955 err = mlx4_multicast_detach(mdev->dev, &mqp->mqp, gid->raw, in mlx4_ib_mcg_detach()
3075 if (mqp->sq.tail != mqp->sq.head) { in mlx4_ib_handle_catas_error()
[all …]
A Dqp.c765 qp->mqp.usage); in mlx4_ib_alloc_wqn()
1164 qp->mqp.usage); in create_qp_common()
1183 qp->mqp.qpn |= (1 << 23); in create_qp_common()
1352 qp->mqp.qpn); in destroy_qp_rss()
1372 qp->mqp.qpn); in destroy_qp_common()
1626 if (is_qp0(dev, mqp)) in _mlx4_ib_destroy_qp()
1630 dev->qp1_proxy[mqp->port - 1] == mqp) { in _mlx4_ib_destroy_qp()
1636 if (mqp->counter_index) in _mlx4_ib_destroy_qp()
1640 destroy_qp_rss(dev, mqp); in _mlx4_ib_destroy_qp()
1645 kfree(mqp->sqp); in _mlx4_ib_destroy_qp()
[all …]
A Dmlx4_ib.h327 struct mlx4_qp mqp; member
718 static inline struct mlx4_ib_qp *to_mibqp(struct mlx4_qp *mqp) in to_mibqp() argument
720 return container_of(mqp, struct mlx4_ib_qp, mqp); in to_mibqp()
850 int mlx4_ib_add_mc(struct mlx4_ib_dev *mdev, struct mlx4_ib_qp *mqp,
908 int mlx4_ib_steer_qp_reg(struct mlx4_ib_dev *mdev, struct mlx4_ib_qp *mqp,
A Dcq.c661 struct mlx4_qp *mqp; in mlx4_ib_poll_one() local
709 (be32_to_cpu(cqe->vlan_my_qpn) & MLX4_CQE_QPN_MASK) != (*cur_qp)->mqp.qpn) { in mlx4_ib_poll_one()
715 mqp = __mlx4_qp_lookup(to_mdev(cq->ibcq.device)->dev, in mlx4_ib_poll_one()
717 *cur_qp = to_mibqp(mqp); in mlx4_ib_poll_one()
/linux/arch/sparc/kernel/
A Dpci_msi.c335 } *mqp; in sparc64_pbm_msi_init() local
355 mqp = of_get_property(pbm->op->dev.of_node, in sparc64_pbm_msi_init()
357 if (!mqp) in sparc64_pbm_msi_init()
358 mqp = of_get_property(pbm->op->dev.of_node, in sparc64_pbm_msi_init()
360 if (!mqp || len != sizeof(struct msiq_prop)) in sparc64_pbm_msi_init()
363 pbm->msiq_first = mqp->first_msiq; in sparc64_pbm_msi_init()
364 pbm->msiq_first_devino = mqp->first_devino; in sparc64_pbm_msi_init()
/linux/drivers/vdpa/mlx5/net/
A Dmlx5_vnet.c81 struct mlx5_core_qp mqp; member
477 vqp->mqp.uid = ndev->mvdev.res.uid; in qp_create()
478 vqp->mqp.qpn = MLX5_GET(create_qp_out, out, qpn); in qp_create()
500 MLX5_SET(destroy_qp_in, in, qpn, vqp->mqp.qpn); in qp_destroy()
503 mlx5_vdpa_warn(&ndev->mvdev, "destroy qp 0x%x\n", vqp->mqp.qpn); in qp_destroy()
864 MLX5_SET(virtio_q, vq_ctx, event_qpn_or_msix, mvq->fwqp.mqp.qpn); in create_virtqueue()
918 return fw ? mvq->vqqp.mqp.qpn : mvq->fwqp.mqp.qpn; in get_rqpn()
923 return fw ? mvq->fwqp.mqp.qpn : mvq->vqqp.mqp.qpn; in get_qpn()
/linux/include/linux/mlx5/
A Dqp.h498 struct mlx5_core_qp mqp; member

Completed in 98 milliseconds