/linux/drivers/net/ethernet/huawei/hinic/ |
A D | hinic_hw_qp.c | 638 void hinic_sq_write_db(struct hinic_sq *sq, u16 prod_idx, unsigned int wqe_size, in hinic_sq_write_db() 661 unsigned int wqe_size, u16 *prod_idx) in hinic_sq_get_wqe() 677 void hinic_sq_return_wqe(struct hinic_sq *sq, unsigned int wqe_size) in hinic_sq_return_wqe() 692 struct sk_buff *skb, unsigned int wqe_size) in hinic_sq_write_wqe() 716 unsigned int *wqe_size, u16 *cons_idx) in hinic_sq_read_wqebb() 754 unsigned int wqe_size, u16 *cons_idx) in hinic_sq_read_wqe() 769 void hinic_sq_put_wqe(struct hinic_sq *sq, unsigned int wqe_size) in hinic_sq_put_wqe() 800 unsigned int wqe_size, u16 *prod_idx) in hinic_rq_get_wqe() 841 unsigned int wqe_size, in hinic_rq_read_wqe() 876 unsigned int wqe_size, in hinic_rq_read_next_wqe() [all …]
|
A D | hinic_hw_wq.c | 743 struct hinic_hw_wqe *hinic_get_wqe(struct hinic_wq *wq, unsigned int wqe_size, in hinic_get_wqe() 792 void hinic_return_wqe(struct hinic_wq *wq, unsigned int wqe_size) in hinic_return_wqe() 806 void hinic_put_wqe(struct hinic_wq *wq, unsigned int wqe_size) in hinic_put_wqe() 824 struct hinic_hw_wqe *hinic_read_wqe(struct hinic_wq *wq, unsigned int wqe_size, in hinic_read_wqe() 889 unsigned int wqe_size) in hinic_write_wqe()
|
A D | hinic_tx.c | 499 unsigned int wqe_size; in hinic_lb_xmit_frame() local 560 unsigned int wqe_size; in hinic_xmit_frame() local 671 unsigned int wqe_size; in free_all_tx_skbs() local 707 unsigned int wqe_size; in free_tx_poll() local
|
A D | hinic_hw_cmdq.c | 160 unsigned int wqe_size = 0; in cmdq_wqe_size_from_bdlen() local 537 unsigned int bufdesc_len, wqe_size; in clear_wqe_complete_bit() local
|
/linux/drivers/infiniband/hw/qedr/ |
A D | qedr_hsi_rdma.h | 310 u8 wqe_size; member 338 u8 wqe_size; member 374 u8 wqe_size; member 420 u8 wqe_size; member 475 u8 wqe_size; member 498 u8 wqe_size; member 548 u8 wqe_size; member 602 u8 wqe_size; member 628 u8 wqe_size; member 663 u8 wqe_size; member [all …]
|
A D | qedr.h | 433 u8 wqe_size; member 444 u8 wqe_size; member
|
A D | verbs.c | 3274 struct qedr_qp *qp, u8 *wqe_size, in qedr_prepare_sq_inline_data() 3359 static u32 qedr_prepare_sq_sges(struct qedr_qp *qp, u8 *wqe_size, in qedr_prepare_sq_sges()
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/en/ |
A D | txrx.h | 83 static inline void *mlx5e_fetch_wqe(struct mlx5_wq_cyc *wq, u16 pi, size_t wqe_size) in mlx5e_fetch_wqe() 429 static inline u16 mlx5e_stop_room_for_wqe(u16 wqe_size) in mlx5e_stop_room_for_wqe() 450 static inline bool mlx5e_icosq_can_post_wqe(struct mlx5e_icosq *sq, u16 wqe_size) in mlx5e_icosq_can_post_wqe()
|
A D | params.c | 476 int wqe_size = BIT(log_stride_sz) * num_strides; in mlx5e_shampo_get_log_cq_size() local 684 int wqe_size = BIT(log_stride_sz) * num_strides; in mlx5e_shampo_hd_per_wqe() local
|
/linux/drivers/infiniband/sw/rxe/ |
A D | rxe_qp.c | 205 int wqe_size; in rxe_qp_init_req() local 278 int wqe_size; in rxe_qp_init_resp() local
|
/linux/drivers/infiniband/hw/ocrdma/ |
A D | ocrdma_verbs.c | 1923 const struct ib_send_wr *wr, u32 wqe_size) in ocrdma_build_inline_sges() 1965 u32 wqe_size = sizeof(*hdr); in ocrdma_build_send() local 1985 u32 wqe_size = sizeof(*hdr) + sizeof(*ext_rw); in ocrdma_build_write() local 2002 u32 wqe_size = ((wr->num_sge + 1) * sizeof(struct ocrdma_sge)) + in ocrdma_build_read() local 2035 u32 wqe_size = sizeof(*fast_reg) + sizeof(*hdr); in ocrdma_build_reg() local 2204 u32 wqe_size = 0; in ocrdma_build_rqe() local
|
A D | ocrdma.h | 110 u32 wqe_size; member
|
/linux/include/uapi/rdma/ |
A D | ocrdma-abi.h | 55 __u32 wqe_size; member
|
A D | ib_user_verbs.h | 826 __u32 wqe_size; member 845 __u32 wqe_size; member 858 __u32 wqe_size; member
|
/linux/drivers/infiniband/hw/bnxt_re/ |
A D | qplib_fp.h | 95 u16 wqe_size; member 254 u16 wqe_size; member 584 static inline u32 bnxt_qplib_set_rq_max_slot(u32 wqe_size) in bnxt_qplib_set_rq_max_slot()
|
A D | roce_hsi.h | 194 u8 wqe_size; member 217 u8 wqe_size; member 237 u8 wqe_size; member 286 u8 wqe_size; member 309 u8 wqe_size; member 324 u8 wqe_size; member 521 u8 wqe_size; member 537 u8 wqe_size; member
|
/linux/drivers/infiniband/hw/vmw_pvrdma/ |
A D | pvrdma.h | 155 int wqe_size; member 170 int wqe_size; member
|
/linux/drivers/infiniband/hw/mlx5/ |
A D | qp.c | 261 size_t wqe_size = 1 << wq->wqe_shift; in mlx5_ib_read_wqe_rq() local 293 size_t wqe_size = 1 << srq->msrq.wqe_shift; in mlx5_ib_read_wqe_srq() local 354 int wqe_size; in set_rq_size() local 486 static int get_send_sge(struct ib_qp_init_attr *attr, int wqe_size) in get_send_sge() 512 int wqe_size; in calc_sq_size() local
|
A D | odp.c | 1115 int wqe_size = 1 << srq->msrq.wqe_shift; in mlx5_ib_mr_responder_pfault_handler_srq() local 1134 int wqe_size = 1 << wq->wqe_shift; in mlx5_ib_mr_responder_pfault_handler_rq() local
|
A D | wr.c | 652 int wqe_size; in set_sig_data_segment() local
|
/linux/drivers/net/ethernet/ibm/ehea/ |
A D | ehea_qmr.c | 358 int nr_pages, int wqe_size, int act_nr_sges, in ehea_qp_alloc_register()
|
/linux/drivers/infiniband/hw/irdma/ |
A D | uk.c | 1591 enum irdma_status_code irdma_fragcnt_to_wqesize_rq(u32 frag_cnt, u16 *wqe_size) in irdma_fragcnt_to_wqesize_rq()
|
/linux/drivers/net/ethernet/microsoft/mana/ |
A D | gdma_main.c | 1025 u32 wqe_size; in mana_gd_post_work_request() local
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/ |
A D | en.h | 170 #define MLX5E_KLM_MAX_ENTRIES_PER_WQE(wqe_size)\ argument 173 #define MLX5E_KLM_ENTRIES_PER_WQE(wqe_size)\ argument
|
/linux/drivers/infiniband/core/ |
A D | uverbs_cmd.c | 2182 u32 wqe_size, u32 sge_count) in ib_uverbs_unmarshall_recv()
|