/linux/include/rdma/ |
A D | ib_sa.h | 135 __be16 slid; member 153 __be32 slid; member 222 (be32_to_cpu(opa->opa.slid) >= in path_conv_opa_to_ib() 235 ib->ib.slid = 0; in path_conv_opa_to_ib() 238 ib->ib.slid = htons(ntohl(opa->opa.slid)); in path_conv_opa_to_ib() 247 __be32 slid, dlid; in path_conv_ib_to_opa() local 254 slid = htonl(ntohs(ib->ib.slid)); in path_conv_ib_to_opa() 257 opa->opa.slid = slid; in path_conv_ib_to_opa() 541 rec->ib.slid = cpu_to_be16(slid); in sa_path_set_slid() 543 rec->opa.slid = cpu_to_be32(slid); in sa_path_set_slid() [all …]
|
A D | opa_addr.h | 58 static inline bool opa_is_extended_lid(__be32 dlid, __be32 slid) in opa_is_extended_lid() argument 62 (be32_to_cpu(slid) >= in opa_is_extended_lid()
|
/linux/drivers/infiniband/hw/hfi1/ |
A D | ud.c | 71 u32 slid; in ud_loopback() local 79 slid, false))) { in ud_loopback() 83 slid, rdma_ah_get_dlid(ah_attr)); in ud_loopback() 265 u16 len, slid, dlid, pkey; in hfi1_make_ud_req_9B() local 311 slid = be16_to_cpu(IB_LID_PERMISSIVE); in hfi1_make_ud_req_9B() 318 slid = lid; in hfi1_make_ud_req_9B() 329 lrh0, len, dlid, slid); in hfi1_make_ud_req_9B() 340 u32 dlid, slid, nwords, extra_bytes; in hfi1_make_ud_req_16B() local 825 u32 slid = packet->slid; in hfi1_ud_rcv() local 884 slid, dlid); in hfi1_ud_rcv() [all …]
|
A D | ruc.c | 34 u32 slid = packet->slid; in hfi1_ruc_check_hdr() local 63 sc5, slid))) { in hfi1_ruc_check_hdr() 65 slid, dlid); in hfi1_ruc_check_hdr() 69 if (slid != rdma_ah_get_dlid(&qp->alt_ah_attr) || in hfi1_ruc_check_hdr() 100 sc5, slid))) { in hfi1_ruc_check_hdr() 102 slid, dlid); in hfi1_ruc_check_hdr() 106 if ((slid != rdma_ah_get_dlid(&qp->remote_ah_attr)) || in hfi1_ruc_check_hdr() 240 u32 slid; in hfi1_make_ruc_header_16B() local 291 slid = be32_to_cpu(OPA_LID_PERMISSIVE); in hfi1_make_ruc_header_16B() 293 slid = ppd->lid | in hfi1_make_ruc_header_16B() [all …]
|
A D | trace_ibhdrs.h | 80 u16 *len, u32 *dlid, u32 *slid); 89 u32 *dlid, u32 *slid); 95 u16 len, u16 pkey, u32 dlid, u32 slid); 144 __field(u32, slid) 170 &__entry->slid); 196 &__entry->slid); 240 __entry->slid), 303 __field(u32, slid) 330 &__entry->slid); 361 &__entry->slid); [all …]
|
A D | hfi.h | 347 u32 slid; member 970 u32 slid; member 1700 u16 pkey, u32 slid, u32 dlid, u8 sc5, 1792 u32 slid) in ingress_pkey_table_fail() argument 1799 dd->err_info_rcv_constraint.slid = slid; in ingress_pkey_table_fail() 1835 ingress_pkey_table_fail(ppd, pkey, slid); in ingress_pkey_check() 1846 u8 sc5, u16 slid) in rcv_pkey_check() argument 2574 struct ib_grh *grh, u32 slid, in hfi1_make_ext_grh() argument 2611 u16 dlid, u16 slid) in hfi1_make_ib_hdr() argument 2616 hdr->lrh[3] = cpu_to_be16(slid); in hfi1_make_ib_hdr() [all …]
|
A D | trace.c | 157 u16 *len, u32 *dlid, u32 *slid) in hfi1_trace_parse_9b_hdr() argument 165 *slid = ib_get_slid(hdr); in hfi1_trace_parse_9b_hdr() 172 u32 *dlid, u32 *slid) in hfi1_trace_parse_16b_hdr() argument 184 *slid = hfi1_16B_get_slid(hdr); in hfi1_trace_parse_16b_hdr() 195 u16 len, u16 pkey, u32 dlid, u32 slid) in hfi1_trace_fmt_lrh() argument 199 trace_seq_printf(p, LRH_PRN, len, sc, dlid, slid); in hfi1_trace_fmt_lrh()
|
A D | ipoib_tx.c | 266 u16 slid; in hfi1_ipoib_build_ib_tx_headers() local 304 slid = be16_to_cpu(IB_LID_PERMISSIVE); in hfi1_ipoib_build_ib_tx_headers() 311 slid = lid; in hfi1_ipoib_build_ib_tx_headers() 313 slid = be16_to_cpu(IB_LID_PERMISSIVE); in hfi1_ipoib_build_ib_tx_headers() 322 hfi1_make_ib_hdr(&sdma_hdr->hdr.ibh, lrh0, dwords, dlid, slid); in hfi1_ipoib_build_ib_tx_headers()
|
A D | driver.c | 416 u32 rlid, slid, dlid = 0; in hfi1_process_ecn_slowpath() local 426 slid = hfi1_16B_get_slid(pkt->hdr); in hfi1_process_ecn_slowpath() 437 slid = ib_get_slid(pkt->hdr); in hfi1_process_ecn_slowpath() 448 rlid = slid; in hfi1_process_ecn_slowpath() 454 rlid = slid; in hfi1_process_ecn_slowpath() 1455 if ((!packet->slid) || (!packet->dlid)) in hfi1_bypass_ingress_pkt_check() 1509 packet->slid = ib_get_slid(hdr); in hfi1_setup_9B_packet() 1590 packet->slid = hfi1_16B_get_slid(packet->hdr); in hfi1_setup_bypass_packet()
|
A D | verbs.c | 450 packet->slid, true); in hfi1_do_pkey_check() 1116 int egress_pkey_check(struct hfi1_pportdata *ppd, u32 slid, u16 pkey, in egress_pkey_check() argument 1160 dd->err_info_xmit_constraint.slid = slid; in egress_pkey_check() 1221 u32 slid; in hfi1_verbs_send() local 1234 slid = hfi1_16B_get_slid(hdr); in hfi1_verbs_send() 1244 slid = ib_get_slid(hdr); in hfi1_verbs_send() 1254 ret = egress_pkey_check(dd->pport, slid, pkey, in hfi1_verbs_send()
|
A D | user_sdma.c | 319 u32 slid; in hfi1_user_sdma_process_request() local 453 slid = be16_to_cpu(req->hdr.lrh[3]); in hfi1_user_sdma_process_request() 454 if (egress_pkey_check(dd->pport, slid, pkey, sc, PKEY_CHECK_INVALID)) { in hfi1_user_sdma_process_request()
|
/linux/drivers/infiniband/core/ |
A D | uverbs_marshall.c | 143 dst->slid = htons(ntohl(sa_path_get_slid(src))); in __ib_copy_path_rec_to_user() 178 u32 slid, dlid; in ib_copy_path_rec_from_user() local 184 slid = opa_get_lid_from_gid((union ib_gid *)src->sgid); in ib_copy_path_rec_from_user() 188 slid = ntohs(src->slid); in ib_copy_path_rec_from_user() 195 sa_path_set_slid(dst, slid); in ib_copy_path_rec_from_user()
|
A D | mad_rmpp.c | 66 u32 slid; member 311 rmpp_recv->slid = mad_recv_wc->wc->slid; in create_rmpp_recv() 332 rmpp_recv->slid == mad_recv_wc->wc->slid && in find_rmpp_recv() 865 if (rmpp_recv->slid == rdma_ah_get_dlid(&ah_attr)) { in init_newwin()
|
/linux/drivers/infiniband/hw/mthca/ |
A D | mthca_mad.c | 205 u16 slid = in_wc ? ib_lid_cpu16(in_wc->slid) : be16_to_cpu(IB_LID_PERMISSIVE); in mthca_process_mad() local 210 if (in->mad_hdr.method == IB_MGMT_METHOD_TRAP && !slid) { in mthca_process_mad()
|
/linux/include/trace/events/ |
A D | ib_mad.h | 205 __field(u32, slid) 224 __entry->slid = wc->slid; 247 __entry->slid, __entry->src_qp, __entry->sl, __entry->pkey
|
/linux/drivers/infiniband/hw/mlx5/ |
A D | mad.c | 227 u16 slid; in mlx5_ib_process_mad() local 230 slid = in_wc ? ib_lid_cpu16(in_wc->slid) : in mlx5_ib_process_mad() 233 if (method == IB_MGMT_METHOD_TRAP && !slid) in mlx5_ib_process_mad()
|
A D | cq.c | 240 wc->slid = be16_to_cpu(cqe->slid); in handle_responder() 245 wc->slid = 0; in handle_responder()
|
/linux/include/uapi/rdma/ |
A D | ib_user_sa.h | 59 __be16 slid; member
|
A D | vmw_pvrdma-abi.h | 301 __u16 slid; member
|
/linux/drivers/infiniband/ulp/opa_vnic/ |
A D | opa_vnic_encap.c | 73 static inline void opa_vnic_make_header(u8 *hdr, u32 slid, u32 dlid, u16 len, in opa_vnic_make_header() argument 85 h[0] |= (slid & OPA_16B_LID_MASK); in opa_vnic_make_header() 86 h[2] |= ((slid >> (20 - OPA_16B_SLID_HIGH_SHFT)) & OPA_16B_SLID_MASK); in opa_vnic_make_header()
|
/linux/drivers/infiniband/hw/vmw_pvrdma/ |
A D | pvrdma_cq.c | 365 wc->slid = cqe->slid; in pvrdma_poll_one()
|
/linux/drivers/infiniband/hw/mlx4/ |
A D | cq.c | 596 wc->slid = 0; in use_tunnel_data() 602 wc->slid = be16_to_cpu(hdr->tun.slid_mac_47_32); in use_tunnel_data() 856 wc->slid = 0; in mlx4_ib_poll_one() 868 wc->slid = be16_to_cpu(cqe->rlid); in mlx4_ib_poll_one()
|
A D | mad.c | 172 in_modifier |= ib_lid_cpu16(in_wc->slid) << 16; in mlx4_MAD_IFC() 646 tun_mad->hdr.slid_mac_47_32 = ib_lid_be16(wc->slid); in mlx4_ib_send_to_slave() 826 u16 slid, prev_lid = 0; in ib_process_mad() local 830 slid = in_wc ? ib_lid_cpu16(in_wc->slid) : be16_to_cpu(IB_LID_PERMISSIVE); in ib_process_mad() 832 if (in_mad->mad_hdr.method == IB_MGMT_METHOD_TRAP && slid == 0) { in ib_process_mad()
|
/linux/drivers/infiniband/sw/rdmavt/ |
A D | cq.c | 90 uqueue[head].slid = ib_lid_cpu16(entry->slid); in rvt_cq_enter()
|
/linux/drivers/infiniband/hw/qib/ |
A D | qib_ud.c | 208 wc.slid = ppd->lid | (rdma_ah_get_path_bits(ah_attr) & in qib_ud_loopback() 568 wc.slid = be16_to_cpu(hdr->lrh[3]); in qib_ud_rcv()
|