/linux/drivers/net/ethernet/intel/ice/ |
A D | ice_flow.c | 717 hdrs = prof->segs[i].hdrs; in ice_flow_proc_seg_hdrs() 1275 if (segs[i].hdrs != p->segs[i].hdrs || in ice_flow_find_prof_conds() 1277 segs[i].match != p->segs[i].match)) in ice_flow_find_prof_conds() 1397 memcpy(¶ms->prof->segs[i], &segs[i], sizeof(*segs)); in ice_flow_add_prof_sync() 1552 if (!segs) in ice_flow_add_prof() 2037 segs = kcalloc(segs_cnt, sizeof(*segs), GFP_KERNEL); in ice_add_rss_cfg_sync() 2038 if (!segs) in ice_add_rss_cfg_sync() 2116 kfree(segs); in ice_add_rss_cfg_sync() 2171 segs = kcalloc(segs_cnt, sizeof(*segs), GFP_KERNEL); in ice_rem_rss_cfg_sync() 2172 if (!segs) in ice_rem_rss_cfg_sync() [all …]
|
/linux/net/ipv4/ |
A D | udp_offload.c | 96 skb = segs; in __skb_udp_tunnel_segment() 148 return segs; in __skb_udp_tunnel_segment() 186 return segs; in skb_udp_tunnel_segment() 223 seg = segs; in __udpv4_gso_segment_list_csum() 231 return segs; in __udpv4_gso_segment_list_csum() 245 return segs; in __udpv4_gso_segment_list_csum() 293 return segs; in __udp_gso_segment() 300 if (skb_is_gso(segs)) in __udp_gso_segment() 303 seg = segs; in __udp_gso_segment() 363 return segs; in __udp_gso_segment() [all …]
|
A D | tcp_offload.c | 57 struct sk_buff *segs = ERR_PTR(-EINVAL); in tcp_gso_segment() local 89 segs = NULL; in tcp_gso_segment() 98 segs = skb_segment(skb, features); in tcp_gso_segment() 99 if (IS_ERR(segs)) in tcp_gso_segment() 103 segs->ooo_okay = ooo_okay; in tcp_gso_segment() 109 if (skb_is_gso(segs)) in tcp_gso_segment() 110 mss *= skb_shinfo(segs)->gso_segs; in tcp_gso_segment() 114 skb = segs; in tcp_gso_segment() 119 tcp_gso_tstamp(segs, skb_shinfo(gso_skb)->tskey, seq, mss); in tcp_gso_segment() 177 return segs; in tcp_gso_segment()
|
A D | gre_offload.c | 19 struct sk_buff *segs = ERR_PTR(-EINVAL); in gre_gso_segment() local 56 segs = skb_mac_gso_segment(skb, features); in gre_gso_segment() 57 if (IS_ERR_OR_NULL(segs)) { in gre_gso_segment() 63 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL); in gre_gso_segment() 67 skb = segs; in gre_gso_segment() 117 return segs; in gre_gso_segment()
|
A D | esp4_offload.c | 121 struct sk_buff *segs = ERR_PTR(-EINVAL); in xfrm4_transport_gso_segment() local 127 segs = ops->callbacks.gso_segment(skb, features); in xfrm4_transport_gso_segment() 129 return segs; in xfrm4_transport_gso_segment() 137 struct sk_buff *segs = ERR_PTR(-EINVAL); in xfrm4_beet_gso_segment() local 165 segs = ops->callbacks.gso_segment(skb, features); in xfrm4_beet_gso_segment() 167 return segs; in xfrm4_beet_gso_segment()
|
/linux/net/sctp/ |
A D | offload.c | 41 struct sk_buff *segs = ERR_PTR(-EINVAL); in sctp_gso_segment() local 67 segs = NULL; in sctp_gso_segment() 71 segs = skb_segment(skb, (features | NETIF_F_HW_CSUM) & ~NETIF_F_SG); in sctp_gso_segment() 72 if (IS_ERR(segs)) in sctp_gso_segment() 77 for (skb = segs; skb; skb = skb->next) { in sctp_gso_segment() 86 return segs; in sctp_gso_segment()
|
/linux/net/mpls/ |
A D | mpls_gso.c | 22 struct sk_buff *segs = ERR_PTR(-EINVAL); in mpls_gso_segment() local 47 segs = skb_mac_gso_segment(skb, mpls_features); in mpls_gso_segment() 48 if (IS_ERR_OR_NULL(segs)) { in mpls_gso_segment() 53 skb = segs; in mpls_gso_segment() 69 return segs; in mpls_gso_segment()
|
/linux/net/nsh/ |
A D | nsh.c | 79 struct sk_buff *segs = ERR_PTR(-EINVAL); in nsh_gso_segment() local 108 segs = skb_mac_gso_segment(skb, features); in nsh_gso_segment() 109 if (IS_ERR_OR_NULL(segs)) { in nsh_gso_segment() 116 for (skb = segs; skb; skb = skb->next) { in nsh_gso_segment() 125 return segs; in nsh_gso_segment()
|
/linux/net/ipv6/ |
A D | ip6_offload.c | 77 struct sk_buff *segs = ERR_PTR(-EINVAL); in ipv6_gso_segment() local 101 segs = ERR_PTR(-EPROTONOSUPPORT); in ipv6_gso_segment() 116 segs = ops->callbacks.gso_segment(skb, features); in ipv6_gso_segment() 119 if (IS_ERR_OR_NULL(segs)) in ipv6_gso_segment() 122 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL); in ipv6_gso_segment() 124 for (skb = segs; skb; skb = skb->next) { in ipv6_gso_segment() 139 kfree_skb_list(segs); in ipv6_gso_segment() 154 return segs; in ipv6_gso_segment()
|
A D | udp_offload.c | 20 struct sk_buff *segs = ERR_PTR(-EINVAL); in udp6_ufo_fragment() local 33 segs = skb_udp_tunnel_segment(skb, features, true); in udp6_ufo_fragment() 107 segs = skb_segment(skb, features); in udp6_ufo_fragment() 111 return segs; in udp6_ufo_fragment()
|
A D | esp6_offload.c | 156 struct sk_buff *segs = ERR_PTR(-EINVAL); in xfrm6_transport_gso_segment() local 162 segs = ops->callbacks.gso_segment(skb, features); in xfrm6_transport_gso_segment() 164 return segs; in xfrm6_transport_gso_segment() 172 struct sk_buff *segs = ERR_PTR(-EINVAL); in xfrm6_beet_gso_segment() local 204 segs = ops->callbacks.gso_segment(skb, features); in xfrm6_beet_gso_segment() 206 return segs; in xfrm6_beet_gso_segment()
|
/linux/net/sched/ |
A D | sch_tbf.c | 208 struct sk_buff *segs, *nskb; in tbf_segment() local 213 segs = skb_gso_segment(skb, features & ~NETIF_F_GSO_MASK); in tbf_segment() 215 if (IS_ERR_OR_NULL(segs)) in tbf_segment() 219 skb_list_walk_safe(segs, segs, nskb) { in tbf_segment() 220 skb_mark_not_on_list(segs); in tbf_segment() 221 qdisc_skb_cb(segs)->pkt_len = segs->len; in tbf_segment() 222 len += segs->len; in tbf_segment() 223 ret = qdisc_enqueue(segs, q->qdisc, to_free); in tbf_segment()
|
A D | sch_netem.c | 415 struct sk_buff *segs; in netem_segment() local 420 if (IS_ERR_OR_NULL(segs)) { in netem_segment() 425 return segs; in netem_segment() 499 segs = skb->next; in netem_enqueue() 522 skb->next = segs; in netem_enqueue() 595 if (segs) { in netem_enqueue() 602 while (segs) { in netem_enqueue() 603 skb2 = segs->next; in netem_enqueue() 605 qdisc_skb_cb(segs)->pkt_len = segs->len; in netem_enqueue() 606 last_len = segs->len; in netem_enqueue() [all …]
|
A D | sch_cake.c | 1355 u16 segs = 1; in cake_overhead() local 1387 segs = shinfo->gso_segs; in cake_overhead() 1742 struct sk_buff *segs, *nskb; in cake_enqueue() local 1747 if (IS_ERR_OR_NULL(segs)) in cake_enqueue() 1750 skb_list_walk_safe(segs, segs, nskb) { in cake_enqueue() 1751 skb_mark_not_on_list(segs); in cake_enqueue() 1752 qdisc_skb_cb(segs)->pkt_len = segs->len; in cake_enqueue() 1755 segs); in cake_enqueue() 1756 flow_queue_add(flow, segs); in cake_enqueue() 1760 slen += segs->len; in cake_enqueue() [all …]
|
/linux/drivers/infiniband/sw/rdmavt/ |
A D | mr.c | 379 mr->mr.map[m]->segs[n].vaddr = vaddr; in rvt_reg_user_mr() 380 mr->mr.map[m]->segs[n].length = PAGE_SIZE; in rvt_reg_user_mr() 569 mr->mr.map[m]->segs[n].vaddr = (void *)addr; in rvt_set_page() 570 mr->mr.map[m]->segs[n].length = ps; in rvt_set_page() 787 while (off >= mr->map[m]->segs[n].length) { in rvt_lkey_ok() 788 off -= mr->map[m]->segs[n].length; in rvt_lkey_ok() 797 isge->vaddr = mr->map[m]->segs[n].vaddr + off; in rvt_lkey_ok() 894 while (off >= mr->map[m]->segs[n].length) { in rvt_rkey_ok() 895 off -= mr->map[m]->segs[n].length; in rvt_rkey_ok() 904 sge->vaddr = mr->map[m]->segs[n].vaddr + off; in rvt_rkey_ok() [all …]
|
/linux/fs/ |
A D | binfmt_elf_fdpic.c | 773 seg = loadmap->segs; in elf_fdpic_map_file() 794 seg = loadmap->segs; in elf_fdpic_map_file() 819 seg = loadmap->segs; in elf_fdpic_map_file() 840 seg = loadmap->segs; in elf_fdpic_map_file() 876 mseg = loadmap->segs; in elf_fdpic_map_file() 904 seg = loadmap->segs; in elf_fdpic_map_file() 935 seg = params->loadmap->segs; in elf_fdpic_map_file_constdisp_on_uclinux() 1255 elf->e_phnum = segs; in fill_elf_fdpic_header() 1468 int vma_count, segs; in elf_fdpic_core_dump() local 1519 segs++; in elf_fdpic_core_dump() [all …]
|
/linux/include/linux/ |
A D | blk-integrity.h | 68 unsigned int segs) in blk_queue_max_integrity_segments() argument 70 q->limits.max_integrity_segments = segs; in blk_queue_max_integrity_segments() 153 unsigned int segs) in blk_queue_max_integrity_segments() argument
|
/linux/include/rdma/ |
A D | rdmavt_mr.h | 28 struct rvt_seg segs[RVT_SEGSZ]; member 133 sge->vaddr = sge->mr->map[sge->m]->segs[sge->n].vaddr; in rvt_update_sge() 134 sge->length = sge->mr->map[sge->m]->segs[sge->n].length; in rvt_update_sge()
|
/linux/net/xfrm/ |
A D | xfrm_output.c | 619 struct sk_buff *segs, *nskb; in xfrm_output_gso() local 623 segs = skb_gso_segment(skb, 0); in xfrm_output_gso() 625 if (IS_ERR(segs)) in xfrm_output_gso() 626 return PTR_ERR(segs); in xfrm_output_gso() 627 if (segs == NULL) in xfrm_output_gso() 630 skb_list_walk_safe(segs, segs, nskb) { in xfrm_output_gso() 633 skb_mark_not_on_list(segs); in xfrm_output_gso() 634 err = xfrm_output2(net, sk, segs); in xfrm_output_gso()
|
A D | xfrm_device.c | 138 struct sk_buff *segs; in validate_xmit_xfrm() local 143 segs = skb_gso_segment(skb, esp_features); in validate_xmit_xfrm() 144 if (IS_ERR(segs)) { in validate_xmit_xfrm() 150 skb = segs; in validate_xmit_xfrm()
|
/linux/drivers/net/wireguard/ |
A D | device.c | 161 struct sk_buff *segs = skb_gso_segment(skb, 0); in wg_xmit() local 163 if (IS_ERR(segs)) { in wg_xmit() 164 ret = PTR_ERR(segs); in wg_xmit() 168 skb = segs; in wg_xmit()
|
/linux/drivers/net/ethernet/marvell/octeontx2/nic/ |
A D | otx2_txrx.c | 260 for (seg = 0; seg < sg->segs; seg++, seg_addr++) in otx2_free_rcv_seg() 326 if (cqe->sg.segs) in otx2_check_rcv_errors() 363 for (seg = 0; seg < sg->segs; seg++, seg_addr++) { in otx2_rcv_pkt_handler() 563 sg->segs = 0; in otx2_sqe_add_sg() 579 sg->segs++; in otx2_sqe_add_sg() 790 sg->segs = 0; in otx2_sqe_tso_add_sg() 803 sg->segs++; in otx2_sqe_tso_add_sg() 1033 if (cqe->sg.segs > 1) { in otx2_cleanup_rx_cqes() 1112 sg->segs = 1; in otx2_xdp_sqe_add_sg()
|
/linux/drivers/net/ethernet/broadcom/bnxt/ |
A D | bnxt_coredump.c | 52 info->segs = le16_to_cpu(*((__le16 *)(resp + in bnxt_hwrm_dbg_dma_data() 54 if (!info->segs) { in bnxt_hwrm_dbg_dma_data() 59 info->dest_buf_size = info->segs * in bnxt_hwrm_dbg_dma_data() 113 coredump->total_segs = info.segs; in bnxt_hwrm_dbg_coredump_list()
|
/linux/include/net/ |
A D | udp.h | 485 struct sk_buff *segs; in udp_rcv_segment() local 507 segs = __skb_gso_segment(skb, features, false); in udp_rcv_segment() 508 if (IS_ERR_OR_NULL(segs)) { in udp_rcv_segment() 518 return segs; in udp_rcv_segment()
|
/linux/include/uapi/linux/ |
A D | elf-fdpic.h | 30 struct elf32_fdpic_loadseg segs[]; member
|