/linux/drivers/net/ethernet/sfc/ |
A D | tx_tso.c | 161 EFX_WARN_ON_ONCE_PARANOID((PTR_DIFF(tcp_hdr(skb), skb->data) + in efx_tso_check_protocol() 162 (tcp_hdr(skb)->doff << 2u)) > in efx_tso_check_protocol() 179 header_len = st->tcp_off + (tcp_hdr(skb)->doff << 2u); in tso_start() 190 st->seqnum = ntohl(tcp_hdr(skb)->seq); in tso_start() 192 EFX_WARN_ON_ONCE_PARANOID(tcp_hdr(skb)->urg); in tso_start() 193 EFX_WARN_ON_ONCE_PARANOID(tcp_hdr(skb)->syn); in tso_start() 194 EFX_WARN_ON_ONCE_PARANOID(tcp_hdr(skb)->rst); in tso_start() 307 tcp_flags = ((u8 *)tcp_hdr(skb))[TCP_FLAGS_OFFSET] & ~tcp_flags_mask; in tso_start_new_packet()
|
A D | tx_common.c | 324 (tcp_hdr(skb)->doff << 2u); in efx_tx_tso_header_length()
|
/linux/net/ipv4/ |
A D | tcp_offload.c | 40 struct tcphdr *th = tcp_hdr(skb); in tcp4_gso_segment() 69 th = tcp_hdr(skb); in tcp_gso_segment() 115 th = tcp_hdr(skb); in tcp_gso_segment() 140 th = tcp_hdr(skb); in tcp_gso_segment() 224 th2 = tcp_hdr(p); in tcp_gro_receive() 290 struct tcphdr *th = tcp_hdr(skb); in tcp_gro_complete() 325 struct tcphdr *th = tcp_hdr(skb); in tcp4_gro_complete()
|
A D | syncookies.c | 179 const struct tcphdr *th = tcp_hdr(skb); in cookie_v4_init_sequence() 328 const struct tcphdr *th = tcp_hdr(skb); in cookie_v4_check()
|
A D | tcp_ipv4.c | 98 tcp_hdr(skb)->dest, in tcp_v4_init_seq() 99 tcp_hdr(skb)->source); in tcp_v4_init_seq() 632 struct tcphdr *th = tcp_hdr(skb); in __tcp_v4_send_check() 669 const struct tcphdr *th = tcp_hdr(skb); in tcp_v4_send_reset() 848 const struct tcphdr *th = tcp_hdr(skb); in tcp_v4_send_ack() 1363 const struct tcphdr *th = tcp_hdr(skb); in tcp_v4_md5_hash_skb() 1423 const struct tcphdr *th = tcp_hdr(skb); in tcp_v4_inbound_md5_hash() 1666 const struct tcphdr *th = tcp_hdr(skb); in tcp_v4_cookie_check() 1779 th = tcp_hdr(skb); in tcp_v4_early_demux()
|
A D | tcp_minisocks.c | 515 newtp->snd_wnd = ntohs(tcp_hdr(skb)->window) << newtp->rx_opt.snd_wscale; in tcp_create_openreq_child() 569 const struct tcphdr *th = tcp_hdr(skb); in tcp_check_req()
|
A D | tcp_input.c | 321 if (tcp_hdr(skb)->cwr) { in tcp_ecn_accept_cwr() 3535 u32 nwin = ntohs(tcp_hdr(skb)->window); in tcp_ack_update_window() 3537 if (likely(!tcp_hdr(skb)->syn)) in tcp_ack_update_window() 3597 !tcp_hdr(skb)->syn) in tcp_oow_rate_limited() 4014 const struct tcphdr *th = tcp_hdr(skb); in tcp_parse_options() 4239 const struct tcphdr *th = tcp_hdr(skb); in tcp_disordered_ack() 4999 __skb_pull(skb, tcp_hdr(skb)->doff * 4); in tcp_data_queue() 6382 const struct tcphdr *th = tcp_hdr(skb); in tcp_rcv_state_process() 6662 const struct tcphdr *th = tcp_hdr(skb); in tcp_ecn_create_request() 6700 ireq->ir_rmt_port = tcp_hdr(skb)->source; in tcp_openreq_init() [all …]
|
A D | tcp_fastopen.c | 268 tp->snd_wnd = ntohs(tcp_hdr(skb)->window); in tcp_fastopen_create_child()
|
/linux/net/ipv6/ |
A D | tcpv6_offload.c | 32 struct tcphdr *th = tcp_hdr(skb); in tcp6_gro_complete() 54 struct tcphdr *th = tcp_hdr(skb); in tcp6_gso_segment()
|
A D | syncookies.c | 112 const struct tcphdr *th = tcp_hdr(skb); in cookie_v6_init_sequence() 135 const struct tcphdr *th = tcp_hdr(skb); in cookie_v6_check()
|
A D | tcp_ipv6.c | 120 tcp_hdr(skb)->dest, in tcp_v6_init_seq() 121 tcp_hdr(skb)->source); in tcp_v6_init_seq() 734 const struct tcphdr *th = tcp_hdr(skb); in tcp_v6_md5_hash_skb() 783 const struct tcphdr *th = tcp_hdr(skb); in tcp_v6_inbound_md5_hash() 896 const struct tcphdr *th = tcp_hdr(skb); in tcp_v6_send_response() 1022 const struct tcphdr *th = tcp_hdr(skb); in tcp_v6_send_reset() 1179 const struct tcphdr *th = tcp_hdr(skb); in tcp_v6_cookie_check() 1866 th = tcp_hdr(skb); in tcp_v6_early_demux()
|
/linux/include/net/ |
A D | ip6_checksum.h | 66 struct tcphdr *th = tcp_hdr(skb); in __tcp_v6_send_check() 82 struct tcphdr *th = tcp_hdr(skb); in tcp_v6_gso_csum_prep()
|
/linux/include/linux/ |
A D | tcp.h | 24 static inline struct tcphdr *tcp_hdr(const struct sk_buff *skb) in tcp_hdr() function 36 return __tcp_hdrlen(tcp_hdr(skb)); in tcp_hdrlen() 51 return (tcp_hdr(skb)->doff - 5) * 4; in tcp_optlen()
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/en_accel/ |
A D | tls_rxtx.c | 162 th = tcp_hdr(nskb); in mlx5e_tls_complete_sync_skb() 191 u32 tcp_seq = ntohl(tcp_hdr(skb)->seq); in mlx5e_tls_handle_ooo() 280 skb_seq = ntohl(tcp_hdr(skb)->seq); in mlx5e_tls_handle_tx_skb()
|
A D | ktls_tx.c | 465 seq = ntohl(tcp_hdr(skb)->seq); in mlx5e_ktls_handle_tx_skb()
|
/linux/net/mptcp/ |
A D | syncookies.c | 41 struct tcphdr *th = tcp_hdr(skb); in mptcp_join_entry_hash()
|
A D | options.c | 37 if (skb->len > tcp_hdr(skb)->doff << 2) in mptcp_parse_option() 360 const struct tcphdr *th = tcp_hdr(skb); in mptcp_get_options()
|
/linux/net/core/ |
A D | tso.c | 79 tso->tcp_seq = (tlen != sizeof(struct udphdr)) ? ntohl(tcp_hdr(skb)->seq) : 0; in tso_start()
|
/linux/net/tls/ |
A D | tls_device_fallback.c | 169 struct tcphdr *th = tcp_hdr(skb); in update_chksum() 237 u32 tcp_seq = ntohl(tcp_hdr(skb)->seq); in fill_sg_in()
|
/linux/net/openvswitch/ |
A D | flow.c | 581 struct tcphdr *tcp = tcp_hdr(skb); in key_extract_l3l4() 705 struct tcphdr *tcp = tcp_hdr(skb); in key_extract_l3l4()
|
A D | actions.c | 354 inet_proto_csum_replace4(&tcp_hdr(skb)->check, skb, in update_ip_l4_checksum() 386 inet_proto_csum_replace16(&tcp_hdr(skb)->check, skb, in update_ipv6_checksum() 692 th = tcp_hdr(skb); in set_tcp()
|
/linux/security/ |
A D | lsm_audit.c | 61 struct tcphdr *th = tcp_hdr(skb); in ipv4_skb_to_auditdata()
|
/linux/drivers/net/ethernet/sun/ |
A D | sunvnet_common.c | 321 struct tcphdr *ptcp = tcp_hdr(skb); in vnet_fullcsum_ipv4() 354 struct tcphdr *ptcp = tcp_hdr(skb); in vnet_fullcsum_ipv6() 1239 hlen += tcp_hdr(skb)->doff * 4; in vnet_handle_offloads()
|
/linux/drivers/net/ |
A D | thunderbolt.c | 980 tucso = dest + ((void *)&(tcp_hdr(skb)->check) - data); in tbnet_xmit_csum_and_map() 990 tucso = dest + ((void *)&(tcp_hdr(skb)->check) - data); in tbnet_xmit_csum_and_map()
|
/linux/drivers/net/ethernet/qlogic/qede/ |
A D | qede_fp.c | 893 th = tcp_hdr(skb); in qede_gro_ip_csum() 907 th = tcp_hdr(skb); in qede_gro_ipv6_csum()
|