/linux/net/ipv4/ |
A D | tcp_fastopen.c | 318 __NET_INC_STATS(sock_net(sk), in tcp_fastopen_queue_check() 335 return (sock_net(sk)->ipv4.sysctl_tcp_fastopen & flag) || in tcp_fastopen_no_cookie() 350 int tcp_fastopen = sock_net(sk)->ipv4.sysctl_tcp_fastopen; in tcp_try_fastopen() 356 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPFASTOPENCOOKIEREQD); in tcp_try_fastopen() 375 NET_INC_STATS(sock_net(sk), in tcp_try_fastopen() 393 NET_INC_STATS(sock_net(sk), in tcp_try_fastopen() 398 NET_INC_STATS(sock_net(sk), in tcp_try_fastopen() 402 NET_INC_STATS(sock_net(sk), in tcp_try_fastopen() 490 struct net *net = sock_net(sk); in tcp_fastopen_active_disable() 492 if (!sock_net(sk)->ipv4.sysctl_tcp_fastopen_blackhole_timeout) in tcp_fastopen_active_disable() [all …]
|
A D | tcp_timer.c | 75 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTONTIMEOUT); in tcp_write_err() 126 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTONMEMORY); in tcp_out_of_resources() 130 if (!check_net(sock_net(sk))) { in tcp_out_of_resources() 162 const struct net *net = sock_net(sk); in tcp_mtu_probing() 235 struct net *net = sock_net(sk); in tcp_write_timeout() 282 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPTIMEOUTREHASH); in tcp_write_timeout() 318 __NET_INC_STATS(sock_net(sk), LINUX_MIB_DELAYEDACKS); in tcp_delack_timer_handler() 383 max_probes = sock_net(sk)->ipv4.sysctl_tcp_retries2; in tcp_probe_timer() 451 struct net *net = sock_net(sk); in tcp_retransmit_timer() 511 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPTIMEOUTS); in tcp_retransmit_timer() [all …]
|
A D | syncookies.c | 338 if (!sock_net(sk)->ipv4.sysctl_tcp_syncookies || !th->ack || th->rst) in cookie_v4_check() 346 __NET_INC_STATS(sock_net(sk), LINUX_MIB_SYNCOOKIESFAILED); in cookie_v4_check() 350 __NET_INC_STATS(sock_net(sk), LINUX_MIB_SYNCOOKIESRECV); in cookie_v4_check() 354 tcp_parse_options(sock_net(sk), skb, &tcp_opt, 0, NULL); in cookie_v4_check() 357 tsoff = secure_tcp_ts_off(sock_net(sk), in cookie_v4_check() 363 if (!cookie_timestamp_decode(sock_net(sk), &tcp_opt)) in cookie_v4_check() 399 RCU_INIT_POINTER(ireq->ireq_opt, tcp_v4_save_options(sock_net(sk), skb)); in cookie_v4_check() 420 rt = ip_route_output_key(sock_net(sk), &fl4); in cookie_v4_check() 440 ireq->ecn_ok = cookie_ecn_ok(&tcp_opt, sock_net(sk), &rt->dst); in cookie_v4_check()
|
A D | tcp_input.c | 576 struct net *net = sock_net(sk); in tcp_clamp_window() 1059 NET_INC_STATS(sock_net(sk), in tcp_check_sack_reordering() 2138 struct net *net = sock_net(sk); in tcp_enter_loss() 2572 NET_INC_STATS(sock_net(sk), in tcp_try_undo_loss() 3610 struct net *net = sock_net(sk); in tcp_send_challenge_ack() 3681 NET_INC_STATS(sock_net(sk), in tcp_process_tlp_ack() 4835 NET_INC_STATS(sock_net(sk), in tcp_data_queue_ofo() 4854 NET_INC_STATS(sock_net(sk), in tcp_data_queue_ofo() 6065 NET_INC_STATS(sock_net(sk), in tcp_rcv_fastopen_synack() 6142 NET_INC_STATS(sock_net(sk), in tcp_rcv_synsent_state_process() [all …]
|
A D | tcp_minisocks.c | 329 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPTIMEWAITOVERFLOW); in tcp_time_wait() 546 __TCP_INC_STATS(sock_net(sk), TCP_MIB_PASSIVEOPENS); in tcp_create_openreq_child() 576 tcp_parse_options(sock_net(sk), skb, &tmp_opt, 0, NULL); in tcp_check_req() 618 if (!tcp_oow_rate_limited(sock_net(sk), skb, in tcp_check_req() 708 !tcp_oow_rate_limited(sock_net(sk), skb, in tcp_check_req() 713 __NET_INC_STATS(sock_net(sk), LINUX_MIB_PAWSESTABREJECTED); in tcp_check_req() 732 __TCP_INC_STATS(sock_net(sk), TCP_MIB_ATTEMPTFAILS); in tcp_check_req() 755 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPDEFERACCEPTDROP); in tcp_check_req() 783 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPMIGRATEREQFAILURE); in tcp_check_req() 785 if (!sock_net(sk)->ipv4.sysctl_tcp_abort_on_overflow) { in tcp_check_req() [all …]
|
A D | inet_hashtables.c | 51 return inet6_ehashfn(sock_net(sk), in sk_ehashfn() 55 return inet_ehashfn(sock_net(sk), in sk_ehashfn() 135 const int bhash = inet_bhashfn(sock_net(sk), port, in __inet_inherit_port() 156 if (net_eq(ib_net(tb), sock_net(sk)) && in __inet_inherit_port() 162 sock_net(sk), head, port, in __inet_inherit_port() 185 hash = ipv6_portaddr_hash(sock_net(sk), in inet_lhash2_bucket_sk() 190 hash = ipv4_portaddr_hash(sock_net(sk), in inet_lhash2_bucket_sk() 450 struct net *net = sock_net(sk); in __inet_check_established() 492 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, 1); in __inet_check_established() 526 struct net *net = sock_net(sk); in inet_ehash_lookup_by_sk() [all …]
|
A D | udp.c | 140 if (net_eq(sock_net(sk2), net) && in udp_lib_lport_inuse() 177 if (net_eq(sock_net(sk2), net) && in udp_lib_lport_inuse2() 200 struct net *net = sock_net(sk); in udp_reuseport_add_sock() 205 if (net_eq(sock_net(sk2), net) && in udp_reuseport_add_sock() 235 struct net *net = sock_net(sk); in udp_lib_get_port() 365 if (!net_eq(sock_net(sk), net) || in compute_score() 969 UDP_INC_STATS(sock_net(sk), in udp_send_skb() 974 UDP_INC_STATS(sock_net(sk), in udp_send_skb() 1305 UDP_INC_STATS(sock_net(sk), in udp_sendmsg() 1901 UDP_INC_STATS(sock_net(sk), in udp_recvmsg() [all …]
|
A D | tcp_output.c | 276 NET_INC_STATS(sock_net(sk), in tcp_select_window() 299 NET_INC_STATS(sock_net(sk), in tcp_select_window() 348 if (sock_net(sk)->ipv4.sysctl_tcp_ecn_fallback) in tcp_ecn_clear_syn() 1385 TCP_ADD_STATS(sock_net(sk), TCP_MIB_OUTSEGS, in __tcp_transmit_skb() 1760 struct net *net = sock_net(sk); in tcp_mtup_init() 1985 sock_net(sk)->ipv4.sysctl_tcp_min_tso_segs; in tcp_tso_segs() 2269 struct net *net = sock_net(sk); in tcp_mtu_check_reprobe() 2324 struct net *net = sock_net(sk); in tcp_mtu_probe() 2776 NET_INC_STATS(sock_net(sk), in skb_still_in_host_queue() 3994 NET_INC_STATS(sock_net(sk), mib); in tcp_xmit_probe_skb() [all …]
|
A D | datagram.c | 41 if (!oif || netif_index_is_l3_master(sock_net(sk), oif)) in __ip4_datagram_connect() 54 IP_INC_STATS(sock_net(sk), IPSTATS_MIB_OUTNOROUTES); in __ip4_datagram_connect() 118 rt = ip_route_output_ports(sock_net(sk), &fl4, sk, daddr, in ip4_datagram_release_cb()
|
A D | raw_diag.c | 99 net = sock_net(in_skb->sk); in raw_diag_dump_one() 143 struct net *net = sock_net(skb->sk); in raw_diag_dump() 164 if (!net_eq(sock_net(sk), net)) in raw_diag_dump() 201 struct net *net = sock_net(in_skb->sk); in raw_diag_destroy()
|
/linux/net/can/ |
A D | raw.c | 275 if (!net_eq(dev_net(dev), sock_net(sk))) in raw_notify() 390 dev = dev_get_by_index(sock_net(sk), ro->ifindex); in raw_release() 396 raw_disable_allfilters(sock_net(sk), NULL, sk); in raw_release() 455 err = raw_enable_allfilters(sock_net(sk), dev, sk); in raw_bind() 470 dev = dev_get_by_index(sock_net(sk), in raw_bind() 478 raw_disable_allfilters(sock_net(sk), NULL, sk); in raw_bind() 565 err = raw_enable_filters(sock_net(sk), dev, sk, in raw_setsockopt() 568 err = raw_enable_filters(sock_net(sk), dev, sk, in raw_setsockopt() 624 err = raw_enable_errfilter(sock_net(sk), dev, sk, in raw_setsockopt() 631 raw_disable_errfilter(sock_net(sk), dev, sk, in raw_setsockopt() [all …]
|
/linux/net/ipv6/ |
A D | syncookies.c | 144 if (!sock_net(sk)->ipv4.sysctl_tcp_syncookies || !th->ack || th->rst) in cookie_v6_check() 152 __NET_INC_STATS(sock_net(sk), LINUX_MIB_SYNCOOKIESFAILED); in cookie_v6_check() 156 __NET_INC_STATS(sock_net(sk), LINUX_MIB_SYNCOOKIESRECV); in cookie_v6_check() 160 tcp_parse_options(sock_net(sk), skb, &tcp_opt, 0, NULL); in cookie_v6_check() 163 tsoff = secure_tcpv6_ts_off(sock_net(sk), in cookie_v6_check() 169 if (!cookie_timestamp_decode(sock_net(sk), &tcp_opt)) in cookie_v6_check() 238 dst = ip6_dst_lookup_flow(sock_net(sk), sk, &fl6, final_p); in cookie_v6_check() 256 ireq->ecn_ok = cookie_ecn_ok(&tcp_opt, sock_net(sk), dst); in cookie_v6_check()
|
/linux/net/vmw_vsock/ |
A D | diag.c | 59 net = sock_net(skb->sk); in vsock_diag_dump() 79 if (!net_eq(sock_net(sk), net)) in vsock_diag_dump() 113 if (!net_eq(sock_net(sk), net)) in vsock_diag_dump() 144 struct net *net = sock_net(skb->sk); in vsock_diag_handler_dump()
|
/linux/net/ieee802154/ |
A D | socket.c | 136 dev_load(sock_net(sk), ifr.ifr_name); in ieee802154_dev_ioctl() 137 dev = dev_get_by_name(sock_net(sk), ifr.ifr_name); in ieee802154_dev_ioctl() 177 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, 1); in raw_hash() 187 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, -1); in raw_unhash() 213 dev = ieee802154_get_dev(sock_net(sk), &addr); in raw_bind() 257 dev = dev_get_by_index(sock_net(sk), sk->sk_bound_dev_if); in raw_sendmsg() 456 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, 1); in dgram_hash() 466 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, -1); in dgram_unhash() 503 dev = ieee802154_get_dev(sock_net(sk), &haddr); in dgram_bind() 623 dev = ieee802154_get_dev(sock_net(sk), &ro->src_addr); in dgram_sendmsg() [all …]
|
/linux/net/tls/ |
A D | tls_main.c | 284 TLS_DEC_STATS(sock_net(sk), LINUX_MIB_TLSCURRTXSW); in tls_sk_proto_cleanup() 287 TLS_DEC_STATS(sock_net(sk), LINUX_MIB_TLSCURRTXDEVICE); in tls_sk_proto_cleanup() 292 TLS_DEC_STATS(sock_net(sk), LINUX_MIB_TLSCURRRXSW); in tls_sk_proto_cleanup() 295 TLS_DEC_STATS(sock_net(sk), LINUX_MIB_TLSCURRRXDEVICE); in tls_sk_proto_cleanup() 639 TLS_INC_STATS(sock_net(sk), LINUX_MIB_TLSTXDEVICE); in do_tls_setsockopt_conf() 640 TLS_INC_STATS(sock_net(sk), LINUX_MIB_TLSCURRTXDEVICE); in do_tls_setsockopt_conf() 645 TLS_INC_STATS(sock_net(sk), LINUX_MIB_TLSTXSW); in do_tls_setsockopt_conf() 646 TLS_INC_STATS(sock_net(sk), LINUX_MIB_TLSCURRTXSW); in do_tls_setsockopt_conf() 653 TLS_INC_STATS(sock_net(sk), LINUX_MIB_TLSRXDEVICE); in do_tls_setsockopt_conf() 659 TLS_INC_STATS(sock_net(sk), LINUX_MIB_TLSRXSW); in do_tls_setsockopt_conf() [all …]
|
/linux/drivers/infiniband/core/ |
A D | nldev.c | 1031 device = ib_device_get_by_index(sock_net(skb->sk), index); in nldev_get_doit() 1075 device = ib_device_get_by_index(sock_net(skb->sk), index); in nldev_set_doit() 1169 device = ib_device_get_by_index(sock_net(skb->sk), index); in nldev_port_get_doit() 1249 if (fill_port_info(skb, device, p, sock_net(skb->sk))) { in nldev_port_get_dumpit() 1278 device = ib_device_get_by_index(sock_net(skb->sk), index); in nldev_res_get_doit() 1421 device = ib_device_get_by_index(sock_net(skb->sk), index); in res_get_common_doit() 1516 device = ib_device_get_by_index(sock_net(skb->sk), index); in res_get_common_dumpit() 1703 ndev = dev_get_by_name(sock_net(skb->sk), ndev_name); in nldev_newlink() 1738 device = ib_device_get_by_index(sock_net(skb->sk), index); in nldev_dellink() 1772 ibdev = ib_device_get_by_index(sock_net(skb->sk), index); in nldev_get_chardev() [all …]
|
/linux/net/mptcp/ |
A D | pm_netlink.c | 371 struct net *net = sock_net(sk); in mptcp_pm_alloc_anno_list() 471 pernet = net_generic(sock_net(sk), pm_nl_pernet_id); in mptcp_pm_create_subflow_or_signal_addr() 550 pernet = net_generic(sock_net(sk), pm_nl_pernet_id); in fill_local_addresses_vec() 677 __MPTCP_INC_STATS(sock_net(sk), MPTCP_MIB_MPPRIOTX); in mptcp_pm_nl_mp_prio_send_ack() 739 __MPTCP_INC_STATS(sock_net(sk), rm_type); in mptcp_pm_nl_rm_addr_or_subflow() 1006 struct net *net = sock_net(sk); in mptcp_pm_nl_subflow_chk_stale() 1564 struct net *net = sock_net(msg->sk); in mptcp_nl_cmd_dump_addrs() 1711 struct net *net = sock_net(skb->sk); in mptcp_nl_cmd_set_flags() 1867 struct net *net = sock_net((const struct sock *)msk); in mptcp_event_addr_removed() 1899 struct net *net = sock_net((const struct sock *)msk); in mptcp_event_addr_announced() [all …]
|
A D | subflow.c | 35 MPTCP_INC_STATS(sock_net(req_to_sk(req)), field); in SUBFLOW_REQ_INC_STATS() 416 MPTCP_INC_STATS(sock_net(sk), in subflow_finish_connect() 451 MPTCP_INC_STATS(sock_net(sk), MPTCP_MIB_JOINACKMAC); in subflow_finish_connect() 466 MPTCP_INC_STATS(sock_net(sk), MPTCP_MIB_JOINSYNACKRX); in subflow_finish_connect() 918 MPTCP_INC_STATS(sock_net(ssk), MPTCP_MIB_DATACSUMERR); in validate_data_csum() 920 MPTCP_INC_STATS(sock_net(ssk), MPTCP_MIB_MPFAILTX); in validate_data_csum() 970 MPTCP_INC_STATS(sock_net(ssk), MPTCP_MIB_INFINITEMAPRX); in get_mapping_status() 1028 MPTCP_INC_STATS(sock_net(ssk), MPTCP_MIB_DSSNOMATCH); in get_mapping_status() 1060 MPTCP_INC_STATS(sock_net(ssk), MPTCP_MIB_DSSTCPMISMATCH); in get_mapping_status() 1081 MPTCP_INC_STATS(sock_net(ssk), MPTCP_MIB_DUPDATA); in mptcp_subflow_discard_data() [all …]
|
/linux/net/phonet/ |
A D | socket.c | 79 if (!net_eq(sock_net(sknode), net)) in pn_find_sock_by_sa() 116 if (!net_eq(sock_net(sknode), net)) in pn_deliver_sock_broadcast() 175 if (saddr && phonet_address_lookup(sock_net(sk), saddr)) in pn_socket_bind() 372 dev = dev_get_by_index(sock_net(sk), in pn_socket_ioctl() 375 dev = phonet_device_get(sock_net(sk)); in pn_socket_ioctl() 473 struct net *net = sock_net(sk); in pn_sock_get_port() 528 if (!net_eq(net, sock_net(sknode))) in pn_sock_get_idx() 545 while (sk && !net_eq(net, sock_net(sk))); in pn_sock_get_next() 635 if (!net_eq(sock_net(sk), &init_net)) in pn_sock_bind_res()
|
/linux/net/unix/ |
A D | diag.c | 199 struct net *net = sock_net(skb->sk); in unix_diag_dump() 214 if (!net_eq(sock_net(sk), net)) in unix_diag_dump() 265 struct net *net = sock_net(in_skb->sk); in unix_diag_get_exact() 274 if (!net_eq(sock_net(sk), net)) in unix_diag_get_exact() 310 struct net *net = sock_net(skb->sk); in unix_diag_handler_dump()
|
/linux/net/netlink/ |
A D | af_netlink.c | 288 if (!net_eq(dev_net(dev), sock_net(sk))) in __netlink_deliver_tap_skb() 343 netlink_deliver_tap(sock_net(dst), skb); in netlink_deliver_tap_kernel() 784 .net = sock_net(sk), in netlink_release() 822 struct net *net = sock_net(sk); in netlink_autobind() 992 struct net *net = sock_net(sk); in netlink_bind() 1263 netlink_deliver_tap(sock_net(sk), skb); in __netlink_sendskb() 1430 if (!net_eq(sock_net(sk), p->net)) { in do_one_broadcast() 1434 if (!peernet_has_id(sock_net(sk), p->net)) in do_one_broadcast() 1493 struct net *net = sock_net(ssk); in netlink_broadcast() 1552 if (!net_eq(sock_net(sk), sock_net(p->exclude_sk))) in do_one_set_err() [all …]
|
A D | diag.c | 94 struct net *net = sock_net(skb->sk); in __netlink_diag_dump() 133 if (!net_eq(sock_net(sk), net)) in __netlink_diag_dump() 159 if (!net_eq(sock_net(sk), net)) in __netlink_diag_dump() 227 struct net *net = sock_net(skb->sk); in netlink_diag_handler_dump()
|
/linux/net/smc/ |
A D | smc_stats.h | 111 struct net *_net = sock_net(&__smc->sk); \ 126 struct net *_net = sock_net(&__smc->sk); \ 154 struct net *_net = sock_net(&(_smc)->sk); \ 172 struct net *net = sock_net(&(_smc)->sk); \ 218 struct net *net = sock_net(&(__smc)->sk); \
|
/linux/net/xfrm/ |
A D | espintcp.c | 18 XFRM_INC_STATS(sock_net(sk), LINUX_MIB_XFRMINERROR); in handle_nonesp() 40 skb->dev = dev_get_by_index_rcu(sock_net(sk), skb->skb_iif); in handle_esp() 67 XFRM_INC_STATS(sock_net(strp->sk), LINUX_MIB_XFRMINHDRERROR); in espintcp_rcv() 80 XFRM_INC_STATS(sock_net(strp->sk), LINUX_MIB_XFRMINHDRERROR); in espintcp_rcv() 88 XFRM_INC_STATS(sock_net(strp->sk), LINUX_MIB_XFRMINHDRERROR); in espintcp_rcv() 95 XFRM_INC_STATS(sock_net(strp->sk), LINUX_MIB_XFRMINERROR); in espintcp_rcv() 101 XFRM_INC_STATS(sock_net(strp->sk), LINUX_MIB_XFRMINERROR); in espintcp_rcv()
|
/linux/include/net/ |
A D | udp.h | 430 ipv4 ? (IS_UDPLITE(sk) ? sock_net(sk)->mib.udplite_statistics : \ 431 sock_net(sk)->mib.udp_statistics) : \ 432 (IS_UDPLITE(sk) ? sock_net(sk)->mib.udplite_stats_in6 : \ 433 sock_net(sk)->mib.udp_stats_in6); \ 438 IS_UDPLITE(sk) ? sock_net(sk)->mib.udplite_statistics : \ 439 sock_net(sk)->mib.udp_statistics; \
|