Lines Matching refs:sock_net
216 dev = dev_get_by_index_rcu(sock_net(sk), skb->skb_iif); in tcp_gro_dev_warn()
429 min(sndmem, sock_net(sk)->ipv4.sysctl_tcp_wmem[2])); in tcp_sndbuf_expand()
464 int window = tcp_win_from_space(sk, sock_net(sk)->ipv4.sysctl_tcp_rmem[2]) >> 1; in __tcp_grow_window()
537 int tcp_app_win = sock_net(sk)->ipv4.sysctl_tcp_app_win; in tcp_init_buffer_space()
576 struct net *net = sock_net(sk); in tcp_clamp_window()
727 if (sock_net(sk)->ipv4.sysctl_tcp_moderate_rcvbuf && in tcp_rcv_space_adjust()
748 sock_net(sk)->ipv4.sysctl_tcp_rmem[2]); in tcp_rcv_space_adjust()
913 rate *= sock_net(sk)->ipv4.sysctl_tcp_pacing_ss_ratio; in tcp_update_pacing_rate()
915 rate *= sock_net(sk)->ipv4.sysctl_tcp_pacing_ca_ratio; in tcp_update_pacing_rate()
1054 sock_net(sk)->ipv4.sysctl_tcp_max_reordering); in tcp_check_sack_reordering()
1059 NET_INC_STATS(sock_net(sk), in tcp_check_sack_reordering()
1099 NET_ADD_STATS(sock_net(sk), LINUX_MIB_TCPLOSTRETRANSMIT, in tcp_mark_skb_lost()
1259 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPDSACKRECV); in tcp_check_dsack()
1266 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPDSACKOFORECV); in tcp_check_dsack()
1273 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPDSACKIGNOREDDUBIOUS); in tcp_check_dsack()
1277 NET_ADD_STATS(sock_net(sk), LINUX_MIB_TCPDSACKRECVSEGS, dup_segs); in tcp_check_dsack()
1480 NET_INC_STATS(sock_net(sk), LINUX_MIB_SACKSHIFTED); in tcp_shifted_skb()
1507 NET_INC_STATS(sock_net(sk), LINUX_MIB_SACKMERGED); in tcp_shifted_skb()
1676 NET_INC_STATS(sock_net(sk), LINUX_MIB_SACKSHIFTFALLBACK); in tcp_shift_skb_data()
1868 NET_INC_STATS(sock_net(sk), mib_idx); in tcp_sacktag_write_queue()
2031 sock_net(sk)->ipv4.sysctl_tcp_max_reordering); in tcp_check_reno_reordering()
2033 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPRENOREORDER); in tcp_check_reno_reordering()
2096 return sock_net(sk)->ipv4.sysctl_tcp_recovery & TCP_RACK_LOSS_DETECTION; in tcp_is_rack()
2112 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPSACKRENEGING); in tcp_timeout_mark_lost()
2138 struct net *net = sock_net(sk); in tcp_enter_loss()
2528 NET_INC_STATS(sock_net(sk), mib_idx); in tcp_try_undo_recovery()
2555 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPDSACKUNDO); in tcp_try_undo_dsack()
2570 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPLOSSUNDO); in tcp_try_undo_loss()
2572 NET_INC_STATS(sock_net(sk), in tcp_try_undo_loss()
2700 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPMTUPFAIL); in tcp_mtup_probe_failed()
2720 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPMTUPSUCCESS); in tcp_mtup_probe_success()
2790 NET_INC_STATS(sock_net(sk), mib_idx); in tcp_enter_recovery()
2892 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPPARTIALUNDO); in tcp_try_undo_partial()
3052 u32 wlen = sock_net(sk)->ipv4.sysctl_tcp_min_rtt_wlen * HZ; in tcp_update_rtt_min()
3462 if (tcp_sk(sk)->reordering > sock_net(sk)->ipv4.sysctl_tcp_reordering) in tcp_may_raise_cwnd()
3610 struct net *net = sock_net(sk); in tcp_send_challenge_ack()
3681 NET_INC_STATS(sock_net(sk), in tcp_process_tlp_ack()
3722 const struct net *net = sock_net(sk); in tcp_newly_delivered()
3810 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPHPACKS); in tcp_ack()
3817 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPPUREACKS); in tcp_ack()
4418 if (tcp_is_sack(tp) && sock_net(sk)->ipv4.sysctl_tcp_dsack) { in tcp_dsack_set()
4426 NET_INC_STATS(sock_net(sk), mib_idx); in tcp_dsack_set()
4453 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPDUPLICATEDATAREHASH); in tcp_rcv_spurious_retrans()
4462 NET_INC_STATS(sock_net(sk), LINUX_MIB_DELAYEDACKLOST); in tcp_send_dupack()
4465 if (tcp_is_sack(tp) && sock_net(sk)->ipv4.sysctl_tcp_dsack) { in tcp_send_dupack()
4521 NET_ADD_STATS(sock_net(sk), LINUX_MIB_TCPACKCOMPRESSED, in tcp_sack_compress_send_ack()
4654 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPRCVCOALESCE); in tcp_try_coalesce()
4772 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPOFODROP); in tcp_data_queue_ofo()
4783 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPOFOQUEUE); in tcp_data_queue_ofo()
4835 NET_INC_STATS(sock_net(sk), in tcp_data_queue_ofo()
4854 NET_INC_STATS(sock_net(sk), in tcp_data_queue_ofo()
4883 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPOFOMERGE); in tcp_data_queue_ofo()
4949 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPRCVQDROP); in tcp_send_rcvq()
5009 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPZEROWINDOWDROP); in tcp_data_queue()
5018 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPRCVQDROP); in tcp_data_queue()
5054 NET_INC_STATS(sock_net(sk), LINUX_MIB_DELAYEDACKLOST); in tcp_data_queue()
5077 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPZEROWINDOWDROP); in tcp_data_queue()
5106 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPRCVCOLLAPSED); in tcp_collapse_one()
5311 NET_INC_STATS(sock_net(sk), LINUX_MIB_OFOPRUNED); in tcp_prune_ofo_queue()
5351 NET_INC_STATS(sock_net(sk), LINUX_MIB_PRUNECALLED); in tcp_prune_queue()
5384 NET_INC_STATS(sock_net(sk), LINUX_MIB_RCVPRUNED); in tcp_prune_queue()
5488 tp->compressed_ack >= sock_net(sk)->ipv4.sysctl_tcp_comp_sack_nr) in __tcp_ack_snd_check()
5509 delay = min_t(unsigned long, sock_net(sk)->ipv4.sysctl_tcp_comp_sack_delay_ns, in __tcp_ack_snd_check()
5513 sock_net(sk)->ipv4.sysctl_tcp_comp_sack_slack_ns, in __tcp_ack_snd_check()
5541 if (ptr && !sock_net(sk)->ipv4.sysctl_tcp_stdurg) in tcp_check_urg()
5654 if (tcp_fast_parse_options(sock_net(sk), skb, th, tp) && in tcp_validate_incoming()
5658 NET_INC_STATS(sock_net(sk), LINUX_MIB_PAWSESTABREJECTED); in tcp_validate_incoming()
5659 if (!tcp_oow_rate_limited(sock_net(sk), skb, in tcp_validate_incoming()
5679 if (!tcp_oow_rate_limited(sock_net(sk), skb, in tcp_validate_incoming()
5742 TCP_INC_STATS(sock_net(sk), TCP_MIB_INERRS); in tcp_validate_incoming()
5743 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPSYNCHALLENGE); in tcp_validate_incoming()
5870 TCP_INC_STATS(sock_net(sk), TCP_MIB_INERRS); in tcp_rcv_established()
5894 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPHPHITS); in tcp_rcv_established()
5953 TCP_INC_STATS(sock_net(sk), TCP_MIB_CSUMERRORS); in tcp_rcv_established()
5954 TCP_INC_STATS(sock_net(sk), TCP_MIB_INERRS); in tcp_rcv_established()
6033 tcp_parse_options(sock_net(sk), synack, &opt, 0, NULL); in tcp_rcv_fastopen_synack()
6065 NET_INC_STATS(sock_net(sk), in tcp_rcv_fastopen_synack()
6071 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPFASTOPENACTIVE); in tcp_rcv_fastopen_synack()
6116 tcp_parse_options(sock_net(sk), skb, &tp->rx_opt, 0, &foc); in tcp_rcv_synsent_state_process()
6142 NET_INC_STATS(sock_net(sk), in tcp_rcv_synsent_state_process()
6527 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTONDATA); in tcp_rcv_state_process()
6536 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTONDATA); in tcp_rcv_state_process()
6600 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTONDATA); in tcp_rcv_state_process()
6663 const struct net *net = sock_net(listen_sk); in tcp_ecn_create_request()
6724 write_pnet(&ireq->ireq_net, sock_net(sk_listener)); in inet_reqsk_alloc()
6740 struct net *net = sock_net(sk); in tcp_syn_flood_action()
6746 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPREQQFULLDOCOOKIES); in tcp_syn_flood_action()
6749 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPREQQFULLDROP); in tcp_syn_flood_action()
6801 if (sock_net(sk)->ipv4.sysctl_tcp_syncookies != 2 && in tcp_get_syncookie_mss()
6809 NET_INC_STATS(sock_net(sk), LINUX_MIB_LISTENOVERFLOWS); in tcp_get_syncookie_mss()
6829 struct net *net = sock_net(sk); in tcp_conn_request()
6848 NET_INC_STATS(sock_net(sk), LINUX_MIB_LISTENOVERFLOWS); in tcp_conn_request()
6866 tcp_parse_options(sock_net(sk), skb, &tmp_opt, 0, in tcp_conn_request()