Lines Matching refs:sock_net
321 __NET_INC_STATS(sock_net(sk), in tcp_fastopen_queue_check()
338 return (READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_fastopen) & flag) || in tcp_fastopen_no_cookie()
353 int tcp_fastopen = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_fastopen); in tcp_try_fastopen()
359 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPFASTOPENCOOKIEREQD); in tcp_try_fastopen()
378 NET_INC_STATS(sock_net(sk), in tcp_try_fastopen()
396 NET_INC_STATS(sock_net(sk), in tcp_try_fastopen()
401 NET_INC_STATS(sock_net(sk), in tcp_try_fastopen()
405 NET_INC_STATS(sock_net(sk), in tcp_try_fastopen()
493 struct net *net = sock_net(sk); in tcp_fastopen_active_disable()
495 if (!READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_fastopen_blackhole_timeout)) in tcp_fastopen_active_disable()
517 READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_fastopen_blackhole_timeout); in tcp_fastopen_active_should_disable()
525 tfo_da_times = atomic_read(&sock_net(sk)->ipv4.tfo_active_disable_times); in tcp_fastopen_active_should_disable()
536 timeout = READ_ONCE(sock_net(sk)->ipv4.tfo_active_disable_stamp) + in tcp_fastopen_active_should_disable()
572 atomic_read(&sock_net(sk)->ipv4.tfo_active_disable_times)) { in tcp_fastopen_active_disable_ofo_check()
575 atomic_set(&sock_net(sk)->ipv4.tfo_active_disable_times, 0); in tcp_fastopen_active_disable_ofo_check()
593 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPFASTOPENACTIVEFAIL); in tcp_fastopen_active_detect_blackhole()