Lines Matching refs:tcp_sk
111 struct tcp_sock *tp = tcp_sk(sk); in tcp_twsk_unique()
145 struct tcp_sock *tp = tcp_sk(sk); in tcp_v4_connect()
276 u32 mtu = tcp_sk(sk)->mtu_info; in tcp_v4_mtu_reduced()
408 tp = tcp_sk(sk); in tcp_v4_err()
687 ip_send_unicast_reply(*this_cpu_ptr(net->ipv4.tcp_sk), in tcp_v4_send_reset()
772 ip_send_unicast_reply(*this_cpu_ptr(net->ipv4.tcp_sk), in tcp_v4_send_ack()
806 tcp_sk(sk)->snd_nxt; in tcp_v4_reqsk_send_ack()
875 const struct tcp_sock *tp = tcp_sk(sk); in tcp_md5_do_lookup()
916 struct tcp_sock *tp = tcp_sk(sk); in tcp_md5_do_add()
975 struct tcp_sock *tp = tcp_sk(sk); in tcp_clear_md5_list()
1271 newtp = tcp_sk(newsk); in tcp_v4_syn_recv_sock()
1302 if (tcp_sk(sk)->rx_opt.user_mss && in tcp_v4_syn_recv_sock()
1303 tcp_sk(sk)->rx_opt.user_mss < newtp->advmss) in tcp_v4_syn_recv_sock()
1304 newtp->advmss = tcp_sk(sk)->rx_opt.user_mss; in tcp_v4_syn_recv_sock()
1497 struct tcp_sock *tp = tcp_sk(sk); in tcp_prequeue()
1768 tcp_sk(sk)->af_specific = &tcp_sock_ipv4_specific; in tcp_v4_init_sock()
1776 struct tcp_sock *tp = tcp_sk(sk); in tcp_v4_destroy_sock()
2191 const struct tcp_sock *tp = tcp_sk(sk); in get_tcp4_sock()
2395 inet_ctl_sock_destroy(*per_cpu_ptr(net->ipv4.tcp_sk, cpu)); in tcp_sk_exit()
2396 free_percpu(net->ipv4.tcp_sk); in tcp_sk_exit()
2403 net->ipv4.tcp_sk = alloc_percpu(struct sock *); in tcp_sk_init()
2404 if (!net->ipv4.tcp_sk) in tcp_sk_init()
2414 *per_cpu_ptr(net->ipv4.tcp_sk, cpu) = sk; in tcp_sk_init()