Lines Matching refs:tcp_sk
111 struct tcp_sock *tp = tcp_sk(sk); in tcp_twsk_unique()
145 struct tcp_sock *tp = tcp_sk(sk); in tcp_v4_connect()
276 u32 mtu = tcp_sk(sk)->mtu_info; in tcp_v4_mtu_reduced()
411 tp = tcp_sk(sk); in tcp_v4_err()
690 ip_send_unicast_reply(*this_cpu_ptr(net->ipv4.tcp_sk), in tcp_v4_send_reset()
775 ip_send_unicast_reply(*this_cpu_ptr(net->ipv4.tcp_sk), in tcp_v4_send_ack()
809 tcp_sk(sk)->snd_nxt; in tcp_v4_reqsk_send_ack()
877 const struct tcp_sock *tp = tcp_sk(sk); in tcp_md5_do_lookup()
918 struct tcp_sock *tp = tcp_sk(sk); in tcp_md5_do_add()
977 struct tcp_sock *tp = tcp_sk(sk); in tcp_clear_md5_list()
1280 newtp = tcp_sk(newsk); in tcp_v4_syn_recv_sock()
1310 if (tcp_sk(sk)->rx_opt.user_mss && in tcp_v4_syn_recv_sock()
1311 tcp_sk(sk)->rx_opt.user_mss < newtp->advmss) in tcp_v4_syn_recv_sock()
1312 newtp->advmss = tcp_sk(sk)->rx_opt.user_mss; in tcp_v4_syn_recv_sock()
1484 struct tcp_sock *tp = tcp_sk(sk); in tcp_prequeue()
1648 tcp_sk(sk)->segs_in += max_t(u16, 1, skb_shinfo(skb)->gso_segs); in tcp_v4_rcv()
1780 tcp_sk(sk)->af_specific = &tcp_sock_ipv4_specific; in tcp_v4_init_sock()
1788 struct tcp_sock *tp = tcp_sk(sk); in tcp_v4_destroy_sock()
2160 const struct tcp_sock *tp = tcp_sk(sk); in get_tcp4_sock()
2360 inet_ctl_sock_destroy(*per_cpu_ptr(net->ipv4.tcp_sk, cpu)); in tcp_sk_exit()
2361 free_percpu(net->ipv4.tcp_sk); in tcp_sk_exit()
2368 net->ipv4.tcp_sk = alloc_percpu(struct sock *); in tcp_sk_init()
2369 if (!net->ipv4.tcp_sk) in tcp_sk_init()
2379 *per_cpu_ptr(net->ipv4.tcp_sk, cpu) = sk; in tcp_sk_init()