Lines Matching refs:sk

35 static void tcp_write_err(struct sock *sk)  in tcp_write_err()  argument
37 sk->sk_err = sk->sk_err_soft ? : ETIMEDOUT; in tcp_write_err()
38 sk->sk_error_report(sk); in tcp_write_err()
40 tcp_done(sk); in tcp_write_err()
41 NET_INC_STATS_BH(sock_net(sk), LINUX_MIB_TCPABORTONTIMEOUT); in tcp_write_err()
55 static int tcp_out_of_resources(struct sock *sk, bool do_reset) in tcp_out_of_resources() argument
57 struct tcp_sock *tp = tcp_sk(sk); in tcp_out_of_resources()
66 if (sk->sk_err_soft) in tcp_out_of_resources()
69 if (tcp_check_oom(sk, shift)) { in tcp_out_of_resources()
77 tcp_send_active_reset(sk, GFP_ATOMIC); in tcp_out_of_resources()
78 tcp_done(sk); in tcp_out_of_resources()
79 NET_INC_STATS_BH(sock_net(sk), LINUX_MIB_TCPABORTONMEMORY); in tcp_out_of_resources()
86 static int tcp_orphan_retries(struct sock *sk, bool alive) in tcp_orphan_retries() argument
91 if (sk->sk_err_soft && !alive) in tcp_orphan_retries()
102 static void tcp_mtu_probing(struct inet_connection_sock *icsk, struct sock *sk) in tcp_mtu_probing() argument
104 struct net *net = sock_net(sk); in tcp_mtu_probing()
111 tcp_sync_mss(sk, icsk->icsk_pmtu_cookie); in tcp_mtu_probing()
113 struct net *net = sock_net(sk); in tcp_mtu_probing()
114 struct tcp_sock *tp = tcp_sk(sk); in tcp_mtu_probing()
117 mss = tcp_mtu_to_mss(sk, icsk->icsk_mtup.search_low) >> 1; in tcp_mtu_probing()
120 icsk->icsk_mtup.search_low = tcp_mss_to_mtu(sk, mss); in tcp_mtu_probing()
121 tcp_sync_mss(sk, icsk->icsk_pmtu_cookie); in tcp_mtu_probing()
131 static bool retransmits_timed_out(struct sock *sk, in retransmits_timed_out() argument
139 if (!inet_csk(sk)->icsk_retransmits) in retransmits_timed_out()
142 start_ts = tcp_sk(sk)->retrans_stamp; in retransmits_timed_out()
144 start_ts = tcp_skb_timestamp(tcp_write_queue_head(sk)); in retransmits_timed_out()
159 static int tcp_write_timeout(struct sock *sk) in tcp_write_timeout() argument
161 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_write_timeout()
162 struct tcp_sock *tp = tcp_sk(sk); in tcp_write_timeout()
166 if ((1 << sk->sk_state) & (TCPF_SYN_SENT | TCPF_SYN_RECV)) { in tcp_write_timeout()
168 dst_negative_advice(sk); in tcp_write_timeout()
170 tcp_fastopen_cache_set(sk, 0, NULL, true, 0); in tcp_write_timeout()
172 NET_INC_STATS_BH(sock_net(sk), in tcp_write_timeout()
178 if (retransmits_timed_out(sk, sysctl_tcp_retries1, 0, 0)) { in tcp_write_timeout()
186 tcp_fastopen_cache_set(sk, 0, NULL, true, 0); in tcp_write_timeout()
188 NET_INC_STATS_BH(sock_net(sk), in tcp_write_timeout()
192 tcp_mtu_probing(icsk, sk); in tcp_write_timeout()
194 dst_negative_advice(sk); in tcp_write_timeout()
198 if (sock_flag(sk, SOCK_DEAD)) { in tcp_write_timeout()
201 retry_until = tcp_orphan_retries(sk, alive); in tcp_write_timeout()
203 !retransmits_timed_out(sk, retry_until, 0, 0); in tcp_write_timeout()
205 if (tcp_out_of_resources(sk, do_reset)) in tcp_write_timeout()
210 if (retransmits_timed_out(sk, retry_until, in tcp_write_timeout()
213 tcp_write_err(sk); in tcp_write_timeout()
219 void tcp_delack_timer_handler(struct sock *sk) in tcp_delack_timer_handler() argument
221 struct tcp_sock *tp = tcp_sk(sk); in tcp_delack_timer_handler()
222 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_delack_timer_handler()
224 sk_mem_reclaim_partial(sk); in tcp_delack_timer_handler()
226 if (sk->sk_state == TCP_CLOSE || !(icsk->icsk_ack.pending & ICSK_ACK_TIMER)) in tcp_delack_timer_handler()
230 sk_reset_timer(sk, &icsk->icsk_delack_timer, icsk->icsk_ack.timeout); in tcp_delack_timer_handler()
238 NET_INC_STATS_BH(sock_net(sk), LINUX_MIB_TCPSCHEDULERFAILED); in tcp_delack_timer_handler()
241 sk_backlog_rcv(sk, skb); in tcp_delack_timer_handler()
246 if (inet_csk_ack_scheduled(sk)) { in tcp_delack_timer_handler()
257 tcp_send_ack(sk); in tcp_delack_timer_handler()
258 NET_INC_STATS_BH(sock_net(sk), LINUX_MIB_DELAYEDACKS); in tcp_delack_timer_handler()
262 if (tcp_under_memory_pressure(sk)) in tcp_delack_timer_handler()
263 sk_mem_reclaim(sk); in tcp_delack_timer_handler()
268 struct sock *sk = (struct sock *)data; in tcp_delack_timer() local
270 bh_lock_sock(sk); in tcp_delack_timer()
271 if (!sock_owned_by_user(sk)) { in tcp_delack_timer()
272 tcp_delack_timer_handler(sk); in tcp_delack_timer()
274 inet_csk(sk)->icsk_ack.blocked = 1; in tcp_delack_timer()
275 NET_INC_STATS_BH(sock_net(sk), LINUX_MIB_DELAYEDACKLOCKED); in tcp_delack_timer()
277 if (!test_and_set_bit(TCP_DELACK_TIMER_DEFERRED, &tcp_sk(sk)->tsq_flags)) in tcp_delack_timer()
278 sock_hold(sk); in tcp_delack_timer()
280 bh_unlock_sock(sk); in tcp_delack_timer()
281 sock_put(sk); in tcp_delack_timer()
284 static void tcp_probe_timer(struct sock *sk) in tcp_probe_timer() argument
286 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_probe_timer()
287 struct tcp_sock *tp = tcp_sk(sk); in tcp_probe_timer()
291 if (tp->packets_out || !tcp_send_head(sk)) { in tcp_probe_timer()
304 start_ts = tcp_skb_timestamp(tcp_send_head(sk)); in tcp_probe_timer()
306 skb_mstamp_get(&tcp_send_head(sk)->skb_mstamp); in tcp_probe_timer()
312 if (sock_flag(sk, SOCK_DEAD)) { in tcp_probe_timer()
315 max_probes = tcp_orphan_retries(sk, alive); in tcp_probe_timer()
318 if (tcp_out_of_resources(sk, true)) in tcp_probe_timer()
323 abort: tcp_write_err(sk); in tcp_probe_timer()
326 tcp_send_probe0(sk); in tcp_probe_timer()
334 static void tcp_fastopen_synack_timer(struct sock *sk) in tcp_fastopen_synack_timer() argument
336 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_fastopen_synack_timer()
341 req = tcp_sk(sk)->fastopen_rsk; in tcp_fastopen_synack_timer()
345 tcp_write_err(sk); in tcp_fastopen_synack_timer()
353 inet_rtx_syn_ack(sk, req); in tcp_fastopen_synack_timer()
355 inet_csk_reset_xmit_timer(sk, ICSK_TIME_RETRANS, in tcp_fastopen_synack_timer()
363 void tcp_retransmit_timer(struct sock *sk) in tcp_retransmit_timer() argument
365 struct tcp_sock *tp = tcp_sk(sk); in tcp_retransmit_timer()
366 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_retransmit_timer()
369 WARN_ON_ONCE(sk->sk_state != TCP_SYN_RECV && in tcp_retransmit_timer()
370 sk->sk_state != TCP_FIN_WAIT1); in tcp_retransmit_timer()
371 tcp_fastopen_synack_timer(sk); in tcp_retransmit_timer()
380 WARN_ON(tcp_write_queue_empty(sk)); in tcp_retransmit_timer()
384 if (!tp->snd_wnd && !sock_flag(sk, SOCK_DEAD) && in tcp_retransmit_timer()
385 !((1 << sk->sk_state) & (TCPF_SYN_SENT | TCPF_SYN_RECV))) { in tcp_retransmit_timer()
391 struct inet_sock *inet = inet_sk(sk); in tcp_retransmit_timer()
392 if (sk->sk_family == AF_INET) { in tcp_retransmit_timer()
400 else if (sk->sk_family == AF_INET6) { in tcp_retransmit_timer()
402 &sk->sk_v6_daddr, in tcp_retransmit_timer()
409 tcp_write_err(sk); in tcp_retransmit_timer()
412 tcp_enter_loss(sk); in tcp_retransmit_timer()
413 tcp_retransmit_skb(sk, tcp_write_queue_head(sk)); in tcp_retransmit_timer()
414 __sk_dst_reset(sk); in tcp_retransmit_timer()
418 if (tcp_write_timeout(sk)) in tcp_retransmit_timer()
440 NET_INC_STATS_BH(sock_net(sk), mib_idx); in tcp_retransmit_timer()
443 tcp_enter_loss(sk); in tcp_retransmit_timer()
445 if (tcp_retransmit_skb(sk, tcp_write_queue_head(sk)) > 0) { in tcp_retransmit_timer()
451 inet_csk_reset_xmit_timer(sk, ICSK_TIME_RETRANS, in tcp_retransmit_timer()
485 if (sk->sk_state == TCP_ESTABLISHED && in tcp_retransmit_timer()
495 inet_csk_reset_xmit_timer(sk, ICSK_TIME_RETRANS, icsk->icsk_rto, TCP_RTO_MAX); in tcp_retransmit_timer()
496 if (retransmits_timed_out(sk, sysctl_tcp_retries1 + 1, 0, 0)) in tcp_retransmit_timer()
497 __sk_dst_reset(sk); in tcp_retransmit_timer()
502 void tcp_write_timer_handler(struct sock *sk) in tcp_write_timer_handler() argument
504 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_write_timer_handler()
507 if (sk->sk_state == TCP_CLOSE || !icsk->icsk_pending) in tcp_write_timer_handler()
511 sk_reset_timer(sk, &icsk->icsk_retransmit_timer, icsk->icsk_timeout); in tcp_write_timer_handler()
519 tcp_resume_early_retransmit(sk); in tcp_write_timer_handler()
522 tcp_send_loss_probe(sk); in tcp_write_timer_handler()
526 tcp_retransmit_timer(sk); in tcp_write_timer_handler()
530 tcp_probe_timer(sk); in tcp_write_timer_handler()
535 sk_mem_reclaim(sk); in tcp_write_timer_handler()
540 struct sock *sk = (struct sock *)data; in tcp_write_timer() local
542 bh_lock_sock(sk); in tcp_write_timer()
543 if (!sock_owned_by_user(sk)) { in tcp_write_timer()
544 tcp_write_timer_handler(sk); in tcp_write_timer()
547 if (!test_and_set_bit(TCP_WRITE_TIMER_DEFERRED, &tcp_sk(sk)->tsq_flags)) in tcp_write_timer()
548 sock_hold(sk); in tcp_write_timer()
550 bh_unlock_sock(sk); in tcp_write_timer()
551 sock_put(sk); in tcp_write_timer()
562 void tcp_set_keepalive(struct sock *sk, int val) in tcp_set_keepalive() argument
564 if ((1 << sk->sk_state) & (TCPF_CLOSE | TCPF_LISTEN)) in tcp_set_keepalive()
567 if (val && !sock_flag(sk, SOCK_KEEPOPEN)) in tcp_set_keepalive()
568 inet_csk_reset_keepalive_timer(sk, keepalive_time_when(tcp_sk(sk))); in tcp_set_keepalive()
570 inet_csk_delete_keepalive_timer(sk); in tcp_set_keepalive()
576 struct sock *sk = (struct sock *) data; in tcp_keepalive_timer() local
577 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_keepalive_timer()
578 struct tcp_sock *tp = tcp_sk(sk); in tcp_keepalive_timer()
582 bh_lock_sock(sk); in tcp_keepalive_timer()
583 if (sock_owned_by_user(sk)) { in tcp_keepalive_timer()
585 inet_csk_reset_keepalive_timer (sk, HZ/20); in tcp_keepalive_timer()
589 if (sk->sk_state == TCP_LISTEN) { in tcp_keepalive_timer()
594 if (sk->sk_state == TCP_FIN_WAIT2 && sock_flag(sk, SOCK_DEAD)) { in tcp_keepalive_timer()
596 const int tmo = tcp_fin_time(sk) - TCP_TIMEWAIT_LEN; in tcp_keepalive_timer()
599 tcp_time_wait(sk, TCP_FIN_WAIT2, tmo); in tcp_keepalive_timer()
603 tcp_send_active_reset(sk, GFP_ATOMIC); in tcp_keepalive_timer()
607 if (!sock_flag(sk, SOCK_KEEPOPEN) || sk->sk_state == TCP_CLOSE) in tcp_keepalive_timer()
613 if (tp->packets_out || tcp_send_head(sk)) in tcp_keepalive_timer()
627 tcp_send_active_reset(sk, GFP_ATOMIC); in tcp_keepalive_timer()
628 tcp_write_err(sk); in tcp_keepalive_timer()
631 if (tcp_write_wakeup(sk, LINUX_MIB_TCPKEEPALIVE) <= 0) { in tcp_keepalive_timer()
645 sk_mem_reclaim(sk); in tcp_keepalive_timer()
648 inet_csk_reset_keepalive_timer (sk, elapsed); in tcp_keepalive_timer()
652 tcp_done(sk); in tcp_keepalive_timer()
655 bh_unlock_sock(sk); in tcp_keepalive_timer()
656 sock_put(sk); in tcp_keepalive_timer()
659 void tcp_init_xmit_timers(struct sock *sk) in tcp_init_xmit_timers() argument
661 inet_csk_init_xmit_timers(sk, &tcp_write_timer, &tcp_delack_timer, in tcp_init_xmit_timers()