Lines Matching refs:inet_csk

134 	struct inet_connection_sock *icsk = inet_csk(sk);  in tcp_measure_rcv_mss()
180 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_incr_quickack()
191 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_enter_quickack_mode()
203 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_in_quickack_mode()
351 return 2 * inet_csk(sk)->icsk_ack.rcv_mss; in __tcp_grow_window()
381 inet_csk(sk)->icsk_ack.quick |= 1; in tcp_grow_window()
447 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_clamp_window()
478 inet_csk(sk)->icsk_ack.rcv_mss = hint; in tcp_initialize_rcv_mss()
548 TCP_SKB_CB(skb)->seq >= inet_csk(sk)->icsk_ack.rcv_mss)) in tcp_rcv_rtt_measure_ts()
635 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_event_data_recv()
810 inet_csk(sk)->icsk_rto = __tcp_set_rto(tp); in tcp_set_rto()
873 tp->rx_opt.sack_ok, inet_csk(sk)->icsk_ca_state, in tcp_update_reordering()
1776 ((inet_csk(sk)->icsk_ca_state != TCP_CA_Loss) || tp->undo_marker)) in tcp_sacktag_write_queue()
1874 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_enter_loss()
1939 !inet_csk(sk)->icsk_mtup.probe_size; in tcp_enter_loss()
2007 if (!time_after(inet_csk(sk)->icsk_timeout, (jiffies + delay))) in tcp_pause_early_retransmit()
2358 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_undo_cwnd_reduction()
2392 DBGUNDO(sk, inet_csk(sk)->icsk_ca_state == TCP_CA_Loss ? "loss" : "retrans"); in tcp_try_undo_recovery()
2394 if (inet_csk(sk)->icsk_ca_state == TCP_CA_Loss) in tcp_try_undo_recovery()
2441 inet_csk(sk)->icsk_retransmits = 0; in tcp_try_undo_loss()
2468 tp->snd_ssthresh = inet_csk(sk)->icsk_ca_ops->ssthresh(sk); in tcp_init_cwnd_reduction()
2506 if (inet_csk(sk)->icsk_ca_state == TCP_CA_CWR || in tcp_end_cwnd_reduction()
2520 if (inet_csk(sk)->icsk_ca_state < TCP_CA_CWR) { in tcp_enter_cwr()
2536 if (inet_csk(sk)->icsk_ca_state != state) { in tcp_try_keep_open()
2554 if (inet_csk(sk)->icsk_ca_state != TCP_CA_CWR) { in tcp_try_to_open()
2563 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_mtup_probe_failed()
2573 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_mtup_probe_success()
2596 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_simple_retransmit()
2763 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_fastretrans_alert()
2975 inet_csk(sk)->icsk_backoff = 0; in tcp_ack_update_rtt()
2997 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_cong_avoid()
3008 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_rearm_rto()
3020 u32 rto = inet_csk(sk)->icsk_rto; in tcp_rearm_rto()
3101 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_clean_rtx_queue()
3250 icsk = inet_csk(sk); in tcp_clean_rtx_queue()
3274 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_ack_probe()
3295 inet_csk(sk)->icsk_ca_state != TCP_CA_Open; in tcp_ack_is_dubious()
3383 tcp_sync_mss(sk, inet_csk(sk)->icsk_pmtu_cookie); in tcp_ack_update_window()
3505 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_in_ack_event()
3514 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_ack()
3930 (s32)(tp->rx_opt.ts_recent - tp->rx_opt.rcv_tsval) <= (inet_csk(sk)->icsk_rto * 1024) / HZ); in tcp_disordered_ack()
4014 inet_csk(sk)->icsk_ack.pingpong = 1; in tcp_fin()
4600 inet_csk(sk)->icsk_ack.pingpong = 0; in tcp_data_queue()
4957 if (((tp->rcv_nxt - tp->rcv_wup) > inet_csk(sk)->icsk_ack.rcv_mss && in __tcp_ack_snd_check()
5234 inet_csk(sk)->icsk_af_ops->sk_rx_dst_set(sk, skb); in tcp_rcv_established()
5426 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_finish_connect()
5520 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_rcv_synsent_state_process()
5748 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_rcv_state_process()
6117 struct request_sock_queue *queue = &inet_csk(sk)->icsk_accept_queue; in tcp_syn_flood_action()