Lines Matching refs:flag
1095 int flag; member
1198 state->flag |= FLAG_ORIG_SACK_ACKED; in tcp_sacktag_one()
1211 state->flag |= FLAG_DATA_SACKED; in tcp_sacktag_one()
1606 state->flag = 0; in tcp_sacktag_write_queue()
1618 state->flag |= FLAG_DSACKING_ACK; in tcp_sacktag_write_queue()
1788 return state->flag; in tcp_sacktag_write_queue()
1952 static bool tcp_check_sack_reneging(struct sock *sk, int flag) in tcp_check_sack_reneging() argument
1954 if (flag & FLAG_SACK_RENEGING) { in tcp_check_sack_reneging()
1991 static bool tcp_pause_early_retransmit(struct sock *sk, int flag) in tcp_pause_early_retransmit() argument
2001 (flag & FLAG_ECE) || !tp->srtt_us) in tcp_pause_early_retransmit()
2108 static bool tcp_time_to_recover(struct sock *sk, int flag) in tcp_time_to_recover() argument
2152 return !tcp_pause_early_retransmit(sk, flag); in tcp_time_to_recover()
2473 int fast_rexmit, int flag) in tcp_cwnd_reduction() argument
2489 } else if ((flag & FLAG_RETRANS_DATA_ACKED) && in tcp_cwnd_reduction()
2490 !(flag & FLAG_LOST_RETRANS)) { in tcp_cwnd_reduction()
2542 static void tcp_try_to_open(struct sock *sk, int flag, const int prior_unsacked) in tcp_try_to_open() argument
2551 if (flag & FLAG_ECE) in tcp_try_to_open()
2557 tcp_cwnd_reduction(sk, prior_unsacked, 0, flag); in tcp_try_to_open()
2667 static void tcp_process_loss(struct sock *sk, int flag, bool is_dupack) in tcp_process_loss() argument
2672 if ((flag & FLAG_SND_UNA_ADVANCED) && in tcp_process_loss()
2680 if ((flag & FLAG_ORIG_SACK_ACKED) && in tcp_process_loss()
2685 if (flag & FLAG_DATA_SACKED || is_dupack) in tcp_process_loss()
2687 } else if (flag & FLAG_SND_UNA_ADVANCED && !recovered) { in tcp_process_loss()
2708 else if (flag & FLAG_SND_UNA_ADVANCED) in tcp_process_loss()
2716 const int prior_unsacked, int flag) in tcp_try_undo_partial() argument
2732 tcp_cwnd_reduction(sk, prior_unsacked, 0, flag); in tcp_try_undo_partial()
2761 bool is_dupack, int flag) in tcp_fastretrans_alert() argument
2765 bool do_lost = is_dupack || ((flag & FLAG_DATA_SACKED) && in tcp_fastretrans_alert()
2776 if (flag & FLAG_ECE) in tcp_fastretrans_alert()
2780 if (tcp_check_sack_reneging(sk, flag)) in tcp_fastretrans_alert()
2815 flag |= FLAG_LOST_RETRANS; in tcp_fastretrans_alert()
2820 if (!(flag & FLAG_SND_UNA_ADVANCED)) { in tcp_fastretrans_alert()
2824 if (tcp_try_undo_partial(sk, acked, prior_unsacked, flag)) in tcp_fastretrans_alert()
2836 tcp_process_loss(sk, flag, is_dupack); in tcp_fastretrans_alert()
2838 !(flag & FLAG_LOST_RETRANS)) in tcp_fastretrans_alert()
2843 if (flag & FLAG_SND_UNA_ADVANCED) in tcp_fastretrans_alert()
2852 if (!tcp_time_to_recover(sk, flag)) { in tcp_fastretrans_alert()
2853 tcp_try_to_open(sk, flag, prior_unsacked); in tcp_fastretrans_alert()
2869 tcp_enter_recovery(sk, (flag & FLAG_ECE)); in tcp_fastretrans_alert()
2875 tcp_cwnd_reduction(sk, prior_unsacked, fast_rexmit, flag); in tcp_fastretrans_alert()
2939 static inline bool tcp_ack_update_rtt(struct sock *sk, const int flag, in tcp_ack_update_rtt() argument
2960 flag & FLAG_ACKED) in tcp_ack_update_rtt()
3113 int flag = 0; in tcp_clean_rtx_queue() local
3144 flag |= FLAG_RETRANS_DATA_ACKED; in tcp_clean_rtx_queue()
3153 flag |= FLAG_ORIG_SACK_ACKED; in tcp_clean_rtx_queue()
3174 flag |= FLAG_DATA_ACKED; in tcp_clean_rtx_queue()
3176 flag |= FLAG_SYN_ACKED; in tcp_clean_rtx_queue()
3195 flag |= FLAG_SACK_RENEGING; in tcp_clean_rtx_queue()
3198 if (likely(first_ackt.v64) && !(flag & FLAG_RETRANS_DATA_ACKED)) { in tcp_clean_rtx_queue()
3207 rtt_update = tcp_ack_update_rtt(sk, flag, seq_rtt_us, sack_rtt_us, in tcp_clean_rtx_queue()
3210 if (flag & FLAG_ACKED) { in tcp_clean_rtx_queue()
3268 return flag; in tcp_clean_rtx_queue()
3292 static inline bool tcp_ack_is_dubious(const struct sock *sk, const int flag) in tcp_ack_is_dubious() argument
3294 return !(flag & FLAG_NOT_DUP) || (flag & FLAG_CA_ALERT) || in tcp_ack_is_dubious()
3299 static inline bool tcp_may_raise_cwnd(const struct sock *sk, const int flag) in tcp_may_raise_cwnd() argument
3311 return flag & FLAG_FORWARD_PROGRESS; in tcp_may_raise_cwnd()
3313 return flag & FLAG_DATA_ACKED; in tcp_may_raise_cwnd()
3359 int flag = 0; in tcp_ack_update_window() local
3366 flag |= FLAG_WIN_UPDATE; in tcp_ack_update_window()
3390 return flag; in tcp_ack_update_window()
3476 static void tcp_process_tlp_ack(struct sock *sk, u32 ack, int flag) in tcp_process_tlp_ack() argument
3483 if (flag & FLAG_DSACKING_ACK) { in tcp_process_tlp_ack()
3496 } else if (!(flag & (FLAG_SND_UNA_ADVANCED | in tcp_process_tlp_ack()
3512 static int tcp_ack(struct sock *sk, const struct sk_buff *skb, int flag) in tcp_ack() argument
3554 flag |= FLAG_SND_UNA_ADVANCED; in tcp_ack()
3563 if (flag & FLAG_UPDATE_TS_RECENT) in tcp_ack()
3566 if (!(flag & FLAG_SLOWPATH) && after(ack, prior_snd_una)) { in tcp_ack()
3573 flag |= FLAG_WIN_UPDATE; in tcp_ack()
3582 flag |= FLAG_DATA; in tcp_ack()
3586 flag |= tcp_ack_update_window(sk, skb, ack, ack_seq); in tcp_ack()
3589 flag |= tcp_sacktag_write_queue(sk, skb, prior_snd_una, in tcp_ack()
3593 flag |= FLAG_ECE; in tcp_ack()
3597 if (flag & FLAG_WIN_UPDATE) in tcp_ack()
3614 flag |= tcp_clean_rtx_queue(sk, prior_fackets, prior_snd_una, in tcp_ack()
3618 if (tcp_ack_is_dubious(sk, flag)) { in tcp_ack()
3619 is_dupack = !(flag & (FLAG_SND_UNA_ADVANCED | FLAG_NOT_DUP)); in tcp_ack()
3621 is_dupack, flag); in tcp_ack()
3624 tcp_process_tlp_ack(sk, ack, flag); in tcp_ack()
3627 if (tcp_may_raise_cwnd(sk, flag)) in tcp_ack()
3630 if ((flag & FLAG_FORWARD_PROGRESS) || !(flag & FLAG_NOT_DUP)) { in tcp_ack()
3643 if (flag & FLAG_DSACKING_ACK) in tcp_ack()
3645 is_dupack, flag); in tcp_ack()
3654 tcp_process_tlp_ack(sk, ack, flag); in tcp_ack()
3666 flag |= tcp_sacktag_write_queue(sk, skb, prior_snd_una, in tcp_ack()
3669 is_dupack, flag); in tcp_ack()