Lines Matching refs:mss_now
68 static bool tcp_write_xmit(struct sock *sk, unsigned int mss_now, int nonagle,
1060 unsigned int mss_now) in tcp_set_skb_tso_segs() argument
1067 if (skb->len <= mss_now || skb->ip_summed == CHECKSUM_NONE) { in tcp_set_skb_tso_segs()
1075 tcp_skb_pcount_set(skb, DIV_ROUND_UP(skb->len, mss_now)); in tcp_set_skb_tso_segs()
1076 shinfo->gso_size = mss_now; in tcp_set_skb_tso_segs()
1147 unsigned int mss_now, gfp_t gfp) in tcp_fragment() argument
1209 tcp_set_skb_tso_segs(sk, skb, mss_now); in tcp_fragment()
1210 tcp_set_skb_tso_segs(sk, buff, mss_now); in tcp_fragment()
1300 int mss_now; in __tcp_mtu_to_mss() local
1305 mss_now = pmtu - icsk->icsk_af_ops->net_header_len - sizeof(struct tcphdr); in __tcp_mtu_to_mss()
1312 mss_now -= icsk->icsk_af_ops->net_frag_header_len; in __tcp_mtu_to_mss()
1316 if (mss_now > tp->rx_opt.mss_clamp) in __tcp_mtu_to_mss()
1317 mss_now = tp->rx_opt.mss_clamp; in __tcp_mtu_to_mss()
1320 mss_now -= icsk->icsk_ext_hdr_len; in __tcp_mtu_to_mss()
1323 if (mss_now < 48) in __tcp_mtu_to_mss()
1324 mss_now = 48; in __tcp_mtu_to_mss()
1325 return mss_now; in __tcp_mtu_to_mss()
1401 int mss_now; in tcp_sync_mss() local
1406 mss_now = tcp_mtu_to_mss(sk, pmtu); in tcp_sync_mss()
1407 mss_now = tcp_bound_to_half_wnd(tp, mss_now); in tcp_sync_mss()
1412 mss_now = min(mss_now, tcp_mtu_to_mss(sk, icsk->icsk_mtup.search_low)); in tcp_sync_mss()
1413 tp->mss_cache = mss_now; in tcp_sync_mss()
1415 return mss_now; in tcp_sync_mss()
1426 u32 mss_now; in tcp_current_mss() local
1431 mss_now = tp->mss_cache; in tcp_current_mss()
1436 mss_now = tcp_sync_mss(sk, mtu); in tcp_current_mss()
1447 mss_now -= delta; in tcp_current_mss()
1450 return mss_now; in tcp_current_mss()
1519 static void tcp_minshall_update(struct tcp_sock *tp, unsigned int mss_now, in tcp_minshall_update() argument
1522 if (skb->len < tcp_skb_pcount(skb) * mss_now) in tcp_minshall_update()
1544 static u32 tcp_tso_autosize(const struct sock *sk, unsigned int mss_now) in tcp_tso_autosize() argument
1556 segs = max_t(u32, bytes / mss_now, sysctl_tcp_min_tso_segs); in tcp_tso_autosize()
1564 unsigned int mss_now, in tcp_mss_split_point() argument
1572 max_len = mss_now * max_segs; in tcp_mss_split_point()
1582 partial = needed % mss_now; in tcp_mss_split_point()
1623 unsigned int mss_now) in tcp_init_tso_segs() argument
1627 if (!tso_segs || (tso_segs > 1 && tcp_skb_mss(skb) != mss_now)) { in tcp_init_tso_segs()
1628 tcp_set_skb_tso_segs(sk, skb, mss_now); in tcp_init_tso_segs()
1715 unsigned int mss_now, gfp_t gfp) in tso_fragment() argument
1723 return tcp_fragment(sk, skb, len, mss_now, gfp); in tso_fragment()
1752 tcp_set_skb_tso_segs(sk, skb, mss_now); in tso_fragment()
1753 tcp_set_skb_tso_segs(sk, buff, mss_now); in tso_fragment()
1890 int mss_now; in tcp_mtu_probe() local
1908 mss_now = tcp_current_mss(sk); in tcp_mtu_probe()
1987 tcp_set_skb_tso_segs(sk, skb, mss_now); in tcp_mtu_probe()
2032 static bool tcp_write_xmit(struct sock *sk, unsigned int mss_now, int nonagle, in tcp_write_xmit() argument
2055 max_segs = tcp_tso_autosize(sk, mss_now); in tcp_write_xmit()
2059 tso_segs = tcp_init_tso_segs(sk, skb, mss_now); in tcp_write_xmit()
2078 if (unlikely(!tcp_snd_wnd_test(tp, skb, mss_now))) in tcp_write_xmit()
2082 if (unlikely(!tcp_nagle_test(tp, skb, mss_now, in tcp_write_xmit()
2093 limit = mss_now; in tcp_write_xmit()
2095 limit = tcp_mss_split_point(sk, skb, mss_now, in tcp_write_xmit()
2102 unlikely(tso_fragment(sk, skb, limit, mss_now, gfp))) in tcp_write_xmit()
2138 tcp_minshall_update(tp, mss_now, skb); in tcp_write_xmit()
2312 void tcp_push_one(struct sock *sk, unsigned int mss_now) in tcp_push_one() argument
2316 BUG_ON(!skb || skb->len < mss_now); in tcp_push_one()
2318 tcp_write_xmit(sk, mss_now, TCP_NAGLE_PUSH, 1, sk->sk_allocation); in tcp_push_one()