mss_cache 114 drivers/crypto/chelsio/chtls/chtls_cm.c tp->mss_cache = tp->rx_opt.mss_clamp; mss_cache 192 drivers/crypto/chelsio/chtls/chtls_io.c FLOWC_PARAM(MSS, tp->mss_cache); mss_cache 667 drivers/infiniband/sw/siw/siw_qp_tx.c c_tx->tcp_seglen = tp->mss_cache * tp->gso_segs; mss_cache 670 drivers/infiniband/sw/siw/siw_qp_tx.c tp->mss_cache * mss_cache 673 drivers/infiniband/sw/siw/siw_qp_tx.c c_tx->tcp_seglen = tp->mss_cache; mss_cache 200 include/linux/tcp.h u32 mss_cache; /* Cached effective mss, not including SACKS */ mss_cache 3085 include/uapi/linux/bpf.h __u32 mss_cache; /* Cached effective mss, not including SACKS */ mss_cache 3340 include/uapi/linux/bpf.h __u32 mss_cache; mss_cache 5633 net/core/filter.c case offsetof(struct bpf_tcp_sock, mss_cache): mss_cache 5634 net/core/filter.c BPF_TCP_SOCK_GET_COMMON(mss_cache); mss_cache 8245 net/core/filter.c case offsetof(struct bpf_sock_ops, mss_cache): mss_cache 8246 net/core/filter.c SOCK_OPS_GET_TCP_SOCK_FIELD(mss_cache); mss_cache 397 net/ipv4/tcp.c rate64 = (u64)rate * tp->mss_cache * USEC_PER_SEC; mss_cache 438 net/ipv4/tcp.c tp->mss_cache = TCP_MSS_DEFAULT; mss_cache 3274 net/ipv4/tcp.c info->tcpi_snd_mss = tp->mss_cache; mss_cache 3429 net/ipv4/tcp.c val = tp->mss_cache; mss_cache 244 net/ipv4/tcp_bbr.c unsigned int mss = tcp_sk(sk)->mss_cache; mss_cache 312 net/ipv4/tcp_bbr.c segs = max_t(u32, bytes / tp->mss_cache, bbr_min_tso_segs(sk)); mss_cache 1113 net/ipv4/tcp_bbr.c bw = bw * tp->mss_cache * USEC_PER_SEC >> BW_SCALE; mss_cache 197 net/ipv4/tcp_dctcp.c info->dctcp.dctcp_ab_ecn = tp->mss_cache * mss_cache 199 net/ipv4/tcp_dctcp.c info->dctcp.dctcp_ab_tot = tp->mss_cache * mss_cache 344 net/ipv4/tcp_input.c per_mss = max_t(u32, tp->rx_opt.mss_clamp, tp->mss_cache) + mss_cache 505 net/ipv4/tcp_input.c unsigned int hint = min_t(unsigned int, tp->advmss, tp->mss_cache); mss_cache 805 net/ipv4/tcp_input.c rate = (u64)tp->mss_cache * ((USEC_PER_SEC / 100) << 3); mss_cache 888 net/ipv4/tcp_input.c const u32 mss = tp->mss_cache; mss_cache 2583 net/ipv4/tcp_input.c tcp_mss_to_mtu(sk, tp->mss_cache) / mss_cache 2776 net/ipv4/tcp_input.c tp->snd_una + tp->reordering * tp->mss_cache); mss_cache 3184 net/ipv4/tcp_input.c if (pkts_acked == 1 && last_in_flight < tp->mss_cache && mss_cache 350 net/ipv4/tcp_nv.c bytes_acked >= (ca->nv_min_cwnd - 1) * tp->mss_cache && mss_cache 356 net/ipv4/tcp_nv.c ca->nv_min_cwnd * tp->mss_cache; mss_cache 368 net/ipv4/tcp_nv.c 80000ULL * tp->mss_cache); mss_cache 1584 net/ipv4/tcp_output.c tp->mss_cache = mss_now; mss_cache 1602 net/ipv4/tcp_output.c mss_now = tp->mss_cache; mss_cache 1971 net/ipv4/tcp_output.c cong_win = (tp->snd_cwnd - in_flight) * tp->mss_cache; mss_cache 1976 net/ipv4/tcp_output.c if (limit >= max_segs * tp->mss_cache) mss_cache 1985 net/ipv4/tcp_output.c u32 chunk = min(tp->snd_wnd, tp->snd_cwnd * tp->mss_cache); mss_cache 1999 net/ipv4/tcp_output.c if (limit > tcp_max_tso_deferred_mss(tp) * tp->mss_cache) mss_cache 2124 net/ipv4/tcp_output.c size_needed = probe_size + (tp->reordering + 1) * tp->mss_cache; mss_cache 188 net/ipv4/tcp_rate.c tp->write_seq - tp->snd_nxt < tp->mss_cache && mss_cache 191 net/ipv4/tcp_westwood.c w->accounted += tp->mss_cache; mss_cache 192 net/ipv4/tcp_westwood.c w->cumul_ack = tp->mss_cache; mss_cache 195 net/ipv4/tcp_westwood.c if (w->cumul_ack > tp->mss_cache) { mss_cache 199 net/ipv4/tcp_westwood.c w->cumul_ack = tp->mss_cache; mss_cache 222 net/ipv4/tcp_westwood.c return max_t(u32, (w->bw_est * w->rtt_min) / tp->mss_cache, 2); mss_cache 3085 tools/include/uapi/linux/bpf.h __u32 mss_cache; /* Cached effective mss, not including SACKS */ mss_cache 3340 tools/include/uapi/linux/bpf.h __u32 mss_cache; mss_cache 116 tools/testing/selftests/bpf/progs/test_sock_fields_kern.c dst->mss_cache = src->mss_cache; mss_cache 114 tools/testing/selftests/bpf/test_sock_fields.c tp->rcv_nxt, tp->snd_nxt, tp->snd_una, tp->mss_cache,