size_goal 695 net/ipv4/tcp.c int size_goal) size_goal 697 net/ipv4/tcp.c return skb->len < size_goal && size_goal 704 net/ipv4/tcp.c int nonagle, int size_goal) size_goal 717 net/ipv4/tcp.c if (tcp_should_autocork(sk, skb, size_goal)) { size_goal 917 net/ipv4/tcp.c u32 new_size_goal, size_goal; size_goal 927 net/ipv4/tcp.c size_goal = tp->gso_segs * mss_now; size_goal 928 net/ipv4/tcp.c if (unlikely(new_size_goal < size_goal || size_goal 929 net/ipv4/tcp.c new_size_goal >= size_goal + mss_now)) { size_goal 932 net/ipv4/tcp.c size_goal = tp->gso_segs * mss_now; size_goal 935 net/ipv4/tcp.c return max(size_goal, mss_now); size_goal 938 net/ipv4/tcp.c static int tcp_send_mss(struct sock *sk, int *size_goal, int flags) size_goal 943 net/ipv4/tcp.c *size_goal = tcp_xmit_size_goal(sk, mss_now, !(flags & MSG_OOB)); size_goal 968 net/ipv4/tcp.c int mss_now, size_goal; size_goal 990 net/ipv4/tcp.c mss_now = tcp_send_mss(sk, &size_goal, flags); size_goal 1002 net/ipv4/tcp.c if (!skb || (copy = size_goal - skb->len) <= 0 || size_goal 1017 net/ipv4/tcp.c copy = size_goal; size_goal 1061 net/ipv4/tcp.c if (skb->len < size_goal || (flags & MSG_OOB)) size_goal 1075 net/ipv4/tcp.c TCP_NAGLE_PUSH, size_goal); size_goal 1081 net/ipv4/tcp.c mss_now = tcp_send_mss(sk, &size_goal, flags); size_goal 1088 net/ipv4/tcp.c tcp_push(sk, flags, mss_now, tp->nonagle, size_goal); size_goal 1193 net/ipv4/tcp.c int mss_now = 0, size_goal, copied_syn = 0; size_goal 1266 net/ipv4/tcp.c mss_now = tcp_send_mss(sk, &size_goal, flags); size_goal 1277 net/ipv4/tcp.c copy = size_goal - skb->len; size_goal 1301 net/ipv4/tcp.c copy = size_goal; size_goal 1385 net/ipv4/tcp.c if (skb->len < size_goal || (flags & MSG_OOB) || unlikely(tp->repair)) size_goal 1400 net/ipv4/tcp.c TCP_NAGLE_PUSH, size_goal); size_goal 1406 net/ipv4/tcp.c mss_now = tcp_send_mss(sk, &size_goal, flags); size_goal 1412 net/ipv4/tcp.c tcp_push(sk, flags, mss_now, tp->nonagle, size_goal);