Lines Matching refs:tp
383 struct tcp_sock *tp = tcp_sk(sk); in tcp_init_sock() local
385 __skb_queue_head_init(&tp->out_of_order_queue); in tcp_init_sock()
387 tcp_prequeue_init(tp); in tcp_init_sock()
388 INIT_LIST_HEAD(&tp->tsq_node); in tcp_init_sock()
391 tp->mdev_us = jiffies_to_usecs(TCP_TIMEOUT_INIT); in tcp_init_sock()
392 tp->rtt_min[0].rtt = ~0U; in tcp_init_sock()
399 tp->snd_cwnd = TCP_INIT_CWND; in tcp_init_sock()
404 tp->snd_ssthresh = TCP_INFINITE_SSTHRESH; in tcp_init_sock()
405 tp->snd_cwnd_clamp = ~0; in tcp_init_sock()
406 tp->mss_cache = TCP_MSS_DEFAULT; in tcp_init_sock()
407 u64_stats_init(&tp->syncp); in tcp_init_sock()
409 tp->reordering = sysctl_tcp_reordering; in tcp_init_sock()
410 tcp_enable_early_retrans(tp); in tcp_init_sock()
413 tp->tsoffset = 0; in tcp_init_sock()
454 const struct tcp_sock *tp = tcp_sk(sk); in tcp_poll() local
506 (state != TCP_SYN_RECV || tp->fastopen_rsk)) { in tcp_poll()
509 if (tp->urg_seq == tp->copied_seq && in tcp_poll()
511 tp->urg_data) in tcp_poll()
514 if (tp->rcv_nxt - tp->copied_seq >= target) in tcp_poll()
536 if (tp->urg_data & TCP_URG_VALID) in tcp_poll()
550 struct tcp_sock *tp = tcp_sk(sk); in tcp_ioctl() local
563 !tp->urg_data || in tcp_ioctl()
564 before(tp->urg_seq, tp->copied_seq) || in tcp_ioctl()
565 !before(tp->urg_seq, tp->rcv_nxt)) { in tcp_ioctl()
567 answ = tp->rcv_nxt - tp->copied_seq; in tcp_ioctl()
573 answ = tp->urg_seq - tp->copied_seq; in tcp_ioctl()
577 answ = tp->urg_data && tp->urg_seq == tp->copied_seq; in tcp_ioctl()
586 answ = tp->write_seq - tp->snd_una; in tcp_ioctl()
595 answ = tp->write_seq - tp->snd_nxt; in tcp_ioctl()
605 static inline void tcp_mark_push(struct tcp_sock *tp, struct sk_buff *skb) in tcp_mark_push() argument
608 tp->pushed_seq = tp->write_seq; in tcp_mark_push()
611 static inline bool forced_push(const struct tcp_sock *tp) in forced_push() argument
613 return after(tp->write_seq, tp->pushed_seq + (tp->max_window >> 1)); in forced_push()
618 struct tcp_sock *tp = tcp_sk(sk); in skb_entail() local
622 tcb->seq = tcb->end_seq = tp->write_seq; in skb_entail()
629 if (tp->nonagle & TCP_NAGLE_PUSH) in skb_entail()
630 tp->nonagle &= ~TCP_NAGLE_PUSH; in skb_entail()
635 static inline void tcp_mark_urg(struct tcp_sock *tp, int flags) in tcp_mark_urg() argument
638 tp->snd_up = tp->write_seq; in tcp_mark_urg()
663 struct tcp_sock *tp = tcp_sk(sk); in tcp_push() local
670 if (!(flags & MSG_MORE) || forced_push(tp)) in tcp_push()
671 tcp_mark_push(tp, skb); in tcp_push()
673 tcp_mark_urg(tp, flags); in tcp_push()
678 if (!test_bit(TSQ_THROTTLED, &tp->tsq_flags)) { in tcp_push()
680 set_bit(TSQ_THROTTLED, &tp->tsq_flags); in tcp_push()
857 struct tcp_sock *tp = tcp_sk(sk); in tcp_xmit_size_goal() local
865 new_size_goal = tcp_bound_to_half_wnd(tp, new_size_goal); in tcp_xmit_size_goal()
868 size_goal = tp->gso_segs * mss_now; in tcp_xmit_size_goal()
871 tp->gso_segs = min_t(u16, new_size_goal / mss_now, in tcp_xmit_size_goal()
873 size_goal = tp->gso_segs * mss_now; in tcp_xmit_size_goal()
892 struct tcp_sock *tp = tcp_sk(sk); in do_tcp_sendpages() local
943 tcp_mark_push(tp, skb); in do_tcp_sendpages()
963 tp->write_seq += copy; in do_tcp_sendpages()
981 if (forced_push(tp)) { in do_tcp_sendpages()
982 tcp_mark_push(tp, skb); in do_tcp_sendpages()
1003 tcp_push(sk, flags, mss_now, tp->nonagle, size_goal); in do_tcp_sendpages()
1035 const struct tcp_sock *tp = tcp_sk(sk); in select_size() local
1036 int tmp = tp->mss_cache; in select_size()
1056 void tcp_free_fastopen_req(struct tcp_sock *tp) in tcp_free_fastopen_req() argument
1058 if (tp->fastopen_req) { in tcp_free_fastopen_req()
1059 kfree(tp->fastopen_req); in tcp_free_fastopen_req()
1060 tp->fastopen_req = NULL; in tcp_free_fastopen_req()
1067 struct tcp_sock *tp = tcp_sk(sk); in tcp_sendmsg_fastopen() local
1072 if (tp->fastopen_req) in tcp_sendmsg_fastopen()
1075 tp->fastopen_req = kzalloc(sizeof(struct tcp_fastopen_request), in tcp_sendmsg_fastopen()
1077 if (unlikely(!tp->fastopen_req)) in tcp_sendmsg_fastopen()
1079 tp->fastopen_req->data = msg; in tcp_sendmsg_fastopen()
1080 tp->fastopen_req->size = size; in tcp_sendmsg_fastopen()
1085 *copied = tp->fastopen_req->copied; in tcp_sendmsg_fastopen()
1086 tcp_free_fastopen_req(tp); in tcp_sendmsg_fastopen()
1092 struct tcp_sock *tp = tcp_sk(sk); in tcp_sendmsg() local
1123 if (unlikely(tp->repair)) { in tcp_sendmsg()
1124 if (tp->repair_queue == TCP_RECV_QUEUE) { in tcp_sendmsg()
1130 if (tp->repair_queue == TCP_NO_QUEUE) in tcp_sendmsg()
1190 if (tp->repair) in tcp_sendmsg()
1216 tcp_mark_push(tp, skb); in tcp_sendmsg()
1248 tp->write_seq += copy; in tcp_sendmsg()
1258 if (skb->len < max || (flags & MSG_OOB) || unlikely(tp->repair)) in tcp_sendmsg()
1261 if (forced_push(tp)) { in tcp_sendmsg()
1262 tcp_mark_push(tp, skb); in tcp_sendmsg()
1284 tcp_push(sk, flags, mss_now, tp->nonagle, size_goal); in tcp_sendmsg()
1319 struct tcp_sock *tp = tcp_sk(sk); in tcp_recv_urg() local
1322 if (sock_flag(sk, SOCK_URGINLINE) || !tp->urg_data || in tcp_recv_urg()
1323 tp->urg_data == TCP_URG_READ) in tcp_recv_urg()
1329 if (tp->urg_data & TCP_URG_VALID) { in tcp_recv_urg()
1331 char c = tp->urg_data; in tcp_recv_urg()
1334 tp->urg_data = TCP_URG_READ; in tcp_recv_urg()
1387 struct tcp_sock *tp = tcp_sk(sk); in tcp_cleanup_rbuf() local
1392 WARN(skb && !before(tp->copied_seq, TCP_SKB_CB(skb)->end_seq), in tcp_cleanup_rbuf()
1394 tp->copied_seq, TCP_SKB_CB(skb)->end_seq, tp->rcv_nxt); in tcp_cleanup_rbuf()
1402 tp->rcv_nxt - tp->rcv_wup > icsk->icsk_ack.rcv_mss || in tcp_cleanup_rbuf()
1424 __u32 rcv_window_now = tcp_receive_window(tp); in tcp_cleanup_rbuf()
1427 if (2*rcv_window_now <= tp->window_clamp) { in tcp_cleanup_rbuf()
1446 struct tcp_sock *tp = tcp_sk(sk); in tcp_prequeue_process() local
1453 while ((skb = __skb_dequeue(&tp->ucopy.prequeue)) != NULL) in tcp_prequeue_process()
1458 tp->ucopy.memory = 0; in tcp_prequeue_process()
1498 struct tcp_sock *tp = tcp_sk(sk); in tcp_read_sock() local
1499 u32 seq = tp->copied_seq; in tcp_read_sock()
1512 if (tp->urg_data) { in tcp_read_sock()
1513 u32 urg_offset = tp->urg_seq - seq; in tcp_read_sock()
1551 tp->copied_seq = seq; in tcp_read_sock()
1553 tp->copied_seq = seq; in tcp_read_sock()
1577 struct tcp_sock *tp = tcp_sk(sk); in tcp_recvmsg() local
1608 if (unlikely(tp->repair)) { in tcp_recvmsg()
1613 if (tp->repair_queue == TCP_SEND_QUEUE) in tcp_recvmsg()
1617 if (tp->repair_queue == TCP_NO_QUEUE) in tcp_recvmsg()
1623 seq = &tp->copied_seq; in tcp_recvmsg()
1625 peek_seq = tp->copied_seq; in tcp_recvmsg()
1635 if (tp->urg_data && tp->urg_seq == *seq) { in tcp_recvmsg()
1654 *seq, TCP_SKB_CB(skb)->seq, tp->rcv_nxt, in tcp_recvmsg()
1667 *seq, TCP_SKB_CB(skb)->seq, tp->rcv_nxt, flags); in tcp_recvmsg()
1718 if (!sysctl_tcp_low_latency && tp->ucopy.task == user_recv) { in tcp_recvmsg()
1722 tp->ucopy.task = user_recv; in tcp_recvmsg()
1723 tp->ucopy.msg = msg; in tcp_recvmsg()
1726 tp->ucopy.len = len; in tcp_recvmsg()
1728 WARN_ON(tp->copied_seq != tp->rcv_nxt && in tcp_recvmsg()
1757 if (!skb_queue_empty(&tp->ucopy.prequeue)) in tcp_recvmsg()
1776 chunk = len - tp->ucopy.len; in tcp_recvmsg()
1783 if (tp->rcv_nxt == tp->copied_seq && in tcp_recvmsg()
1784 !skb_queue_empty(&tp->ucopy.prequeue)) { in tcp_recvmsg()
1788 chunk = len - tp->ucopy.len; in tcp_recvmsg()
1797 (peek_seq - copied - urg_hole != tp->copied_seq)) { in tcp_recvmsg()
1801 peek_seq = tp->copied_seq; in tcp_recvmsg()
1812 if (tp->urg_data) { in tcp_recvmsg()
1813 u32 urg_offset = tp->urg_seq - *seq; in tcp_recvmsg()
1846 if (tp->urg_data && after(tp->copied_seq, tp->urg_seq)) { in tcp_recvmsg()
1847 tp->urg_data = 0; in tcp_recvmsg()
1868 if (!skb_queue_empty(&tp->ucopy.prequeue)) { in tcp_recvmsg()
1871 tp->ucopy.len = copied > 0 ? len : 0; in tcp_recvmsg()
1875 if (copied > 0 && (chunk = len - tp->ucopy.len) != 0) { in tcp_recvmsg()
1882 tp->ucopy.task = NULL; in tcp_recvmsg()
1883 tp->ucopy.len = 0; in tcp_recvmsg()
2145 struct tcp_sock *tp = tcp_sk(sk); in tcp_close() local
2146 if (tp->linger2 < 0) { in tcp_close()
2205 struct tcp_sock *tp = tcp_sk(sk); in tcp_disconnect() local
2215 } else if (unlikely(tp->repair)) { in tcp_disconnect()
2218 (tp->snd_nxt != tp->write_seq && in tcp_disconnect()
2231 __skb_queue_purge(&tp->out_of_order_queue); in tcp_disconnect()
2240 tp->srtt_us = 0; in tcp_disconnect()
2241 tp->write_seq += tp->max_window + 2; in tcp_disconnect()
2242 if (tp->write_seq == 0) in tcp_disconnect()
2243 tp->write_seq = 1; in tcp_disconnect()
2245 tp->snd_cwnd = 2; in tcp_disconnect()
2247 tp->packets_out = 0; in tcp_disconnect()
2248 tp->snd_ssthresh = TCP_INFINITE_SSTHRESH; in tcp_disconnect()
2249 tp->snd_cwnd_cnt = 0; in tcp_disconnect()
2250 tp->window_clamp = 0; in tcp_disconnect()
2252 tcp_clear_retrans(tp); in tcp_disconnect()
2255 memset(&tp->rx_opt, 0, sizeof(tp->rx_opt)); in tcp_disconnect()
2271 static int tcp_repair_options_est(struct tcp_sock *tp, in tcp_repair_options_est() argument
2285 tp->rx_opt.mss_clamp = opt.opt_val; in tcp_repair_options_est()
2295 tp->rx_opt.snd_wscale = snd_wscale; in tcp_repair_options_est()
2296 tp->rx_opt.rcv_wscale = rcv_wscale; in tcp_repair_options_est()
2297 tp->rx_opt.wscale_ok = 1; in tcp_repair_options_est()
2304 tp->rx_opt.sack_ok |= TCP_SACK_SEEN; in tcp_repair_options_est()
2306 tcp_enable_fack(tp); in tcp_repair_options_est()
2312 tp->rx_opt.tstamp_ok = 1; in tcp_repair_options_est()
2326 struct tcp_sock *tp = tcp_sk(sk); in do_tcp_setsockopt() local
2372 tp->rx_opt.user_mss = val; in do_tcp_setsockopt()
2385 tp->nonagle |= TCP_NAGLE_OFF|TCP_NAGLE_PUSH; in do_tcp_setsockopt()
2388 tp->nonagle &= ~TCP_NAGLE_OFF; in do_tcp_setsockopt()
2396 tp->thin_lto = val; in do_tcp_setsockopt()
2403 tp->thin_dupack = val; in do_tcp_setsockopt()
2404 if (tp->thin_dupack) in do_tcp_setsockopt()
2405 tcp_disable_early_retrans(tp); in do_tcp_setsockopt()
2413 tp->repair = 1; in do_tcp_setsockopt()
2415 tp->repair_queue = TCP_NO_QUEUE; in do_tcp_setsockopt()
2417 tp->repair = 0; in do_tcp_setsockopt()
2426 if (!tp->repair) in do_tcp_setsockopt()
2429 tp->repair_queue = val; in do_tcp_setsockopt()
2437 else if (tp->repair_queue == TCP_SEND_QUEUE) in do_tcp_setsockopt()
2438 tp->write_seq = val; in do_tcp_setsockopt()
2439 else if (tp->repair_queue == TCP_RECV_QUEUE) in do_tcp_setsockopt()
2440 tp->rcv_nxt = val; in do_tcp_setsockopt()
2446 if (!tp->repair) in do_tcp_setsockopt()
2449 err = tcp_repair_options_est(tp, in do_tcp_setsockopt()
2469 tp->nonagle |= TCP_NAGLE_CORK; in do_tcp_setsockopt()
2471 tp->nonagle &= ~TCP_NAGLE_CORK; in do_tcp_setsockopt()
2472 if (tp->nonagle&TCP_NAGLE_OFF) in do_tcp_setsockopt()
2473 tp->nonagle |= TCP_NAGLE_PUSH; in do_tcp_setsockopt()
2482 tp->keepalive_time = val * HZ; in do_tcp_setsockopt()
2486 u32 elapsed = keepalive_time_elapsed(tp); in do_tcp_setsockopt()
2487 if (tp->keepalive_time > elapsed) in do_tcp_setsockopt()
2488 elapsed = tp->keepalive_time - elapsed; in do_tcp_setsockopt()
2499 tp->keepalive_intvl = val * HZ; in do_tcp_setsockopt()
2505 tp->keepalive_probes = val; in do_tcp_setsockopt()
2518 tp->save_syn = val; in do_tcp_setsockopt()
2523 tp->linger2 = -1; in do_tcp_setsockopt()
2525 tp->linger2 = 0; in do_tcp_setsockopt()
2527 tp->linger2 = val * HZ; in do_tcp_setsockopt()
2543 tp->window_clamp = 0; in do_tcp_setsockopt()
2545 tp->window_clamp = val < SOCK_MIN_RCVBUF / 2 ? in do_tcp_setsockopt()
2568 err = tp->af_specific->md5_parse(sk, optval, optlen); in do_tcp_setsockopt()
2592 if (!tp->repair) in do_tcp_setsockopt()
2595 tp->tsoffset = val - tcp_time_stamp; in do_tcp_setsockopt()
2598 tp->notsent_lowat = val; in do_tcp_setsockopt()
2637 const struct tcp_sock *tp = tcp_sk(sk); /* iff sk_type == SOCK_STREAM */ in tcp_get_info() local
2655 if (tp->rx_opt.tstamp_ok) in tcp_get_info()
2657 if (tcp_is_sack(tp)) in tcp_get_info()
2659 if (tp->rx_opt.wscale_ok) { in tcp_get_info()
2661 info->tcpi_snd_wscale = tp->rx_opt.snd_wscale; in tcp_get_info()
2662 info->tcpi_rcv_wscale = tp->rx_opt.rcv_wscale; in tcp_get_info()
2665 if (tp->ecn_flags & TCP_ECN_OK) in tcp_get_info()
2667 if (tp->ecn_flags & TCP_ECN_SEEN) in tcp_get_info()
2669 if (tp->syn_data_acked) in tcp_get_info()
2674 info->tcpi_snd_mss = tp->mss_cache; in tcp_get_info()
2681 info->tcpi_unacked = tp->packets_out; in tcp_get_info()
2682 info->tcpi_sacked = tp->sacked_out; in tcp_get_info()
2684 info->tcpi_lost = tp->lost_out; in tcp_get_info()
2685 info->tcpi_retrans = tp->retrans_out; in tcp_get_info()
2686 info->tcpi_fackets = tp->fackets_out; in tcp_get_info()
2688 info->tcpi_last_data_sent = jiffies_to_msecs(now - tp->lsndtime); in tcp_get_info()
2690 info->tcpi_last_ack_recv = jiffies_to_msecs(now - tp->rcv_tstamp); in tcp_get_info()
2693 info->tcpi_rcv_ssthresh = tp->rcv_ssthresh; in tcp_get_info()
2694 info->tcpi_rtt = tp->srtt_us >> 3; in tcp_get_info()
2695 info->tcpi_rttvar = tp->mdev_us >> 2; in tcp_get_info()
2696 info->tcpi_snd_ssthresh = tp->snd_ssthresh; in tcp_get_info()
2697 info->tcpi_snd_cwnd = tp->snd_cwnd; in tcp_get_info()
2698 info->tcpi_advmss = tp->advmss; in tcp_get_info()
2699 info->tcpi_reordering = tp->reordering; in tcp_get_info()
2701 info->tcpi_rcv_rtt = jiffies_to_usecs(tp->rcv_rtt_est.rtt)>>3; in tcp_get_info()
2702 info->tcpi_rcv_space = tp->rcvq_space.space; in tcp_get_info()
2704 info->tcpi_total_retrans = tp->total_retrans; in tcp_get_info()
2715 start = u64_stats_fetch_begin_irq(&tp->syncp); in tcp_get_info()
2716 put_unaligned(tp->bytes_acked, &info->tcpi_bytes_acked); in tcp_get_info()
2717 put_unaligned(tp->bytes_received, &info->tcpi_bytes_received); in tcp_get_info()
2718 } while (u64_stats_fetch_retry_irq(&tp->syncp, start)); in tcp_get_info()
2719 info->tcpi_segs_out = tp->segs_out; in tcp_get_info()
2720 info->tcpi_segs_in = tp->segs_in; in tcp_get_info()
2728 struct tcp_sock *tp = tcp_sk(sk); in do_tcp_getsockopt() local
2741 val = tp->mss_cache; in do_tcp_getsockopt()
2743 val = tp->rx_opt.user_mss; in do_tcp_getsockopt()
2744 if (tp->repair) in do_tcp_getsockopt()
2745 val = tp->rx_opt.mss_clamp; in do_tcp_getsockopt()
2748 val = !!(tp->nonagle&TCP_NAGLE_OFF); in do_tcp_getsockopt()
2751 val = !!(tp->nonagle&TCP_NAGLE_CORK); in do_tcp_getsockopt()
2754 val = keepalive_time_when(tp) / HZ; in do_tcp_getsockopt()
2757 val = keepalive_intvl_when(tp) / HZ; in do_tcp_getsockopt()
2760 val = keepalive_probes(tp); in do_tcp_getsockopt()
2766 val = tp->linger2; in do_tcp_getsockopt()
2775 val = tp->window_clamp; in do_tcp_getsockopt()
2827 val = tp->thin_lto; in do_tcp_getsockopt()
2830 val = tp->thin_dupack; in do_tcp_getsockopt()
2834 val = tp->repair; in do_tcp_getsockopt()
2838 if (tp->repair) in do_tcp_getsockopt()
2839 val = tp->repair_queue; in do_tcp_getsockopt()
2845 if (tp->repair_queue == TCP_SEND_QUEUE) in do_tcp_getsockopt()
2846 val = tp->write_seq; in do_tcp_getsockopt()
2847 else if (tp->repair_queue == TCP_RECV_QUEUE) in do_tcp_getsockopt()
2848 val = tp->rcv_nxt; in do_tcp_getsockopt()
2862 val = tcp_time_stamp + tp->tsoffset; in do_tcp_getsockopt()
2865 val = tp->notsent_lowat; in do_tcp_getsockopt()
2868 val = tp->save_syn; in do_tcp_getsockopt()
2875 if (tp->saved_syn) { in do_tcp_getsockopt()
2876 if (len < tp->saved_syn[0]) { in do_tcp_getsockopt()
2877 if (put_user(tp->saved_syn[0], optlen)) { in do_tcp_getsockopt()
2884 len = tp->saved_syn[0]; in do_tcp_getsockopt()
2889 if (copy_to_user(optval, tp->saved_syn + 1, len)) { in do_tcp_getsockopt()
2893 tcp_saved_syn_free(tp); in do_tcp_getsockopt()
3022 const struct tcphdr *tp = tcp_hdr(skb); in tcp_md5_hash_skb_data() local
3032 sg_set_buf(&sg, ((u8 *) tp) + header_len, head_data_len); in tcp_md5_hash_skb_data()