inet_csk 241 drivers/crypto/chelsio/chtls/chtls_cm.c if (inet_csk(sk)->icsk_bind_hash) inet_csk 568 drivers/crypto/chelsio/chtls/chtls_cm.c reqsk_queue_removed(&inet_csk(parent)->icsk_accept_queue, req); inet_csk 898 drivers/crypto/chelsio/chtls/chtls_cm.c inet_csk(sk)->icsk_pmtu_cookie = pmtu; inet_csk 1137 drivers/crypto/chelsio/chtls/chtls_cm.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 1395 drivers/crypto/chelsio/chtls/chtls_cm.c reqsk_queue_removed(&inet_csk(lsk)->icsk_accept_queue, oreq); inet_csk 91 drivers/crypto/chelsio/chtls/chtls_cm.h #define ACCEPT_QUEUE(sk) (&inet_csk(sk)->icsk_accept_queue.rskq_accept_head) inet_csk 776 drivers/xen/pvcalls-back.c icsk = inet_csk(mappass->sock->sk); inet_csk 361 include/linux/skmsg.h struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 456 include/linux/tcp.h struct request_sock_queue *queue = &inet_csk(sk)->icsk_accept_queue; inet_csk 156 include/net/inet_connection_sock.h return (void *)inet_csk(sk)->icsk_ca_priv; inet_csk 179 include/net/inet_connection_sock.h inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_SCHED; inet_csk 184 include/net/inet_connection_sock.h return inet_csk(sk)->icsk_ack.pending & ICSK_ACK_SCHED; inet_csk 189 include/net/inet_connection_sock.h memset(&inet_csk(sk)->icsk_ack, 0, sizeof(inet_csk(sk)->icsk_ack)); inet_csk 197 include/net/inet_connection_sock.h struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 221 include/net/inet_connection_sock.h struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 274 include/net/inet_connection_sock.h reqsk_queue_added(&inet_csk(sk)->icsk_accept_queue); inet_csk 279 include/net/inet_connection_sock.h return reqsk_queue_len(&inet_csk(sk)->icsk_accept_queue); inet_csk 298 include/net/inet_connection_sock.h return !reqsk_queue_empty(&inet_csk(sk)->icsk_accept_queue) ? inet_csk 318 include/net/inet_connection_sock.h inet_csk(sk)->icsk_ack.pingpong = TCP_PINGPONG_THRESH; inet_csk 323 include/net/inet_connection_sock.h inet_csk(sk)->icsk_ack.pingpong = 0; inet_csk 328 include/net/inet_connection_sock.h return inet_csk(sk)->icsk_ack.pingpong >= TCP_PINGPONG_THRESH; inet_csk 333 include/net/inet_connection_sock.h struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 349 include/net/tcp.h struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 662 include/net/tcp.h if (inet_csk(sk)->icsk_rto > TCP_RTO_MAX) inet_csk 663 include/net/tcp.h inet_csk(sk)->icsk_rto = TCP_RTO_MAX; inet_csk 1111 include/net/tcp.h const struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 1118 include/net/tcp.h struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 1127 include/net/tcp.h const struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 1197 include/net/tcp.h (1 << inet_csk(sk)->icsk_ca_state); inet_csk 1302 include/net/tcp.h return max_t(unsigned long, inet_csk(sk)->icsk_rto, TCP_RTO_MIN); inet_csk 1309 include/net/tcp.h u64 when = (u64)tcp_probe0_base(sk) << inet_csk(sk)->icsk_backoff; inet_csk 1316 include/net/tcp.h if (!tcp_sk(sk)->packets_out && !inet_csk(sk)->icsk_pending) inet_csk 1364 include/net/tcp.h const struct tcp_congestion_ops *ca_ops = inet_csk(sk)->icsk_ca_ops; inet_csk 1372 include/net/tcp.h if (delta > inet_csk(sk)->icsk_rto) inet_csk 1456 include/net/tcp.h const int rto = inet_csk(sk)->icsk_rto; inet_csk 1686 include/net/tcp.h ctx = rcu_dereference(inet_csk(sk)->icsk_accept_queue.fastopenq.ctx); inet_csk 2041 include/net/tcp.h u32 rto = inet_csk(sk)->icsk_rto; inet_csk 483 include/net/tls.h struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 4413 net/core/filter.c icsk = inet_csk(sk); inet_csk 97 net/core/request_sock.c fastopenq = &inet_csk(lsk)->icsk_accept_queue.fastopenq; inet_csk 343 net/core/sock_map.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 667 net/core/sock_map.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 19 net/dccp/diag.c const struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 401 net/dccp/input.c const struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 609 net/dccp/input.c acceptable = inet_csk(sk)->icsk_af_ops->conn_request(sk, skb) >= 0; inet_csk 92 net/dccp/ipv4.c inet_csk(sk)->icsk_ext_hdr_len = 0; inet_csk 94 net/dccp/ipv4.c inet_csk(sk)->icsk_ext_hdr_len = inet_opt->opt.optlen; inet_csk 172 net/dccp/ipv4.c inet_csk(sk)->icsk_pmtu_cookie > mtu) { inet_csk 928 net/dccp/ipv4.c inet_csk(sk)->icsk_af_ops = &dccp_ipv4_af_ops; inet_csk 149 net/dccp/ipv6.c if (inet_csk(sk)->icsk_pmtu_cookie > dst_mtu(dst)) inet_csk 429 net/dccp/ipv6.c inet_csk(newsk)->icsk_af_ops = &dccp_ipv6_mapped; inet_csk 449 net/dccp/ipv6.c dccp_sync_mss(newsk, inet_csk(newsk)->icsk_pmtu_cookie); inet_csk 521 net/dccp/ipv6.c inet_csk(newsk)->icsk_ext_hdr_len = 0; inet_csk 523 net/dccp/ipv6.c inet_csk(newsk)->icsk_ext_hdr_len = opt->opt_nflen + inet_csk 807 net/dccp/ipv6.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 1010 net/dccp/ipv6.c inet_csk(sk)->icsk_af_ops = &dccp_ipv6_af_ops; inet_csk 38 net/dccp/minisocks.c const struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 91 net/dccp/minisocks.c struct inet_connection_sock *newicsk = inet_csk(newsk); inet_csk 192 net/dccp/minisocks.c child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL, inet_csk 46 net/dccp/output.c const struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 161 net/dccp/output.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 268 net/dccp/output.c inet_csk(sk)->icsk_rto, inet_csk 381 net/dccp/output.c if (inet_csk(sk)->icsk_af_ops->rebuild_header(sk) != 0) inet_csk 385 net/dccp/output.c inet_csk(sk)->icsk_retransmits++; inet_csk 512 net/dccp/output.c int err = inet_csk(sk)->icsk_af_ops->rebuild_header(sk); inet_csk 537 net/dccp/output.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 581 net/dccp/output.c inet_csk(sk)->icsk_ack.ato = TCP_ATO_MIN; inet_csk 601 net/dccp/output.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 101 net/dccp/proto.c if (inet_csk(sk)->icsk_bind_hash != NULL && inet_csk 186 net/dccp/proto.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 222 net/dccp/proto.c if (inet_csk(sk)->icsk_bind_hash != NULL) inet_csk 260 net/dccp/proto.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 570 net/dccp/proto.c return inet_csk(sk)->icsk_af_ops->setsockopt(sk, level, inet_csk 691 net/dccp/proto.c return inet_csk(sk)->icsk_af_ops->getsockopt(sk, level, inet_csk 33 net/dccp/timer.c const struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 85 net/dccp/timer.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 225 net/ipv4/af_inet.c !inet_csk(sk)->icsk_accept_queue.fastopenq.max_qlen) { inet_csk 1891 net/ipv4/cipso_ipv4.c sk_conn = inet_csk(sk); inet_csk 2061 net/ipv4/cipso_ipv4.c struct inet_connection_sock *sk_conn = inet_csk(sk); inet_csk 379 net/ipv4/inet_connection_sock.c if (!inet_csk(sk)->icsk_bind_hash) inet_csk 381 net/ipv4/inet_connection_sock.c WARN_ON(inet_csk(sk)->icsk_bind_hash != tb); inet_csk 396 net/ipv4/inet_connection_sock.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 444 net/ipv4/inet_connection_sock.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 533 net/ipv4/inet_connection_sock.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 544 net/ipv4/inet_connection_sock.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 697 net/ipv4/inet_connection_sock.c reqsk_queue_removed(&inet_csk(sk)->icsk_accept_queue, req); inet_csk 715 net/ipv4/inet_connection_sock.c struct inet_connection_sock *icsk = inet_csk(sk_listener); inet_csk 813 net/ipv4/inet_connection_sock.c struct inet_connection_sock *newicsk = inet_csk(newsk); inet_csk 859 net/ipv4/inet_connection_sock.c WARN_ON(inet_sk(sk)->inet_num && !inet_csk(sk)->icsk_bind_hash); inet_csk 894 net/ipv4/inet_connection_sock.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 952 net/ipv4/inet_connection_sock.c struct request_sock_queue *queue = &inet_csk(sk)->icsk_accept_queue; inet_csk 978 net/ipv4/inet_connection_sock.c reqsk_queue_removed(&inet_csk(sk)->icsk_accept_queue, req); inet_csk 995 net/ipv4/inet_connection_sock.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 1054 net/ipv4/inet_connection_sock.c const struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 1067 net/ipv4/inet_connection_sock.c const struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 319 net/ipv4/inet_diag.c return inet_sk_diag_fill(sk, inet_csk(sk), skb, req, user_ns, inet_csk 97 net/ipv4/inet_hashtables.c inet_csk(sk)->icsk_bind_hash = tb; inet_csk 112 net/ipv4/inet_hashtables.c tb = inet_csk(sk)->icsk_bind_hash; inet_csk 114 net/ipv4/inet_hashtables.c inet_csk(sk)->icsk_bind_hash = NULL; inet_csk 139 net/ipv4/inet_hashtables.c tb = inet_csk(sk)->icsk_bind_hash; inet_csk 203 net/ipv4/inet_hashtables.c hlist_add_tail_rcu(&inet_csk(sk)->icsk_listen_portaddr_node, inet_csk 206 net/ipv4/inet_hashtables.c hlist_add_head_rcu(&inet_csk(sk)->icsk_listen_portaddr_node, inet_csk 217 net/ipv4/inet_hashtables.c WARN_ON_ONCE(hlist_unhashed(&inet_csk(sk)->icsk_listen_portaddr_node))) inet_csk 223 net/ipv4/inet_hashtables.c hlist_del_init_rcu(&inet_csk(sk)->icsk_listen_portaddr_node); inet_csk 518 net/ipv4/inet_hashtables.c struct inet_bind_bucket *tb = inet_csk(sk)->icsk_bind_hash; inet_csk 528 net/ipv4/inet_hashtables.c inet_csk(sk2)->icsk_bind_hash == tb && inet_csk 638 net/ipv4/inet_hashtables.c tb = inet_csk(sk)->icsk_bind_hash; inet_csk 105 net/ipv4/inet_timewait_sock.c const struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 661 net/ipv4/ip_sockglue.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 205 net/ipv4/syncookies.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 410 net/ipv4/tcp.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 1541 net/ipv4/tcp.c const struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 2257 net/ipv4/tcp.c if (inet_csk(sk)->icsk_bind_hash && inet_csk 2566 net/ipv4/tcp.c inet_csk(sk)->icsk_backoff = 0; inet_csk 2572 net/ipv4/tcp.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 2806 net/ipv4/tcp.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 3170 net/ipv4/tcp.c const struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 3214 net/ipv4/tcp.c const struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 3391 net/ipv4/tcp.c nla_put_u8(stats, TCP_NLA_RECUR_RETRANS, inet_csk(sk)->icsk_retransmits); inet_csk 3398 net/ipv4/tcp.c nla_put_u8(stats, TCP_NLA_CA_STATE, inet_csk(sk)->icsk_ca_state); inet_csk 3414 net/ipv4/tcp.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 3701 net/ipv4/tcp.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 482 net/ipv4/tcp_bbr.c u8 prev_state = bbr->prev_ca_state, state = inet_csk(sk)->icsk_ca_state; inet_csk 191 net/ipv4/tcp_bic.c const struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 160 net/ipv4/tcp_cong.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 179 net/ipv4/tcp_cong.c const struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 193 net/ipv4/tcp_cong.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 207 net/ipv4/tcp_cong.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 339 net/ipv4/tcp_cong.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 98 net/ipv4/tcp_dctcp.c inet_csk(sk)->icsk_ca_ops = &dctcp_reno; inet_csk 156 net/ipv4/tcp_dctcp.c new_state != inet_csk(sk)->icsk_ca_state) inet_csk 193 net/ipv4/tcp_dctcp.c if (inet_csk(sk)->icsk_ca_ops != &dctcp_reno) { inet_csk 29 net/ipv4/tcp_dctcp.h if (inet_csk(sk)->icsk_ack.pending & ICSK_ACK_TIMER) { inet_csk 33 net/ipv4/tcp_dctcp.h inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_NOW; inet_csk 115 net/ipv4/tcp_diag.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 146 net/ipv4/tcp_diag.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 49 net/ipv4/tcp_fastopen.c inet_csk(sk)->icsk_accept_queue.fastopenq.ctx, 1); inet_csk 94 net/ipv4/tcp_fastopen.c q = &inet_csk(sk)->icsk_accept_queue.fastopenq; inet_csk 237 net/ipv4/tcp_fastopen.c struct request_sock_queue *queue = &inet_csk(sk)->icsk_accept_queue; inet_csk 241 net/ipv4/tcp_fastopen.c child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL, inet_csk 303 net/ipv4/tcp_fastopen.c fastopenq = &inet_csk(sk)->icsk_accept_queue.fastopenq; inet_csk 550 net/ipv4/tcp_fastopen.c u32 timeouts = inet_csk(sk)->icsk_retransmits; inet_csk 84 net/ipv4/tcp_htcp.c const struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 104 net/ipv4/tcp_htcp.c const struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 164 net/ipv4/tcp_input.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 215 net/ipv4/tcp_input.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 227 net/ipv4/tcp_input.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 241 net/ipv4/tcp_input.c const struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 263 net/ipv4/tcp_input.c inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_NOW; inet_csk 337 net/ipv4/tcp_input.c const struct tcp_congestion_ops *ca_ops = inet_csk(sk)->icsk_ca_ops; inet_csk 401 net/ipv4/tcp_input.c return 2 * inet_csk(sk)->icsk_ack.rcv_mss; inet_csk 431 net/ipv4/tcp_input.c inet_csk(sk)->icsk_ack.quick |= 1; inet_csk 478 net/ipv4/tcp_input.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 511 net/ipv4/tcp_input.c inet_csk(sk)->icsk_ack.rcv_mss = hint; inet_csk 586 net/ipv4/tcp_input.c TCP_SKB_CB(skb)->seq >= inet_csk(sk)->icsk_ack.rcv_mss) { inet_csk 679 net/ipv4/tcp_input.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 849 net/ipv4/tcp_input.c inet_csk(sk)->icsk_rto = __tcp_set_rto(tp); inet_csk 899 net/ipv4/tcp_input.c tp->rx_opt.sack_ok, inet_csk(sk)->icsk_ca_state, inet_csk 1843 net/ipv4/tcp_input.c if (inet_csk(sk)->icsk_ca_state != TCP_CA_Loss || tp->undo_marker) inet_csk 1991 net/ipv4/tcp_input.c const struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 2029 net/ipv4/tcp_input.c !inet_csk(sk)->icsk_mtup.probe_size; inet_csk 2362 net/ipv4/tcp_input.c const struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 2392 net/ipv4/tcp_input.c DBGUNDO(sk, inet_csk(sk)->icsk_ca_state == TCP_CA_Loss ? "loss" : "retrans"); inet_csk 2394 net/ipv4/tcp_input.c if (inet_csk(sk)->icsk_ca_state == TCP_CA_Loss) inet_csk 2445 net/ipv4/tcp_input.c inet_csk(sk)->icsk_retransmits = 0; inet_csk 2474 net/ipv4/tcp_input.c tp->snd_ssthresh = inet_csk(sk)->icsk_ca_ops->ssthresh(sk); inet_csk 2509 net/ipv4/tcp_input.c if (inet_csk(sk)->icsk_ca_ops->cong_control) inet_csk 2514 net/ipv4/tcp_input.c (inet_csk(sk)->icsk_ca_state == TCP_CA_CWR || tp->undo_marker)) { inet_csk 2527 net/ipv4/tcp_input.c if (inet_csk(sk)->icsk_ca_state < TCP_CA_CWR) { inet_csk 2543 net/ipv4/tcp_input.c if (inet_csk(sk)->icsk_ca_state != state) { inet_csk 2561 net/ipv4/tcp_input.c if (inet_csk(sk)->icsk_ca_state != TCP_CA_CWR) { inet_csk 2568 net/ipv4/tcp_input.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 2578 net/ipv4/tcp_input.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 2601 net/ipv4/tcp_input.c const struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 2794 net/ipv4/tcp_input.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 2963 net/ipv4/tcp_input.c inet_csk(sk)->icsk_backoff = 0; inet_csk 2982 net/ipv4/tcp_input.c const struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 2993 net/ipv4/tcp_input.c const struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 3005 net/ipv4/tcp_input.c u32 rto = inet_csk(sk)->icsk_rto; inet_csk 3074 net/ipv4/tcp_input.c const struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 3253 net/ipv4/tcp_input.c icsk = inet_csk(sk); inet_csk 3276 net/ipv4/tcp_input.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 3300 net/ipv4/tcp_input.c inet_csk(sk)->icsk_ca_state != TCP_CA_Open; inet_csk 3326 net/ipv4/tcp_input.c const struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 3408 net/ipv4/tcp_input.c tcp_sync_mss(sk, inet_csk(sk)->icsk_pmtu_cookie); inet_csk 3541 net/ipv4/tcp_input.c const struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 3587 net/ipv4/tcp_input.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 4077 net/ipv4/tcp_input.c (s32)(tp->rx_opt.ts_recent - tp->rx_opt.rcv_tsval) <= (inet_csk(sk)->icsk_rto * 1024) / HZ); inet_csk 4810 net/ipv4/tcp_input.c inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_NOW; inet_csk 5236 net/ipv4/tcp_input.c if (((tp->rcv_nxt - tp->rcv_wup) > inet_csk(sk)->icsk_ack.rcv_mss && inet_csk 5247 net/ipv4/tcp_input.c inet_csk(sk)->icsk_ack.pending & ICSK_ACK_NOW) { inet_csk 5561 net/ipv4/tcp_input.c inet_csk(sk)->icsk_af_ops->sk_rx_dst_set(sk, skb); inet_csk 5730 net/ipv4/tcp_input.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 5757 net/ipv4/tcp_input.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 5873 net/ipv4/tcp_input.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 6103 net/ipv4/tcp_input.c if (inet_csk(sk)->icsk_ca_state == TCP_CA_Loss) inet_csk 6108 net/ipv4/tcp_input.c inet_csk(sk)->icsk_retransmits = 0; inet_csk 6138 net/ipv4/tcp_input.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 6251 net/ipv4/tcp_input.c if (!inet_csk(sk)->icsk_ca_ops->cong_control) inet_csk 6488 net/ipv4/tcp_input.c struct request_sock_queue *queue = &inet_csk(sk)->icsk_accept_queue; inet_csk 264 net/ipv4/tcp_ipv4.c inet_csk(sk)->icsk_ext_hdr_len = 0; inet_csk 266 net/ipv4/tcp_ipv4.c inet_csk(sk)->icsk_ext_hdr_len = inet_opt->opt.optlen; inet_csk 361 net/ipv4/tcp_ipv4.c inet_csk(sk)->icsk_pmtu_cookie > mtu) { inet_csk 481 net/ipv4/tcp_ipv4.c icsk = inet_csk(sk); inet_csk 1450 net/ipv4/tcp_ipv4.c inet_csk(newsk)->icsk_ext_hdr_len = 0; inet_csk 1452 net/ipv4/tcp_ipv4.c inet_csk(newsk)->icsk_ext_hdr_len = inet_opt->opt.optlen; inet_csk 2080 net/ipv4/tcp_ipv4.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 2124 net/ipv4/tcp_ipv4.c if (inet_csk(sk)->icsk_bind_hash) inet_csk 2429 net/ipv4/tcp_ipv4.c const struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 322 net/ipv4/tcp_metrics.c const struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 502 net/ipv4/tcp_metrics.c inet_csk(sk)->icsk_rto = crtt + max(2 * crtt, tcp_rto_min(sk)); inet_csk 513 net/ipv4/tcp_metrics.c inet_csk(sk)->icsk_rto = TCP_TIMEOUT_FALLBACK; inet_csk 255 net/ipv4/tcp_minisocks.c const struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 408 net/ipv4/tcp_minisocks.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 470 net/ipv4/tcp_minisocks.c newicsk = inet_csk(newsk); inet_csk 757 net/ipv4/tcp_minisocks.c if (req->num_timeout < inet_csk(sk)->icsk_accept_queue.rskq_defer_accept && inet_csk 770 net/ipv4/tcp_minisocks.c child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL, inet_csk 242 net/ipv4/tcp_nv.c const struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 66 net/ipv4/tcp_output.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 151 net/ipv4/tcp_output.c while ((delta -= inet_csk(sk)->icsk_rto) > 0 && cwnd > restart_cwnd) inet_csk 162 net/ipv4/tcp_output.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 896 net/ipv4/tcp_output.c inet_csk(sk)->icsk_af_ops->mtu_reduced(sk); inet_csk 1020 net/ipv4/tcp_output.c const struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 1470 net/ipv4/tcp_output.c const struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 1510 net/ipv4/tcp_output.c const struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 1533 net/ipv4/tcp_output.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 1571 net/ipv4/tcp_output.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 1606 net/ipv4/tcp_output.c if (mtu != inet_csk(sk)->icsk_pmtu_cookie) inet_csk 1632 net/ipv4/tcp_output.c if (inet_csk(sk)->icsk_ca_state == TCP_CA_Open && inet_csk 1648 net/ipv4/tcp_output.c const struct tcp_congestion_ops *ca_ops = inet_csk(sk)->icsk_ca_ops; inet_csk 1671 net/ipv4/tcp_output.c (s32)(tcp_jiffies32 - tp->snd_cwnd_stamp) >= inet_csk(sk)->icsk_rto && inet_csk 1753 net/ipv4/tcp_output.c const struct tcp_congestion_ops *ca_ops = inet_csk(sk)->icsk_ca_ops; inet_csk 1944 net/ipv4/tcp_output.c const struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 2043 net/ipv4/tcp_output.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 2095 net/ipv4/tcp_output.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 2112 net/ipv4/tcp_output.c inet_csk(sk)->icsk_ca_state != TCP_CA_Open || inet_csk 2493 net/ipv4/tcp_output.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 2530 net/ipv4/tcp_output.c jiffies_to_usecs(inet_csk(sk)->icsk_rto) : inet_csk 2578 net/ipv4/tcp_output.c inet_csk(sk)->icsk_pending = 0; inet_csk 2613 net/ipv4/tcp_output.c inet_csk(sk)->icsk_pending = 0; inet_csk 2703 net/ipv4/tcp_output.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 2899 net/ipv4/tcp_output.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 2929 net/ipv4/tcp_output.c if (inet_csk(sk)->icsk_af_ops->rebuild_header(sk)) inet_csk 3048 net/ipv4/tcp_output.c const struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 3111 net/ipv4/tcp_output.c inet_csk(sk)->icsk_rto, inet_csk 3370 net/ipv4/tcp_output.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 3459 net/ipv4/tcp_output.c inet_csk(sk)->icsk_rto = tcp_timeout_init(sk); inet_csk 3460 net/ipv4/tcp_output.c inet_csk(sk)->icsk_retransmits = 0; inet_csk 3501 net/ipv4/tcp_output.c space = __tcp_mtu_to_mss(sk, inet_csk(sk)->icsk_pmtu_cookie) - inet_csk 3580 net/ipv4/tcp_output.c if (inet_csk(sk)->icsk_af_ops->rebuild_header(sk)) inet_csk 3621 net/ipv4/tcp_output.c inet_csk(sk)->icsk_rto, TCP_RTO_MAX); inet_csk 3632 net/ipv4/tcp_output.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 3700 net/ipv4/tcp_output.c inet_csk(sk)->icsk_ack.ato = TCP_ATO_MIN; inet_csk 3818 net/ipv4/tcp_output.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 166 net/ipv4/tcp_rate.c inet_csk(sk)->icsk_ca_state, inet_csk 32 net/ipv4/tcp_recovery.c if (inet_csk(sk)->icsk_ca_state >= TCP_CA_Recovery) inet_csk 127 net/ipv4/tcp_recovery.c timeout, inet_csk(sk)->icsk_rto); inet_csk 174 net/ipv4/tcp_recovery.c if (inet_csk(sk)->icsk_ca_state != TCP_CA_Recovery) { inet_csk 176 net/ipv4/tcp_recovery.c if (!inet_csk(sk)->icsk_ca_ops->cong_control) inet_csk 181 net/ipv4/tcp_recovery.c if (inet_csk(sk)->icsk_pending != ICSK_TIME_RETRANS) inet_csk 233 net/ipv4/tcp_recovery.c const u8 state = inet_csk(sk)->icsk_ca_state; inet_csk 28 net/ipv4/tcp_timer.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 197 net/ipv4/tcp_timer.c if (!inet_csk(sk)->icsk_retransmits) inet_csk 215 net/ipv4/tcp_timer.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 273 net/ipv4/tcp_timer.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 340 net/ipv4/tcp_timer.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 391 net/ipv4/tcp_timer.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 435 net/ipv4/tcp_timer.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 575 net/ipv4/tcp_timer.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 653 net/ipv4/tcp_timer.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 102 net/ipv4/tcp_ulp.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 116 net/ipv4/tcp_ulp.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 134 net/ipv4/tcp_ulp.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 63 net/ipv4/tcp_yeah.c const struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 107 net/ipv6/ipv6_sockglue.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 217 net/ipv6/ipv6_sockglue.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 149 net/ipv6/tcp_ipv6.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 354 net/ipv6/tcp_ipv6.c if (inet_csk(sk)->icsk_pmtu_cookie > dst_mtu(dst)) { inet_csk 1153 net/ipv6/tcp_ipv6.c inet_csk(newsk)->icsk_af_ops = &ipv6_mapped; inet_csk 1180 net/ipv6/tcp_ipv6.c tcp_sync_mss(newsk, inet_csk(newsk)->icsk_pmtu_cookie); inet_csk 1256 net/ipv6/tcp_ipv6.c inet_csk(newsk)->icsk_ext_hdr_len = 0; inet_csk 1258 net/ipv6/tcp_ipv6.c inet_csk(newsk)->icsk_ext_hdr_len = opt->opt_nflen + inet_csk 1805 net/ipv6/tcp_ipv6.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 1866 net/ipv6/tcp_ipv6.c const struct inet_connection_sock *icsk = inet_csk(sp); inet_csk 191 net/tls/tls_device.c clean_acked_data_disable(inet_csk(sk)); inet_csk 1001 net/tls/tls_device.c clean_acked_data_enable(inet_csk(sk), &tls_icsk_clean_acked); inet_csk 1062 net/tls/tls_device.c clean_acked_data_disable(inet_csk(sk)); inet_csk 293 net/tls/tls_main.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 600 net/tls/tls_main.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 642 net/tls/tls_main.c struct inet_connection_sock *icsk = inet_csk(sk); inet_csk 828 net/tls/tls_main.c ctx = rcu_dereference(inet_csk(sk)->icsk_ulp_data);