/linux-4.1.27/include/net/ |
D | inet_connection_sock.h | 145 static inline struct inet_connection_sock *inet_csk(const struct sock *sk) in inet_csk() function 152 return (void *)inet_csk(sk)->icsk_ca_priv; in inet_csk_ca() 174 inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_SCHED; in inet_csk_schedule_ack() 179 return inet_csk(sk)->icsk_ack.pending & ICSK_ACK_SCHED; in inet_csk_ack_scheduled() 184 memset(&inet_csk(sk)->icsk_ack, 0, sizeof(inet_csk(sk)->icsk_ack)); in inet_csk_delack_init() 196 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_clear_xmit_timer() 223 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_reset_xmit_timer() 278 reqsk_queue_add(&inet_csk(sk)->icsk_accept_queue, req, sk, child); in inet_csk_reqsk_queue_add() 287 reqsk_queue_added(&inet_csk(sk)->icsk_accept_queue); in inet_csk_reqsk_queue_added() 292 return reqsk_queue_len(&inet_csk(sk)->icsk_accept_queue); in inet_csk_reqsk_queue_len() [all …]
|
D | tcp.h | 382 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_dec_quickack_mode() 596 if (inet_csk(sk)->icsk_rto > TCP_RTO_MAX) in tcp_bound_rto() 597 inet_csk(sk)->icsk_rto = TCP_RTO_MAX; in tcp_bound_rto() 873 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_ca_needs_ecn() 880 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_set_ca_state() 889 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_ca_event() 972 (1 << inet_csk(sk)->icsk_ca_state); in tcp_in_cwnd_reduction() 1049 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_check_probe_timer() 1174 const int rto = inet_csk(sk)->icsk_rto; in tcp_fin_time()
|
/linux-4.1.27/net/ipv4/ |
D | inet_connection_sock.c | 131 !inet_csk(sk)->icsk_af_ops->bind_conflict(sk, tb, false)) { in inet_csk_get_port() 136 if (!inet_csk(sk)->icsk_af_ops->bind_conflict(sk, tb, false)) { in inet_csk_get_port() 192 if (inet_csk(sk)->icsk_af_ops->bind_conflict(sk, tb, true)) { in inet_csk_get_port() 229 if (!inet_csk(sk)->icsk_bind_hash) in inet_csk_get_port() 231 WARN_ON(inet_csk(sk)->icsk_bind_hash != tb); in inet_csk_get_port() 248 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_wait_for_connect() 296 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_accept() 367 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_init_xmit_timers() 380 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_clear_xmit_timers() 492 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_search_req() [all …]
|
D | tcp_timer.c | 139 if (!inet_csk(sk)->icsk_retransmits) in retransmits_timed_out() 161 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_write_timeout() 222 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_delack_timer_handler() 274 inet_csk(sk)->icsk_ack.blocked = 1; in tcp_delack_timer() 286 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_probe_timer() 336 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_fastopen_synack_timer() 366 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_retransmit_timer() 504 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_write_timer_handler() 577 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_keepalive_timer()
|
D | tcp_cong.c | 152 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_assign_congestion_control() 177 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_init_congestion_control() 186 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_reinit_congestion_control() 199 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_cleanup_congestion_control() 329 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_set_congestion_control()
|
D | tcp_output.c | 74 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_event_new_data_sent() 153 while ((delta -= inet_csk(sk)->icsk_rto) > 0 && cwnd > restart_cwnd) in tcp_cwnd_restart() 164 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_event_data_sent() 832 inet_csk(sk)->icsk_af_ops->mtu_reduced(sk); in tcp_release_cb() 909 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_transmit_skb() 1299 const struct inet_connection_sock *icsk = inet_csk(sk); in __tcp_mtu_to_mss() 1340 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_mss_to_mtu() 1362 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_mtup_init() 1400 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_sync_mss() 1435 if (mtu != inet_csk(sk)->icsk_pmtu_cookie) in tcp_current_mss() [all …]
|
D | tcp_input.c | 132 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_measure_rcv_mss() 178 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_incr_quickack() 189 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_enter_quickack_mode() 201 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_in_quickack_mode() 347 return 2 * inet_csk(sk)->icsk_ack.rcv_mss; in __tcp_grow_window() 377 inet_csk(sk)->icsk_ack.quick |= 1; in tcp_grow_window() 443 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_clamp_window() 474 inet_csk(sk)->icsk_ack.rcv_mss = hint; in tcp_initialize_rcv_mss() 544 TCP_SKB_CB(skb)->seq >= inet_csk(sk)->icsk_ack.rcv_mss)) in tcp_rcv_rtt_measure_ts() 631 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_event_data_recv() [all …]
|
D | tcp_fastopen.c | 133 struct request_sock_queue *queue = &inet_csk(sk)->icsk_accept_queue; in tcp_fastopen_create_child() 141 child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL); in tcp_fastopen_create_child() 175 inet_csk(child)->icsk_af_ops->rebuild_header(child); in tcp_fastopen_create_child() 238 fastopenq = inet_csk(sk)->icsk_accept_queue.fastopenq; in tcp_fastopen_queue_check()
|
D | tcp_dctcp.c | 108 inet_csk(sk)->icsk_ca_ops = &dctcp_reno; in dctcp_init() 196 acked_bytes = inet_csk(sk)->icsk_ack.rcv_mss; in dctcp_update_alpha() 291 if (inet_csk(sk)->icsk_ca_ops != &dctcp_reno) { in dctcp_get_info()
|
D | tcp_minisocks.c | 271 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_time_wait() 405 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_ca_openreq_child() 445 struct inet_connection_sock *newicsk = inet_csk(newsk); in tcp_create_openreq_child() 743 if (req->num_timeout < inet_csk(sk)->icsk_accept_queue.rskq_defer_accept && in tcp_check_req() 756 child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL); in tcp_check_req()
|
D | inet_hashtables.c | 100 inet_csk(sk)->icsk_bind_hash = tb; in inet_bind_hash() 117 tb = inet_csk(sk)->icsk_bind_hash; in __inet_put_port() 120 inet_csk(sk)->icsk_bind_hash = NULL; in __inet_put_port() 144 tb = inet_csk(sk)->icsk_bind_hash; in __inet_inherit_port() 580 tb = inet_csk(sk)->icsk_bind_hash; in __inet_hash_connect()
|
D | tcp_ipv4.c | 209 inet_csk(sk)->icsk_ext_hdr_len = 0; in tcp_v4_connect() 211 inet_csk(sk)->icsk_ext_hdr_len = inet_opt->opt.optlen; in tcp_v4_connect() 292 inet_csk(sk)->icsk_pmtu_cookie > mtu) { in tcp_v4_mtu_reduced() 407 icsk = inet_csk(sk); in tcp_v4_err() 1283 inet_csk(newsk)->icsk_ext_hdr_len = 0; in tcp_v4_syn_recv_sock() 1286 inet_csk(newsk)->icsk_ext_hdr_len = inet_opt->opt.optlen; in tcp_v4_syn_recv_sock() 1761 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_v4_init_sock() 1801 if (inet_csk(sk)->icsk_bind_hash) in tcp_v4_destroy_sock() 1845 icsk = inet_csk(st->syn_wait_sk); in listening_get_next() 1864 icsk = inet_csk(sk); in listening_get_next() [all …]
|
D | tcp_htcp.c | 84 const struct inet_connection_sock *icsk = inet_csk(sk); in measure_rtt() 104 const struct inet_connection_sock *icsk = inet_csk(sk); in measure_achieved_throughput()
|
D | tcp.c | 382 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_init_sock() 1369 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_cleanup_rbuf() 1892 if (inet_csk(sk)->icsk_bind_hash && in tcp_set_state() 2171 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_disconnect() 2235 kfree(inet_csk(sk)->icsk_accept_queue.fastopenq); in tcp_sock_destruct() 2300 struct inet_connection_sock *icsk = inet_csk(sk); in do_tcp_setsockopt() 2579 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_setsockopt() 2604 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_get_info() 2688 struct inet_connection_sock *icsk = inet_csk(sk); in do_tcp_getsockopt() 2845 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_getsockopt()
|
D | tcp_bic.c | 202 const struct inet_connection_sock *icsk = inet_csk(sk); in bictcp_acked()
|
D | tcp_yeah.c | 61 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_yeah_pkts_acked()
|
D | inet_timewait_sock.c | 135 const struct inet_connection_sock *icsk = inet_csk(sk); in __inet_twsk_hashdance()
|
D | tcp_metrics.c | 385 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_update_metrics() 570 inet_csk(sk)->icsk_rto = crtt + max(2 * crtt, tcp_rto_min(sk)); in tcp_init_metrics() 581 inet_csk(sk)->icsk_rto = TCP_TIMEOUT_FALLBACK; in tcp_init_metrics()
|
D | syncookies.c | 226 struct inet_connection_sock *icsk = inet_csk(sk); in get_cookie_sock()
|
D | inet_diag.c | 257 return inet_sk_diag_fill(sk, inet_csk(sk), skb, req, in inet_csk_diag_fill() 734 struct inet_connection_sock *icsk = inet_csk(sk); in inet_diag_dump_reqs()
|
D | cipso_ipv4.c | 1938 sk_conn = inet_csk(sk); in cipso_v4_sock_setattr() 2110 struct inet_connection_sock *sk_conn = inet_csk(sk); in cipso_v4_sock_delattr()
|
D | af_inet.c | 220 !inet_csk(sk)->icsk_accept_queue.fastopenq) { in inet_listen()
|
D | ip_sockglue.c | 638 struct inet_connection_sock *icsk = inet_csk(sk); in do_ip_setsockopt()
|
/linux-4.1.27/net/dccp/ |
D | timer.c | 37 const struct inet_connection_sock *icsk = inet_csk(sk); in dccp_write_timeout() 89 struct inet_connection_sock *icsk = inet_csk(sk); in dccp_retransmit_timer() 131 struct inet_connection_sock *icsk = inet_csk(sk); in dccp_write_timer() 176 struct inet_connection_sock *icsk = inet_csk(sk); in dccp_delack_timer()
|
D | output.c | 49 const struct inet_connection_sock *icsk = inet_csk(sk); in dccp_transmit_skb() 164 struct inet_connection_sock *icsk = inet_csk(sk); in dccp_sync_mss() 271 inet_csk(sk)->icsk_rto, in dccp_xmit_packet() 384 if (inet_csk(sk)->icsk_af_ops->rebuild_header(sk) != 0) in dccp_retransmit_skb() 388 inet_csk(sk)->icsk_retransmits++; in dccp_retransmit_skb() 510 int err = inet_csk(sk)->icsk_af_ops->rebuild_header(sk); in dccp_send_reset() 535 struct inet_connection_sock *icsk = inet_csk(sk); in dccp_connect() 579 inet_csk(sk)->icsk_ack.ato = TCP_ATO_MIN; in dccp_send_ack() 599 struct inet_connection_sock *icsk = inet_csk(sk);
|
D | minisocks.c | 42 const struct inet_connection_sock *icsk = inet_csk(sk); in dccp_time_wait() 89 struct inet_connection_sock *newicsk = inet_csk(newsk); in dccp_create_openreq_child() 185 child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL); in dccp_check_req()
|
D | ipv6.c | 146 if (inet_csk(sk)->icsk_pmtu_cookie > dst_mtu(dst)) in dccp_v6_err() 445 inet_csk(newsk)->icsk_af_ops = &dccp_ipv6_mapped; in dccp_v6_request_recv_sock() 462 dccp_sync_mss(newsk, inet_csk(newsk)->icsk_pmtu_cookie); in dccp_v6_request_recv_sock() 548 inet_csk(newsk)->icsk_ext_hdr_len = 0; in dccp_v6_request_recv_sock() 550 inet_csk(newsk)->icsk_ext_hdr_len = opt->opt_nflen + in dccp_v6_request_recv_sock() 799 struct inet_connection_sock *icsk = inet_csk(sk); in dccp_v6_connect() 1003 inet_csk(sk)->icsk_af_ops = &dccp_ipv6_af_ops; in dccp_v6_init_sock()
|
D | proto.c | 100 if (inet_csk(sk)->icsk_bind_hash != NULL && in dccp_set_state() 176 struct inet_connection_sock *icsk = inet_csk(sk); in dccp_init_sock() 214 if (inet_csk(sk)->icsk_bind_hash != NULL) in dccp_destroy_sock() 253 struct inet_connection_sock *icsk = inet_csk(sk); in dccp_disconnect() 562 return inet_csk(sk)->icsk_af_ops->setsockopt(sk, level, in dccp_setsockopt() 683 return inet_csk(sk)->icsk_af_ops->getsockopt(sk, level, in dccp_getsockopt()
|
D | diag.c | 22 const struct inet_connection_sock *icsk = inet_csk(sk); in dccp_get_info()
|
D | ipv4.c | 96 inet_csk(sk)->icsk_ext_hdr_len = 0; in dccp_v4_connect() 98 inet_csk(sk)->icsk_ext_hdr_len = inet_opt->opt.optlen; in dccp_v4_connect() 176 inet_csk(sk)->icsk_pmtu_cookie > mtu) { in dccp_do_pmtu_discovery() 941 inet_csk(sk)->icsk_af_ops = &dccp_ipv4_af_ops; in dccp_v4_init_sock()
|
D | input.c | 405 const struct inet_connection_sock *icsk = inet_csk(sk); in dccp_rcv_request_sent_state_process() 606 if (inet_csk(sk)->icsk_af_ops->conn_request(sk, in dccp_rcv_state_process()
|
/linux-4.1.27/net/ipv6/ |
D | inet6_connection_sock.c | 123 struct inet_connection_sock *icsk = inet_csk(sk); in inet6_csk_search_req() 153 struct inet_connection_sock *icsk = inet_csk(sk); in inet6_csk_reqsk_queue_hash_add()
|
D | tcp_ipv6.c | 119 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_v6_connect() 317 if (inet_csk(sk)->icsk_pmtu_cookie > dst_mtu(dst)) { in tcp_v6_mtu_reduced() 1031 inet_csk(newsk)->icsk_af_ops = &ipv6_mapped; in tcp_v6_syn_recv_sock() 1057 tcp_sync_mss(newsk, inet_csk(newsk)->icsk_pmtu_cookie); in tcp_v6_syn_recv_sock() 1142 inet_csk(newsk)->icsk_ext_hdr_len = 0; in tcp_v6_syn_recv_sock() 1144 inet_csk(newsk)->icsk_ext_hdr_len = opt->opt_nflen + in tcp_v6_syn_recv_sock() 1626 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_v6_init_sock() 1686 const struct inet_connection_sock *icsk = inet_csk(sp); in get_tcp6_sock()
|
D | syncookies.c | 48 struct inet_connection_sock *icsk = inet_csk(sk); in get_cookie_sock()
|
D | ipv6_sockglue.c | 109 struct inet_connection_sock *icsk = inet_csk(sk); in ipv6_update_options() 212 struct inet_connection_sock *icsk = inet_csk(sk); in do_ipv6_setsockopt()
|
/linux-4.1.27/net/core/ |
D | request_sock.c | 177 fastopenq = inet_csk(lsk)->icsk_accept_queue.fastopenq; in reqsk_fastopen_remove()
|
/linux-4.1.27/include/linux/ |
D | tcp.h | 382 &inet_csk(sk)->icsk_accept_queue; in fastopen_init_queue()
|