inet_csk          241 drivers/crypto/chelsio/chtls/chtls_cm.c 	if (inet_csk(sk)->icsk_bind_hash)
inet_csk          568 drivers/crypto/chelsio/chtls/chtls_cm.c 	reqsk_queue_removed(&inet_csk(parent)->icsk_accept_queue, req);
inet_csk          898 drivers/crypto/chelsio/chtls/chtls_cm.c 	inet_csk(sk)->icsk_pmtu_cookie = pmtu;
inet_csk         1137 drivers/crypto/chelsio/chtls/chtls_cm.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         1395 drivers/crypto/chelsio/chtls/chtls_cm.c 	reqsk_queue_removed(&inet_csk(lsk)->icsk_accept_queue, oreq);
inet_csk           91 drivers/crypto/chelsio/chtls/chtls_cm.h #define ACCEPT_QUEUE(sk) (&inet_csk(sk)->icsk_accept_queue.rskq_accept_head)
inet_csk          776 drivers/xen/pvcalls-back.c 	icsk = inet_csk(mappass->sock->sk);
inet_csk          361 include/linux/skmsg.h 		struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          456 include/linux/tcp.h 	struct request_sock_queue *queue = &inet_csk(sk)->icsk_accept_queue;
inet_csk          156 include/net/inet_connection_sock.h 	return (void *)inet_csk(sk)->icsk_ca_priv;
inet_csk          179 include/net/inet_connection_sock.h 	inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_SCHED;
inet_csk          184 include/net/inet_connection_sock.h 	return inet_csk(sk)->icsk_ack.pending & ICSK_ACK_SCHED;
inet_csk          189 include/net/inet_connection_sock.h 	memset(&inet_csk(sk)->icsk_ack, 0, sizeof(inet_csk(sk)->icsk_ack));
inet_csk          197 include/net/inet_connection_sock.h 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          221 include/net/inet_connection_sock.h 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          274 include/net/inet_connection_sock.h 	reqsk_queue_added(&inet_csk(sk)->icsk_accept_queue);
inet_csk          279 include/net/inet_connection_sock.h 	return reqsk_queue_len(&inet_csk(sk)->icsk_accept_queue);
inet_csk          298 include/net/inet_connection_sock.h 	return !reqsk_queue_empty(&inet_csk(sk)->icsk_accept_queue) ?
inet_csk          318 include/net/inet_connection_sock.h 	inet_csk(sk)->icsk_ack.pingpong = TCP_PINGPONG_THRESH;
inet_csk          323 include/net/inet_connection_sock.h 	inet_csk(sk)->icsk_ack.pingpong = 0;
inet_csk          328 include/net/inet_connection_sock.h 	return inet_csk(sk)->icsk_ack.pingpong >= TCP_PINGPONG_THRESH;
inet_csk          333 include/net/inet_connection_sock.h 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          349 include/net/tcp.h 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          662 include/net/tcp.h 	if (inet_csk(sk)->icsk_rto > TCP_RTO_MAX)
inet_csk          663 include/net/tcp.h 		inet_csk(sk)->icsk_rto = TCP_RTO_MAX;
inet_csk         1111 include/net/tcp.h 	const struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         1118 include/net/tcp.h 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         1127 include/net/tcp.h 	const struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         1197 include/net/tcp.h 	       (1 << inet_csk(sk)->icsk_ca_state);
inet_csk         1302 include/net/tcp.h 	return max_t(unsigned long, inet_csk(sk)->icsk_rto, TCP_RTO_MIN);
inet_csk         1309 include/net/tcp.h 	u64 when = (u64)tcp_probe0_base(sk) << inet_csk(sk)->icsk_backoff;
inet_csk         1316 include/net/tcp.h 	if (!tcp_sk(sk)->packets_out && !inet_csk(sk)->icsk_pending)
inet_csk         1364 include/net/tcp.h 	const struct tcp_congestion_ops *ca_ops = inet_csk(sk)->icsk_ca_ops;
inet_csk         1372 include/net/tcp.h 	if (delta > inet_csk(sk)->icsk_rto)
inet_csk         1456 include/net/tcp.h 	const int rto = inet_csk(sk)->icsk_rto;
inet_csk         1686 include/net/tcp.h 	ctx = rcu_dereference(inet_csk(sk)->icsk_accept_queue.fastopenq.ctx);
inet_csk         2041 include/net/tcp.h 	u32 rto = inet_csk(sk)->icsk_rto;
inet_csk          483 include/net/tls.h 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         4413 net/core/filter.c 			icsk = inet_csk(sk);
inet_csk           97 net/core/request_sock.c 	fastopenq = &inet_csk(lsk)->icsk_accept_queue.fastopenq;
inet_csk          343 net/core/sock_map.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          667 net/core/sock_map.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk           19 net/dccp/diag.c 	const struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          401 net/dccp/input.c 		const struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          609 net/dccp/input.c 			acceptable = inet_csk(sk)->icsk_af_ops->conn_request(sk, skb) >= 0;
inet_csk           92 net/dccp/ipv4.c 	inet_csk(sk)->icsk_ext_hdr_len = 0;
inet_csk           94 net/dccp/ipv4.c 		inet_csk(sk)->icsk_ext_hdr_len = inet_opt->opt.optlen;
inet_csk          172 net/dccp/ipv4.c 	    inet_csk(sk)->icsk_pmtu_cookie > mtu) {
inet_csk          928 net/dccp/ipv4.c 		inet_csk(sk)->icsk_af_ops = &dccp_ipv4_af_ops;
inet_csk          149 net/dccp/ipv6.c 		if (inet_csk(sk)->icsk_pmtu_cookie > dst_mtu(dst))
inet_csk          429 net/dccp/ipv6.c 		inet_csk(newsk)->icsk_af_ops = &dccp_ipv6_mapped;
inet_csk          449 net/dccp/ipv6.c 		dccp_sync_mss(newsk, inet_csk(newsk)->icsk_pmtu_cookie);
inet_csk          521 net/dccp/ipv6.c 	inet_csk(newsk)->icsk_ext_hdr_len = 0;
inet_csk          523 net/dccp/ipv6.c 		inet_csk(newsk)->icsk_ext_hdr_len = opt->opt_nflen +
inet_csk          807 net/dccp/ipv6.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         1010 net/dccp/ipv6.c 		inet_csk(sk)->icsk_af_ops = &dccp_ipv6_af_ops;
inet_csk           38 net/dccp/minisocks.c 		const struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk           91 net/dccp/minisocks.c 		struct inet_connection_sock *newicsk = inet_csk(newsk);
inet_csk          192 net/dccp/minisocks.c 	child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL,
inet_csk           46 net/dccp/output.c 		const struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          161 net/dccp/output.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          268 net/dccp/output.c 					      inet_csk(sk)->icsk_rto,
inet_csk          381 net/dccp/output.c 	if (inet_csk(sk)->icsk_af_ops->rebuild_header(sk) != 0)
inet_csk          385 net/dccp/output.c 	inet_csk(sk)->icsk_retransmits++;
inet_csk          512 net/dccp/output.c 	int err = inet_csk(sk)->icsk_af_ops->rebuild_header(sk);
inet_csk          537 net/dccp/output.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          581 net/dccp/output.c 			inet_csk(sk)->icsk_ack.ato = TCP_ATO_MIN;
inet_csk          601 net/dccp/output.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          101 net/dccp/proto.c 		if (inet_csk(sk)->icsk_bind_hash != NULL &&
inet_csk          186 net/dccp/proto.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          222 net/dccp/proto.c 	if (inet_csk(sk)->icsk_bind_hash != NULL)
inet_csk          260 net/dccp/proto.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          570 net/dccp/proto.c 		return inet_csk(sk)->icsk_af_ops->setsockopt(sk, level,
inet_csk          691 net/dccp/proto.c 		return inet_csk(sk)->icsk_af_ops->getsockopt(sk, level,
inet_csk           33 net/dccp/timer.c 	const struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk           85 net/dccp/timer.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          225 net/ipv4/af_inet.c 		    !inet_csk(sk)->icsk_accept_queue.fastopenq.max_qlen) {
inet_csk         1891 net/ipv4/cipso_ipv4.c 		sk_conn = inet_csk(sk);
inet_csk         2061 net/ipv4/cipso_ipv4.c 		struct inet_connection_sock *sk_conn = inet_csk(sk);
inet_csk          379 net/ipv4/inet_connection_sock.c 	if (!inet_csk(sk)->icsk_bind_hash)
inet_csk          381 net/ipv4/inet_connection_sock.c 	WARN_ON(inet_csk(sk)->icsk_bind_hash != tb);
inet_csk          396 net/ipv4/inet_connection_sock.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          444 net/ipv4/inet_connection_sock.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          533 net/ipv4/inet_connection_sock.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          544 net/ipv4/inet_connection_sock.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          697 net/ipv4/inet_connection_sock.c 		reqsk_queue_removed(&inet_csk(sk)->icsk_accept_queue, req);
inet_csk          715 net/ipv4/inet_connection_sock.c 	struct inet_connection_sock *icsk = inet_csk(sk_listener);
inet_csk          813 net/ipv4/inet_connection_sock.c 		struct inet_connection_sock *newicsk = inet_csk(newsk);
inet_csk          859 net/ipv4/inet_connection_sock.c 	WARN_ON(inet_sk(sk)->inet_num && !inet_csk(sk)->icsk_bind_hash);
inet_csk          894 net/ipv4/inet_connection_sock.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          952 net/ipv4/inet_connection_sock.c 	struct request_sock_queue *queue = &inet_csk(sk)->icsk_accept_queue;
inet_csk          978 net/ipv4/inet_connection_sock.c 		reqsk_queue_removed(&inet_csk(sk)->icsk_accept_queue, req);
inet_csk          995 net/ipv4/inet_connection_sock.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         1054 net/ipv4/inet_connection_sock.c 	const struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         1067 net/ipv4/inet_connection_sock.c 	const struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          319 net/ipv4/inet_diag.c 	return inet_sk_diag_fill(sk, inet_csk(sk), skb, req, user_ns,
inet_csk           97 net/ipv4/inet_hashtables.c 	inet_csk(sk)->icsk_bind_hash = tb;
inet_csk          112 net/ipv4/inet_hashtables.c 	tb = inet_csk(sk)->icsk_bind_hash;
inet_csk          114 net/ipv4/inet_hashtables.c 	inet_csk(sk)->icsk_bind_hash = NULL;
inet_csk          139 net/ipv4/inet_hashtables.c 	tb = inet_csk(sk)->icsk_bind_hash;
inet_csk          203 net/ipv4/inet_hashtables.c 		hlist_add_tail_rcu(&inet_csk(sk)->icsk_listen_portaddr_node,
inet_csk          206 net/ipv4/inet_hashtables.c 		hlist_add_head_rcu(&inet_csk(sk)->icsk_listen_portaddr_node,
inet_csk          217 net/ipv4/inet_hashtables.c 	    WARN_ON_ONCE(hlist_unhashed(&inet_csk(sk)->icsk_listen_portaddr_node)))
inet_csk          223 net/ipv4/inet_hashtables.c 	hlist_del_init_rcu(&inet_csk(sk)->icsk_listen_portaddr_node);
inet_csk          518 net/ipv4/inet_hashtables.c 	struct inet_bind_bucket *tb = inet_csk(sk)->icsk_bind_hash;
inet_csk          528 net/ipv4/inet_hashtables.c 		    inet_csk(sk2)->icsk_bind_hash == tb &&
inet_csk          638 net/ipv4/inet_hashtables.c 		tb = inet_csk(sk)->icsk_bind_hash;
inet_csk          105 net/ipv4/inet_timewait_sock.c 	const struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          661 net/ipv4/ip_sockglue.c 			struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          205 net/ipv4/syncookies.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          410 net/ipv4/tcp.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         1541 net/ipv4/tcp.c 		const struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         2257 net/ipv4/tcp.c 		if (inet_csk(sk)->icsk_bind_hash &&
inet_csk         2566 net/ipv4/tcp.c 	inet_csk(sk)->icsk_backoff = 0;
inet_csk         2572 net/ipv4/tcp.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         2806 net/ipv4/tcp.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         3170 net/ipv4/tcp.c 	const struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         3214 net/ipv4/tcp.c 	const struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         3391 net/ipv4/tcp.c 	nla_put_u8(stats, TCP_NLA_RECUR_RETRANS, inet_csk(sk)->icsk_retransmits);
inet_csk         3398 net/ipv4/tcp.c 	nla_put_u8(stats, TCP_NLA_CA_STATE, inet_csk(sk)->icsk_ca_state);
inet_csk         3414 net/ipv4/tcp.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         3701 net/ipv4/tcp.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          482 net/ipv4/tcp_bbr.c 	u8 prev_state = bbr->prev_ca_state, state = inet_csk(sk)->icsk_ca_state;
inet_csk          191 net/ipv4/tcp_bic.c 	const struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          160 net/ipv4/tcp_cong.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          179 net/ipv4/tcp_cong.c 	const struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          193 net/ipv4/tcp_cong.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          207 net/ipv4/tcp_cong.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          339 net/ipv4/tcp_cong.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk           98 net/ipv4/tcp_dctcp.c 	inet_csk(sk)->icsk_ca_ops = &dctcp_reno;
inet_csk          156 net/ipv4/tcp_dctcp.c 	    new_state != inet_csk(sk)->icsk_ca_state)
inet_csk          193 net/ipv4/tcp_dctcp.c 		if (inet_csk(sk)->icsk_ca_ops != &dctcp_reno) {
inet_csk           29 net/ipv4/tcp_dctcp.h 		if (inet_csk(sk)->icsk_ack.pending & ICSK_ACK_TIMER) {
inet_csk           33 net/ipv4/tcp_dctcp.h 		inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_NOW;
inet_csk          115 net/ipv4/tcp_diag.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          146 net/ipv4/tcp_diag.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk           49 net/ipv4/tcp_fastopen.c 			inet_csk(sk)->icsk_accept_queue.fastopenq.ctx, 1);
inet_csk           94 net/ipv4/tcp_fastopen.c 		q = &inet_csk(sk)->icsk_accept_queue.fastopenq;
inet_csk          237 net/ipv4/tcp_fastopen.c 	struct request_sock_queue *queue = &inet_csk(sk)->icsk_accept_queue;
inet_csk          241 net/ipv4/tcp_fastopen.c 	child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL,
inet_csk          303 net/ipv4/tcp_fastopen.c 	fastopenq = &inet_csk(sk)->icsk_accept_queue.fastopenq;
inet_csk          550 net/ipv4/tcp_fastopen.c 	u32 timeouts = inet_csk(sk)->icsk_retransmits;
inet_csk           84 net/ipv4/tcp_htcp.c 	const struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          104 net/ipv4/tcp_htcp.c 	const struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          164 net/ipv4/tcp_input.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          215 net/ipv4/tcp_input.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          227 net/ipv4/tcp_input.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          241 net/ipv4/tcp_input.c 	const struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          263 net/ipv4/tcp_input.c 		inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_NOW;
inet_csk          337 net/ipv4/tcp_input.c 	const struct tcp_congestion_ops *ca_ops = inet_csk(sk)->icsk_ca_ops;
inet_csk          401 net/ipv4/tcp_input.c 			return 2 * inet_csk(sk)->icsk_ack.rcv_mss;
inet_csk          431 net/ipv4/tcp_input.c 			inet_csk(sk)->icsk_ack.quick |= 1;
inet_csk          478 net/ipv4/tcp_input.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          511 net/ipv4/tcp_input.c 	inet_csk(sk)->icsk_ack.rcv_mss = hint;
inet_csk          586 net/ipv4/tcp_input.c 	    TCP_SKB_CB(skb)->seq >= inet_csk(sk)->icsk_ack.rcv_mss) {
inet_csk          679 net/ipv4/tcp_input.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          849 net/ipv4/tcp_input.c 	inet_csk(sk)->icsk_rto = __tcp_set_rto(tp);
inet_csk          899 net/ipv4/tcp_input.c 			 tp->rx_opt.sack_ok, inet_csk(sk)->icsk_ca_state,
inet_csk         1843 net/ipv4/tcp_input.c 	if (inet_csk(sk)->icsk_ca_state != TCP_CA_Loss || tp->undo_marker)
inet_csk         1991 net/ipv4/tcp_input.c 	const struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         2029 net/ipv4/tcp_input.c 		   !inet_csk(sk)->icsk_mtup.probe_size;
inet_csk         2362 net/ipv4/tcp_input.c 		const struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         2392 net/ipv4/tcp_input.c 		DBGUNDO(sk, inet_csk(sk)->icsk_ca_state == TCP_CA_Loss ? "loss" : "retrans");
inet_csk         2394 net/ipv4/tcp_input.c 		if (inet_csk(sk)->icsk_ca_state == TCP_CA_Loss)
inet_csk         2445 net/ipv4/tcp_input.c 		inet_csk(sk)->icsk_retransmits = 0;
inet_csk         2474 net/ipv4/tcp_input.c 	tp->snd_ssthresh = inet_csk(sk)->icsk_ca_ops->ssthresh(sk);
inet_csk         2509 net/ipv4/tcp_input.c 	if (inet_csk(sk)->icsk_ca_ops->cong_control)
inet_csk         2514 net/ipv4/tcp_input.c 	    (inet_csk(sk)->icsk_ca_state == TCP_CA_CWR || tp->undo_marker)) {
inet_csk         2527 net/ipv4/tcp_input.c 	if (inet_csk(sk)->icsk_ca_state < TCP_CA_CWR) {
inet_csk         2543 net/ipv4/tcp_input.c 	if (inet_csk(sk)->icsk_ca_state != state) {
inet_csk         2561 net/ipv4/tcp_input.c 	if (inet_csk(sk)->icsk_ca_state != TCP_CA_CWR) {
inet_csk         2568 net/ipv4/tcp_input.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         2578 net/ipv4/tcp_input.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         2601 net/ipv4/tcp_input.c 	const struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         2794 net/ipv4/tcp_input.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         2963 net/ipv4/tcp_input.c 	inet_csk(sk)->icsk_backoff = 0;
inet_csk         2982 net/ipv4/tcp_input.c 	const struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         2993 net/ipv4/tcp_input.c 	const struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         3005 net/ipv4/tcp_input.c 		u32 rto = inet_csk(sk)->icsk_rto;
inet_csk         3074 net/ipv4/tcp_input.c 	const struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         3253 net/ipv4/tcp_input.c 		icsk = inet_csk(sk);
inet_csk         3276 net/ipv4/tcp_input.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         3300 net/ipv4/tcp_input.c 		inet_csk(sk)->icsk_ca_state != TCP_CA_Open;
inet_csk         3326 net/ipv4/tcp_input.c 	const struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         3408 net/ipv4/tcp_input.c 				tcp_sync_mss(sk, inet_csk(sk)->icsk_pmtu_cookie);
inet_csk         3541 net/ipv4/tcp_input.c 	const struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         3587 net/ipv4/tcp_input.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         4077 net/ipv4/tcp_input.c 		(s32)(tp->rx_opt.ts_recent - tp->rx_opt.rcv_tsval) <= (inet_csk(sk)->icsk_rto * 1024) / HZ);
inet_csk         4810 net/ipv4/tcp_input.c 				inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_NOW;
inet_csk         5236 net/ipv4/tcp_input.c 	if (((tp->rcv_nxt - tp->rcv_wup) > inet_csk(sk)->icsk_ack.rcv_mss &&
inet_csk         5247 net/ipv4/tcp_input.c 	    inet_csk(sk)->icsk_ack.pending & ICSK_ACK_NOW) {
inet_csk         5561 net/ipv4/tcp_input.c 		inet_csk(sk)->icsk_af_ops->sk_rx_dst_set(sk, skb);
inet_csk         5730 net/ipv4/tcp_input.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         5757 net/ipv4/tcp_input.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         5873 net/ipv4/tcp_input.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         6103 net/ipv4/tcp_input.c 	if (inet_csk(sk)->icsk_ca_state == TCP_CA_Loss)
inet_csk         6108 net/ipv4/tcp_input.c 	inet_csk(sk)->icsk_retransmits = 0;
inet_csk         6138 net/ipv4/tcp_input.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         6251 net/ipv4/tcp_input.c 		if (!inet_csk(sk)->icsk_ca_ops->cong_control)
inet_csk         6488 net/ipv4/tcp_input.c 	struct request_sock_queue *queue = &inet_csk(sk)->icsk_accept_queue;
inet_csk          264 net/ipv4/tcp_ipv4.c 	inet_csk(sk)->icsk_ext_hdr_len = 0;
inet_csk          266 net/ipv4/tcp_ipv4.c 		inet_csk(sk)->icsk_ext_hdr_len = inet_opt->opt.optlen;
inet_csk          361 net/ipv4/tcp_ipv4.c 	    inet_csk(sk)->icsk_pmtu_cookie > mtu) {
inet_csk          481 net/ipv4/tcp_ipv4.c 	icsk = inet_csk(sk);
inet_csk         1450 net/ipv4/tcp_ipv4.c 	inet_csk(newsk)->icsk_ext_hdr_len = 0;
inet_csk         1452 net/ipv4/tcp_ipv4.c 		inet_csk(newsk)->icsk_ext_hdr_len = inet_opt->opt.optlen;
inet_csk         2080 net/ipv4/tcp_ipv4.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         2124 net/ipv4/tcp_ipv4.c 	if (inet_csk(sk)->icsk_bind_hash)
inet_csk         2429 net/ipv4/tcp_ipv4.c 	const struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          322 net/ipv4/tcp_metrics.c 	const struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          502 net/ipv4/tcp_metrics.c 		inet_csk(sk)->icsk_rto = crtt + max(2 * crtt, tcp_rto_min(sk));
inet_csk          513 net/ipv4/tcp_metrics.c 		inet_csk(sk)->icsk_rto = TCP_TIMEOUT_FALLBACK;
inet_csk          255 net/ipv4/tcp_minisocks.c 	const struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          408 net/ipv4/tcp_minisocks.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          470 net/ipv4/tcp_minisocks.c 	newicsk = inet_csk(newsk);
inet_csk          757 net/ipv4/tcp_minisocks.c 	if (req->num_timeout < inet_csk(sk)->icsk_accept_queue.rskq_defer_accept &&
inet_csk          770 net/ipv4/tcp_minisocks.c 	child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL,
inet_csk          242 net/ipv4/tcp_nv.c 	const struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk           66 net/ipv4/tcp_output.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          151 net/ipv4/tcp_output.c 	while ((delta -= inet_csk(sk)->icsk_rto) > 0 && cwnd > restart_cwnd)
inet_csk          162 net/ipv4/tcp_output.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          896 net/ipv4/tcp_output.c 		inet_csk(sk)->icsk_af_ops->mtu_reduced(sk);
inet_csk         1020 net/ipv4/tcp_output.c 	const struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         1470 net/ipv4/tcp_output.c 	const struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         1510 net/ipv4/tcp_output.c 	const struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         1533 net/ipv4/tcp_output.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         1571 net/ipv4/tcp_output.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         1606 net/ipv4/tcp_output.c 		if (mtu != inet_csk(sk)->icsk_pmtu_cookie)
inet_csk         1632 net/ipv4/tcp_output.c 	if (inet_csk(sk)->icsk_ca_state == TCP_CA_Open &&
inet_csk         1648 net/ipv4/tcp_output.c 	const struct tcp_congestion_ops *ca_ops = inet_csk(sk)->icsk_ca_ops;
inet_csk         1671 net/ipv4/tcp_output.c 		    (s32)(tcp_jiffies32 - tp->snd_cwnd_stamp) >= inet_csk(sk)->icsk_rto &&
inet_csk         1753 net/ipv4/tcp_output.c 	const struct tcp_congestion_ops *ca_ops = inet_csk(sk)->icsk_ca_ops;
inet_csk         1944 net/ipv4/tcp_output.c 	const struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         2043 net/ipv4/tcp_output.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         2095 net/ipv4/tcp_output.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         2112 net/ipv4/tcp_output.c 		   inet_csk(sk)->icsk_ca_state != TCP_CA_Open ||
inet_csk         2493 net/ipv4/tcp_output.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         2530 net/ipv4/tcp_output.c 			jiffies_to_usecs(inet_csk(sk)->icsk_rto) :
inet_csk         2578 net/ipv4/tcp_output.c 		inet_csk(sk)->icsk_pending = 0;
inet_csk         2613 net/ipv4/tcp_output.c 	inet_csk(sk)->icsk_pending = 0;
inet_csk         2703 net/ipv4/tcp_output.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         2899 net/ipv4/tcp_output.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         2929 net/ipv4/tcp_output.c 	if (inet_csk(sk)->icsk_af_ops->rebuild_header(sk))
inet_csk         3048 net/ipv4/tcp_output.c 	const struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         3111 net/ipv4/tcp_output.c 					     inet_csk(sk)->icsk_rto,
inet_csk         3370 net/ipv4/tcp_output.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         3459 net/ipv4/tcp_output.c 	inet_csk(sk)->icsk_rto = tcp_timeout_init(sk);
inet_csk         3460 net/ipv4/tcp_output.c 	inet_csk(sk)->icsk_retransmits = 0;
inet_csk         3501 net/ipv4/tcp_output.c 	space = __tcp_mtu_to_mss(sk, inet_csk(sk)->icsk_pmtu_cookie) -
inet_csk         3580 net/ipv4/tcp_output.c 	if (inet_csk(sk)->icsk_af_ops->rebuild_header(sk))
inet_csk         3621 net/ipv4/tcp_output.c 				  inet_csk(sk)->icsk_rto, TCP_RTO_MAX);
inet_csk         3632 net/ipv4/tcp_output.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         3700 net/ipv4/tcp_output.c 		inet_csk(sk)->icsk_ack.ato = TCP_ATO_MIN;
inet_csk         3818 net/ipv4/tcp_output.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          166 net/ipv4/tcp_rate.c 				 inet_csk(sk)->icsk_ca_state,
inet_csk           32 net/ipv4/tcp_recovery.c 		if (inet_csk(sk)->icsk_ca_state >= TCP_CA_Recovery)
inet_csk          127 net/ipv4/tcp_recovery.c 					  timeout, inet_csk(sk)->icsk_rto);
inet_csk          174 net/ipv4/tcp_recovery.c 		if (inet_csk(sk)->icsk_ca_state != TCP_CA_Recovery) {
inet_csk          176 net/ipv4/tcp_recovery.c 			if (!inet_csk(sk)->icsk_ca_ops->cong_control)
inet_csk          181 net/ipv4/tcp_recovery.c 	if (inet_csk(sk)->icsk_pending != ICSK_TIME_RETRANS)
inet_csk          233 net/ipv4/tcp_recovery.c 	const u8 state = inet_csk(sk)->icsk_ca_state;
inet_csk           28 net/ipv4/tcp_timer.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          197 net/ipv4/tcp_timer.c 	if (!inet_csk(sk)->icsk_retransmits)
inet_csk          215 net/ipv4/tcp_timer.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          273 net/ipv4/tcp_timer.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          340 net/ipv4/tcp_timer.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          391 net/ipv4/tcp_timer.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          435 net/ipv4/tcp_timer.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          575 net/ipv4/tcp_timer.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          653 net/ipv4/tcp_timer.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          102 net/ipv4/tcp_ulp.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          116 net/ipv4/tcp_ulp.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          134 net/ipv4/tcp_ulp.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk           63 net/ipv4/tcp_yeah.c 	const struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          107 net/ipv6/ipv6_sockglue.c 			struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          217 net/ipv6/ipv6_sockglue.c 				struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          149 net/ipv6/tcp_ipv6.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          354 net/ipv6/tcp_ipv6.c 	if (inet_csk(sk)->icsk_pmtu_cookie > dst_mtu(dst)) {
inet_csk         1153 net/ipv6/tcp_ipv6.c 		inet_csk(newsk)->icsk_af_ops = &ipv6_mapped;
inet_csk         1180 net/ipv6/tcp_ipv6.c 		tcp_sync_mss(newsk, inet_csk(newsk)->icsk_pmtu_cookie);
inet_csk         1256 net/ipv6/tcp_ipv6.c 	inet_csk(newsk)->icsk_ext_hdr_len = 0;
inet_csk         1258 net/ipv6/tcp_ipv6.c 		inet_csk(newsk)->icsk_ext_hdr_len = opt->opt_nflen +
inet_csk         1805 net/ipv6/tcp_ipv6.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk         1866 net/ipv6/tcp_ipv6.c 	const struct inet_connection_sock *icsk = inet_csk(sp);
inet_csk          191 net/tls/tls_device.c 		clean_acked_data_disable(inet_csk(sk));
inet_csk         1001 net/tls/tls_device.c 	clean_acked_data_enable(inet_csk(sk), &tls_icsk_clean_acked);
inet_csk         1062 net/tls/tls_device.c 	clean_acked_data_disable(inet_csk(sk));
inet_csk          293 net/tls/tls_main.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          600 net/tls/tls_main.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          642 net/tls/tls_main.c 	struct inet_connection_sock *icsk = inet_csk(sk);
inet_csk          828 net/tls/tls_main.c 	ctx = rcu_dereference(inet_csk(sk)->icsk_ulp_data);