Home
last modified time | relevance | path

Searched refs:sk_receive_queue (Results 1 – 77 of 77) sorted by relevance

/linux-4.4.14/net/unix/
Dgarbage.c165 spin_lock(&x->sk_receive_queue.lock); in scan_inflight()
166 skb_queue_walk_safe(&x->sk_receive_queue, skb, next) { in scan_inflight()
193 __skb_unlink(skb, &x->sk_receive_queue); in scan_inflight()
198 spin_unlock(&x->sk_receive_queue.lock); in scan_inflight()
215 spin_lock(&x->sk_receive_queue.lock); in scan_children()
216 skb_queue_walk_safe(&x->sk_receive_queue, skb, next) { in scan_children()
225 spin_unlock(&x->sk_receive_queue.lock); in scan_children()
Ddiag.c64 spin_lock(&sk->sk_receive_queue.lock); in sk_diag_dump_icons()
67 sk->sk_receive_queue.qlen * sizeof(u32)); in sk_diag_dump_icons()
73 skb_queue_walk(&sk->sk_receive_queue, skb) { in sk_diag_dump_icons()
87 spin_unlock(&sk->sk_receive_queue.lock); in sk_diag_dump_icons()
93 spin_unlock(&sk->sk_receive_queue.lock); in sk_diag_dump_icons()
102 rql.udiag_rqueue = sk->sk_receive_queue.qlen; in sk_diag_show_rqlen()
Daf_unix.c196 return skb_queue_len(&sk->sk_receive_queue) > sk->sk_max_ack_backlog; in unix_recvq_full()
468 if (!skb_queue_empty(&sk->sk_receive_queue)) { in unix_dgram_disconnected()
469 skb_queue_purge(&sk->sk_receive_queue); in unix_dgram_disconnected()
487 skb_queue_purge(&sk->sk_receive_queue); in unix_sock_destructor()
540 if (!skb_queue_empty(&sk->sk_receive_queue) || embrion) in unix_release_sock()
554 while ((skb = skb_dequeue(&sk->sk_receive_queue)) != NULL) { in unix_release_sock()
769 lockdep_set_class(&sk->sk_receive_queue.lock, in unix_create1()
1360 spin_lock(&other->sk_receive_queue.lock); in unix_stream_connect()
1361 __skb_queue_tail(&other->sk_receive_queue, skb); in unix_stream_connect()
1362 spin_unlock(&other->sk_receive_queue.lock); in unix_stream_connect()
[all …]
/linux-4.4.14/net/rxrpc/
Dar-recvmsg.c86 skb = skb_peek(&rx->sk.sk_receive_queue); in rxrpc_recvmsg()
101 if (skb_queue_empty(&rx->sk.sk_receive_queue)) { in rxrpc_recvmsg()
124 if (skb_dequeue(&rx->sk.sk_receive_queue) != skb) in rxrpc_recvmsg()
215 if (skb_dequeue(&rx->sk.sk_receive_queue) != in rxrpc_recvmsg()
235 if (skb == (struct sk_buff *) &rx->sk.sk_receive_queue) in rxrpc_recvmsg()
241 if (skb_dequeue(&rx->sk.sk_receive_queue) != skb) in rxrpc_recvmsg()
268 if (skb_dequeue(&rx->sk.sk_receive_queue) != skb) in rxrpc_recvmsg()
320 if (skb_dequeue(&rx->sk.sk_receive_queue) != skb) in rxrpc_recvmsg()
Dar-input.c85 spin_lock_bh(&sk->sk_receive_queue.lock); in rxrpc_queue_rcv_skb()
102 spin_unlock_bh(&sk->sk_receive_queue.lock); in rxrpc_queue_rcv_skb()
105 __skb_queue_tail(&sk->sk_receive_queue, skb); in rxrpc_queue_rcv_skb()
106 spin_unlock_bh(&sk->sk_receive_queue.lock); in rxrpc_queue_rcv_skb()
113 spin_unlock_bh(&sk->sk_receive_queue.lock); in rxrpc_queue_rcv_skb()
Daf_rxrpc.c599 if (!skb_queue_empty(&sk->sk_receive_queue)) in rxrpc_poll()
666 rxrpc_purge_queue(&sk->sk_receive_queue); in rxrpc_sock_destructor()
691 spin_lock_bh(&sk->sk_receive_queue.lock); in rxrpc_release_sock()
693 spin_unlock_bh(&sk->sk_receive_queue.lock); in rxrpc_release_sock()
706 rxrpc_purge_queue(&sk->sk_receive_queue); in rxrpc_release_sock()
/linux-4.4.14/include/net/
Dbusy_poll.h111 } while (!nonblock && skb_queue_empty(&sk->sk_receive_queue) && in sk_busy_loop()
114 rc = !skb_queue_empty(&sk->sk_receive_queue); in sk_busy_loop()
Dsock.h359 struct sk_buff_head sk_receive_queue; member
2206 __skb_unlink(skb, &sk->sk_receive_queue); in sk_eat_skb()
/linux-4.4.14/net/atm/
Dsignaling.c32 skb_queue_tail(&sk_atm(sigd)->sk_receive_queue, skb); in sigd_put_skb()
112 skb_queue_tail(&sk->sk_receive_queue, skb); in sigd_send()
204 if (skb_peek(&sk_atm(vcc)->sk_receive_queue)) in sigd_close()
206 skb_queue_purge(&sk_atm(vcc)->sk_receive_queue); in sigd_close()
Draw.c27 skb_queue_tail(&sk->sk_receive_queue, skb); in atm_push_raw()
Dlec.c154 skb_queue_tail(&sk->sk_receive_queue, skb2); in lec_handle_bridge()
451 skb_queue_tail(&sk->sk_receive_queue, skb2); in lec_atm_send()
478 if (skb_peek(&sk_atm(vcc)->sk_receive_queue)) in lec_atm_close()
480 while ((skb = skb_dequeue(&sk_atm(vcc)->sk_receive_queue))) { in lec_atm_close()
534 skb_queue_tail(&sk->sk_receive_queue, skb); in send_to_lecd()
540 skb_queue_tail(&sk->sk_receive_queue, data); in send_to_lecd()
620 skb_queue_tail(&sk->sk_receive_queue, skb); in lec_push()
Dcommon.c191 while ((skb = skb_dequeue(&sk->sk_receive_queue)) != NULL) { in vcc_destroy_socket()
236 rq = &sk_atm(vcc)->sk_receive_queue; in vcc_process_recv_queue()
667 if (!skb_queue_empty(&sk->sk_receive_queue)) in vcc_poll()
Dsvc.c64 while ((skb = skb_dequeue(&sk->sk_receive_queue)) != NULL) { in svc_disconnect()
343 while (!(skb = skb_dequeue(&sk->sk_receive_queue)) && in svc_accept()
Dmpc.c708 skb_queue_tail(&sk->sk_receive_queue, skb); in mpc_push()
896 while ((skb = skb_dequeue(&sk_atm(vcc)->sk_receive_queue))) { in mpoad_close()
994 skb_queue_tail(&sk->sk_receive_queue, skb); in msg_to_mpoad()
1275 skb_queue_tail(&sk->sk_receive_queue, skb); in purge_egress_shortcut()
Dioctl.c78 skb = skb_peek(&sk->sk_receive_queue); in do_vcc_ioctl()
Dclip.c70 skb_queue_tail(&sk->sk_receive_queue, skb); in to_atmarpd()
613 skb_queue_purge(&sk_atm(vcc)->sk_receive_queue); in atmarpd_close()
/linux-4.4.14/net/core/
Ddatagram.c99 if (sk->sk_receive_queue.prev != skb) in wait_for_more_packets()
197 struct sk_buff_head *queue = &sk->sk_receive_queue; in __skb_recv_datagram()
331 spin_lock_bh(&sk->sk_receive_queue.lock); in skb_kill_datagram()
332 if (skb == skb_peek(&sk->sk_receive_queue)) { in skb_kill_datagram()
333 __skb_unlink(skb, &sk->sk_receive_queue); in skb_kill_datagram()
337 spin_unlock_bh(&sk->sk_receive_queue.lock); in skb_kill_datagram()
772 if (!skb_queue_empty(&sk->sk_receive_queue)) in datagram_poll()
Dstream.c193 __skb_queue_purge(&sk->sk_receive_queue); in sk_stream_kill_queues()
Dsock.c451 struct sk_buff_head *list = &sk->sk_receive_queue; in sock_queue_rcv_skb()
1529 skb_queue_head_init(&newsk->sk_receive_queue); in sk_clone_lock()
2050 rc = sk_wait_event(sk, timeo, skb_peek_tail(&sk->sk_receive_queue) != skb); in sk_wait_data()
2366 skb_queue_head_init(&sk->sk_receive_queue); in sock_init_data()
/linux-4.4.14/net/bluetooth/
Daf_bluetooth.c265 if (!skb_queue_empty(&sk->sk_receive_queue)) in bt_sock_data_wait()
308 skb = skb_dequeue(&sk->sk_receive_queue); in bt_sock_stream_recvmsg()
334 skb_queue_head(&sk->sk_receive_queue, skb); in bt_sock_stream_recvmsg()
373 skb_queue_head(&sk->sk_receive_queue, skb); in bt_sock_stream_recvmsg()
380 skb_queue_head(&sk->sk_receive_queue, skb); in bt_sock_stream_recvmsg()
430 if (!skb_queue_empty(&sk->sk_receive_queue)) in bt_sock_poll()
475 skb = skb_peek(&sk->sk_receive_queue); in bt_sock_ioctl()
Dsco.c370 skb_queue_purge(&sk->sk_receive_queue); in sco_sock_destruct()
Dhci_sock.c590 skb_queue_purge(&sk->sk_receive_queue); in hci_sock_release()
Dl2cap_sock.c1484 skb_queue_purge(&sk->sk_receive_queue); in l2cap_sock_destruct()
/linux-4.4.14/net/sctp/
Dulpqueue.c144 sctp_skb_list_tail(&sp->pd_lobby, &sk->sk_receive_queue); in sctp_clear_pd()
163 __skb_queue_tail(&sk->sk_receive_queue, in sctp_clear_pd()
222 queue = &sk->sk_receive_queue; in sctp_ulpq_tail_event()
236 queue = &sk->sk_receive_queue; in sctp_ulpq_tail_event()
245 queue = &sk->sk_receive_queue; in sctp_ulpq_tail_event()
266 if (queue == &sk->sk_receive_queue) in sctp_ulpq_tail_event()
1096 if (skb_queue_empty(&asoc->base.sk->sk_receive_queue)) { in sctp_ulpq_renege()
1139 __skb_queue_tail(&sk->sk_receive_queue, sctp_event2skb(ev)); in sctp_ulpq_abort_pd()
Dsocket.c1493 data_was_unread = sctp_queue_purge_ulpevents(&sk->sk_receive_queue); in sctp_close()
2133 skb_queue_head(&sk->sk_receive_queue, skb); in sctp_recvmsg()
3963 skb = skb_peek(&sk->sk_receive_queue); in sctp_ioctl()
6453 if (!skb_queue_empty(&sk->sk_receive_queue)) in sctp_poll()
6701 if (!skb_queue_empty(&sk->sk_receive_queue)) in sctp_wait_for_packet()
6768 spin_lock_bh(&sk->sk_receive_queue.lock); in sctp_skb_recv_datagram()
6769 skb = skb_peek(&sk->sk_receive_queue); in sctp_skb_recv_datagram()
6772 spin_unlock_bh(&sk->sk_receive_queue.lock); in sctp_skb_recv_datagram()
6774 skb = skb_dequeue(&sk->sk_receive_queue); in sctp_skb_recv_datagram()
7278 sctp_skb_for_each(skb, &oldsk->sk_receive_queue, tmp) { in sctp_sock_migrate()
[all …]
/linux-4.4.14/net/caif/
Dcaif_socket.c128 struct sk_buff_head *list = &sk->sk_receive_queue; in caif_queue_rcv_skb()
317 if (!skb_queue_empty(&sk->sk_receive_queue) || in caif_stream_data_wait()
381 skb = skb_dequeue(&sk->sk_receive_queue); in caif_stream_recvmsg()
426 skb_queue_head(&sk->sk_receive_queue, skb); in caif_stream_recvmsg()
440 skb_queue_head(&sk->sk_receive_queue, skb); in caif_stream_recvmsg()
450 skb_queue_head(&sk->sk_receive_queue, skb); in caif_stream_recvmsg()
909 spin_lock_bh(&sk->sk_receive_queue.lock); in caif_release()
911 spin_unlock_bh(&sk->sk_receive_queue.lock); in caif_release()
952 if (!skb_queue_empty(&sk->sk_receive_queue) || in caif_poll()
/linux-4.4.14/net/tipc/
Dsocket.c249 kfree_skb(__skb_dequeue(&sk->sk_receive_queue)); in tsk_advance_rx_queue()
277 while ((skb = __skb_dequeue(&sk->sk_receive_queue))) in tsk_rej_rx_queue()
446 skb = __skb_dequeue(&sk->sk_receive_queue); in tipc_release()
647 if (!skb_queue_empty(&sk->sk_receive_queue)) in tipc_poll()
1253 if (timeo && skb_queue_empty(&sk->sk_receive_queue)) { in tipc_wait_for_rcvmsg()
1263 if (!skb_queue_empty(&sk->sk_receive_queue)) in tipc_wait_for_rcvmsg()
1320 buf = skb_peek(&sk->sk_receive_queue); in tipc_recvmsg()
1417 buf = skb_peek(&sk->sk_receive_queue); in tipc_recv_stream()
1478 (!skb_queue_empty(&sk->sk_receive_queue) || in tipc_recv_stream()
1524 __skb_queue_purge(&sk->sk_receive_queue); in tipc_sock_destruct()
[all …]
/linux-4.4.14/net/phonet/
Ddatagram.c52 skb = skb_peek(&sk->sk_receive_queue); in pn_ioctl()
77 skb_queue_purge(&sk->sk_receive_queue); in pn_destruct()
Dpep.c417 queue = &sk->sk_receive_queue; in pipe_do_rcv()
476 skb_queue_purge(&sk->sk_receive_queue); in pipe_destruct()
589 skb_queue_tail(&sk->sk_receive_queue, skb); in pipe_handler_do_rcv()
696 skb_queue_head(&sk->sk_receive_queue, skb); in pep_do_rcv()
945 else if (!skb_queue_empty(&sk->sk_receive_queue)) in pep_ioctl()
946 answ = skb_peek(&sk->sk_receive_queue)->len; in pep_ioctl()
1241 struct sk_buff *skb = skb_dequeue(&sk->sk_receive_queue); in pep_read()
Dsocket.c351 if (!skb_queue_empty(&sk->sk_receive_queue)) in pn_socket_poll()
/linux-4.4.14/net/llc/
Daf_llc.c663 if (skb_queue_empty(&sk->sk_receive_queue)) { in llc_ui_accept()
670 skb = skb_dequeue(&sk->sk_receive_queue); in llc_ui_accept()
757 skb = skb_peek(&sk->sk_receive_queue); in llc_ui_recvmsg()
841 spin_lock_irqsave(&sk->sk_receive_queue.lock, cpu_flags); in llc_ui_recvmsg()
843 spin_unlock_irqrestore(&sk->sk_receive_queue.lock, cpu_flags); in llc_ui_recvmsg()
864 spin_lock_irqsave(&sk->sk_receive_queue.lock, cpu_flags); in llc_ui_recvmsg()
866 spin_unlock_irqrestore(&sk->sk_receive_queue.lock, cpu_flags); in llc_ui_recvmsg()
Dllc_conn.c109 skb_queue_tail(&sk->sk_receive_queue, skb); in llc_conn_state_process()
969 skb_queue_purge(&sk->sk_receive_queue); in llc_sk_free()
/linux-4.4.14/net/irda/
Daf_irda.c288 skb_queue_tail(&sk->sk_receive_queue, skb); in irda_connect_indication()
874 skb = skb_dequeue(&sk->sk_receive_queue); in irda_accept()
884 skb_peek(&sk->sk_receive_queue)); in irda_accept()
1226 skb_queue_purge(&sk->sk_receive_queue); in irda_release()
1431 struct sk_buff *skb = skb_dequeue(&sk->sk_receive_queue); in irda_recvmsg_stream()
1456 else if (skb_peek(&sk->sk_receive_queue) == NULL) in irda_recvmsg_stream()
1472 skb_queue_head(&sk->sk_receive_queue, skb); in irda_recvmsg_stream()
1488 skb_queue_head(&sk->sk_receive_queue, skb); in irda_recvmsg_stream()
1497 skb_queue_head(&sk->sk_receive_queue, skb); in irda_recvmsg_stream()
1758 if (!skb_queue_empty(&sk->sk_receive_queue)) { in irda_poll()
[all …]
/linux-4.4.14/net/packet/
Daf_packet.c678 spin_lock(&po->sk.sk_receive_queue.lock); in prb_retire_rx_blk_timer_expired()
741 spin_unlock(&po->sk.sk_receive_queue.lock); in prb_retire_rx_blk_timer_expired()
1301 spin_lock_bh(&po->sk.sk_receive_queue.lock); in packet_rcv_has_room()
1306 spin_unlock_bh(&po->sk.sk_receive_queue.lock); in packet_rcv_has_room()
2067 spin_lock(&sk->sk_receive_queue.lock); in packet_rcv()
2070 __skb_queue_tail(&sk->sk_receive_queue, skb); in packet_rcv()
2071 spin_unlock(&sk->sk_receive_queue.lock); in packet_rcv()
2076 spin_lock(&sk->sk_receive_queue.lock); in packet_rcv()
2079 spin_unlock(&sk->sk_receive_queue.lock); in packet_rcv()
2188 spin_lock(&sk->sk_receive_queue.lock); in tpacket_rcv()
[all …]
/linux-4.4.14/drivers/net/
Dmacvtap.c354 if (skb_queue_len(&q->sk.sk_receive_queue) >= dev->tx_queue_len) in macvtap_handle_frame()
372 skb_queue_tail(&q->sk.sk_receive_queue, skb); in macvtap_handle_frame()
381 skb_queue_tail(&q->sk.sk_receive_queue, segs); in macvtap_handle_frame()
394 skb_queue_tail(&q->sk.sk_receive_queue, skb); in macvtap_handle_frame()
511 skb_queue_purge(&sk->sk_receive_queue); in macvtap_sock_destruct()
584 if (!skb_queue_empty(&q->sk.sk_receive_queue)) in macvtap_poll()
927 skb = skb_dequeue(&q->sk.sk_receive_queue); in macvtap_do_read()
Dtun.c505 skb_queue_purge(&tfile->sk.sk_receive_queue); in tun_queue_purge()
858 if (skb_queue_len(&tfile->socket.sk->sk_receive_queue) * numqueues in tun_net_xmit()
878 skb_queue_tail(&tfile->socket.sk->sk_receive_queue, skb); in tun_net_xmit()
1041 if (!skb_queue_empty(&sk->sk_receive_queue)) in tun_chr_poll()
/linux-4.4.14/net/nfc/
Dllcp_sock.c560 if (!skb_queue_empty(&sk->sk_receive_queue)) in llcp_sock_poll()
810 skb_queue_empty(&sk->sk_receive_queue)) { in llcp_sock_recvmsg()
837 skb_queue_head(&sk->sk_receive_queue, skb); in llcp_sock_recvmsg()
868 skb_queue_head(&sk->sk_receive_queue, skb); in llcp_sock_recvmsg()
935 skb_queue_purge(&sk->sk_receive_queue); in llcp_sock_destruct()
Drawsock.c329 skb_queue_purge(&sk->sk_receive_queue); in rawsock_destruct()
/linux-4.4.14/net/netlink/
Daf_netlink.c297 if (skb_queue_empty(&sk->sk_receive_queue)) in netlink_rcv_wake()
387 queue = tx_ring ? &sk->sk_write_queue : &sk->sk_receive_queue; in __netlink_set_ring()
700 spin_lock_bh(&sk->sk_receive_queue.lock); in netlink_poll()
705 spin_unlock_bh(&sk->sk_receive_queue.lock); in netlink_poll()
845 spin_lock_bh(&sk->sk_receive_queue.lock); in netlink_ring_set_copied()
848 spin_unlock_bh(&sk->sk_receive_queue.lock); in netlink_ring_set_copied()
854 __skb_queue_tail(&sk->sk_receive_queue, skb); in netlink_ring_set_copied()
855 spin_unlock_bh(&sk->sk_receive_queue.lock); in netlink_ring_set_copied()
939 skb_queue_purge(&sk->sk_receive_queue); in netlink_sock_destruct()
1777 skb_queue_tail(&sk->sk_receive_queue, skb); in __netlink_sendskb()
[all …]
/linux-4.4.14/net/dccp/
Dproto.c275 __skb_queue_purge(&sk->sk_receive_queue); in dccp_disconnect()
373 skb = skb_peek(&sk->sk_receive_queue); in dccp_ioctl()
823 struct sk_buff *skb = skb_peek(&sk->sk_receive_queue); in dccp_recvmsg()
1007 while ((skb = __skb_dequeue(&sk->sk_receive_queue)) != NULL) { in dccp_close()
Dinput.c29 __skb_queue_tail(&sk->sk_receive_queue, skb); in dccp_enqueue_skb()
/linux-4.4.14/net/x25/
Daf_x25.c399 while ((skb = skb_dequeue(&sk->sk_receive_queue)) != NULL) { in __x25_destroy_socket()
843 if (skb_queue_empty(&sk->sk_receive_queue)) { in x25_wait_for_data()
877 skb = skb_dequeue(&sk->sk_receive_queue); in x25_accept()
1062 skb_queue_head(&sk->sk_receive_queue, skb); in x25_rx_call_request()
1382 if ((skb = skb_peek(&sk->sk_receive_queue)) != NULL) in x25_ioctl()
Dx25_in.c80 skb_queue_tail(&sk->sk_receive_queue, skbn); in x25_queue_rx_frame()
/linux-4.4.14/net/netrom/
Daf_netrom.c274 while ((skb = skb_dequeue(&sk->sk_receive_queue)) != NULL) { in nr_destroy_socket()
796 skb = skb_dequeue(&sk->sk_receive_queue); in nr_accept()
1011 skb_queue_head(&sk->sk_receive_queue, skb); in nr_rx_frame()
1215 if ((skb = skb_peek(&sk->sk_receive_queue)) != NULL) in nr_ioctl()
/linux-4.4.14/net/iucv/
Daf_iucv.c394 skb_queue_purge(&sk->sk_receive_queue); in iucv_sock_destruct()
1333 skb_queue_empty(&sk->sk_receive_queue) && in iucv_sock_recvmsg()
1358 skb_queue_head(&sk->sk_receive_queue, skb); in iucv_sock_recvmsg()
1378 skb_queue_head(&sk->sk_receive_queue, skb); in iucv_sock_recvmsg()
1389 skb_queue_head(&sk->sk_receive_queue, skb); in iucv_sock_recvmsg()
1476 if (!skb_queue_empty(&sk->sk_receive_queue) || in iucv_sock_poll()
1550 skb_queue_purge(&sk->sk_receive_queue); in iucv_sock_shutdown()
/linux-4.4.14/net/ipv4/
Dtcp.c1390 struct sk_buff *skb = skb_peek(&sk->sk_receive_queue); in tcp_cleanup_rbuf()
1466 while ((skb = skb_peek(&sk->sk_receive_queue)) != NULL) { in tcp_recv_skb()
1592 if (sk_can_busy_loop(sk) && skb_queue_empty(&sk->sk_receive_queue) && in tcp_recvmsg()
1646 last = skb_peek_tail(&sk->sk_receive_queue); in tcp_recvmsg()
1647 skb_queue_walk(&sk->sk_receive_queue, skb) { in tcp_recvmsg()
2040 while ((skb = __skb_dequeue(&sk->sk_receive_queue)) != NULL) { in tcp_close()
2229 __skb_queue_purge(&sk->sk_receive_queue); in tcp_disconnect()
Dtcp_fastopen.c198 __skb_queue_tail(&child->sk_receive_queue, skb2); in tcp_fastopen_create_child()
Draw.c884 spin_lock_bh(&sk->sk_receive_queue.lock); in raw_ioctl()
885 skb = skb_peek(&sk->sk_receive_queue); in raw_ioctl()
888 spin_unlock_bh(&sk->sk_receive_queue.lock); in raw_ioctl()
Dtcp_input.c4306 tail = skb_peek_tail(&sk->sk_receive_queue); in tcp_ofo_queue()
4310 __skb_queue_tail(&sk->sk_receive_queue, skb); in tcp_ofo_queue()
4471 struct sk_buff *tail = skb_peek_tail(&sk->sk_receive_queue); in tcp_queue_rcv()
4478 __skb_queue_tail(&sk->sk_receive_queue, skb); in tcp_queue_rcv()
4580 if (skb_queue_len(&sk->sk_receive_queue) == 0) in tcp_data_queue()
4859 if (!skb_queue_empty(&sk->sk_receive_queue)) in tcp_prune_queue()
4860 tcp_collapse(sk, &sk->sk_receive_queue, in tcp_prune_queue()
4861 skb_peek(&sk->sk_receive_queue), in tcp_prune_queue()
5043 struct sk_buff *skb = skb_peek(&sk->sk_receive_queue); in tcp_check_urg()
5046 __skb_unlink(skb, &sk->sk_receive_queue); in tcp_check_urg()
Daf_inet.c137 __skb_queue_purge(&sk->sk_receive_queue); in inet_sock_destruct()
Dudp.c1196 struct sk_buff_head list_kill, *rcvq = &sk->sk_receive_queue; in first_packet_length()
/linux-4.4.14/net/decnet/
Ddn_nsp_in.c340 skb_queue_tail(&sk->sk_receive_queue, skb); in dn_nsp_conn_init()
653 if (dn_queue_skb(sk, skb, SIGIO, &sk->sk_receive_queue) == 0) { in dn_nsp_data()
Daf_decnet.c1049 skb = skb_dequeue(&sk->sk_receive_queue); in dn_wait_for_connect()
1052 skb = skb_dequeue(&sk->sk_receive_queue); in dn_wait_for_connect()
1091 skb = skb_dequeue(&sk->sk_receive_queue); in dn_accept()
1259 skb_queue_walk(&sk->sk_receive_queue, skb) in dn_ioctl()
1680 struct sk_buff_head *queue = &sk->sk_receive_queue; in dn_recvmsg()
/linux-4.4.14/net/ieee802154/
Dsocket.c552 spin_lock_bh(&sk->sk_receive_queue.lock); in dgram_ioctl()
553 skb = skb_peek(&sk->sk_receive_queue); in dgram_ioctl()
561 spin_unlock_bh(&sk->sk_receive_queue.lock); in dgram_ioctl()
/linux-4.4.14/net/rose/
Daf_rose.c343 while ((skb = skb_dequeue(&sk->sk_receive_queue)) != NULL) { in rose_destroy_socket()
903 skb = skb_dequeue(&sk->sk_receive_queue); in rose_accept()
1040 skb_queue_head(&sk->sk_receive_queue, skb); in rose_rx_call_request()
1296 if ((skb = skb_peek(&sk->sk_receive_queue)) != NULL) in rose_ioctl()
/linux-4.4.14/net/rds/
Dtcp_recv.c256 skb_queue_len(&tc->t_sock->sk->sk_receive_queue)); in rds_tcp_data_recv()
/linux-4.4.14/drivers/vhost/
Dnet.c426 spin_lock_irqsave(&sk->sk_receive_queue.lock, flags); in peek_head_len()
427 head = skb_peek(&sk->sk_receive_queue); in peek_head_len()
434 spin_unlock_irqrestore(&sk->sk_receive_queue.lock, flags); in peek_head_len()
/linux-4.4.14/net/ipv6/
Draw.c1140 spin_lock_bh(&sk->sk_receive_queue.lock); in rawv6_ioctl()
1141 skb = skb_peek(&sk->sk_receive_queue); in rawv6_ioctl()
1145 spin_unlock_bh(&sk->sk_receive_queue.lock); in rawv6_ioctl()
/linux-4.4.14/net/ax25/
Dax25_in.c358 skb_queue_head(&sk->sk_receive_queue, skb); in ax25_rcv()
Daf_ax25.c306 while ((skb = skb_dequeue(&ax25->sk->sk_receive_queue)) != NULL) { in ax25_destroy_socket()
1353 skb = skb_dequeue(&sk->sk_receive_queue); in ax25_accept()
1703 if ((skb = skb_peek(&sk->sk_receive_queue)) != NULL) in ax25_ioctl()
/linux-4.4.14/drivers/isdn/mISDN/
Dsocket.c160 skb_queue_head(&sk->sk_receive_queue, skb); in mISDN_sock_recvmsg()
280 skb_queue_purge(&sk->sk_receive_queue); in data_sock_release()
/linux-4.4.14/net/bluetooth/cmtp/
Dcore.c300 while ((skb = skb_dequeue(&sk->sk_receive_queue))) { in cmtp_session()
/linux-4.4.14/net/bluetooth/rfcomm/
Dsock.c56 skb_queue_tail(&sk->sk_receive_queue, skb); in rfcomm_sk_data_ready()
161 skb_queue_purge(&sk->sk_receive_queue); in rfcomm_sock_destruct()
Dtty.c286 while ((skb = skb_dequeue(&sk->sk_receive_queue))) { in __rfcomm_dev_add()
Dcore.c1910 BT_DBG("session %p state %ld qlen %d", s, s->state, skb_queue_len(&sk->sk_receive_queue)); in rfcomm_process_rx()
1913 while ((skb = skb_dequeue(&sk->sk_receive_queue))) { in rfcomm_process_rx()
/linux-4.4.14/net/bluetooth/hidp/
Dcore.c1205 while ((skb = skb_dequeue(&intr_sk->sk_receive_queue))) { in hidp_session_run()
1218 while ((skb = skb_dequeue(&ctrl_sk->sk_receive_queue))) { in hidp_session_run()
/linux-4.4.14/net/bluetooth/bnep/
Dcore.c501 while ((skb = skb_dequeue(&sk->sk_receive_queue))) { in bnep_session()
/linux-4.4.14/net/can/
Daf_can.c115 skb_queue_purge(&sk->sk_receive_queue); in can_sock_destruct()
/linux-4.4.14/net/vmw_vsock/
Daf_vsock.c667 while ((skb = skb_dequeue(&sk->sk_receive_queue))) in __vsock_release()
875 if (!skb_queue_empty(&sk->sk_receive_queue) || in vsock_poll()
/linux-4.4.14/drivers/net/ppp/
Dpptp.c580 skb_queue_purge(&sk->sk_receive_queue); in pptp_sock_destruct()
Dpppoe.c620 skb_queue_purge(&sk->sk_receive_queue); in pppoe_release()
/linux-4.4.14/net/appletalk/
Dddp.c175 skb_queue_purge(&sk->sk_receive_queue); in atalk_destroy_socket()
1803 struct sk_buff *skb = skb_peek(&sk->sk_receive_queue); in atalk_ioctl()
/linux-4.4.14/net/sched/
Dem_meta.c429 dst->value = sk->sk_receive_queue.qlen; in META_COLLECTOR()
/linux-4.4.14/net/ipx/
Daf_ipx.c134 skb_queue_purge(&sk->sk_receive_queue); in ipx_destroy_socket()
1851 struct sk_buff *skb = skb_peek(&sk->sk_receive_queue); in ipx_ioctl()
/linux-4.4.14/net/netfilter/ipvs/
Dip_vs_sync.c1749 !skb_queue_empty(&tinfo->sock->sk->sk_receive_queue) in sync_thread_backup()
1753 while (!skb_queue_empty(&(tinfo->sock->sk->sk_receive_queue))) { in sync_thread_backup()
/linux-4.4.14/net/l2tp/
Dl2tp_ppp.c508 skb_queue_purge(&sk->sk_receive_queue); in pppol2tp_release()
/linux-4.4.14/net/key/
Daf_key.c99 skb_queue_purge(&sk->sk_receive_queue); in pfkey_sock_destruct()
207 skb_queue_tail(&sk->sk_receive_queue, *skb2); in pfkey_broadcast_one()