/linux-4.4.14/net/unix/ |
D | garbage.c | 165 spin_lock(&x->sk_receive_queue.lock); in scan_inflight() 166 skb_queue_walk_safe(&x->sk_receive_queue, skb, next) { in scan_inflight() 193 __skb_unlink(skb, &x->sk_receive_queue); in scan_inflight() 198 spin_unlock(&x->sk_receive_queue.lock); in scan_inflight() 215 spin_lock(&x->sk_receive_queue.lock); in scan_children() 216 skb_queue_walk_safe(&x->sk_receive_queue, skb, next) { in scan_children() 225 spin_unlock(&x->sk_receive_queue.lock); in scan_children()
|
D | diag.c | 64 spin_lock(&sk->sk_receive_queue.lock); in sk_diag_dump_icons() 67 sk->sk_receive_queue.qlen * sizeof(u32)); in sk_diag_dump_icons() 73 skb_queue_walk(&sk->sk_receive_queue, skb) { in sk_diag_dump_icons() 87 spin_unlock(&sk->sk_receive_queue.lock); in sk_diag_dump_icons() 93 spin_unlock(&sk->sk_receive_queue.lock); in sk_diag_dump_icons() 102 rql.udiag_rqueue = sk->sk_receive_queue.qlen; in sk_diag_show_rqlen()
|
D | af_unix.c | 196 return skb_queue_len(&sk->sk_receive_queue) > sk->sk_max_ack_backlog; in unix_recvq_full() 468 if (!skb_queue_empty(&sk->sk_receive_queue)) { in unix_dgram_disconnected() 469 skb_queue_purge(&sk->sk_receive_queue); in unix_dgram_disconnected() 487 skb_queue_purge(&sk->sk_receive_queue); in unix_sock_destructor() 540 if (!skb_queue_empty(&sk->sk_receive_queue) || embrion) in unix_release_sock() 554 while ((skb = skb_dequeue(&sk->sk_receive_queue)) != NULL) { in unix_release_sock() 769 lockdep_set_class(&sk->sk_receive_queue.lock, in unix_create1() 1360 spin_lock(&other->sk_receive_queue.lock); in unix_stream_connect() 1361 __skb_queue_tail(&other->sk_receive_queue, skb); in unix_stream_connect() 1362 spin_unlock(&other->sk_receive_queue.lock); in unix_stream_connect() [all …]
|
/linux-4.4.14/net/rxrpc/ |
D | ar-recvmsg.c | 86 skb = skb_peek(&rx->sk.sk_receive_queue); in rxrpc_recvmsg() 101 if (skb_queue_empty(&rx->sk.sk_receive_queue)) { in rxrpc_recvmsg() 124 if (skb_dequeue(&rx->sk.sk_receive_queue) != skb) in rxrpc_recvmsg() 215 if (skb_dequeue(&rx->sk.sk_receive_queue) != in rxrpc_recvmsg() 235 if (skb == (struct sk_buff *) &rx->sk.sk_receive_queue) in rxrpc_recvmsg() 241 if (skb_dequeue(&rx->sk.sk_receive_queue) != skb) in rxrpc_recvmsg() 268 if (skb_dequeue(&rx->sk.sk_receive_queue) != skb) in rxrpc_recvmsg() 320 if (skb_dequeue(&rx->sk.sk_receive_queue) != skb) in rxrpc_recvmsg()
|
D | ar-input.c | 85 spin_lock_bh(&sk->sk_receive_queue.lock); in rxrpc_queue_rcv_skb() 102 spin_unlock_bh(&sk->sk_receive_queue.lock); in rxrpc_queue_rcv_skb() 105 __skb_queue_tail(&sk->sk_receive_queue, skb); in rxrpc_queue_rcv_skb() 106 spin_unlock_bh(&sk->sk_receive_queue.lock); in rxrpc_queue_rcv_skb() 113 spin_unlock_bh(&sk->sk_receive_queue.lock); in rxrpc_queue_rcv_skb()
|
D | af_rxrpc.c | 599 if (!skb_queue_empty(&sk->sk_receive_queue)) in rxrpc_poll() 666 rxrpc_purge_queue(&sk->sk_receive_queue); in rxrpc_sock_destructor() 691 spin_lock_bh(&sk->sk_receive_queue.lock); in rxrpc_release_sock() 693 spin_unlock_bh(&sk->sk_receive_queue.lock); in rxrpc_release_sock() 706 rxrpc_purge_queue(&sk->sk_receive_queue); in rxrpc_release_sock()
|
/linux-4.4.14/include/net/ |
D | busy_poll.h | 111 } while (!nonblock && skb_queue_empty(&sk->sk_receive_queue) && in sk_busy_loop() 114 rc = !skb_queue_empty(&sk->sk_receive_queue); in sk_busy_loop()
|
D | sock.h | 359 struct sk_buff_head sk_receive_queue; member 2206 __skb_unlink(skb, &sk->sk_receive_queue); in sk_eat_skb()
|
/linux-4.4.14/net/atm/ |
D | signaling.c | 32 skb_queue_tail(&sk_atm(sigd)->sk_receive_queue, skb); in sigd_put_skb() 112 skb_queue_tail(&sk->sk_receive_queue, skb); in sigd_send() 204 if (skb_peek(&sk_atm(vcc)->sk_receive_queue)) in sigd_close() 206 skb_queue_purge(&sk_atm(vcc)->sk_receive_queue); in sigd_close()
|
D | raw.c | 27 skb_queue_tail(&sk->sk_receive_queue, skb); in atm_push_raw()
|
D | lec.c | 154 skb_queue_tail(&sk->sk_receive_queue, skb2); in lec_handle_bridge() 451 skb_queue_tail(&sk->sk_receive_queue, skb2); in lec_atm_send() 478 if (skb_peek(&sk_atm(vcc)->sk_receive_queue)) in lec_atm_close() 480 while ((skb = skb_dequeue(&sk_atm(vcc)->sk_receive_queue))) { in lec_atm_close() 534 skb_queue_tail(&sk->sk_receive_queue, skb); in send_to_lecd() 540 skb_queue_tail(&sk->sk_receive_queue, data); in send_to_lecd() 620 skb_queue_tail(&sk->sk_receive_queue, skb); in lec_push()
|
D | common.c | 191 while ((skb = skb_dequeue(&sk->sk_receive_queue)) != NULL) { in vcc_destroy_socket() 236 rq = &sk_atm(vcc)->sk_receive_queue; in vcc_process_recv_queue() 667 if (!skb_queue_empty(&sk->sk_receive_queue)) in vcc_poll()
|
D | svc.c | 64 while ((skb = skb_dequeue(&sk->sk_receive_queue)) != NULL) { in svc_disconnect() 343 while (!(skb = skb_dequeue(&sk->sk_receive_queue)) && in svc_accept()
|
D | mpc.c | 708 skb_queue_tail(&sk->sk_receive_queue, skb); in mpc_push() 896 while ((skb = skb_dequeue(&sk_atm(vcc)->sk_receive_queue))) { in mpoad_close() 994 skb_queue_tail(&sk->sk_receive_queue, skb); in msg_to_mpoad() 1275 skb_queue_tail(&sk->sk_receive_queue, skb); in purge_egress_shortcut()
|
D | ioctl.c | 78 skb = skb_peek(&sk->sk_receive_queue); in do_vcc_ioctl()
|
D | clip.c | 70 skb_queue_tail(&sk->sk_receive_queue, skb); in to_atmarpd() 613 skb_queue_purge(&sk_atm(vcc)->sk_receive_queue); in atmarpd_close()
|
/linux-4.4.14/net/core/ |
D | datagram.c | 99 if (sk->sk_receive_queue.prev != skb) in wait_for_more_packets() 197 struct sk_buff_head *queue = &sk->sk_receive_queue; in __skb_recv_datagram() 331 spin_lock_bh(&sk->sk_receive_queue.lock); in skb_kill_datagram() 332 if (skb == skb_peek(&sk->sk_receive_queue)) { in skb_kill_datagram() 333 __skb_unlink(skb, &sk->sk_receive_queue); in skb_kill_datagram() 337 spin_unlock_bh(&sk->sk_receive_queue.lock); in skb_kill_datagram() 772 if (!skb_queue_empty(&sk->sk_receive_queue)) in datagram_poll()
|
D | stream.c | 193 __skb_queue_purge(&sk->sk_receive_queue); in sk_stream_kill_queues()
|
D | sock.c | 451 struct sk_buff_head *list = &sk->sk_receive_queue; in sock_queue_rcv_skb() 1529 skb_queue_head_init(&newsk->sk_receive_queue); in sk_clone_lock() 2050 rc = sk_wait_event(sk, timeo, skb_peek_tail(&sk->sk_receive_queue) != skb); in sk_wait_data() 2366 skb_queue_head_init(&sk->sk_receive_queue); in sock_init_data()
|
/linux-4.4.14/net/bluetooth/ |
D | af_bluetooth.c | 265 if (!skb_queue_empty(&sk->sk_receive_queue)) in bt_sock_data_wait() 308 skb = skb_dequeue(&sk->sk_receive_queue); in bt_sock_stream_recvmsg() 334 skb_queue_head(&sk->sk_receive_queue, skb); in bt_sock_stream_recvmsg() 373 skb_queue_head(&sk->sk_receive_queue, skb); in bt_sock_stream_recvmsg() 380 skb_queue_head(&sk->sk_receive_queue, skb); in bt_sock_stream_recvmsg() 430 if (!skb_queue_empty(&sk->sk_receive_queue)) in bt_sock_poll() 475 skb = skb_peek(&sk->sk_receive_queue); in bt_sock_ioctl()
|
D | sco.c | 370 skb_queue_purge(&sk->sk_receive_queue); in sco_sock_destruct()
|
D | hci_sock.c | 590 skb_queue_purge(&sk->sk_receive_queue); in hci_sock_release()
|
D | l2cap_sock.c | 1484 skb_queue_purge(&sk->sk_receive_queue); in l2cap_sock_destruct()
|
/linux-4.4.14/net/sctp/ |
D | ulpqueue.c | 144 sctp_skb_list_tail(&sp->pd_lobby, &sk->sk_receive_queue); in sctp_clear_pd() 163 __skb_queue_tail(&sk->sk_receive_queue, in sctp_clear_pd() 222 queue = &sk->sk_receive_queue; in sctp_ulpq_tail_event() 236 queue = &sk->sk_receive_queue; in sctp_ulpq_tail_event() 245 queue = &sk->sk_receive_queue; in sctp_ulpq_tail_event() 266 if (queue == &sk->sk_receive_queue) in sctp_ulpq_tail_event() 1096 if (skb_queue_empty(&asoc->base.sk->sk_receive_queue)) { in sctp_ulpq_renege() 1139 __skb_queue_tail(&sk->sk_receive_queue, sctp_event2skb(ev)); in sctp_ulpq_abort_pd()
|
D | socket.c | 1493 data_was_unread = sctp_queue_purge_ulpevents(&sk->sk_receive_queue); in sctp_close() 2133 skb_queue_head(&sk->sk_receive_queue, skb); in sctp_recvmsg() 3963 skb = skb_peek(&sk->sk_receive_queue); in sctp_ioctl() 6453 if (!skb_queue_empty(&sk->sk_receive_queue)) in sctp_poll() 6701 if (!skb_queue_empty(&sk->sk_receive_queue)) in sctp_wait_for_packet() 6768 spin_lock_bh(&sk->sk_receive_queue.lock); in sctp_skb_recv_datagram() 6769 skb = skb_peek(&sk->sk_receive_queue); in sctp_skb_recv_datagram() 6772 spin_unlock_bh(&sk->sk_receive_queue.lock); in sctp_skb_recv_datagram() 6774 skb = skb_dequeue(&sk->sk_receive_queue); in sctp_skb_recv_datagram() 7278 sctp_skb_for_each(skb, &oldsk->sk_receive_queue, tmp) { in sctp_sock_migrate() [all …]
|
/linux-4.4.14/net/caif/ |
D | caif_socket.c | 128 struct sk_buff_head *list = &sk->sk_receive_queue; in caif_queue_rcv_skb() 317 if (!skb_queue_empty(&sk->sk_receive_queue) || in caif_stream_data_wait() 381 skb = skb_dequeue(&sk->sk_receive_queue); in caif_stream_recvmsg() 426 skb_queue_head(&sk->sk_receive_queue, skb); in caif_stream_recvmsg() 440 skb_queue_head(&sk->sk_receive_queue, skb); in caif_stream_recvmsg() 450 skb_queue_head(&sk->sk_receive_queue, skb); in caif_stream_recvmsg() 909 spin_lock_bh(&sk->sk_receive_queue.lock); in caif_release() 911 spin_unlock_bh(&sk->sk_receive_queue.lock); in caif_release() 952 if (!skb_queue_empty(&sk->sk_receive_queue) || in caif_poll()
|
/linux-4.4.14/net/tipc/ |
D | socket.c | 249 kfree_skb(__skb_dequeue(&sk->sk_receive_queue)); in tsk_advance_rx_queue() 277 while ((skb = __skb_dequeue(&sk->sk_receive_queue))) in tsk_rej_rx_queue() 446 skb = __skb_dequeue(&sk->sk_receive_queue); in tipc_release() 647 if (!skb_queue_empty(&sk->sk_receive_queue)) in tipc_poll() 1253 if (timeo && skb_queue_empty(&sk->sk_receive_queue)) { in tipc_wait_for_rcvmsg() 1263 if (!skb_queue_empty(&sk->sk_receive_queue)) in tipc_wait_for_rcvmsg() 1320 buf = skb_peek(&sk->sk_receive_queue); in tipc_recvmsg() 1417 buf = skb_peek(&sk->sk_receive_queue); in tipc_recv_stream() 1478 (!skb_queue_empty(&sk->sk_receive_queue) || in tipc_recv_stream() 1524 __skb_queue_purge(&sk->sk_receive_queue); in tipc_sock_destruct() [all …]
|
/linux-4.4.14/net/phonet/ |
D | datagram.c | 52 skb = skb_peek(&sk->sk_receive_queue); in pn_ioctl() 77 skb_queue_purge(&sk->sk_receive_queue); in pn_destruct()
|
D | pep.c | 417 queue = &sk->sk_receive_queue; in pipe_do_rcv() 476 skb_queue_purge(&sk->sk_receive_queue); in pipe_destruct() 589 skb_queue_tail(&sk->sk_receive_queue, skb); in pipe_handler_do_rcv() 696 skb_queue_head(&sk->sk_receive_queue, skb); in pep_do_rcv() 945 else if (!skb_queue_empty(&sk->sk_receive_queue)) in pep_ioctl() 946 answ = skb_peek(&sk->sk_receive_queue)->len; in pep_ioctl() 1241 struct sk_buff *skb = skb_dequeue(&sk->sk_receive_queue); in pep_read()
|
D | socket.c | 351 if (!skb_queue_empty(&sk->sk_receive_queue)) in pn_socket_poll()
|
/linux-4.4.14/net/llc/ |
D | af_llc.c | 663 if (skb_queue_empty(&sk->sk_receive_queue)) { in llc_ui_accept() 670 skb = skb_dequeue(&sk->sk_receive_queue); in llc_ui_accept() 757 skb = skb_peek(&sk->sk_receive_queue); in llc_ui_recvmsg() 841 spin_lock_irqsave(&sk->sk_receive_queue.lock, cpu_flags); in llc_ui_recvmsg() 843 spin_unlock_irqrestore(&sk->sk_receive_queue.lock, cpu_flags); in llc_ui_recvmsg() 864 spin_lock_irqsave(&sk->sk_receive_queue.lock, cpu_flags); in llc_ui_recvmsg() 866 spin_unlock_irqrestore(&sk->sk_receive_queue.lock, cpu_flags); in llc_ui_recvmsg()
|
D | llc_conn.c | 109 skb_queue_tail(&sk->sk_receive_queue, skb); in llc_conn_state_process() 969 skb_queue_purge(&sk->sk_receive_queue); in llc_sk_free()
|
/linux-4.4.14/net/irda/ |
D | af_irda.c | 288 skb_queue_tail(&sk->sk_receive_queue, skb); in irda_connect_indication() 874 skb = skb_dequeue(&sk->sk_receive_queue); in irda_accept() 884 skb_peek(&sk->sk_receive_queue)); in irda_accept() 1226 skb_queue_purge(&sk->sk_receive_queue); in irda_release() 1431 struct sk_buff *skb = skb_dequeue(&sk->sk_receive_queue); in irda_recvmsg_stream() 1456 else if (skb_peek(&sk->sk_receive_queue) == NULL) in irda_recvmsg_stream() 1472 skb_queue_head(&sk->sk_receive_queue, skb); in irda_recvmsg_stream() 1488 skb_queue_head(&sk->sk_receive_queue, skb); in irda_recvmsg_stream() 1497 skb_queue_head(&sk->sk_receive_queue, skb); in irda_recvmsg_stream() 1758 if (!skb_queue_empty(&sk->sk_receive_queue)) { in irda_poll() [all …]
|
/linux-4.4.14/net/packet/ |
D | af_packet.c | 678 spin_lock(&po->sk.sk_receive_queue.lock); in prb_retire_rx_blk_timer_expired() 741 spin_unlock(&po->sk.sk_receive_queue.lock); in prb_retire_rx_blk_timer_expired() 1301 spin_lock_bh(&po->sk.sk_receive_queue.lock); in packet_rcv_has_room() 1306 spin_unlock_bh(&po->sk.sk_receive_queue.lock); in packet_rcv_has_room() 2067 spin_lock(&sk->sk_receive_queue.lock); in packet_rcv() 2070 __skb_queue_tail(&sk->sk_receive_queue, skb); in packet_rcv() 2071 spin_unlock(&sk->sk_receive_queue.lock); in packet_rcv() 2076 spin_lock(&sk->sk_receive_queue.lock); in packet_rcv() 2079 spin_unlock(&sk->sk_receive_queue.lock); in packet_rcv() 2188 spin_lock(&sk->sk_receive_queue.lock); in tpacket_rcv() [all …]
|
/linux-4.4.14/drivers/net/ |
D | macvtap.c | 354 if (skb_queue_len(&q->sk.sk_receive_queue) >= dev->tx_queue_len) in macvtap_handle_frame() 372 skb_queue_tail(&q->sk.sk_receive_queue, skb); in macvtap_handle_frame() 381 skb_queue_tail(&q->sk.sk_receive_queue, segs); in macvtap_handle_frame() 394 skb_queue_tail(&q->sk.sk_receive_queue, skb); in macvtap_handle_frame() 511 skb_queue_purge(&sk->sk_receive_queue); in macvtap_sock_destruct() 584 if (!skb_queue_empty(&q->sk.sk_receive_queue)) in macvtap_poll() 927 skb = skb_dequeue(&q->sk.sk_receive_queue); in macvtap_do_read()
|
D | tun.c | 505 skb_queue_purge(&tfile->sk.sk_receive_queue); in tun_queue_purge() 858 if (skb_queue_len(&tfile->socket.sk->sk_receive_queue) * numqueues in tun_net_xmit() 878 skb_queue_tail(&tfile->socket.sk->sk_receive_queue, skb); in tun_net_xmit() 1041 if (!skb_queue_empty(&sk->sk_receive_queue)) in tun_chr_poll()
|
/linux-4.4.14/net/nfc/ |
D | llcp_sock.c | 560 if (!skb_queue_empty(&sk->sk_receive_queue)) in llcp_sock_poll() 810 skb_queue_empty(&sk->sk_receive_queue)) { in llcp_sock_recvmsg() 837 skb_queue_head(&sk->sk_receive_queue, skb); in llcp_sock_recvmsg() 868 skb_queue_head(&sk->sk_receive_queue, skb); in llcp_sock_recvmsg() 935 skb_queue_purge(&sk->sk_receive_queue); in llcp_sock_destruct()
|
D | rawsock.c | 329 skb_queue_purge(&sk->sk_receive_queue); in rawsock_destruct()
|
/linux-4.4.14/net/netlink/ |
D | af_netlink.c | 297 if (skb_queue_empty(&sk->sk_receive_queue)) in netlink_rcv_wake() 387 queue = tx_ring ? &sk->sk_write_queue : &sk->sk_receive_queue; in __netlink_set_ring() 700 spin_lock_bh(&sk->sk_receive_queue.lock); in netlink_poll() 705 spin_unlock_bh(&sk->sk_receive_queue.lock); in netlink_poll() 845 spin_lock_bh(&sk->sk_receive_queue.lock); in netlink_ring_set_copied() 848 spin_unlock_bh(&sk->sk_receive_queue.lock); in netlink_ring_set_copied() 854 __skb_queue_tail(&sk->sk_receive_queue, skb); in netlink_ring_set_copied() 855 spin_unlock_bh(&sk->sk_receive_queue.lock); in netlink_ring_set_copied() 939 skb_queue_purge(&sk->sk_receive_queue); in netlink_sock_destruct() 1777 skb_queue_tail(&sk->sk_receive_queue, skb); in __netlink_sendskb() [all …]
|
/linux-4.4.14/net/dccp/ |
D | proto.c | 275 __skb_queue_purge(&sk->sk_receive_queue); in dccp_disconnect() 373 skb = skb_peek(&sk->sk_receive_queue); in dccp_ioctl() 823 struct sk_buff *skb = skb_peek(&sk->sk_receive_queue); in dccp_recvmsg() 1007 while ((skb = __skb_dequeue(&sk->sk_receive_queue)) != NULL) { in dccp_close()
|
D | input.c | 29 __skb_queue_tail(&sk->sk_receive_queue, skb); in dccp_enqueue_skb()
|
/linux-4.4.14/net/x25/ |
D | af_x25.c | 399 while ((skb = skb_dequeue(&sk->sk_receive_queue)) != NULL) { in __x25_destroy_socket() 843 if (skb_queue_empty(&sk->sk_receive_queue)) { in x25_wait_for_data() 877 skb = skb_dequeue(&sk->sk_receive_queue); in x25_accept() 1062 skb_queue_head(&sk->sk_receive_queue, skb); in x25_rx_call_request() 1382 if ((skb = skb_peek(&sk->sk_receive_queue)) != NULL) in x25_ioctl()
|
D | x25_in.c | 80 skb_queue_tail(&sk->sk_receive_queue, skbn); in x25_queue_rx_frame()
|
/linux-4.4.14/net/netrom/ |
D | af_netrom.c | 274 while ((skb = skb_dequeue(&sk->sk_receive_queue)) != NULL) { in nr_destroy_socket() 796 skb = skb_dequeue(&sk->sk_receive_queue); in nr_accept() 1011 skb_queue_head(&sk->sk_receive_queue, skb); in nr_rx_frame() 1215 if ((skb = skb_peek(&sk->sk_receive_queue)) != NULL) in nr_ioctl()
|
/linux-4.4.14/net/iucv/ |
D | af_iucv.c | 394 skb_queue_purge(&sk->sk_receive_queue); in iucv_sock_destruct() 1333 skb_queue_empty(&sk->sk_receive_queue) && in iucv_sock_recvmsg() 1358 skb_queue_head(&sk->sk_receive_queue, skb); in iucv_sock_recvmsg() 1378 skb_queue_head(&sk->sk_receive_queue, skb); in iucv_sock_recvmsg() 1389 skb_queue_head(&sk->sk_receive_queue, skb); in iucv_sock_recvmsg() 1476 if (!skb_queue_empty(&sk->sk_receive_queue) || in iucv_sock_poll() 1550 skb_queue_purge(&sk->sk_receive_queue); in iucv_sock_shutdown()
|
/linux-4.4.14/net/ipv4/ |
D | tcp.c | 1390 struct sk_buff *skb = skb_peek(&sk->sk_receive_queue); in tcp_cleanup_rbuf() 1466 while ((skb = skb_peek(&sk->sk_receive_queue)) != NULL) { in tcp_recv_skb() 1592 if (sk_can_busy_loop(sk) && skb_queue_empty(&sk->sk_receive_queue) && in tcp_recvmsg() 1646 last = skb_peek_tail(&sk->sk_receive_queue); in tcp_recvmsg() 1647 skb_queue_walk(&sk->sk_receive_queue, skb) { in tcp_recvmsg() 2040 while ((skb = __skb_dequeue(&sk->sk_receive_queue)) != NULL) { in tcp_close() 2229 __skb_queue_purge(&sk->sk_receive_queue); in tcp_disconnect()
|
D | tcp_fastopen.c | 198 __skb_queue_tail(&child->sk_receive_queue, skb2); in tcp_fastopen_create_child()
|
D | raw.c | 884 spin_lock_bh(&sk->sk_receive_queue.lock); in raw_ioctl() 885 skb = skb_peek(&sk->sk_receive_queue); in raw_ioctl() 888 spin_unlock_bh(&sk->sk_receive_queue.lock); in raw_ioctl()
|
D | tcp_input.c | 4306 tail = skb_peek_tail(&sk->sk_receive_queue); in tcp_ofo_queue() 4310 __skb_queue_tail(&sk->sk_receive_queue, skb); in tcp_ofo_queue() 4471 struct sk_buff *tail = skb_peek_tail(&sk->sk_receive_queue); in tcp_queue_rcv() 4478 __skb_queue_tail(&sk->sk_receive_queue, skb); in tcp_queue_rcv() 4580 if (skb_queue_len(&sk->sk_receive_queue) == 0) in tcp_data_queue() 4859 if (!skb_queue_empty(&sk->sk_receive_queue)) in tcp_prune_queue() 4860 tcp_collapse(sk, &sk->sk_receive_queue, in tcp_prune_queue() 4861 skb_peek(&sk->sk_receive_queue), in tcp_prune_queue() 5043 struct sk_buff *skb = skb_peek(&sk->sk_receive_queue); in tcp_check_urg() 5046 __skb_unlink(skb, &sk->sk_receive_queue); in tcp_check_urg()
|
D | af_inet.c | 137 __skb_queue_purge(&sk->sk_receive_queue); in inet_sock_destruct()
|
D | udp.c | 1196 struct sk_buff_head list_kill, *rcvq = &sk->sk_receive_queue; in first_packet_length()
|
/linux-4.4.14/net/decnet/ |
D | dn_nsp_in.c | 340 skb_queue_tail(&sk->sk_receive_queue, skb); in dn_nsp_conn_init() 653 if (dn_queue_skb(sk, skb, SIGIO, &sk->sk_receive_queue) == 0) { in dn_nsp_data()
|
D | af_decnet.c | 1049 skb = skb_dequeue(&sk->sk_receive_queue); in dn_wait_for_connect() 1052 skb = skb_dequeue(&sk->sk_receive_queue); in dn_wait_for_connect() 1091 skb = skb_dequeue(&sk->sk_receive_queue); in dn_accept() 1259 skb_queue_walk(&sk->sk_receive_queue, skb) in dn_ioctl() 1680 struct sk_buff_head *queue = &sk->sk_receive_queue; in dn_recvmsg()
|
/linux-4.4.14/net/ieee802154/ |
D | socket.c | 552 spin_lock_bh(&sk->sk_receive_queue.lock); in dgram_ioctl() 553 skb = skb_peek(&sk->sk_receive_queue); in dgram_ioctl() 561 spin_unlock_bh(&sk->sk_receive_queue.lock); in dgram_ioctl()
|
/linux-4.4.14/net/rose/ |
D | af_rose.c | 343 while ((skb = skb_dequeue(&sk->sk_receive_queue)) != NULL) { in rose_destroy_socket() 903 skb = skb_dequeue(&sk->sk_receive_queue); in rose_accept() 1040 skb_queue_head(&sk->sk_receive_queue, skb); in rose_rx_call_request() 1296 if ((skb = skb_peek(&sk->sk_receive_queue)) != NULL) in rose_ioctl()
|
/linux-4.4.14/net/rds/ |
D | tcp_recv.c | 256 skb_queue_len(&tc->t_sock->sk->sk_receive_queue)); in rds_tcp_data_recv()
|
/linux-4.4.14/drivers/vhost/ |
D | net.c | 426 spin_lock_irqsave(&sk->sk_receive_queue.lock, flags); in peek_head_len() 427 head = skb_peek(&sk->sk_receive_queue); in peek_head_len() 434 spin_unlock_irqrestore(&sk->sk_receive_queue.lock, flags); in peek_head_len()
|
/linux-4.4.14/net/ipv6/ |
D | raw.c | 1140 spin_lock_bh(&sk->sk_receive_queue.lock); in rawv6_ioctl() 1141 skb = skb_peek(&sk->sk_receive_queue); in rawv6_ioctl() 1145 spin_unlock_bh(&sk->sk_receive_queue.lock); in rawv6_ioctl()
|
/linux-4.4.14/net/ax25/ |
D | ax25_in.c | 358 skb_queue_head(&sk->sk_receive_queue, skb); in ax25_rcv()
|
D | af_ax25.c | 306 while ((skb = skb_dequeue(&ax25->sk->sk_receive_queue)) != NULL) { in ax25_destroy_socket() 1353 skb = skb_dequeue(&sk->sk_receive_queue); in ax25_accept() 1703 if ((skb = skb_peek(&sk->sk_receive_queue)) != NULL) in ax25_ioctl()
|
/linux-4.4.14/drivers/isdn/mISDN/ |
D | socket.c | 160 skb_queue_head(&sk->sk_receive_queue, skb); in mISDN_sock_recvmsg() 280 skb_queue_purge(&sk->sk_receive_queue); in data_sock_release()
|
/linux-4.4.14/net/bluetooth/cmtp/ |
D | core.c | 300 while ((skb = skb_dequeue(&sk->sk_receive_queue))) { in cmtp_session()
|
/linux-4.4.14/net/bluetooth/rfcomm/ |
D | sock.c | 56 skb_queue_tail(&sk->sk_receive_queue, skb); in rfcomm_sk_data_ready() 161 skb_queue_purge(&sk->sk_receive_queue); in rfcomm_sock_destruct()
|
D | tty.c | 286 while ((skb = skb_dequeue(&sk->sk_receive_queue))) { in __rfcomm_dev_add()
|
D | core.c | 1910 BT_DBG("session %p state %ld qlen %d", s, s->state, skb_queue_len(&sk->sk_receive_queue)); in rfcomm_process_rx() 1913 while ((skb = skb_dequeue(&sk->sk_receive_queue))) { in rfcomm_process_rx()
|
/linux-4.4.14/net/bluetooth/hidp/ |
D | core.c | 1205 while ((skb = skb_dequeue(&intr_sk->sk_receive_queue))) { in hidp_session_run() 1218 while ((skb = skb_dequeue(&ctrl_sk->sk_receive_queue))) { in hidp_session_run()
|
/linux-4.4.14/net/bluetooth/bnep/ |
D | core.c | 501 while ((skb = skb_dequeue(&sk->sk_receive_queue))) { in bnep_session()
|
/linux-4.4.14/net/can/ |
D | af_can.c | 115 skb_queue_purge(&sk->sk_receive_queue); in can_sock_destruct()
|
/linux-4.4.14/net/vmw_vsock/ |
D | af_vsock.c | 667 while ((skb = skb_dequeue(&sk->sk_receive_queue))) in __vsock_release() 875 if (!skb_queue_empty(&sk->sk_receive_queue) || in vsock_poll()
|
/linux-4.4.14/drivers/net/ppp/ |
D | pptp.c | 580 skb_queue_purge(&sk->sk_receive_queue); in pptp_sock_destruct()
|
D | pppoe.c | 620 skb_queue_purge(&sk->sk_receive_queue); in pppoe_release()
|
/linux-4.4.14/net/appletalk/ |
D | ddp.c | 175 skb_queue_purge(&sk->sk_receive_queue); in atalk_destroy_socket() 1803 struct sk_buff *skb = skb_peek(&sk->sk_receive_queue); in atalk_ioctl()
|
/linux-4.4.14/net/sched/ |
D | em_meta.c | 429 dst->value = sk->sk_receive_queue.qlen; in META_COLLECTOR()
|
/linux-4.4.14/net/ipx/ |
D | af_ipx.c | 134 skb_queue_purge(&sk->sk_receive_queue); in ipx_destroy_socket() 1851 struct sk_buff *skb = skb_peek(&sk->sk_receive_queue); in ipx_ioctl()
|
/linux-4.4.14/net/netfilter/ipvs/ |
D | ip_vs_sync.c | 1749 !skb_queue_empty(&tinfo->sock->sk->sk_receive_queue) in sync_thread_backup() 1753 while (!skb_queue_empty(&(tinfo->sock->sk->sk_receive_queue))) { in sync_thread_backup()
|
/linux-4.4.14/net/l2tp/ |
D | l2tp_ppp.c | 508 skb_queue_purge(&sk->sk_receive_queue); in pppol2tp_release()
|
/linux-4.4.14/net/key/ |
D | af_key.c | 99 skb_queue_purge(&sk->sk_receive_queue); in pfkey_sock_destruct() 207 skb_queue_tail(&sk->sk_receive_queue, *skb2); in pfkey_broadcast_one()
|