Lines Matching refs:sk
134 int sctp_clear_pd(struct sock *sk, struct sctp_association *asoc) in sctp_clear_pd() argument
136 struct sctp_sock *sp = sctp_sk(sk); in sctp_clear_pd()
144 sctp_skb_list_tail(&sp->pd_lobby, &sk->sk_receive_queue); in sctp_clear_pd()
145 list = (struct list_head *)&sctp_sk(sk)->pd_lobby; in sctp_clear_pd()
163 __skb_queue_tail(&sk->sk_receive_queue, in sctp_clear_pd()
176 struct sctp_sock *sp = sctp_sk(ulpq->asoc->base.sk); in sctp_ulpq_set_pd()
187 return sctp_clear_pd(ulpq->asoc->base.sk, ulpq->asoc); in sctp_ulpq_clear_pd()
195 struct sock *sk = ulpq->asoc->base.sk; in sctp_ulpq_tail_event() local
205 if (sock_flag(sk, SOCK_DEAD) || (sk->sk_shutdown & RCV_SHUTDOWN)) in sctp_ulpq_tail_event()
209 sk_mark_napi_id(sk, skb); in sctp_ulpq_tail_event()
210 sk_incoming_cpu_update(sk); in sctp_ulpq_tail_event()
213 if (!sctp_ulpevent_is_enabled(event, &sctp_sk(sk)->subscribe)) in sctp_ulpq_tail_event()
221 if (atomic_read(&sctp_sk(sk)->pd_mode) == 0) { in sctp_ulpq_tail_event()
222 queue = &sk->sk_receive_queue; in sctp_ulpq_tail_event()
233 queue = &sctp_sk(sk)->pd_lobby; in sctp_ulpq_tail_event()
236 queue = &sk->sk_receive_queue; in sctp_ulpq_tail_event()
244 if (sctp_sk(sk)->frag_interleave) in sctp_ulpq_tail_event()
245 queue = &sk->sk_receive_queue; in sctp_ulpq_tail_event()
247 queue = &sctp_sk(sk)->pd_lobby; in sctp_ulpq_tail_event()
266 if (queue == &sk->sk_receive_queue) in sctp_ulpq_tail_event()
267 sk->sk_data_ready(sk); in sctp_ulpq_tail_event()
363 sctp_skb_set_owner_r(new, f_frag->sk); in sctp_make_reassembled_event()
490 if (!sctp_sk(asoc->base.sk)->frag_interleave && in sctp_ulpq_retrieve_reassembled()
491 atomic_read(&sctp_sk(asoc->base.sk)->pd_mode)) in sctp_ulpq_retrieve_reassembled()
495 pd_point = sctp_sk(asoc->base.sk)->pd_point; in sctp_ulpq_retrieve_reassembled()
497 retval = sctp_make_reassembled_event(sock_net(asoc->base.sk), in sctp_ulpq_retrieve_reassembled()
508 retval = sctp_make_reassembled_event(sock_net(ulpq->asoc->base.sk), in sctp_ulpq_retrieve_reassembled()
574 retval = sctp_make_reassembled_event(sock_net(ulpq->asoc->base.sk), in sctp_ulpq_retrieve_partial()
675 retval = sctp_make_reassembled_event(sock_net(ulpq->asoc->base.sk), in sctp_ulpq_retrieve_first()
1044 sp = sctp_sk(asoc->base.sk); in sctp_ulpq_partial_delivery()
1096 if (skb_queue_empty(&asoc->base.sk->sk_receive_queue)) { in sctp_ulpq_renege()
1116 sk_mem_reclaim(asoc->base.sk); in sctp_ulpq_renege()
1127 struct sock *sk; in sctp_ulpq_abort_pd() local
1132 sk = ulpq->asoc->base.sk; in sctp_ulpq_abort_pd()
1134 &sctp_sk(sk)->subscribe)) in sctp_ulpq_abort_pd()
1139 __skb_queue_tail(&sk->sk_receive_queue, sctp_event2skb(ev)); in sctp_ulpq_abort_pd()
1143 sk->sk_data_ready(sk); in sctp_ulpq_abort_pd()