Searched refs:send (Results 1 - 200 of 3139) sorted by relevance

1234567891011>>

/linux-4.1.27/net/rds/
H A Diw_send.c78 struct rds_iw_send_work *send, rds_iw_send_unmap_rm()
81 struct rds_message *rm = send->s_rm; rds_iw_send_unmap_rm()
83 rdsdebug("ic %p send %p rm %p\n", ic, send, rm); rds_iw_send_unmap_rm()
125 send->s_rm = NULL; rds_iw_send_unmap_rm()
130 struct rds_iw_send_work *send; rds_iw_send_init_ring() local
133 for (i = 0, send = ic->i_sends; i < ic->i_send_ring.w_nr; i++, send++) { rds_iw_send_init_ring()
136 send->s_rm = NULL; rds_iw_send_init_ring()
137 send->s_op = NULL; rds_iw_send_init_ring()
138 send->s_mapping = NULL; rds_iw_send_init_ring()
140 send->s_wr.next = NULL; rds_iw_send_init_ring()
141 send->s_wr.wr_id = i; rds_iw_send_init_ring()
142 send->s_wr.sg_list = send->s_sge; rds_iw_send_init_ring()
143 send->s_wr.num_sge = 1; rds_iw_send_init_ring()
144 send->s_wr.opcode = IB_WR_SEND; rds_iw_send_init_ring()
145 send->s_wr.send_flags = 0; rds_iw_send_init_ring()
146 send->s_wr.ex.imm_data = 0; rds_iw_send_init_ring()
148 sge = rds_iw_data_sge(ic, send->s_sge); rds_iw_send_init_ring()
151 sge = rds_iw_header_sge(ic, send->s_sge); rds_iw_send_init_ring()
156 send->s_mr = ib_alloc_fast_reg_mr(ic->i_pd, fastreg_message_size); rds_iw_send_init_ring()
157 if (IS_ERR(send->s_mr)) { rds_iw_send_init_ring()
162 send->s_page_list = ib_alloc_fast_reg_page_list( rds_iw_send_init_ring()
164 if (IS_ERR(send->s_page_list)) { rds_iw_send_init_ring()
173 struct rds_iw_send_work *send; rds_iw_send_clear_ring() local
176 for (i = 0, send = ic->i_sends; i < ic->i_send_ring.w_nr; i++, send++) { rds_iw_send_clear_ring()
177 BUG_ON(!send->s_mr); rds_iw_send_clear_ring()
178 ib_dereg_mr(send->s_mr); rds_iw_send_clear_ring()
179 BUG_ON(!send->s_page_list); rds_iw_send_clear_ring()
180 ib_free_fast_reg_page_list(send->s_page_list); rds_iw_send_clear_ring()
181 if (send->s_wr.opcode == 0xdead) rds_iw_send_clear_ring()
183 if (send->s_rm) rds_iw_send_clear_ring()
184 rds_iw_send_unmap_rm(ic, send, IB_WC_WR_FLUSH_ERR); rds_iw_send_clear_ring()
185 if (send->s_op) rds_iw_send_clear_ring()
186 rds_iw_send_unmap_rdma(ic, send->s_op); rds_iw_send_clear_ring()
192 * operations performed in the send path. As the sender allocs and potentially
201 struct rds_iw_send_work *send; rds_iw_send_cq_comp_handler() local
211 rdsdebug("ib_req_notify_cq send failed: %d\n", ret); rds_iw_send_cq_comp_handler()
246 send = &ic->i_sends[oldest]; rds_iw_send_cq_comp_handler()
249 switch (send->s_wr.opcode) { rds_iw_send_cq_comp_handler()
251 if (send->s_rm) rds_iw_send_cq_comp_handler()
252 rds_iw_send_unmap_rm(ic, send, wc.status); rds_iw_send_cq_comp_handler()
264 __func__, send->s_wr.opcode); rds_iw_send_cq_comp_handler()
268 send->s_wr.opcode = 0xdead; rds_iw_send_cq_comp_handler()
269 send->s_wr.num_sge = 1; rds_iw_send_cq_comp_handler()
270 if (time_after(jiffies, send->s_queued + HZ/2)) rds_iw_send_cq_comp_handler()
277 if (unlikely(wc.status == IB_WC_REM_ACCESS_ERR && send->s_op)) { rds_iw_send_cq_comp_handler()
280 rm = rds_send_get_message(conn, send->s_op); rds_iw_send_cq_comp_handler()
297 "send completion on %pI4 " rds_iw_send_cq_comp_handler()
309 * - send credits: this tells us how many WRs we're allowed
320 * exhausted their send credits, and are unable to send new credits
321 * to the peer. We achieve this by requiring that we send at least
327 * The RDS send code is essentially single-threaded; rds_send_xmit
328 * grabs c_send_lock to ensure exclusive access to the send ring.
332 * In the send path, we need to update the counters for send credits
338 * Spinlocks shared between the send and the receive path are bad,
367 /* The last credit must be used to send a credit update. */ rds_iw_send_grab_credits()
385 * the posted regardless of whether any send credits are rds_iw_send_grab_credits()
431 /* Decide whether to send an update to the peer now. rds_iw_advertise_credits()
432 * If we would send a credit update for every single buffer we rds_iw_advertise_credits()
434 * consumes buffer, we refill the ring, send ACK to remote rds_iw_advertise_credits()
437 * Performance pretty much depends on how often we send rds_iw_advertise_credits()
449 struct rds_iw_send_work *send, unsigned int pos, rds_iw_xmit_populate_wr()
455 WARN_ON(pos != send - ic->i_sends); rds_iw_xmit_populate_wr()
457 send->s_wr.send_flags = send_flags; rds_iw_xmit_populate_wr()
458 send->s_wr.opcode = IB_WR_SEND; rds_iw_xmit_populate_wr()
459 send->s_wr.num_sge = 2; rds_iw_xmit_populate_wr()
460 send->s_wr.next = NULL; rds_iw_xmit_populate_wr()
461 send->s_queued = jiffies; rds_iw_xmit_populate_wr()
462 send->s_op = NULL; rds_iw_xmit_populate_wr()
465 sge = rds_iw_data_sge(ic, send->s_sge); rds_iw_xmit_populate_wr()
470 sge = rds_iw_header_sge(ic, send->s_sge); rds_iw_xmit_populate_wr()
474 send->s_wr.num_sge = 1; rds_iw_xmit_populate_wr()
475 sge = &send->s_sge[0]; rds_iw_xmit_populate_wr()
490 * once we send the final fragment.
501 struct rds_iw_send_work *send = NULL; rds_iw_xmit() local
609 * sticking the header into the send ring. Which is why we rds_iw_xmit()
622 send = &ic->i_sends[pos]; rds_iw_xmit()
623 first = send; rds_iw_xmit()
642 * use a second sge and our long-lived ring of mapped headers. We send rds_iw_xmit()
649 rds_iw_xmit_populate_wr(ic, send, pos, 0, 0, send_flags); rds_iw_xmit()
657 send = &ic->i_sends[pos]; rds_iw_xmit()
660 rds_iw_xmit_populate_wr(ic, send, pos, rds_iw_xmit()
671 send->s_wr.send_flags |= IB_SEND_SIGNALED | IB_SEND_SOLICITED; rds_iw_xmit()
677 send->s_wr.send_flags |= IB_SEND_SIGNALED | IB_SEND_SOLICITED; rds_iw_xmit()
684 send->s_wr.send_flags |= IB_SEND_SIGNALED | IB_SEND_SOLICITED; rds_iw_xmit()
686 rdsdebug("send %p wr %p num_sge %u next %p\n", send, rds_iw_xmit()
687 &send->s_wr, send->s_wr.num_sge, send->s_wr.next); rds_iw_xmit()
704 printk(KERN_NOTICE "send WR dport=%u flags=0x%x len=%d\n", rds_iw_xmit()
720 prev->s_wr.next = &send->s_wr; rds_iw_xmit()
721 prev = send; rds_iw_xmit()
731 /* if we finished the message then send completion owns it */ rds_iw_xmit()
768 static void rds_iw_build_send_fastreg(struct rds_iw_device *rds_iwdev, struct rds_iw_connection *ic, struct rds_iw_send_work *send, int nent, int len, u64 sg_addr) rds_iw_build_send_fastreg() argument
770 BUG_ON(nent > send->s_page_list->max_page_list_len); rds_iw_build_send_fastreg()
776 send->s_wr.opcode = IB_WR_FAST_REG_MR; rds_iw_build_send_fastreg()
777 send->s_wr.wr.fast_reg.length = len; rds_iw_build_send_fastreg()
778 send->s_wr.wr.fast_reg.rkey = send->s_mr->rkey; rds_iw_build_send_fastreg()
779 send->s_wr.wr.fast_reg.page_list = send->s_page_list; rds_iw_build_send_fastreg()
780 send->s_wr.wr.fast_reg.page_list_len = nent; rds_iw_build_send_fastreg()
781 send->s_wr.wr.fast_reg.page_shift = PAGE_SHIFT; rds_iw_build_send_fastreg()
782 send->s_wr.wr.fast_reg.access_flags = IB_ACCESS_REMOTE_WRITE; rds_iw_build_send_fastreg()
783 send->s_wr.wr.fast_reg.iova_start = sg_addr; rds_iw_build_send_fastreg()
785 ib_update_fast_reg_key(send->s_mr, send->s_remap_count++); rds_iw_build_send_fastreg()
791 struct rds_iw_send_work *send = NULL; rds_iw_xmit_rdma() local
825 /* Alloc space on the send queue for the fastreg */ rds_iw_xmit_rdma()
837 * be enough work requests to send the entire message. rds_iw_xmit_rdma()
849 send = &ic->i_sends[pos]; rds_iw_xmit_rdma()
853 first = send; rds_iw_xmit_rdma()
861 send->s_wr.send_flags = 0; rds_iw_xmit_rdma()
862 send->s_queued = jiffies; rds_iw_xmit_rdma()
870 send->s_wr.send_flags = IB_SEND_SIGNALED; rds_iw_xmit_rdma()
878 send->s_wr.opcode = IB_WR_RDMA_WRITE; rds_iw_xmit_rdma()
880 send->s_wr.opcode = IB_WR_RDMA_READ_WITH_INV; rds_iw_xmit_rdma()
882 send->s_wr.wr.rdma.remote_addr = remote_addr; rds_iw_xmit_rdma()
883 send->s_wr.wr.rdma.rkey = op->op_rkey; rds_iw_xmit_rdma()
884 send->s_op = op; rds_iw_xmit_rdma()
887 send->s_wr.num_sge = rds_iwdev->max_sge; rds_iw_xmit_rdma()
890 send->s_wr.num_sge = num_sge; rds_iw_xmit_rdma()
892 send->s_wr.next = NULL; rds_iw_xmit_rdma()
895 prev->s_wr.next = &send->s_wr; rds_iw_xmit_rdma()
897 for (j = 0; j < send->s_wr.num_sge && scat != &op->op_sg[op->op_count]; j++) { rds_iw_xmit_rdma()
900 if (send->s_wr.opcode == IB_WR_RDMA_READ_WITH_INV) rds_iw_xmit_rdma()
901 send->s_page_list->page_list[j] = ib_sg_dma_address(ic->i_cm_id->device, scat); rds_iw_xmit_rdma()
903 send->s_sge[j].addr = ib_sg_dma_address(ic->i_cm_id->device, scat); rds_iw_xmit_rdma()
904 send->s_sge[j].length = len; rds_iw_xmit_rdma()
905 send->s_sge[j].lkey = rds_iw_local_dma_lkey(ic); rds_iw_xmit_rdma()
915 if (send->s_wr.opcode == IB_WR_RDMA_READ_WITH_INV) { rds_iw_xmit_rdma()
916 send->s_wr.num_sge = 1; rds_iw_xmit_rdma()
917 send->s_sge[0].addr = conn->c_xmit_rm->m_rs->rs_user_addr; rds_iw_xmit_rdma()
918 send->s_sge[0].length = conn->c_xmit_rm->m_rs->rs_user_bytes; rds_iw_xmit_rdma()
919 send->s_sge[0].lkey = ic->i_sends[fr_pos].s_mr->lkey; rds_iw_xmit_rdma()
922 rdsdebug("send %p wr %p num_sge %u next %p\n", send, rds_iw_xmit_rdma()
923 &send->s_wr, send->s_wr.num_sge, send->s_wr.next); rds_iw_xmit_rdma()
925 prev = send; rds_iw_xmit_rdma()
926 if (++send == &ic->i_sends[ic->i_send_ring.w_nr]) rds_iw_xmit_rdma()
927 send = ic->i_sends; rds_iw_xmit_rdma()
930 /* if we finished the message then send completion owns it */ rds_iw_xmit_rdma()
972 * to send previously (due to flow control). Try again. */ rds_iw_xmit_complete()
77 rds_iw_send_unmap_rm(struct rds_iw_connection *ic, struct rds_iw_send_work *send, int wc_status) rds_iw_send_unmap_rm() argument
448 rds_iw_xmit_populate_wr(struct rds_iw_connection *ic, struct rds_iw_send_work *send, unsigned int pos, unsigned long buffer, unsigned int length, int send_flags) rds_iw_xmit_populate_wr() argument
H A Dib_send.c183 struct rds_ib_send_work *send, rds_ib_send_unmap_op()
189 switch (send->s_wr.opcode) { rds_ib_send_unmap_op()
191 if (send->s_op) { rds_ib_send_unmap_op()
192 rm = container_of(send->s_op, struct rds_message, data); rds_ib_send_unmap_op()
193 rds_ib_send_unmap_data(ic, send->s_op, wc_status); rds_ib_send_unmap_op()
198 if (send->s_op) { rds_ib_send_unmap_op()
199 rm = container_of(send->s_op, struct rds_message, rdma); rds_ib_send_unmap_op()
200 rds_ib_send_unmap_rdma(ic, send->s_op, wc_status); rds_ib_send_unmap_op()
205 if (send->s_op) { rds_ib_send_unmap_op()
206 rm = container_of(send->s_op, struct rds_message, atomic); rds_ib_send_unmap_op()
207 rds_ib_send_unmap_atomic(ic, send->s_op, wc_status); rds_ib_send_unmap_op()
213 __func__, send->s_wr.opcode); rds_ib_send_unmap_op()
217 send->s_wr.opcode = 0xdead; rds_ib_send_unmap_op()
224 struct rds_ib_send_work *send; rds_ib_send_init_ring() local
227 for (i = 0, send = ic->i_sends; i < ic->i_send_ring.w_nr; i++, send++) { rds_ib_send_init_ring()
230 send->s_op = NULL; rds_ib_send_init_ring()
232 send->s_wr.wr_id = i; rds_ib_send_init_ring()
233 send->s_wr.sg_list = send->s_sge; rds_ib_send_init_ring()
234 send->s_wr.ex.imm_data = 0; rds_ib_send_init_ring()
236 sge = &send->s_sge[0]; rds_ib_send_init_ring()
241 send->s_sge[1].lkey = ic->i_mr->lkey; rds_ib_send_init_ring()
247 struct rds_ib_send_work *send; rds_ib_send_clear_ring() local
250 for (i = 0, send = ic->i_sends; i < ic->i_send_ring.w_nr; i++, send++) { rds_ib_send_clear_ring()
251 if (send->s_op && send->s_wr.opcode != 0xdead) rds_ib_send_clear_ring()
252 rds_ib_send_unmap_op(ic, send, IB_WC_WR_FLUSH_ERR); rds_ib_send_clear_ring()
270 * operations performed in the send path. As the sender allocs and potentially
280 struct rds_ib_send_work *send; rds_ib_send_cq_comp_handler() local
291 rdsdebug("ib_req_notify_cq send failed: %d\n", ret); rds_ib_send_cq_comp_handler()
312 send = &ic->i_sends[oldest]; rds_ib_send_cq_comp_handler()
313 if (send->s_wr.send_flags & IB_SEND_SIGNALED) rds_ib_send_cq_comp_handler()
316 rm = rds_ib_send_unmap_op(ic, send, wc.status); rds_ib_send_cq_comp_handler()
318 if (time_after(jiffies, send->s_queued + HZ/2)) rds_ib_send_cq_comp_handler()
321 if (send->s_op) { rds_ib_send_cq_comp_handler()
322 if (send->s_op == rm->m_final_op) { rds_ib_send_cq_comp_handler()
328 send->s_op = NULL; rds_ib_send_cq_comp_handler()
344 rds_ib_conn_error(conn, "send completion on %pI4 had status " rds_ib_send_cq_comp_handler()
357 * - send credits: this tells us how many WRs we're allowed
368 * exhausted their send credits, and are unable to send new credits
369 * to the peer. We achieve this by requiring that we send at least
375 * The RDS send code is essentially single-threaded; rds_send_xmit
376 * sets RDS_IN_XMIT to ensure exclusive access to the send ring.
380 * In the send path, we need to update the counters for send credits
386 * Spinlocks shared between the send and the receive path are bad,
415 /* The last credit must be used to send a credit update. */ rds_ib_send_grab_credits()
433 * the posted regardless of whether any send credits are rds_ib_send_grab_credits()
479 /* Decide whether to send an update to the peer now. rds_ib_advertise_credits()
480 * If we would send a credit update for every single buffer we rds_ib_advertise_credits()
482 * consumes buffer, we refill the ring, send ACK to remote rds_ib_advertise_credits()
485 * Performance pretty much depends on how often we send rds_ib_advertise_credits()
496 struct rds_ib_send_work *send, rds_ib_set_wr_signal_state()
506 send->s_wr.send_flags |= IB_SEND_SIGNALED; rds_ib_set_wr_signal_state()
519 * once we send the final fragment.
530 struct rds_ib_send_work *send = NULL; rds_ib_xmit() local
550 /* Do not send cong updates to IB loopback */ rds_ib_xmit()
633 * sticking the header into the send ring. Which is why we rds_ib_xmit()
658 send = &ic->i_sends[pos]; rds_ib_xmit()
659 first = send; rds_ib_xmit()
667 send->s_wr.send_flags = send_flags; rds_ib_xmit()
668 send->s_wr.opcode = IB_WR_SEND; rds_ib_xmit()
669 send->s_wr.num_sge = 1; rds_ib_xmit()
670 send->s_wr.next = NULL; rds_ib_xmit()
671 send->s_queued = jiffies; rds_ib_xmit()
672 send->s_op = NULL; rds_ib_xmit()
674 send->s_sge[0].addr = ic->i_send_hdrs_dma rds_ib_xmit()
676 send->s_sge[0].length = sizeof(struct rds_header); rds_ib_xmit()
684 send->s_wr.num_sge = 2; rds_ib_xmit()
686 send->s_sge[1].addr = ib_sg_dma_address(dev, scat) + off; rds_ib_xmit()
687 send->s_sge[1].length = len; rds_ib_xmit()
697 rds_ib_set_wr_signal_state(ic, send, 0); rds_ib_xmit()
703 send->s_wr.send_flags |= IB_SEND_SIGNALED | IB_SEND_SOLICITED; rds_ib_xmit()
705 if (send->s_wr.send_flags & IB_SEND_SIGNALED) rds_ib_xmit()
708 rdsdebug("send %p wr %p num_sge %u next %p\n", send, rds_ib_xmit()
709 &send->s_wr, send->s_wr.num_sge, send->s_wr.next); rds_ib_xmit()
722 prev->s_wr.next = &send->s_wr; rds_ib_xmit()
723 prev = send; rds_ib_xmit()
726 send = &ic->i_sends[pos]; rds_ib_xmit()
737 /* if we finished the message then send completion owns it */ rds_ib_xmit()
789 struct rds_ib_send_work *send = NULL; rds_ib_xmit_atomic() local
807 /* address of send request in ring */ rds_ib_xmit_atomic()
808 send = &ic->i_sends[pos]; rds_ib_xmit_atomic()
809 send->s_queued = jiffies; rds_ib_xmit_atomic()
812 send->s_wr.opcode = IB_WR_MASKED_ATOMIC_CMP_AND_SWP; rds_ib_xmit_atomic()
813 send->s_wr.wr.atomic.compare_add = op->op_m_cswp.compare; rds_ib_xmit_atomic()
814 send->s_wr.wr.atomic.swap = op->op_m_cswp.swap; rds_ib_xmit_atomic()
815 send->s_wr.wr.atomic.compare_add_mask = op->op_m_cswp.compare_mask; rds_ib_xmit_atomic()
816 send->s_wr.wr.atomic.swap_mask = op->op_m_cswp.swap_mask; rds_ib_xmit_atomic()
818 send->s_wr.opcode = IB_WR_MASKED_ATOMIC_FETCH_AND_ADD; rds_ib_xmit_atomic()
819 send->s_wr.wr.atomic.compare_add = op->op_m_fadd.add; rds_ib_xmit_atomic()
820 send->s_wr.wr.atomic.swap = 0; rds_ib_xmit_atomic()
821 send->s_wr.wr.atomic.compare_add_mask = op->op_m_fadd.nocarry_mask; rds_ib_xmit_atomic()
822 send->s_wr.wr.atomic.swap_mask = 0; rds_ib_xmit_atomic()
824 nr_sig = rds_ib_set_wr_signal_state(ic, send, op->op_notify); rds_ib_xmit_atomic()
825 send->s_wr.num_sge = 1; rds_ib_xmit_atomic()
826 send->s_wr.next = NULL; rds_ib_xmit_atomic()
827 send->s_wr.wr.atomic.remote_addr = op->op_remote_addr; rds_ib_xmit_atomic()
828 send->s_wr.wr.atomic.rkey = op->op_rkey; rds_ib_xmit_atomic()
829 send->s_op = op; rds_ib_xmit_atomic()
830 rds_message_addref(container_of(send->s_op, struct rds_message, atomic)); rds_ib_xmit_atomic()
843 send->s_sge[0].addr = ib_sg_dma_address(ic->i_cm_id->device, op->op_sg); rds_ib_xmit_atomic()
844 send->s_sge[0].length = ib_sg_dma_len(ic->i_cm_id->device, op->op_sg); rds_ib_xmit_atomic()
845 send->s_sge[0].lkey = ic->i_mr->lkey; rds_ib_xmit_atomic()
848 send->s_sge[0].addr, send->s_sge[0].length); rds_ib_xmit_atomic()
853 failed_wr = &send->s_wr; rds_ib_xmit_atomic()
854 ret = ib_post_send(ic->i_cm_id->qp, &send->s_wr, &failed_wr); rds_ib_xmit_atomic()
855 rdsdebug("ic %p send %p (wr %p) ret %d wr %p\n", ic, rds_ib_xmit_atomic()
856 send, &send->s_wr, ret, failed_wr); rds_ib_xmit_atomic()
857 BUG_ON(failed_wr != &send->s_wr); rds_ib_xmit_atomic()
866 if (unlikely(failed_wr != &send->s_wr)) { rds_ib_xmit_atomic()
868 BUG_ON(failed_wr != &send->s_wr); rds_ib_xmit_atomic()
878 struct rds_ib_send_work *send = NULL; rds_ib_xmit_rdma() local
912 * be enough work requests to send the entire message. rds_ib_xmit_rdma()
924 send = &ic->i_sends[pos]; rds_ib_xmit_rdma()
925 first = send; rds_ib_xmit_rdma()
932 send->s_wr.send_flags = 0; rds_ib_xmit_rdma()
933 send->s_queued = jiffies; rds_ib_xmit_rdma()
934 send->s_op = NULL; rds_ib_xmit_rdma()
936 nr_sig += rds_ib_set_wr_signal_state(ic, send, op->op_notify); rds_ib_xmit_rdma()
938 send->s_wr.opcode = op->op_write ? IB_WR_RDMA_WRITE : IB_WR_RDMA_READ; rds_ib_xmit_rdma()
939 send->s_wr.wr.rdma.remote_addr = remote_addr; rds_ib_xmit_rdma()
940 send->s_wr.wr.rdma.rkey = op->op_rkey; rds_ib_xmit_rdma()
943 send->s_wr.num_sge = max_sge; rds_ib_xmit_rdma()
946 send->s_wr.num_sge = num_sge; rds_ib_xmit_rdma()
949 send->s_wr.next = NULL; rds_ib_xmit_rdma()
952 prev->s_wr.next = &send->s_wr; rds_ib_xmit_rdma()
954 for (j = 0; j < send->s_wr.num_sge && scat != &op->op_sg[op->op_count]; j++) { rds_ib_xmit_rdma()
956 send->s_sge[j].addr = rds_ib_xmit_rdma()
958 send->s_sge[j].length = len; rds_ib_xmit_rdma()
959 send->s_sge[j].lkey = ic->i_mr->lkey; rds_ib_xmit_rdma()
968 rdsdebug("send %p wr %p num_sge %u next %p\n", send, rds_ib_xmit_rdma()
969 &send->s_wr, send->s_wr.num_sge, send->s_wr.next); rds_ib_xmit_rdma()
971 prev = send; rds_ib_xmit_rdma()
972 if (++send == &ic->i_sends[ic->i_send_ring.w_nr]) rds_ib_xmit_rdma()
973 send = ic->i_sends; rds_ib_xmit_rdma()
1018 * to send previously (due to flow control). Try again. */ rds_ib_xmit_complete()
182 rds_ib_send_unmap_op(struct rds_ib_connection *ic, struct rds_ib_send_work *send, int wc_status) rds_ib_send_unmap_op() argument
495 rds_ib_set_wr_signal_state(struct rds_ib_connection *ic, struct rds_ib_send_work *send, bool notify) rds_ib_set_wr_signal_state() argument
H A DMakefile3 recv.o send.o stats.o sysctl.o threads.o transport.o \
H A Dsend.c56 MODULE_PARM_DESC(send_batch_count, " batch factor when working the send queue");
61 * Reset the send state. Callers must ensure that this doesn't race with
92 /* Mark messages as retransmissions, and move them to the send q */ rds_send_reset()
122 * We're making the conscious trade-off here to only send one message
150 * sendmsg calls here after having queued its message on the send rds_send_xmit()
163 * we record the send generation after doing the xmit acquire. rds_send_xmit()
195 * If between sending messages, we can send a pending congestion rds_send_xmit()
238 * Move the message from the send queue to the retransmit rds_send_xmit()
291 * You can't unmap it while it's on the send queue */ rds_send_xmit()
303 * You can't unmap it while it's on the send queue */ rds_send_xmit()
312 * we skip the hdr/data send, to enable silent operation. rds_send_xmit()
398 * Other senders can queue a message after we last test the send queue rds_send_xmit()
400 * not try and send their newly queued message. We need to check the rds_send_xmit()
401 * send queue after having cleared RDS_IN_XMIT so that their message rds_send_xmit()
402 * doesn't get stuck on the send queue. rds_send_xmit()
448 * the IB send completion on the RDMA op and the accompanying
549 * This is called from the IB send completion when we detect
669 * moved to the retrans queue when rds_send_xmit picks them off the send
801 * and userspace gets -EAGAIN. But poll() indicates there's send rds_send_queue_rm()
810 /* let recv side know we are close to send space exhaustion. rds_send_queue_rm()
1095 * By now we've committed to the send. We reuse rds_send_worker() rds_sendmsg()
H A Dtcp_send.c146 printk(KERN_WARNING "RDS/tcp: send to %pI4 " rds_tcp_xmit()
200 * write_space is only called when data leaves tcp's send queue if rds_tcp_write_space()
202 * data in tcp's send queue because we use write_space to parse the rds_tcp_write_space()
206 * tcp's write_space clears SOCK_NOSPACE if the send queue has more rds_tcp_write_space()
/linux-4.1.27/net/ipv6/
H A Dip6_icmp.c35 ip6_icmp_send_t *send; icmpv6_send() local
38 send = rcu_dereference(ip6_icmp_send); icmpv6_send()
40 if (!send) icmpv6_send()
42 send(skb, type, code, info); icmpv6_send()
/linux-4.1.27/net/atm/
H A Draw.c56 return vcc->dev->ops->send(vcc, skb); atm_send_aal0()
64 vcc->send = atm_send_aal0; atm_init_aal0()
73 vcc->send = vcc->dev->ops->send; atm_init_aal34()
82 vcc->send = vcc->dev->ops->send; atm_init_aal5()
/linux-4.1.27/drivers/staging/fsl-mc/bus/
H A Ddpbp.c47 /* send command to mc*/ dpbp_open()
67 /* send command to mc*/ dpbp_close()
85 /* send command to mc*/ dpbp_create()
104 /* send command to mc*/ dpbp_destroy()
116 /* send command to mc*/ dpbp_enable()
129 /* send command to mc*/ dpbp_disable()
142 /* send command to mc*/ dpbp_is_enabled()
161 /* send command to mc*/ dpbp_reset()
182 /* send command to mc*/ dpbp_set_irq()
202 /* send command to mc*/ dpbp_get_irq()
228 /* send command to mc*/ dpbp_set_irq_enable()
245 /* send command to mc*/ dpbp_get_irq_enable()
268 /* send command to mc*/ dpbp_set_irq_mask()
285 /* send command to mc*/ dpbp_get_irq_mask()
308 /* send command to mc*/ dpbp_get_irq_status()
331 /* send command to mc*/ dpbp_clear_irq_status()
346 /* send command to mc*/ dpbp_get_attributes()
H A Ddpmcp.c47 /* send command to mc*/ dpmcp_open()
66 /* send command to mc*/ dpmcp_close()
82 /* send command to mc*/ dpmcp_create()
101 /* send command to mc*/ dpmcp_destroy()
113 /* send command to mc*/ dpmcp_reset()
134 /* send command to mc*/ dpmcp_set_irq()
154 /* send command to mc*/ dpmcp_get_irq()
180 /* send command to mc*/ dpmcp_set_irq_enable()
197 /* send command to mc*/ dpmcp_get_irq_enable()
220 /* send command to mc*/ dpmcp_set_irq_mask()
237 /* send command to mc*/ dpmcp_get_irq_mask()
260 /* send command to mc*/ dpmcp_get_irq_status()
283 /* send command to mc*/ dpmcp_clear_irq_status()
298 /* send command to mc*/ dpmcp_get_attributes()
H A Ddprc.c47 /* send command to mc*/ dprc_open()
67 /* send command to mc*/ dprc_close()
89 /* send command to mc*/ dprc_create_container()
112 /* send command to mc*/ dprc_destroy_container()
127 /* send command to mc*/ dprc_reset_container()
148 /* send command to mc*/ dprc_get_irq()
180 /* send command to mc*/ dprc_set_irq()
197 /* send command to mc*/ dprc_get_irq_enable()
221 /* send command to mc*/ dprc_set_irq_enable()
238 /* send command to mc*/ dprc_get_irq_mask()
262 /* send command to mc*/ dprc_set_irq_mask()
279 /* send command to mc*/ dprc_get_irq_status()
303 /* send command to mc*/ dprc_clear_irq_status()
319 /* send command to mc*/ dprc_get_attributes()
365 /* send command to mc*/ dprc_set_res_quota()
399 /* send command to mc*/ dprc_get_res_quota()
441 /* send command to mc*/ dprc_assign()
477 /* send command to mc*/ dprc_unassign()
492 /* send command to mc*/ dprc_get_pool_count()
517 /* send command to mc*/ dprc_get_pool()
552 /* send command to mc*/ dprc_get_obj_count()
578 /* send command to mc*/ dprc_get_obj()
642 /* send command to mc*/ dprc_get_res_count()
685 /* send command to mc*/ dprc_get_res_ids()
712 /* send command to mc*/ dprc_get_portal_paddr()
756 /* send command to mc*/ dprc_get_obj_region()
817 /* send command to mc*/ dprc_connect()
850 /* send command to mc*/ dprc_disconnect()
886 /* send command to mc*/ dprc_get_connection()
H A Ddpmng.c46 /* send command to mc*/ mc_get_version()
68 /* send command to mc*/ dpmng_get_container_id()
/linux-4.1.27/drivers/isdn/hardware/mISDN/
H A Dnetjet.c89 struct tiger_dma send; member in struct:tiger_hw
194 bc->bch.nr, fill, cnt, idx, card->send.idx); fill_mem()
201 val = card->send.start[idx]; fill_mem()
204 card->send.start[idx++] = val; fill_mem()
205 if (idx >= card->send.size) fill_mem()
221 fill_mem(bc, 0, card->send.size, 0xff); mode_tiger()
240 bc->free = card->send.size / 2; mode_tiger()
254 bc->free = card->send.size / 2; mode_tiger()
271 card->send.dmacur = inl(card->base + NJ_DMA_READ_ADR); mode_tiger()
273 card->send.idx = (card->send.dmacur - card->send.dmastart) >> 2; mode_tiger()
280 card->send.idx, mode_tiger()
326 pr_info("%s: no B%d send buffer\n", card->name, i + 1); inittiger()
337 card->send.start = card->dma_p; inittiger()
338 card->send.dmastart = (u32)card->dma; inittiger()
339 card->send.dmaend = card->send.dmastart + inittiger()
341 card->send.dmairq = card->send.dmastart + inittiger()
343 card->send.size = NJ_DMA_TXSIZE; inittiger()
346 pr_notice("%s: send buffer phy %#x - %#x - %#x virt %p" inittiger()
348 card->send.dmastart, card->send.dmairq, inittiger()
349 card->send.dmaend, card->send.start, card->send.size); inittiger()
351 outl(card->send.dmastart, card->base + NJ_DMA_READ_START); inittiger()
352 outl(card->send.dmairq, card->base + NJ_DMA_READ_IRQ); inittiger()
353 outl(card->send.dmaend, card->base + NJ_DMA_READ_END); inittiger()
480 card->send.dmacur = inl(card->base | NJ_DMA_READ_ADR); resync()
481 card->send.idx = (card->send.dmacur - card->send.dmastart) >> 2; resync()
482 if (bc->free > card->send.size / 2) resync()
483 bc->free = card->send.size / 2; resync()
488 if (card->send.idx < ((card->send.size / 2) - 1)) resync()
494 __func__, bc->bch.nr, bc->free, bc->idx, card->send.idx); resync()
511 bc->idx, card->send.idx); fill_hdlc_flag()
522 if (bc->idx >= card->send.size) fill_hdlc_flag()
524 v = card->send.start[bc->idx]; fill_hdlc_flag()
527 card->send.start[bc->idx++] = v; fill_hdlc_flag()
530 snprintf(card->log, LOG_SIZE, "B%1d-send %s %d ", fill_hdlc_flag()
550 count = card->send.size >> 1; fill_dma()
559 bc->idx, card->send.idx); fill_dma()
585 if (bc->idx >= card->send.size) fill_dma()
587 v = card->send.start[bc->idx]; fill_dma()
590 card->send.start[bc->idx++] = v; fill_dma()
594 if (bc->idx >= card->send.size) fill_dma()
596 v = card->send.start[bc->idx]; fill_dma()
600 card->send.start[bc->idx++] = v; fill_dma()
604 snprintf(card->log, LOG_SIZE, "B%1d-send %s %d ", fill_dma()
643 bc->free += card->send.size / 2; send_tiger_bc()
644 if (bc->free >= card->send.size) { send_tiger_bc()
650 bc->free = card->send.size; send_tiger_bc()
659 bc->bch.nr, bc->free, bc->idx, card->send.idx); send_tiger_bc()
662 if (bc->free == card->send.size) send_tiger_bc()
673 /* Note send is via the READ DMA channel */ send_tiger()
724 card->send.dmacur = inl(card->base | NJ_DMA_READ_ADR); nj_irq()
725 card->send.idx = (card->send.dmacur - card->send.dmastart) >> 2; nj_irq()
726 if (card->send.dmacur < card->send.dmairq) nj_irq()
733 card->recv.idx, card->send.idx); nj_irq()
1037 card->bc[i].bch.ch.send = nj_l2l1B; setup_instance()
/linux-4.1.27/drivers/isdn/hysdn/
H A Dhysdn_sched.c68 /* If the routine wants to send data it must fill buf, len and chan with the */
70 /* data to send is assumed. maxlen specifies the buffer size available for */
82 hysdn_tx_netack(card); /* acknowledge packet send */ hysdn_sched_tx()
101 return (1); /* tell that data should be send */ hysdn_sched_tx()
102 } /* error log start and able to send */ hysdn_sched_tx()
109 return (1); /* tell that data should be send */ hysdn_sched_tx()
110 } /* error log start and able to send */ hysdn_sched_tx()
121 return (1); /* go and send the data */ hysdn_sched_tx()
124 } /* send a network packet if available */ hysdn_sched_tx()
134 return (1); /* go and send the data */ hysdn_sched_tx()
138 return (0); /* nothing to send */ hysdn_sched_tx()
143 /* send one config line to the card and return 0 if successful, otherwise a */
194 hysdn_addlog(card, "async tx-cfg data send"); hysdn_tx_cfgline()
196 return (0); /* line send correctly */ hysdn_tx_cfgline()
H A Dhysdn_net.c116 /* send a packet on this interface. */
134 * layer to send no more. net_send_packet()
154 /* acknowlegde a packet send. The network layer will be informed about */
215 /* return the pointer to a network packet to be send */
228 return (lp->skbs[lp->out_idx]); /* next packet to send */ hysdn_tx_netget()
/linux-4.1.27/include/uapi/linux/
H A Dif_pppol2tp.h29 struct sockaddr_in addr; /* IP address and port to send to */
46 struct sockaddr_in6 addr; /* IP address and port to send to */
57 struct sockaddr_in addr; /* IP address and port to send to */
71 struct sockaddr_in6 addr; /* IP address and port to send to */
76 * SENDSEQ - 0 => don't send packets with sequence numbers
77 * 1 => send packets with sequence numbers
H A Dudp.h30 #define UDP_CORK 1 /* Never send partially complete segments */
H A Dneighbour.h76 * To retrieve the neighbour tables send RTM_GETNEIGHTBL with the
90 * To change neighbour table attributes, send RTM_SETNEIGHTBL
H A Dserial.h124 __u32 delay_rts_before_send; /* Delay before send (milliseconds) */
125 __u32 delay_rts_after_send; /* Delay after send (milliseconds) */
H A Dvt.h34 unsigned short v_signal; /* signal to send */
38 #define VT_SENDSIG 0x5604 /* signal to send to bitmask of vts */
H A Datmlec.h89 unsigned char receive; /* 1= receive vcc, 0 = send vcc */
H A Dhdlcdrv.h31 /* this just makes them send even if DCD is on */
H A Dipmi.h52 * send commands, receive responses, etc. The driver keeps track of
54 * will go back to the application that send the command. If the
181 * it allows you to get the message results when you send a response
190 response. When you send a
225 * In general, you send a command down to the interface and receive
230 * send. Asynchronous events, however, go to all open users, so you
240 * not, for instance, have to format a send command, you just send
242 * the send command, automatically issue receive command and get even
253 unsigned char __user *addr; /* Address to send the message to. */
/linux-4.1.27/net/batman-adv/
H A Dsend.h52 * batadv_send_skb_via_tt - send an skb via TT lookup
54 * @skb: the payload to send
60 * unicast header. Then send this frame to the according destination node.
73 * batadv_send_skb_via_tt_4addr - send an skb via TT lookup
75 * @skb: the payload to send
82 * unicast-4addr header. Then send this frame to the according destination
H A Dsend.c20 #include "send.h"
34 /* send out an already prepared packet to the given address via the
51 pr_warn("Interface %s is not up - can't send packet via that interface!\n", batadv_send_skb_packet()
112 /* Check if the skb is too large to send in one piece and fragment batadv_send_skb_to_orig()
117 /* Fragment and send packet. */ batadv_send_skb_to_orig()
126 * network coding fails, then send the packet as usual. batadv_send_skb_to_orig()
236 * batadv_send_skb_unicast - encapsulate and send an skb via unicast
238 * @skb: payload to send
242 * @orig_node: the originator to send the packet to
247 * as packet_type. Then send this frame to the given orig_node and release a
309 * batadv_send_skb_via_tt_generic - send an skb via TT lookup
311 * @skb: payload to send
321 * BATADV_UNICAST_4ADDR was supplied as packet_type. Then send this frame
338 /* if we got an hint! let's send the packet to this client (if any) */ batadv_send_skb_via_tt_generic()
350 * batadv_send_skb_via_gw - send an skb via gateway lookup
352 * @skb: payload to send
356 * unicast header and send this frame to this gateway node.
463 /* how often did we send the bcast packet ? */ batadv_add_bcast_packet_to_list()
516 /* send a copy of the saved skb */ batadv_send_outstanding_bcast_packet()
526 /* if we still have some more bcasts to send */ batadv_send_outstanding_bcast_packet()
H A Dicmp_socket.c22 #include "send.h"
165 "Error - can't send packet from char device: invalid packet size\n"); batadv_socket_write()
198 "Error - can't send packet from char device: got bogus packet type (expected: BAT_ICMP)\n"); batadv_socket_write()
207 "Error - can't send packet from char device: invalid packet size\n"); batadv_socket_write()
239 "Error - can't send packet from char device: got unknown message type\n"); batadv_socket_write()
/linux-4.1.27/net/llc/
H A Dllc_s_ac.c46 * @skb: the event to send
67 * llc_sap_action_send_xid_c - send XID PDU as response to XID REQ
69 * @skb: the event to send
90 * llc_sap_action_send_xid_r - send XID PDU resp to MAC for received XID
92 * @skb: the event to send
121 * llc_sap_action_send_test_c - send TEST PDU to MAC in resp to TEST REQ
123 * @skb: the event to send
172 * @skb: the event to send
183 * llc_sap_action_xid_ind - send XID PDU resp to net layer via XID IND
185 * @skb: the event to send
197 * llc_sap_action_test_ind - send TEST PDU to net layer via TEST IND
199 * @skb: the event to send
H A Dllc_output.c54 * @skb: packet to send
58 * Upper layers calls this function when upper layer wants to send data
63 * package primitive as an event and send to SAP event handler
H A Dllc_if.c32 * @skb: packet to send
34 * This function is called when upper layer wants to send data using
39 * LLC has send an I pdu with p bit set to 1 and is waiting for it's
/linux-4.1.27/drivers/net/wireless/ti/wl18xx/
H A Dcmd.c77 wl1271_error("failed to send channel switch command"); wl18xx_cmd_channel_switch()
105 wl1271_error("failed to send smart config start command"); wl18xx_cmd_smart_config_start()
130 wl1271_error("failed to send smart config stop command"); wl18xx_cmd_smart_config_stop()
166 wl1271_error("failed to send smart config set group key cmd"); wl18xx_cmd_smart_config_set_group_key()
198 wl1271_error("failed to send cac command"); wl18xx_cmd_set_cac()
224 wl1271_error("failed to send radar detection debug command"); wl18xx_cmd_radar_detection_debug()
250 wl1271_error("failed to send dfs master restart command"); wl18xx_cmd_dfs_master_restart()
/linux-4.1.27/drivers/media/usb/as102/
H A Das10x_cmd.c22 * as10x_cmd_turn_on - send turn on command to AS10x
42 /* send command */ as10x_cmd_turn_on()
63 * as10x_cmd_turn_off - send turn off command to AS10x
83 /* send command */ as10x_cmd_turn_off()
103 * as10x_cmd_set_tune - send set tune command to AS10x
136 /* send command */ as10x_cmd_set_tune()
158 * as10x_cmd_get_tune_status - send get tune status command to AS10x
181 /* send command */ as10x_cmd_get_tune_status()
211 * as10x_cmd_get_tps - send get TPS command to AS10x
233 /* send command */ as10x_cmd_get_tps()
269 * as10x_cmd_get_demod_stats - send get demod stats command to AS10x
292 /* send command */ as10x_cmd_get_demod_stats()
328 * as10x_cmd_get_impulse_resp - send get impulse response command to AS10x
352 /* send command */ as10x_cmd_get_impulse_resp()
H A Das10x_cmd_cfg.c50 /* send command */ as10x_cmd_get_context()
81 * as10x_cmd_set_context - send set context command to AS10x
108 /* send command */ as10x_cmd_set_context()
133 * as10x_cmd_eLNA_change_mode - send eLNA change mode command to AS10x
160 /* send command */ as10x_cmd_eLNA_change_mode()
H A Das10x_cmd_stream.c21 * as10x_cmd_add_PID_filter - send add filter command to AS10x
51 /* send command */ as10x_cmd_add_PID_filter()
102 /* send command */ as10x_cmd_del_PID_filter()
145 /* send command */ as10x_cmd_start_streaming()
188 /* send command */ as10x_cmd_stop_streaming()
H A Das102_fw.c128 /* send EOF command */ as102_firmware_upload()
136 /* prepare command to send */ as102_firmware_upload()
143 /* send cmd to device */ as102_firmware_upload()
/linux-4.1.27/drivers/misc/vmw_vmci/
H A Dvmci_route.c45 * them again when we do the actual send to ensure that we do vmci_route()
58 * cannot send it to the hypervisor. It must come vmci_route()
65 * We must be acting as a guest in order to send to vmci_route()
71 /* And we cannot send if the source is the host context. */ vmci_route()
96 * guest, then we need to send it down to the host. vmci_route()
106 * may send vmci event datagrams to the host vmci_route()
107 * itself, but it will never send datagrams to vmci_route()
194 * send it down, since we have no guest vmci_route()
203 * We must be a guest trying to send to another guest, which means vmci_route()
204 * we need to send it down to the host. We do not filter out VM to vmci_route()
/linux-4.1.27/arch/mips/include/uapi/asm/
H A Dsocket.h23 #define SO_KEEPALIVE 0x0008 /* Keep connections alive and send
38 #define SO_SNDLOWAT 0x1003 /* send low-water mark */
40 #define SO_SNDTIMEO 0x1005 /* send timeout */
H A Dtermios.h66 #define TIOCM_RTS 0x004 /* request to send */
69 #define TIOCM_CTS 0x040 /* clear to send */
/linux-4.1.27/include/linux/i2c/
H A Dpxa-i2c.h42 * 4. MA 0 (Only send stop with the ICR stop bit)
45 * 1. STOP 0 (Do not send a STOP)
46 * 0. START 0 (Do not send a START)
H A Dlm8323.h24 * Largest keycode that the chip can send, plus one,
/linux-4.1.27/drivers/gpu/drm/nouveau/include/nvkm/core/
H A Devent.h22 void (*send)(void *data, u32 size, struct nvkm_notify *); member in struct:nvkm_event_func
/linux-4.1.27/drivers/isdn/pcbit/
H A Dcallbacks.c80 * send CONN_ACTIVE_RESP
81 * send Select protocol request
187 * send CONNECT message CONNECT_ACTIVE_REQ in CAPI
239 * send disconnect resp
240 * send msg to user
269 * send disc.req
291 * Disc confirm received send BHUP
312 * send activate b-chan protocol
/linux-4.1.27/drivers/media/usb/gspca/m5602/
H A Dm5602_sensor.h63 /* Executed when the camera starts to send data */
66 /* Executed when the camera ends to send data */
/linux-4.1.27/drivers/thunderbolt/
H A Dctl.h37 TB_CFG_ERROR_ACK_PLUG_EVENT = 7, /* send as reply to TB_CFG_PKG_EVENT */
44 * If err = 1 then this is the port that send the
/linux-4.1.27/include/net/
H A Datmclip.h25 int xoff; /* 1 if send buffer is full */
27 unsigned long last_use; /* last send or receive operation */
/linux-4.1.27/arch/m32r/include/asm/opsput/
H A Dopsput_lan.h36 * ICUCR15: control register for SC send interrupt
38 * ICUCR17: control register for SIO0 send interrupt
H A Dopsput_pld.h68 * ICUCR15: control register for SC send interrupt
70 * ICUCR17: control register for SIO0 send interrupt
88 #define PLD_IRQ_SC_SND (OPSPUT_PLD_IRQ_BASE + 15) /* SC send */
90 #define PLD_IRQ_SIO0_SND (OPSPUT_PLD_IRQ_BASE + 17) /* SIO send */
/linux-4.1.27/arch/m32r/platforms/oaks32r/
H A Dsetup.c95 /* SIO0_S : uart send data */ init_IRQ()
107 /* SIO1_S : uart send data */ init_IRQ()
/linux-4.1.27/drivers/net/ethernet/intel/i40e/
H A Di40e_adminq.h88 struct i40e_adminq_ring asq; /* send queue */
89 u32 asq_cmd_timeout; /* send queue cmd write back timeout*/
91 u16 num_asq_entries; /* send queue depth */
93 u16 asq_buf_size; /* send queue buffer size */
104 /* last status values on send and receive queues */
/linux-4.1.27/drivers/net/ethernet/intel/i40evf/
H A Di40e_adminq.h88 struct i40e_adminq_ring asq; /* send queue */
89 u32 asq_cmd_timeout; /* send queue cmd write back timeout*/
91 u16 num_asq_entries; /* send queue depth */
93 u16 asq_buf_size; /* send queue buffer size */
104 /* last status values on send and receive queues */
/linux-4.1.27/drivers/usb/usbip/
H A Dusbip_common.c455 static void correct_endian_basic(struct usbip_header_basic *base, int send) correct_endian_basic() argument
457 if (send) { correct_endian_basic()
473 int send) correct_endian_cmd_submit()
475 if (send) { correct_endian_cmd_submit()
493 int send) correct_endian_ret_submit()
495 if (send) { correct_endian_ret_submit()
511 int send) correct_endian_cmd_unlink()
513 if (send) correct_endian_cmd_unlink()
520 int send) correct_endian_ret_unlink()
522 if (send) correct_endian_ret_unlink()
528 void usbip_header_correct_endian(struct usbip_header *pdu, int send) usbip_header_correct_endian() argument
532 if (send) usbip_header_correct_endian()
535 correct_endian_basic(&pdu->base, send); usbip_header_correct_endian()
537 if (!send) usbip_header_correct_endian()
542 correct_endian_cmd_submit(&pdu->u.cmd_submit, send); usbip_header_correct_endian()
545 correct_endian_ret_submit(&pdu->u.ret_submit, send); usbip_header_correct_endian()
548 correct_endian_cmd_unlink(&pdu->u.cmd_unlink, send); usbip_header_correct_endian()
551 correct_endian_ret_unlink(&pdu->u.ret_unlink, send); usbip_header_correct_endian()
562 struct usbip_iso_packet_descriptor *iso, int send) usbip_iso_packet_correct_endian()
565 if (send) { usbip_iso_packet_correct_endian()
472 correct_endian_cmd_submit(struct usbip_header_cmd_submit *pdu, int send) correct_endian_cmd_submit() argument
492 correct_endian_ret_submit(struct usbip_header_ret_submit *pdu, int send) correct_endian_ret_submit() argument
510 correct_endian_cmd_unlink(struct usbip_header_cmd_unlink *pdu, int send) correct_endian_cmd_unlink() argument
519 correct_endian_ret_unlink(struct usbip_header_ret_unlink *pdu, int send) correct_endian_ret_unlink() argument
561 usbip_iso_packet_correct_endian( struct usbip_iso_packet_descriptor *iso, int send) usbip_iso_packet_correct_endian() argument
/linux-4.1.27/drivers/isdn/hisax/
H A Dnetjet.c82 p = bcs->hw.tiger.send; fill_mem()
99 fill_mem(bcs, bcs->hw.tiger.send, mode_tiger()
102 debugl1(cs, "Tiger stat rec %d/%d send %d", mode_tiger()
126 fill_mem(bcs, bcs->hw.tiger.send, mode_tiger()
135 fill_mem(bcs, bcs->hw.tiger.send, mode_tiger()
143 bcs->hw.tiger.sendp = bcs->hw.tiger.send; mode_tiger()
636 p = bcs->hw.tiger.send - 1; netjet_fill_dma()
638 sp = bcs->hw.tiger.send - 1; netjet_fill_dma()
646 p = bcs->hw.tiger.send; netjet_fill_dma()
650 p = bcs->hw.tiger.send; netjet_fill_dma()
657 p = bcs->hw.tiger.send + 1; netjet_fill_dma()
698 p = bcs->hw.tiger.send; write_raw()
740 p = bcs->hw.tiger.send; write_raw()
782 p = cs->bcs[0].hw.tiger.send + NETJET_DMA_TXSIZE - 1; write_tiger()
784 p = cs->bcs[0].hw.tiger.send + cnt - 1; write_tiger()
915 if (!(cs->bcs[0].hw.tiger.send = kmalloc(NETJET_DMA_TXSIZE * sizeof(unsigned int), inittiger()
918 "HiSax: No memory for tiger.send\n"); inittiger()
921 cs->bcs[0].hw.tiger.s_irq = cs->bcs[0].hw.tiger.send + NETJET_DMA_TXSIZE / 2 - 1; inittiger()
922 cs->bcs[0].hw.tiger.s_end = cs->bcs[0].hw.tiger.send + NETJET_DMA_TXSIZE - 1; inittiger()
923 cs->bcs[1].hw.tiger.send = cs->bcs[0].hw.tiger.send; inittiger()
927 memset(cs->bcs[0].hw.tiger.send, 0xff, NETJET_DMA_TXSIZE * sizeof(unsigned int)); inittiger()
928 debugl1(cs, "tiger: send buf %p - %p", cs->bcs[0].hw.tiger.send, inittiger()
929 cs->bcs[0].hw.tiger.send + NETJET_DMA_TXSIZE - 1); inittiger()
930 outl(virt_to_bus(cs->bcs[0].hw.tiger.send), inittiger()
966 kfree(cs->bcs[0].hw.tiger.send); releasetiger()
967 cs->bcs[0].hw.tiger.send = NULL; releasetiger()
968 cs->bcs[1].hw.tiger.send = NULL; releasetiger()
H A Dhfc_2bds0.c149 s = bcs->hw.hfc.send[bcs->hw.hfc.f1] - bcs->hw.hfc.send[bcs->hw.hfc.f2]; GetFreeFifoBytes_B()
163 s = cs->hw.hfcD.send[cs->hw.hfcD.f1] - cs->hw.hfcD.send[cs->hw.hfcD.f2]; GetFreeFifoBytes_D()
280 bcs->hw.hfc.send[bcs->hw.hfc.f1] = ReadZReg(cs, HFCB_FIFO | HFCB_Z1 | HFCB_SEND | HFCB_CHANNEL(bcs->channel)); hfc_fill_fifo()
284 bcs->hw.hfc.send[bcs->hw.hfc.f1]); hfc_fill_fifo()
707 cs->hw.hfcD.send[cs->hw.hfcD.f1] = ReadZReg(cs, HFCD_FIFO | HFCD_Z1 | HFCD_SEND); hfc_fill_dfifo()
711 cs->hw.hfcD.send[cs->hw.hfcD.f1]); hfc_fill_dfifo()
1025 unsigned *send; init_send_hfcd() local
1027 if (!(send = kmalloc(cnt * sizeof(unsigned int), GFP_ATOMIC))) { init_send_hfcd()
1029 "HiSax: No memory for hfcd.send\n"); init_send_hfcd()
1033 send[i] = 0x1fff; init_send_hfcd()
1034 return (send); init_send_hfcd()
1041 if (!cs->hw.hfcD.send) init2bds0()
1042 cs->hw.hfcD.send = init_send_hfcd(16); init2bds0()
1043 if (!cs->bcs[0].hw.hfc.send) init2bds0()
1044 cs->bcs[0].hw.hfc.send = init_send_hfcd(32); init2bds0()
1045 if (!cs->bcs[1].hw.hfc.send) init2bds0()
1046 cs->bcs[1].hw.hfc.send = init_send_hfcd(32); init2bds0()
1059 kfree(cs->bcs[0].hw.hfc.send); release2bds0()
1060 cs->bcs[0].hw.hfc.send = NULL; release2bds0()
1061 kfree(cs->bcs[1].hw.hfc.send); release2bds0()
1062 cs->bcs[1].hw.hfc.send = NULL; release2bds0()
1063 kfree(cs->hw.hfcD.send); release2bds0()
1064 cs->hw.hfcD.send = NULL; release2bds0()
H A Dhfc_2bs0.c63 s = bcs->hw.hfc.send[bcs->hw.hfc.f1] - bcs->hw.hfc.send[bcs->hw.hfc.f2]; GetFreeFifoBytes()
261 bcs->hw.hfc.send[bcs->hw.hfc.f1] = ReadZReg(bcs, HFC_Z1 | HFC_SEND | HFC_CHANNEL(bcs->channel)); hfc_fill_fifo()
265 bcs->hw.hfc.send[bcs->hw.hfc.f1]); hfc_fill_fifo()
560 if (!(bcs->hw.hfc.send = kmalloc(32 * sizeof(unsigned int), GFP_ATOMIC))) { init_send()
562 "HiSax: No memory for hfc.send\n"); init_send()
566 bcs->hw.hfc.send[i] = 0x1fff; init_send()
586 kfree(cs->bcs[0].hw.hfc.send); releasehfc()
587 cs->bcs[0].hw.hfc.send = NULL; releasehfc()
588 kfree(cs->bcs[1].hw.hfc.send); releasehfc()
589 cs->bcs[1].hw.hfc.send = NULL; releasehfc()
/linux-4.1.27/arch/powerpc/mm/
H A Dicswx.c47 * HFI immediate send function uses icswx instruction. The immediate
48 * send function allows small (single cache-line) packets be sent
49 * without using the regular HFI send FIFO and doorbell, which are
50 * much slower than immediate send.
52 * For each task intending to use HFI immediate send, the HFI driver
58 * The HFI driver repeatedly creates immediate send packets and
59 * issues icswx instruction to send data through the HFI window.
110 * running. We need to send an IPI to force them to pick up any use_cop()
150 * running. We need to send an IPI to force them to pick up any drop_cop()
/linux-4.1.27/arch/m32r/include/asm/
H A Dm32102.h256 #define M32R_IRQ_SIO0_R (40) /* SIO0 send */
258 #define M32R_IRQ_SIO1_R (42) /* SIO1 send */
260 #define M32R_IRQ_SIO2_R (44) /* SIO2 send */
262 #define M32R_IRQ_SIO3_R (46) /* SIO3 send */
269 #define M32R_IRQ_SIO0_R (48) /* SIO0 send */
271 #define M32R_IRQ_SIO1_R (50) /* SIO1 send */
273 #define M32R_IRQ_SIO2_R (52) /* SIO2 send */
275 #define M32R_IRQ_SIO3_R (54) /* SIO3 send */
277 #define M32R_IRQ_SIO4_R (56) /* SIO4 send */
/linux-4.1.27/arch/x86/kernel/apic/
H A Dipi.c27 * Hack. The clustered APIC addressing mode doesn't allow us to send default_send_IPI_mask_sequence_phys()
67 * Hack. The clustered APIC addressing mode doesn't allow us to send default_send_IPI_mask_sequence_logical()
120 * if there are no other CPUs in the system then we get an APIC send default_send_IPI_allbutself()
/linux-4.1.27/net/rxrpc/
H A Dar-skbuff.c87 * been consumed, so send a hard-ACK if no more packets are rxrpc_hard_ACK_data()
91 _debug("send Rx idle ACK"); rxrpc_hard_ACK_data()
113 /* send the final ACK on a client call */ rxrpc_packet_destructor()
H A Dar-output.c127 * send a message forming part of a client call through an RxRPC socket
201 * rxrpc_kernel_send_data - Allow a kernel service to send data on a call
202 * @call: The call to send data through
203 * @msg: The data to send
204 * @len: The amount of data to send
206 * Allow a kernel service to send data on a call. The call must be in an state
269 * send a message through a server socket
330 * send a packet through the transport endpoint
349 /* send the packet with the don't fragment bit set if we currently rxrpc_send_packet()
353 /* send the packet by UDP rxrpc_send_packet()
371 /* attempt to send this message with fragmentation enabled */ rxrpc_send_packet()
372 _debug("send fragment"); rxrpc_send_packet()
448 * to send the packet immediately
516 * send data through a socket
634 /* add the packet to the send queue if it's now full */ rxrpc_send_data()
/linux-4.1.27/drivers/net/hyperv/
H A Dnetvsc.c112 * to send a revoke msg here netvsc_destroy_buf()
134 netdev_err(ndev, "unable to send " netvsc_destroy_buf()
168 /* Deal with the send buffer we may have setup. netvsc_destroy_buf()
169 * If we got a send section size, it means we received a netvsc_destroy_buf()
172 * to send a revoke msg here netvsc_destroy_buf()
193 netdev_err(ndev, "unable to send " netvsc_destroy_buf()
194 "revoke send buffer to netvsp\n"); netvsc_destroy_buf()
208 "unable to teardown send buffer's gpadl\n"); netvsc_destroy_buf()
214 /* Free up the send buffer */ netvsc_destroy_buf()
278 "unable to send receive buffer's gpadl to netvsp\n"); netvsc_init_buf()
322 /* Now setup the send buffer. netvsc_init_buf()
326 netdev_err(ndev, "unable to allocate send " netvsc_init_buf()
341 "unable to establish send buffer's gpadl\n"); netvsc_init_buf()
361 "unable to send send buffer's gpadl to netvsp\n"); netvsc_init_buf()
371 netdev_err(ndev, "Unable to complete send buffer " netvsc_init_buf()
391 /* Setup state for managing the send buffer. */ netvsc_init_buf()
633 /* Get the send context */ netvsc_send_completion()
663 netdev_err(ndev, "Unknown send completion packet type- " netvsc_send_completion()
813 netdev_err(ndev, "Unable to send packet %p ret %d\n", netvsc_send_pkt()
850 /* batch packets in send buffer if possible */ netvsc_send()
960 netdev_err(ndev, "unable to send receive completion pkt" netvsc_send_recv_completion()
967 netdev_err(ndev, "unable to send receive " netvsc_send_recv_completion()
972 netdev_err(ndev, "unable to send receive " netvsc_send_recv_completion()
994 * All inbound packets other than send completion should be xfer page netvsc_receive()
1069 netdev_err(ndev, "Received wrong send-table size:%u\n", count); netvsc_send_table()
H A Dhyperv_net.h134 bool cp_partial; /* partial copy into send buffer */
154 /* Points to the send/receive buffer where the ethernet frame is */
338 * This message is used by the VSC to send the NDIS version to the VSP. The VSP
347 * This message is used by the VSC to send a receive buffer to the VSP. The VSP
348 * can then use the receive buffer to send data to the VSC.
363 * This message is used by the VSP to acknowledge a receive buffer send by the
403 * This message is used by the VSC to send a send buffer to the VSP. The VSC
404 * can then use the send buffer to send data to the VSP.
412 * This message is used by the VSP to acknowledge a send buffer sent by the
420 * The VSC gets to choose the size of the send buffer and the VSP gets
429 * This message is sent by the VSC to revoke the send buffer. After the VSP
430 * completes this transaction, the vsp should never use the send buffer again.
437 * This message is used by both the VSP and the VSC to send a RNDIS message to
449 * This field is used to send part or all of the data through a send
450 * buffer. This values specifies an index into the send buffer. If the
451 * index is 0xFFFFFFFF, then the send buffer is not being used and all
560 /* The number of entries in the send indirection table */
563 /* The offset of the send indireciton table from top of this struct.
564 * The send indirection table tells which channel to put the send
676 u32 max_pkt; /* max number of pkt in one send, e.g. 8 */
/linux-4.1.27/drivers/isdn/sc/
H A Dmessage.c81 * send a message to the board
102 * Make sure we only send CEPID messages when the engine is up sendmessage()
106 pr_debug("%s: Attempt to send CM message with engine up\n", sendmessage()
112 pr_debug("%s: Attempt to send CE message with engine down\n", sendmessage()
H A Dpacket.c76 pr_debug("%s: failed to send packet, status = %d\n", sndpkt()
157 * Calculate the buffer offsets (send/recv/send/recv) setup_buffers()
179 pr_debug("%s: send buffer setup complete: first=0x%lx n=%d f=%d, nxt=%d\n", setup_buffers()
H A Dcard.h61 unsigned long first_sendbuf; /* Offset of first send buffer */
62 unsigned int num_sendbufs; /* Number of send buffers */
89 unsigned char seq_no; /* Next send seq. number */
/linux-4.1.27/drivers/infiniband/hw/cxgb3/
H A Diwch_qp.c51 wqe->send.rdmaop = T3_SEND_WITH_SE; build_rdma_send()
53 wqe->send.rdmaop = T3_SEND; build_rdma_send()
54 wqe->send.rem_stag = 0; build_rdma_send()
58 wqe->send.rdmaop = T3_SEND_WITH_SE_INV; build_rdma_send()
60 wqe->send.rdmaop = T3_SEND_WITH_INV; build_rdma_send()
61 wqe->send.rem_stag = cpu_to_be32(wr->ex.invalidate_rkey); build_rdma_send()
68 wqe->send.reserved[0] = 0; build_rdma_send()
69 wqe->send.reserved[1] = 0; build_rdma_send()
70 wqe->send.reserved[2] = 0; build_rdma_send()
77 wqe->send.sgl[i].stag = cpu_to_be32(wr->sg_list[i].lkey); build_rdma_send()
78 wqe->send.sgl[i].len = cpu_to_be32(wr->sg_list[i].length); build_rdma_send()
79 wqe->send.sgl[i].to = cpu_to_be64(wr->sg_list[i].addr); build_rdma_send()
81 wqe->send.num_sgle = cpu_to_be32(wr->num_sge); build_rdma_send()
83 wqe->send.plen = cpu_to_be32(plen); build_rdma_send()
435 wqe->send.wrid.id0.hi = qhp->wq.sq_wptr; iwch_post_send()
588 wqe->send.wrid.id0.hi = qhp->wq.sq_wptr; iwch_bind_mw()
751 printk(KERN_ERR "%s cannot send zb_read!!\n", __func__); iwch_post_zb_read()
764 wqe->send.wrh.op_seop_flags = cpu_to_be32(V_FW_RIWR_OP(T3_WR_READ)); iwch_post_zb_read()
765 wqe->send.wrh.gen_tid_len = cpu_to_be32(V_FW_RIWR_TID(ep->hwtid)| iwch_post_zb_read()
783 printk(KERN_ERR "%s cannot send TERMINATE!\n", __func__); iwch_post_terminate()
788 wqe->send.rdmaop = T3_TERMINATE; iwch_post_terminate()
791 wqe->send.plen = htonl(4); iwch_post_terminate()
794 term = (struct terminate_message *)wqe->send.sgl; iwch_post_terminate()
796 wqe->send.wrh.op_seop_flags = cpu_to_be32(V_FW_RIWR_OP(T3_WR_SEND) | iwch_post_terminate()
798 wqe->send.wrh.gen_tid_len = cpu_to_be32(V_FW_RIWR_TID(qhp->ep->hwtid)); iwch_post_terminate()
/linux-4.1.27/drivers/gpu/drm/nouveau/include/nvkm/subdev/
H A Dpmu.h11 } send; member in struct:nvkm_pmu
/linux-4.1.27/arch/mips/cavium-octeon/
H A Dcpu.c41 return NOTIFY_OK; /* Let default notifier send signals */ cnmips_cu2_call()
/linux-4.1.27/drivers/tty/
H A Dmxser.h129 /* send Xon1/Xoff1 */
131 /* send Xon2/Xoff2 */
133 /* send Xon1,Xon2/Xoff1,Xoff2 */
135 /* don't send Xon/Xoff */
H A Dn_tracesink.h25 * This header file is used by n_tracerouter to be able to send the
/linux-4.1.27/drivers/video/fbdev/mmp/hw/
H A Dmmp_spi.c3 * using the spi in LCD controler for commands send
67 /* SPI start to send command */ lcd_spi_write()
79 dev_err(&spi->dev, "spi cmd send time out\n"); lcd_spi_write()
108 * or it would cause the wrong waveform when send spi command, lcd_spi_setup()
/linux-4.1.27/include/rdma/
H A Dib_mad.h237 * @seg_count: The number of RMPP segments allocated for this send.
324 * only for snooping that occurs on a send completion.
338 * MADs received in response to a send request operation will be handed to
339 * the user before the send operation completes. All data buffers given
381 * ib_mad_send_wc - MAD send completion information.
382 * @send_buf: Send MAD data buffer associated with the send MAD request.
415 * for the corresponding send request.
445 * ib_register_mad_agent - Register to send/receive MADs.
453 * @rmpp_version: If set, indicates that the client will send
456 * @send_handler: The completion callback routine invoked after a send
490 * @send_handler: The callback routine invoked for a snooped send.
512 * ib_post_send_mad - Posts MAD(s) to the send queue of the QP associated
514 * @send_buf: Specifies the information needed to send the MAD(s).
544 * ib_cancel_mad - Cancels an outstanding send MAD operation.
555 * ib_modify_mad - Modifies an outstanding send MAD operation.
569 * @rmpp_version: If set, indicates that the client will send
572 * @send_handler: The completion callback routine invoked after a send
579 * on user-owned QPs. After calling this routine, users may send
596 * If the work completion is associated with a send operation, calling
613 * @rmpp_active: Indicates if the send will enable RMPP.
622 * This routine allocates a MAD for sending. The returned MAD send buffer
625 * MAD data buffer before posting the send.
659 * @send_buf: Previously allocated send data buffer.
668 * ib_free_send_mad - Returns data buffers used to send a MAD.
669 * @send_buf: Previously allocated send data buffer.
H A Drdma_netlink.h62 * @nlh: Header of the netlink message to send
72 * @nlh: Header of the netlink message to send
H A Diw_portmap.h87 * @pm_msg: Contains driver info to send to the userspace port mapper
95 * @pm_msg: Contains the local ip/tcp address info to send
106 * @pm_msg: Contains the local and remote ip/tcp address info to send
174 * Using the received port mapper pid, send all the local mapping
/linux-4.1.27/arch/powerpc/boot/
H A Dmv64x60_i2c.c107 /* send reset */ mv64x60_i2c_read()
117 /* send start */ mv64x60_i2c_read()
130 /* send offset of data */ mv64x60_i2c_read()
174 /* send stop */ mv64x60_i2c_read()
/linux-4.1.27/arch/arm/mach-sa1100/include/mach/
H A Duncompress.h40 /* send the character out. */ putc()
/linux-4.1.27/sound/usb/line6/
H A Dmidi.h31 /* Number of currently active MIDI send URBs */
/linux-4.1.27/drivers/media/usb/dvb-usb/
H A Dm920x.h46 always send the 7-bit slave I2C address as the 7 MSB, followed by
55 the master should send an ACK, that is pull SDA low during the 9th
/linux-4.1.27/drivers/usb/serial/
H A Dcypress_m8.h66 #define CONTROL_RTS 0x10 /* request to send */
73 #define UART_CTS 0x10 /* clear to send */
/linux-4.1.27/drivers/video/fbdev/omap/
H A Dlcd_h3.c44 /* GPIO1 and GPIO2 of TPS65010 send LCD_ENBKL and LCD_ENVDD signals */ h3_panel_enable()
58 /* GPIO1 and GPIO2 of TPS65010 send LCD_ENBKL and LCD_ENVDD signals */ h3_panel_disable()
/linux-4.1.27/drivers/platform/x86/
H A Dasus-wmi.h47 * For machines with AMD graphic chips, it will send out WMI event
50 * and let the ACPI interrupt to send out the key event.
H A Dintel_mid_powerbtn.c97 * SCU firmware might send power button interrupts to IA core before mfld_pb_probe()
99 * MSIC reg 0x21 is kept masked, and SCU firmware doesn't send new mfld_pb_probe()
/linux-4.1.27/include/scsi/
H A Dscsi_devinfo.h20 #define BLIST_MS_SKIP_PAGE_08 0x2000 /* do not send ms page 0x08 */
21 #define BLIST_MS_SKIP_PAGE_3F 0x4000 /* do not send ms page 0x3f */
/linux-4.1.27/arch/powerpc/platforms/pseries/
H A Dhvconsole.c61 * hvc_put_chars: send characters to firmware for denoted vterm adapter
64 * @buf: The character buffer that contains the character data to send to
/linux-4.1.27/samples/rpmsg/
H A Drpmsg_client_sample.c44 /* send a new message now */ rpmsg_sample_cb()
57 /* send a message to our remote processor */ rpmsg_sample_probe()
/linux-4.1.27/include/linux/isdn/
H A Dhdlc.h48 /* set if D channel (send idle instead of flags) */
52 /* set if in closing phase (need to send CRC + flag) */
/linux-4.1.27/include/net/irda/
H A Dirttp.h59 * If the peer has less credits than 9 frames, we will explicitly send
63 * has something to send while waiting for credits (one LAP window == 7
67 * default maximum number of frames he can send us before needing flow
69 * We want to minimise the number of times we have to explicitly send some
71 * particular, it doesn't make sense for us to send credit more than once
143 unsigned long disconnect_pend; /* Disconnect, but still data to send */
H A Dirlap.h65 * 2 buffers is the minimum we can work with, one that we send while polling
66 * IrTTP, and another to know that we should not send the pf bit.
161 /* To send a faster RR if tx queue empty */
178 int window; /* Nr of I-frames allowed to send */
182 __u32 line_capacity; /* Number of bytes allowed to send */
/linux-4.1.27/fs/cifs/
H A Dtransport.c125 * smb_send_kvec - send an array of kvecs to the server
126 * @server: Server to send the data to
131 * Our basic "send data to server" function. Should be called with srv_mutex
163 * If blocking send, we try 3 times, since each can block smb_send_kvec()
167 * case to send on the socket is about 15 seconds. smb_send_kvec()
169 * the server in SendReceive[2] for the server to send smb_send_kvec()
176 * send a packet. In most cases if we fail to send smb_send_kvec()
197 /* send was at least partially successful */ smb_send_kvec()
235 i = 0; /* in case we get ENOSPC on the next send */ smb_send_kvec()
266 /* if last page, don't send beyond this offset into page */ cifs_rqst_page_to_kvec()
309 /* sanity check send length */ smb_send_rqst()
330 /* now walk the page array and send each page in it */ smb_send_rqst()
350 cifs_dbg(FYI, "partial send (wanted=%u sent=%zu): terminating session\n", smb_send_rqst()
445 /* Since an echo is already inflight, no need to wait to send another */ wait_for_free_request()
737 * Ensure that we do not send more than 50 overlapping requests SendReceive2()
749 * Make sure that we sign in the same order that we send on this socket SendReceive2()
856 /* Ensure that we do not send more than 50 overlapping requests SendReceive()
871 /* make sure that we sign in the same order that we send on this socket SendReceive()
946 /* We send a LOCKINGX_CANCEL_LOCK to cause the Windows
995 /* Ensure that we do not send more than 50 overlapping requests SendReceiveBlockingLock()
1010 /* make sure that we sign in the same order that we send on this socket SendReceiveBlockingLock()
1058 /* POSIX lock. We send a NT_CANCEL SMB to cause the SendReceiveBlockingLock()
1066 /* Windows lock. We send a LOCKINGX_CANCEL_LOCK SendReceiveBlockingLock()
H A Dsmb2glob.h61 /* Maximum buffer size value we can send with 1 credit */
/linux-4.1.27/drivers/net/wireless/iwlwifi/dvm/
H A Drxon.c100 struct iwl_rxon_cmd *send) iwlagn_disable_bss()
102 __le32 old_filter = send->filter_flags; iwlagn_disable_bss()
105 send->filter_flags &= ~RXON_FILTER_ASSOC_MSK; iwlagn_disable_bss()
107 0, sizeof(*send), send); iwlagn_disable_bss()
109 send->filter_flags = old_filter; iwlagn_disable_bss()
120 struct iwl_rxon_cmd *send) iwlagn_disable_pan()
123 __le32 old_filter = send->filter_flags; iwlagn_disable_pan()
124 u8 old_dev_type = send->dev_type; iwlagn_disable_pan()
134 send->filter_flags &= ~RXON_FILTER_ASSOC_MSK; iwlagn_disable_pan()
135 send->dev_type = RXON_DEV_TYPE_P2P; iwlagn_disable_pan()
137 0, sizeof(*send), send); iwlagn_disable_pan()
139 send->filter_flags = old_filter; iwlagn_disable_pan()
140 send->dev_type = old_dev_type; iwlagn_disable_pan()
157 struct iwl_rxon_cmd *send) iwlagn_disconn_pan()
159 __le32 old_filter = send->filter_flags; iwlagn_disconn_pan()
162 send->filter_flags &= ~RXON_FILTER_ASSOC_MSK; iwlagn_disconn_pan()
164 sizeof(*send), send); iwlagn_disconn_pan()
166 send->filter_flags = old_filter; iwlagn_disconn_pan()
188 IWL_DEBUG_INFO(priv, "send QoS cmd with Qos active=%d FLAGS=0x%X\n", iwlagn_update_qos()
374 IWL_ERR(priv, "Failed to send timing (%d)!\n", ret); iwlagn_rxon_disconn()
469 IWL_ERR(priv, "Failed to send timing (%d)!\n", ret); iwlagn_rxon_connect()
478 * enabled, but then we also need to send the beacon iwlagn_rxon_connect()
514 * we must send a new TXPOWER command or we won't be able to iwlagn_rxon_connect()
1039 * 1. make sure send RXON command with association bit unset if not connect
1043 * 3. RXON_timing has to send before RXON for connection
1103 * If we don't need to send a full RXON, we can use iwlagn_commit_rxon()
1401 "Could not send REPLY_PHY_CALIBRATION_CMD\n"); iwlagn_chain_noise_reset()
98 iwlagn_disable_bss(struct iwl_priv *priv, struct iwl_rxon_context *ctx, struct iwl_rxon_cmd *send) iwlagn_disable_bss() argument
118 iwlagn_disable_pan(struct iwl_priv *priv, struct iwl_rxon_context *ctx, struct iwl_rxon_cmd *send) iwlagn_disable_pan() argument
155 iwlagn_disconn_pan(struct iwl_priv *priv, struct iwl_rxon_context *ctx, struct iwl_rxon_cmd *send) iwlagn_disconn_pan() argument
/linux-4.1.27/drivers/char/tpm/st33zp24/
H A Dst33zp24.c94 tpm_dev->ops->send(tpm_dev->phy_id, TPM_INT_STATUS, &interrupt, 1); clear_interruption()
111 tpm_dev->ops->send(tpm_dev->phy_id, TPM_STS, &data, 1); st33zp24_cancel()
170 ret = tpm_dev->ops->send(tpm_dev->phy_id, TPM_ACCESS, &data, 1); request_locality()
199 tpm_dev->ops->send(tpm_dev->phy_id, TPM_ACCESS, &data, 1); release_locality()
385 * st33zp24_send send TPM commands through the I2C bus.
388 * @param: buf, the buffer to send.
389 * @param: count, the number of bytes to send.
429 ret = tpm_dev->ops->send(tpm_dev->phy_id, TPM_DATA_FIFO, st33zp24_send()
443 ret = tpm_dev->ops->send(tpm_dev->phy_id, TPM_DATA_FIFO, st33zp24_send()
455 ret = tpm_dev->ops->send(tpm_dev->phy_id, TPM_STS, &data, 1); st33zp24_send()
480 * @param: count, the number of bytes to send.
535 .send = st33zp24_send,
603 ret = tpm_dev->ops->send(tpm_dev->phy_id, TPM_INT_ENABLE, st33zp24_probe()
609 ret = tpm_dev->ops->send(tpm_dev->phy_id, (TPM_INT_ENABLE + 3), st33zp24_probe()
H A Dst33zp24.h25 int (*send)(void *phy_id, u8 tpm_register, u8 *tpm_data, int tpm_size); member in struct:st33zp24_phy_ops
/linux-4.1.27/drivers/net/fddi/skfp/h/
H A Dhwmtm.h65 #define LOC_TX 0x04 /* send frame to the local SMT */
68 #define LAN_TX 0x20 /* send frame to network if set */
69 #define RING_DOWN 0x40 /* error: unable to send, ring down */
236 * para queue the number of the send queue: Can be specified by
239 * return number of used TxDs for this send queue
254 * para queue the number of the send queue: Can be specified by
391 #define HWM_E0010_MSG "HWM: A protocol layer has tried to send a frame with an invalid frame control"
/linux-4.1.27/drivers/net/wireless/iwlwifi/mvm/
H A Dquota.c186 bool send = false; iwl_mvm_update_quotas() local
300 send = true; iwl_mvm_update_quotas()
302 send = true; iwl_mvm_update_quotas()
306 send = true; iwl_mvm_update_quotas()
313 if (!send && !force_update) { iwl_mvm_update_quotas()
314 /* don't send a practically unchanged command, the firmware has iwl_mvm_update_quotas()
324 IWL_ERR(mvm, "Failed to send quota: %d\n", err); iwl_mvm_update_quotas()
/linux-4.1.27/drivers/staging/comedi/drivers/
H A Dcb_pcidda.c77 #define SELECT_EEPROM_BIT 0x1 /* send serial data in to eeprom */
78 /* don't send serial data to MAX542 reference dac */
80 /* don't send serial data to caldac n */
186 /* send bits most significant bit first */ cb_pcidda_serial_out()
205 /* bits to send to tell eeprom we want to read */ cb_pcidda_read_eeprom()
210 /* send serial output stream to eeprom */ cb_pcidda_read_eeprom()
219 /* send address we want to read from */ cb_pcidda_read_eeprom()
/linux-4.1.27/arch/cris/arch-v32/drivers/
H A Di2c.c405 * send slave address i2c_write()
414 * send data i2c_write()
464 * send slave address i2c_read()
519 * send slave address i2c_writereg()
538 * send register register data i2c_writereg()
583 * send slave address i2c_readreg()
607 * send slave address i2c_readreg()
/linux-4.1.27/net/irda/
H A Dirlap_event.c163 * Poll timer has expired. Normally we must now send a RR frame to the
177 * Calculate and set time before we will have to send back the pf bit
233 * then send queued data frames.
258 * try to disconnect link if we send any data frames, since irlap_do_event()
269 * We send frames up to when we fill the window or irlap_do_event()
272 * send them. irlap_do_event()
273 * After each frame that we send, we poll the higher irlap_do_event()
276 * and then send the first frame, so we can afford irlap_do_event()
277 * to send a bit of time in kernel space. irlap_do_event()
283 /* Try to send away all queued data frames */ irlap_do_event()
463 /* We don't send the frame, just post an event. irlap_state_ndm()
534 * will only be used to send out the same info as the cmd irlap_state_ndm()
665 * are waiting for the right time slot to send a response XID frame
696 /* If it's our slot, send our reply */ irlap_state_reply()
768 * We are allowed to send two frames, but this may increase irlap_state_conn()
781 * immediately send a RR that will likely collide with the irlap_state_conn()
986 * Only send frame if send-window > 0. irlap_state_xmit_p()
995 * At 115k, we can send only 2 packets of 2048 bytes irlap_state_xmit_p()
998 * option, if we send smaller packets, we can send irlap_state_xmit_p()
1042 * we send 2000B packets, we may wait another irlap_state_xmit_p()
1062 /* More packet to send in current window */ irlap_state_xmit_p()
1070 * Make sure state machine does not try to send irlap_state_xmit_p()
1080 pr_debug("%s(), Unable to send! remote busy?\n", irlap_state_xmit_p()
1114 /* Nothing to do, irlap_do_event() will send the packet irlap_state_xmit_p()
1201 * and send an rr:cmd immediately. This happens before irlap_state_nrm_p()
1255 * to give them a chance to send data in the irlap_state_nrm_p()
1278 /* Unexpected next to send (Ns) */ irlap_state_nrm_p()
1356 * Unexpected next to send (Ns) and next to receive (Nr) irlap_state_nrm_p()
1751 * Send frame only if send window > 0 irlap_state_xmit_s()
1820 * Make sure state machine does not try to send irlap_state_xmit_s()
1826 pr_debug("%s(), Unable to send!\n", __func__); irlap_state_xmit_s()
1838 /* Nothing to do, irlap_do_event() will send the packet irlap_state_xmit_s()
1922 * we decide if we should send a RR frame irlap_state_nrm_s()
1948 * Check for Unexpected next to send (Ns) irlap_state_nrm_s()
1952 /* Unexpected next to send, with final bit cleared */ irlap_state_nrm_s()
2088 /* Just send back pf bit */ irlap_state_nrm_s()
/linux-4.1.27/drivers/infiniband/hw/qib/
H A Dqib_tx.c129 * Return true if send buffer is being used by a user context.
165 * Disarm a set of send buffers. If the buffer might be actively being
222 * called whenever our local copy indicates we have run out of send buffers
238 * critical send path, but that's not really possible, given the update_send_bufs()
279 * Common code for normal driver send buffer allocation, and reserved
379 * qib_chg_pioavailkernel - change which send buffers are available for kernel
381 * @start: the starting send buffer number
382 * @len: the number of send buffers
392 /* There are two bits per send buffer (busy and generation) */ qib_chg_pioavailkernel()
449 * Flush all sends that might be in the ready to send state, as well as any
451 * sure the send side is idle. Cleans up all buffer state by canceling
455 * PIOAvail bits are updated by the chip as if a normal send had happened.
522 * at INIT when we might be trying to send SMI packets. qib_hol_down()
H A Dqib_verbs.h344 u32 ssn; /* send sequence number */
386 /* send signal when number of RWQEs < limit */
398 * This structure holds the information that the send tasklet needs
399 * to send a RDMA read response or atomic operation.
426 struct qib_swqe *s_wq; /* send work queue */
428 struct qib_ib_header *s_hdr; /* next packet header to send */
435 u32 s_size; /* send work queue size */
448 u8 s_max_rd_atomic; /* max number of RDMA read/atomic to send */
488 struct qib_sge_state s_sge; /* current send request data */
491 u32 s_cur_size; /* size of send packet in bytes */
549 * QIB_S_SIGNAL_REQ_WR - set if QP send WRs contain completion signaled
550 * QIB_S_BUSY - send tasklet is processing the QP
559 * QIB_S_WAIT_DMA - waiting for send DMA queue to drain before generating
560 * next send completion entry not via send DMA
561 * QIB_S_WAIT_PIO - waiting for a send buffer to be available
565 * QIB_S_WAIT_PSN - waiting for a packet to exit the send DMA queue
567 * QIB_S_SEND_ONE - send one packet, request ACK, then wait for ACK
595 * Wait flags that would prevent send work requests from making progress.
845 * a RC response is pending or we can process send work requests.
H A Dqib_rc.c86 /* Don't send an ACK if we aren't supposed to. */ qib_make_rc_ack()
311 /* Check if send work queue is empty. */ qib_make_rc_req()
641 * qib_send_rc_ack - Construct an ACK packet and send it
646 * send side QP state and tasklet.
669 /* Don't send ACK or NAK if a RDMA read or atomic is pending. */ qib_send_rc_ack()
706 /* Don't try to send ACKs if the link isn't ACTIVE */ qib_send_rc_ack()
720 * send tasklet so that when a PIO buffer becomes qib_send_rc_ack()
765 /* Schedule the send tasklet. */ qib_send_rc_ack()
775 * reset_psn - reset the QP state to send starting from PSN
793 * let the normal send code handle initialization. reset_psn()
816 * let the normal send code handle initialization. reset_psn()
856 * asynchronously before the send tasklet can get scheduled. reset_psn()
1017 /* Post a send completion queue entry if requested. */ qib_rc_send_complete()
1063 * completion if the SWQE is being resent until the send do_rc_completion()
1073 /* Post a send completion queue entry if requested. */ do_rc_completion()
1423 * If send tasklet not running attempt to progress qib_rc_rcv_resp()
1648 * Wait to send the sequence NAK until all packets qib_rc_rcv_error()
1669 * send the earliest so that RDMA reads can be restarted at qib_rc_rcv_error()
1761 * or the send tasklet is already backed up to send an OP()
1788 * Try to send a simple ACK to work around a Mellanox bug
1896 * packet sequence number will be for something in the send work qib_rc_rcv()
1942 * Note that it is up to the requester to not send a new qib_rc_rcv()
2157 /* Schedule the send tasklet. */ OP()
2221 /* Schedule the send tasklet. */ OP()
/linux-4.1.27/drivers/net/phy/
H A Dmdio-bitbang.c94 /* Utility to send the preamble, address, and
116 /* send the start bit (01) and the read opcode (10) or write (10). mdiobb_cmd()
143 /* send the turnaround (10) */ mdiobb_cmd_addr()
171 * may be trying to send. mdiobb_read()
194 /* send the turnaround (10) */ mdiobb_write()
/linux-4.1.27/arch/powerpc/platforms/pasemi/
H A Dgpio_mdio.c97 /* Utility to send the preamble, address, and register (common to read and write). */ bitbang_pre()
108 /* send the start bit (01) and the read opcode (10) or write (10) */ bitbang_pre()
115 /* send the PHY address */ bitbang_pre()
121 /* send the register address */ bitbang_pre()
177 /* send the turnaround (10) */ gpio_mdio_write()
/linux-4.1.27/drivers/mfd/
H A Dsi476x-cmd.c288 /* First send the command and its arguments */ si476x_core_send_command()
302 /* Set CTS to zero only after the command is send to avoid si476x_core_send_command()
402 * si476x_cmd_func_info() - send 'FUNC_INFO' command to the device
403 * @core: device to send the command to
436 * si476x_cmd_set_property() - send 'SET_PROPERTY' command to the device
437 * @core: device to send the command to
464 * si476x_cmd_get_property() - send 'GET_PROPERTY' command to the device
465 * @core: device to send the command to
493 * si476x_cmd_dig_audio_pin_cfg() - send 'DIG_AUDIO_PIN_CFG' command to
495 * @core: device to send the command to
551 * si476x_cmd_zif_pin_cfg - send 'ZIF_PIN_CFG_COMMAND'
552 * @core - device to send the command to
600 * si476x_cmd_ic_link_gpo_ctl_pin_cfg - send
602 * @core - device to send the command to
657 * si476x_cmd_ana_audio_pin_cfg - send 'ANA_AUDIO_PIN_CFG' to the
659 * @core - device to send the command to
686 * si476x_cmd_intb_pin_cfg - send 'INTB_PIN_CFG' command to the device
687 * @core - device to send the command to
738 * si476x_cmd_am_rsq_status - send 'AM_RSQ_STATUS' command to the
740 * @core - device to send the command to
872 * si476x_cmd_fm_seek_start - send 'FM_SEEK_START' command to the
874 * @core - device to send the command to
900 * si476x_cmd_fm_rds_status - send 'FM_RDS_STATUS' command to the
902 * @core - device to send the command to
1046 * si476x_cmd_am_seek_start - send 'FM_SEEK_START' command to the
1048 * @core - device to send the command to
/linux-4.1.27/drivers/isdn/mISDN/
H A Ddsp_core.c129 * receive data ist disabled, the card may not send/receive any data at all.
590 /* send indication if it worked to set it */ dsp_control_req()
595 if (dsp->up->send(dsp->up, nskb)) dsp_control_req()
710 return dsp->up->send(dsp->up, skb); dsp_function()
739 /* send dtmf result, if any */ dsp_function()
754 if (dsp->up->send( dsp_function()
769 return dsp->up->send(dsp->up, skb); dsp_function()
800 if (dsp->up->send( dsp_function()
853 /* send activation to upper layer */ dsp_function()
856 return dsp->up->send(dsp->up, skb); dsp_function()
872 return dsp->up->send(dsp->up, skb); dsp_function()
894 /* send data to tx-buffer (if no tone is played) */ dsp_function()
917 /* send ph_activate */ dsp_function()
971 * must lock here, or we may hit send-process currently dsp_ctrl()
1016 /* send queued data */ dsp_send_bh()
1029 /* send packet up */ dsp_send_bh()
1031 if (dsp->up->send(dsp->up, skb)) dsp_send_bh()
1036 /* send packet down */ dsp_send_bh()
1069 ndsp->ch.send = dsp_function; dspcreate()
H A Dstack.c108 ret = ch->send(ch, cskb); send_layer2()
127 ret = ch->send(ch, skb); send_layer2()
163 return st->layer1->send(st->layer1, skb); send_msg_to_layer()
172 return ch->send(ch, skb); send_msg_to_layer()
182 return ch->send(ch, skb); send_msg_to_layer()
248 "send call(%d)\n", mISDNStackd()
404 newst->own.send = mISDN_queue_message; create_stack()
483 ch->recv = rq.ch->send; connect_Bstack()
485 rq.ch->recv = ch->send; connect_Bstack()
498 ch->recv = rq2.ch->send; connect_Bstack()
508 rq2.ch->recv = rq.ch->send; connect_Bstack()
510 rq.ch->recv = rq2.ch->send; connect_Bstack()
/linux-4.1.27/drivers/scsi/
H A Dmesh.h57 #define SEQ_COMMAND 3 /* send a command */
59 #define SEQ_DATAOUT 5 /* send data */
61 #define SEQ_MSGOUT 7 /* send a message */
H A Discsi_tcp.h31 /* Socket connection send helper */
/linux-4.1.27/drivers/staging/lustre/lnet/klnds/socklnd/
H A Dsocklnd_modparams.c69 MODULE_PARM_DESC(eager_ack, "send tcp ack packets eagerly");
98 MODULE_PARM_DESC(keepalive, "# seconds before send keepalive");
123 MODULE_PARM_DESC(nonblk_zcack, "always send ZC-ACK on non-blocking connection");
/linux-4.1.27/include/linux/platform_data/
H A Dbfin_rotary.h27 * bfin-rotary to send EV_KEY otherwise set 0
31 * bfin-rotary to send EV_KEY otherwise set 0
37 * bfin-rotary to send EV_REL otherwise set 0
/linux-4.1.27/arch/mips/sgi-ip27/
H A Dip27-nmi.c202 * If a cpu has not responded after 10 sec, send it 1 additional NMI. cont_nmi_dump()
205 * - on 512p SN0 system, the MMSC will only send NMIs to cont_nmi_dump()
210 * send NMIs to all cpus on a 256p system. cont_nmi_dump()
/linux-4.1.27/net/dccp/ccids/
H A Dccid3.h86 * @tx_t_ipi: Interpacket (send) interval (RFC 3448, 4.6) in usecs
93 * @tx_t_nom: Nominal send time of next packet
132 * @rx_x_recv: Receiver estimate of send rate (RFC 3448, sec. 4.3)
/linux-4.1.27/net/dccp/
H A Doutput.c30 /* enqueue @skb on sk_send_head for retransmission, return clone to send now */ dccp_skb_entail()
214 * dccp_wait_for_ccid - Await CCID send permission
218 * This is used by CCIDs which need to delay the send time in process context.
319 * If the CCID determines when to send, the next sending dccp_flush_write_queue()
320 * time is unknown or the CCID may not even send again dccp_flush_write_queue()
323 DCCP_WARN("CCID did not manage to send all packets\n"); dccp_flush_write_queue()
333 /* check again if we can send now */ dccp_flush_write_queue()
467 /* Swap the send and the receive. */ dccp_ctl_make_reset()
502 /* send Reset on established socket, to close or abort the connection */ dccp_send_reset()
572 /* If we have been reset, we may not send again. */ dccp_send_ack()
610 * send ACK now.
640 DCCP_CRIT("could not send %s", dccp_packet_name(pkt_type)); dccp_send_sync()
/linux-4.1.27/sound/core/seq/
H A Dseq_system.c34 * - send tempo /start/stop etc. events to this port to manipulate the
40 * The modifier client/port is not send.
45 * detaching from the system an announcement is send to the subscribed
/linux-4.1.27/tools/usb/usbip/src/
H A Dusbip_attach.c130 /* send a request */ query_import_device()
133 err("send op_common"); query_import_device()
143 err("send op_import_request"); query_import_device()
/linux-4.1.27/include/linux/
H A Dhdlcdrv.h124 * 0 = send flags
125 * 1 = send txtail (flags)
126 * 2 = send packet
H A Dtty_ldisc.h82 * This function is called by the low-level tty driver to send
93 * that line discpline should try to send more characters to the
95 * not have any more data to send, it can just return. If the line
96 * discipline does have some data to send, please arise a tasklet
97 * or workqueue to do the real data transfer. Do not send data in
120 * This function is called by the low-level tty driver to send
H A Dppp_channel.h7 * A PPP channel provides a way for the generic PPP code to send
49 /* Called by the channel when it can send some more data. */
H A Dtfrc.h34 * @tfrctx_x_recv: receiver estimate of send rate (4.3)
H A Dadb.h43 #define ADBREQ_NOSEND 4 /* build the request, but don't send it */
H A Drpmsg.h181 * rpmsg_send() - send a message across to the remote processor
205 * rpmsg_sendto() - send a message across to the remote processor, specify dst
231 * rpmsg_send_offchannel() - send a message using explicit src/dst addresses
258 * rpmsg_send() - send a message across to the remote processor
282 * rpmsg_sendto() - send a message across to the remote processor, specify dst
307 * rpmsg_send_offchannel() - send a message using explicit src/dst addresses
/linux-4.1.27/drivers/net/wimax/i2400m/
H A Dusb-tx.c51 * For sending we just obtain a FIFO buffer to send, send it to the
79 * Get the next TX message in the TX FIFO and send it to the device
154 dev_err(dev, "TX: cannot send URB; retrying. " i2400mu_tx()
169 * Get the next TX message in the TX FIFO and send it to the device
226 * it will see there is data in the FIFO and send it. Else, just
/linux-4.1.27/drivers/tty/ipwireless/
H A Dnetwork.c88 * Called by the ppp system when it has a packet to send to the hardware.
144 * needs to be unblocked once we are ready to send. ipwireless_ppp_start_xmit()
327 * when we're online), then send the data to that tty. The RAS ipwireless_network_notify_control_line_change()
379 * when we're online), then send the data to that tty. The RAS ipwireless_network_packet_received()
389 * the modem tty, and we are online, then we send it to ipwireless_network_packet_received()
409 /* Otherwise we send it out the tty. */ ipwireless_network_packet_received()
/linux-4.1.27/drivers/isdn/isdnloop/
H A Disdnloop.h59 #define ISDNLOOP_MAX_SQUEUE 65536 /* Max. outstanding send-data */
93 int sndcount[ISDNLOOP_BCH]; /* Byte-counters for B-Ch.-send */
/linux-4.1.27/drivers/net/ethernet/intel/igb/
H A De1000_mbx.h29 #define E1000_P2VMAILBOX_STS 0x00000001 /* Initiate message send to VF */
50 /* Indicates that VF is still clear to send requests */
/linux-4.1.27/drivers/misc/ibmasm/
H A Ddot_command.c66 * send the 4.3.5.10 dot command (driver VPD) to the service processor
117 * send the 4.3.6 dot command (os state) to the service processor
H A Dmodule.c33 * to send periodic heartbeats to the OS.
45 * 5) Allows user space applications to send heartbeats to the service
131 dev_err(sp->dev, "Failed to send driver VPD to service processor\n"); ibmasm_init_one()
136 dev_err(sp->dev, "Failed to send OS state to service processor\n"); ibmasm_init_one()
/linux-4.1.27/drivers/scsi/bnx2i/
H A Dbnx2i_sysfs.c31 * bnx2i_show_sq_info - return(s currently configured send queue (SQ) size
48 * bnx2i_set_sq_info - update send queue (SQ) size parameter
/linux-4.1.27/drivers/staging/wlan-ng/
H A Dp80211ioctl.h55 * via ioctl is that we're reserving the right to be able to send
57 * if we ever need to send request messages when there aren't any
/linux-4.1.27/drivers/nfc/nfcmrvl/
H A Dmain.c59 nfc_info(priv->dev, "send entry, len %d\n", skb->len); nfcmrvl_nci_send()
86 .send = nfcmrvl_nci_send,
/linux-4.1.27/arch/x86/platform/iris/
H A Diris.c36 #define IRIS_GIO_PULSE 0x80 /* First byte to send */
37 #define IRIS_GIO_REST 0x00 /* Second byte to send */
/linux-4.1.27/drivers/gpu/drm/nouveau/nvkm/core/
H A Devent.c64 if (event->func->send) { nvkm_event_send()
65 event->func->send(data, size, notify); nvkm_event_send()
/linux-4.1.27/fs/quota/
H A Dnetlink.c40 * dquot) to send a message to userspace relating to quota limits.
60 "VFS: Not enough memory to send quota warning.\n"); quota_send_warning()
/linux-4.1.27/include/scsi/fc/
H A Dfc_fcoe.h53 __u8 fcoe_resvd[12]; /* reserved - send zero and ignore */
66 __u8 fcoe_resvd[3]; /* reserved - send zero and ignore */
/linux-4.1.27/arch/mn10300/kernel/
H A Dgdb-io-serial.c148 * send a character to the debugger
167 * send a character to the debugger
/linux-4.1.27/arch/m32r/include/asm/m32700ut/
H A Dm32700ut_lan.h36 * ICUCR15: control register for SC send interrupt
38 * ICUCR17: control register for SIO0 send interrupt
H A Dm32700ut_pld.h72 * ICUCR15: control register for SC send interrupt
74 * ICUCR17: control register for SIO0 send interrupt
92 #define PLD_IRQ_SC_SND (M32700UT_PLD_IRQ_BASE + 15) /* SC send */
94 #define PLD_IRQ_SIO0_SND (M32700UT_PLD_IRQ_BASE + 17) /* SIO send */
/linux-4.1.27/arch/m32r/platforms/mappi/
H A Dsetup.c96 /* SIO0_S : uart send data */ init_IRQ()
108 /* SIO1_S : uart send data */ init_IRQ()
/linux-4.1.27/arch/m32r/platforms/mappi2/
H A Dsetup.c97 /* SIO0_S : uart send data */ init_IRQ()
108 /* SIO1_S : uart send data */ init_IRQ()
/linux-4.1.27/arch/m32r/platforms/mappi3/
H A Dsetup.c96 /* SIO0_S : uart send data */ init_IRQ()
107 /* SIO1_S : uart send data */ init_IRQ()
/linux-4.1.27/net/lapb/
H A Dlapb_timer.c103 * Awaiting connection state, send SABM(E), up to N2 times. lapb_t1timer_expiry()
127 * Awaiting disconnection state, send DISC, up to N2 times. lapb_t1timer_expiry()
/linux-4.1.27/net/vmw_vsock/
H A Dvmci_transport_notify.c280 /* Notify the peer that we have read, retrying the send on vmci_transport_send_read_notification()
298 pr_err("%p unable to send read notify to peer\n", sk); vmci_transport_send_read_notification()
381 /* This is a connected socket but we can't currently send data. vmci_transport_notify_pkt_poll_out()
383 * only send a waiting write if the queue is full because vmci_transport_notify_pkt_poll_out()
385 * WAITING_WRITE, READ, etc. loop. Treat failing to send the vmci_transport_notify_pkt_poll_out()
420 * send a notification to the sender before we block. vmci_transport_notify_pkt_recv_init()
578 /* Notify the peer that we have written, retrying the send on vmci_transport_notify_pkt_send_post_enqueue()
594 pr_err("%p unable to send wrote notify to peer\n", sk); vmci_transport_notify_pkt_send_post_enqueue()
H A Dvmci_transport_notify_qstate.c121 /* Notify the peer that we have read, retrying the send on vmci_transport_send_read_notification()
139 pr_err("%p unable to send read notification to peer\n", vmci_transport_send_read_notification()
199 /* This is a connected socket but we can't currently send data. vmci_transport_notify_pkt_poll_out()
228 * send a notification to the sender before we block. vmci_transport_notify_pkt_recv_init()
341 pr_err("%p unable to send wrote notification to peer\n", vmci_transport_notify_pkt_send_post_enqueue()
/linux-4.1.27/security/selinux/include/
H A Dclassmap.h100 { "msg", { "send", "receive", NULL } },
139 { "send", "recv", "relabelto", "forward_in", "forward_out", NULL } },
/linux-4.1.27/sound/drivers/vx/
H A Dvx_core.c116 * @cmd: IRQ message to send (IRQ_MESS_XXX_END)
233 * vx_send_msg_nolock - send a DSP message and read back the status
234 * @rmh: the rmh record to send and receive
282 snd_printd(KERN_DEBUG "vx_send_msg: send IRQ_MESSAGE error\n"); vx_send_msg_nolock()
340 * vx_send_msg - send a DSP message with mutex
341 * @rmh: the rmh record to send and receive
358 * vx_send_rih_nolock - send an RIH to xilinx
359 * @cmd: the command to send
380 /* send the IRQ */ vx_send_rih_nolock()
400 * vx_send_rih - send an RIH with mutex
401 * @cmd: the command to send
/linux-4.1.27/drivers/i2c/busses/
H A Di2c-puv3.c61 /* send addr */ xfer_read()
67 /* send read CMD */ xfer_read()
100 /* send addr */ xfer_write()
103 /* send write CMD */ xfer_write()
/linux-4.1.27/drivers/misc/sgi-xp/
H A Dxp_main.c97 enum xp_retval (*send) (short, int, u32, void *, u16), xpc_set_interface()
105 xpc_interface.send = send; xpc_set_interface()
120 xpc_interface.send = (enum xp_retval(*)(short, int, u32, void *, u16)) xpc_clear_interface()
/linux-4.1.27/drivers/media/usb/pwc/
H A Dpwc-dec1.c7 Please send bug reports and support requests to <luc@saillard.org>.
H A Dpwc-dec1.h6 Please send bug reports and support requests to <luc@saillard.org>.
H A Dpwc-kiara.h6 Please send bug reports and support requests to <luc@saillard.org>.
/linux-4.1.27/drivers/scsi/megaraid/
H A Dmegaraid_mm.h61 * : new uioc_t format and send it to the hba drivers.
/linux-4.1.27/arch/um/drivers/
H A Dslip_common.h60 * For each byte in the packet, send the appropriate slip_esc()
/linux-4.1.27/arch/mips/include/asm/octeon/
H A Dcvmx-address.h67 /* send out the write buffer entry to DRAM */
70 /* send out the write buffer entry as an IOBDMA command */
73 /* send out the write buffer entry as an IO write */
76 /* send out a single-tick command on the NCB bus */
H A Dcvmx-pko.h149 * The output queue to send the packet to (0-127 are
215 /* If set, the hardware will send a response when done */
346 * Prepare to send a packet. This may initiate a tag switch to
348 * performs other prep work for the packet send operation.
372 * @port: Port to send it on
408 * @port: Port to send it on
412 * @packet: Packet to send
449 * @port: Port to send it on
453 * @packet: Packet to send
/linux-4.1.27/drivers/media/dvb-core/
H A Ddemux.h79 #define TS_PACKET 1 /* send TS packets (188 bytes) to callback (default) */
80 #define TS_PAYLOAD_ONLY 2 /* in case TS_PACKET is set, only send the TS
82 #define TS_DECODER 4 /* send stream to built-in decoder (if present) */
83 #define TS_DEMUX 8 /* in case TS_PACKET is set, send the TS to
/linux-4.1.27/include/net/caif/
H A Dcaif_device.h19 * Link Layer to send flow-stop to CAIF Core.
/linux-4.1.27/arch/mips/loongson/loongson-3/
H A Dcop2-ex.c56 return NOTIFY_OK; /* Let default notifier send signals */ loongson_cu2_call()
/linux-4.1.27/arch/mips/netlogic/xlp/
H A Dcop2-ex.c111 return NOTIFY_OK; /* Let default notifier send signals */ nlm_cu2_call()
/linux-4.1.27/arch/powerpc/include/asm/
H A Dhvconsole.h27 * PSeries firmware will only send/recv up to 16 bytes of character data per
/linux-4.1.27/arch/blackfin/include/asm/
H A Dbfin5xx_spi.h81 /* Value to send if no TX value is supplied, usually 0x0 or 0xFFFF */
/linux-4.1.27/fs/btrfs/
H A Dsend.h94 /* attributes in send stream */
/linux-4.1.27/net/netfilter/ipvs/
H A Dip_vs_fo.c35 * Find virtual server with highest weight and send it traffic ip_vs_fo_schedule()
/linux-4.1.27/net/sctp/
H A Dprimitive.c29 * Please send any bug reports or fixes you make to the
129 * This is the main method to send user data via SCTP.
146 * o stream id - to indicate which stream to send the data on. If not
154 * send primitive) within the life time variable. However, the
/linux-4.1.27/net/sunrpc/
H A Dbc_svc.c26 * They implement the transport level processing required to send the
/linux-4.1.27/sound/firewire/
H A Dlib.c17 * snd_fw_transaction - send a request and wait for its completion
/linux-4.1.27/include/crypto/
H A Dvmac.h26 * Please send bug reports to the authors.
/linux-4.1.27/drivers/bluetooth/
H A Dhci_ll.c114 /* send packet */ send_hcill_cmd()
214 * perfectly safe to always send one. ll_device_want_to_wakeup()
231 /* send pending packets and change state to HCILL_AWAKE */ ll_device_want_to_wakeup()
237 /* actually send the packets */ ll_device_want_to_wakeup()
270 /* actually send the sleep ack packet */ ll_device_want_to_sleep()
291 /* send pending packets and change state to HCILL_AWAKE */ ll_device_woke_up()
296 /* actually send the packets */ ll_device_woke_up()
/linux-4.1.27/net/mac80211/
H A Doffchannel.c25 * hardware unconditionally awake, but still let the AP send us normal frames.
51 * If power save was enabled, no need to send a nullfunc ieee80211_offchannel_ps_enable()
57 * to send a new nullfunc frame to inform the AP that we ieee80211_offchannel_ps_enable()
73 * will send a nullfunc frame with the powersave bit set ieee80211_offchannel_ps_disable()
80 * to send a nullfunc frame because AP already knows that ieee80211_offchannel_ps_disable()
93 * restart the timer now and send a nullfunc frame to inform ieee80211_offchannel_ps_disable()
380 /* tell userspace or send frame */ ieee80211_sw_roc_work()
/linux-4.1.27/sound/core/seq/oss/
H A Dseq_oss_synth.c396 * call reset callback. if no callback is defined, send a heartbeat
499 int i, send; snd_seq_oss_synth_sysex() local
514 send = 0; snd_seq_oss_synth_sysex()
519 send = 1; snd_seq_oss_synth_sysex()
531 if (sysex->len && send) { snd_seq_oss_synth_sysex()
537 /* copy the data to event record and send it */ snd_seq_oss_synth_sysex()
587 * send OSS raw events - SEQ_PRIVATE and SEQ_VOLUME
/linux-4.1.27/sound/pci/mixart/
H A Dmixart_core.c142 * send a message to miXart. return: the msg_frame used for this message
248 /* send the message */ snd_mixart_send_msg()
249 err = send_msg(mgr, request, max_resp_size, 1, &msg_frame); /* send and mark the answer pending */ snd_mixart_send_msg()
300 /* send the message */ snd_mixart_send_msg_wait_notif()
301 err = send_msg(mgr, request, MSG_DEFAULT_SIZE, 1, &notif_event); /* send and mark the notification event pending */ snd_mixart_send_msg_wait_notif()
329 /* just send the message (do not mark it as a pending one) */ snd_mixart_send_msg_nonblock()
341 /* common buffer of interrupt to send/receive messages */
/linux-4.1.27/drivers/staging/rtl8192u/ieee80211/
H A Drtl819x_BAProc.c3 * related to TS, this part need some structure defined in QOS side code. Also TX RX is going to be resturctured, so how to send
36 *function: deactivete BA entry in Tx Ts, and send DELBA.
66 *function: deactivete BA entry in Tx Ts, and send DELBA.
152 printk("=====>to send ADDBARSP\n"); ieee80211_ADDBA()
248 *function: send ADDBAReq frame out
275 *function: send ADDBARSP frame out
301 *function: send ADDBARSP frame out
425 return 0; //we send RSP out. ieee80211_rx_ADDBAReq()
523 // If it is OK, then admitted. Or we can send DELBA to cancel BA mechanism. ieee80211_rx_ADDBARsp()
706 * input: unsigned long data //acturally we send TX_TS_RECORD or RX_TS_RECORD to these timer
/linux-4.1.27/drivers/net/usb/
H A Dcx82310_eth.c57 * - optionally send some data (command parameters)
76 /* send command packet */ cx82310_cmd()
81 dev_err(&dev->udev->dev, "send command %#x: error %d\n", cx82310_cmd()
154 * this must not include ethernet header as the device can send partial cx82310_bind()
158 /* we can send at most 1514 bytes of data (+ 2-byte header) per URB */ cx82310_bind()
H A Dint51x1.c79 * usbnet would send a ZLP if packetlength mod urbsize == 0 for us, int51x1_tx_fixup()
81 * we send down to the device... int51x1_tx_fixup()
/linux-4.1.27/lib/
H A Dkobject_uevent.c155 * kobject_uevent_env - send an uevent with environmental data
189 pr_debug("kobject: '%s' (%p): %s: attempted to send uevent " kobject_uevent_env()
272 * send an "add" event, "remove" will automatically generated by kobject_uevent_env()
281 /* we will send an event, so request a new sequence number */ kobject_uevent_env()
289 /* send netlink message */ kobject_uevent_env()
/linux-4.1.27/drivers/block/drbd/
H A Ddrbd_worker.c142 * Do not remove from the write_requests tree here: we did not send the __releases()
363 /* Free peer_req and pages before send. w_e_send_csum()
622 /* Stop generating RS requests when half of the send buffer is filled, make_resync_request()
774 * We need to send at least one request out. */ make_ov_request()
1013 * w_e_end_data_req() - Worker callback, to send a P_DATA_REPLY packet in response to a P_DATA_REQUEST
1051 * w_e_end_rsdata_req() - Worker callback to send a P_RS_DATA_REPLY packet in response to a P_RS_DATA_REQUEST
1195 /* Free e and pages before send. w_e_end_ov_req()
1266 /* Free peer_req and pages before send. w_e_end_ov_reply()
1313 p->barrier = connection->send.current_epoch_nr; drbd_send_barrier()
1315 connection->send.current_epoch_writes = 0; drbd_send_barrier()
1336 if (!connection->send.seen_any_write_yet) { re_init_if_first_write()
1337 connection->send.seen_any_write_yet = true; re_init_if_first_write()
1338 connection->send.current_epoch_nr = epoch; re_init_if_first_write()
1339 connection->send.current_epoch_writes = 0; re_init_if_first_write()
1346 if (!connection->send.seen_any_write_yet) maybe_send_barrier()
1348 if (connection->send.current_epoch_nr != epoch) { maybe_send_barrier()
1349 if (connection->send.current_epoch_writes) maybe_send_barrier()
1351 connection->send.current_epoch_nr = epoch; maybe_send_barrier()
1369 /* this time, no connection->send.current_epoch_writes++; w_send_out_of_sync()
1382 * w_send_dblock() - Worker callback to send a P_DATA packet in order to mirror a write request
1402 connection->send.current_epoch_writes++; w_send_dblock()
1411 * w_send_read_req() - Worker callback to send a read request (P_DATA_REQUEST) packet
2045 * safe to send the epoch separating barrier now. wait_for_work()
2049 connection->send.current_epoch_nr; wait_for_work()
2054 connection->send.current_epoch_nr + 1); wait_for_work()
2066 * In which case we send the barrier above. */ wait_for_work()
/linux-4.1.27/drivers/net/ethernet/smsc/
H A Dsmc9194.c178 If I have to wait until memory is available to send
180 desired memory. Then, I'll send it out and free it.
279 /* this is called to actually send the packet to the chip */
312 . 1. send a SOFT RESET
477 /* Well, I want to send the packet.. but I don't know smc_wait_to_send_packet()
478 if I can send it right now... */ smc_wait_to_send_packet()
526 . wait a short amount of time.. if I can send a packet now, I send smc_wait_to_send_packet()
554 /* or YES! I can send the packet now.. */ smc_wait_to_send_packet()
574 . Tell the card to send it
590 PRINTK((CARDNAME": In XMIT with no packet to send\n")); smc_hardware_send_packet()
618 /* send the packet length ( +6 for status, length and ctl byte ) smc_hardware_send_packet()
624 /* send the packet length ( +6 for status words, length, and ctl*/ smc_hardware_send_packet()
629 /* send the actual data smc_hardware_send_packet()
630 . I _think_ it's faster to send the longs first, and then smc_hardware_send_packet()
668 /* we can send another packet */ smc_hardware_send_packet()
1120 . Called by the kernel to send a packet out into the void
1217 to send the DWORDs or the bytes first, or some smc_rcv()
1422 /* and let the card send more packets to me */ smc_interrupt()
/linux-4.1.27/net/can/
H A Dbcm.c2 * bcm.c - Broadcast Manager to filter/send (cyclic) CAN content
64 * To send multiple CAN frame content within TX_SETUP or to filter
239 * bcm_can_tx - send the (next) CAN frame to the appropriate CAN interface
268 /* send with loopback */ bcm_can_tx()
285 * bcm_send_to_user - send a BCM message to the userspace
432 * 2. send a notification to the user (if possible)
445 /* send RX_CHANGED to the user immediately */ bcm_rx_update_and_send()
461 /* do not send the saved data - only start throttle timer */ bcm_rx_update_and_send()
487 /* received data for the first time => send update to user */ bcm_rx_cmp_to_index()
576 * bcm_rx_thr_flush - Check for throttled data and send it to the userspace
579 * update == 1 : check and send throttled data to userspace (soft_irq context)
610 * Check for throttled data and send it to the userspace
651 /* send reply for RTR-request (placed in op->frames[0]) */ bcm_rx_handler()
803 * bcm_read_op - read out a bcm_op and send it to the user (for bcm_sendmsg)
836 /* we need a real device to send frames */ bcm_tx_setup()
966 /* spec: send can_frame when starting timer */ bcm_tx_setup()
1190 * bcm_tx_send - send a single CAN frame to the CAN interface (for bcm_sendmsg)
1198 /* we need a real device to send frames */ bcm_tx_send()
1224 err = can_send(skb, 1); /* send with loopback */ bcm_tx_send()
/linux-4.1.27/arch/m68k/mac/
H A Diop.c31 * Sending _definitely_ works; my adb-bus.c mods can send
39 * It turns out that replies are placed back in the send buffer
68 * MSG_RCVD, MSG_COMPLETE). To send a message you copy the message into the
73 * reply back out of the send channel buffer and reset the channel state back
85 * register. The IOP will raise INT0 when one or more messages in the send
370 * Actually put a message into a send channel buffer
472 * The message is placed at the end of the send queue. Afterwards if the
473 * channel is idle we force an immediate send of the next message in the
586 printk("iop_ism_irq: new status = %02X, send states", iop_ism_irq()
/linux-4.1.27/arch/arm/mach-ixp4xx/
H A Dixp4xx_npe.c414 const u32 *send = msg; npe_send_message() local
417 debug_msg(npe, "Trying to send message %s [%08X:%08X]\n", npe_send_message()
418 what, send[0], send[1]); npe_send_message()
425 __raw_writel(send[0], &npe->regs->in_out_fifo); npe_send_message()
432 __raw_writel(send[1], &npe->regs->in_out_fifo); npe_send_message()
492 u32 *send = msg, recv[2]; npe_send_recv_message() local
499 if ((recv[0] != send[0]) || (recv[1] != send[1])) { npe_send_recv_message()
/linux-4.1.27/drivers/infiniband/hw/ipath/
H A Dipath_verbs.h283 u32 ssn; /* send sequence number */
324 /* send signal when number of RWQEs < limit */
336 * This structure holds the information that the send tasklet needs
337 * to send a RDMA read response or atomic operation.
365 struct ipath_ib_header s_hdr; /* next packet header to send */
373 struct ipath_sge_state s_sge; /* current send request data */
382 u32 s_cur_size; /* size of send packet in bytes */
415 u8 s_max_rd_atomic; /* max number of RDMA read/atomic to send */
425 u32 s_size; /* send work queue size */
432 struct ipath_swqe *s_wq; /* send work queue */
457 * IPATH_S_WAITING - waiting for RNR timeout or send buffer available.
459 * IPATH_S_WAIT_DMA - waiting for send DMA queue to drain before generating
460 * next send completion entry not via send DMA.
H A Dipath_rc.c95 /* Don't send an ACK if we aren't supposed to. */ ipath_make_rc_ack()
287 /* Check if send work queue is empty. */ ipath_make_rc_req()
492 * This case can only happen if a send is restarted. ipath_make_rc_req()
604 * send_rc_ack - Construct an ACK packet and send it
610 * send side QP state and tasklet.
626 /* Don't send ACK or NAK if a RDMA read or atomic is pending. */ send_rc_ack()
634 /* Don't try to send ACKs if the link isn't ACTIVE */ send_rc_ack()
644 * send tasklet so that when a PIO buffer becomes send_rc_ack()
707 /* Schedule the send tasklet. */ send_rc_ack()
716 * reset_psn - reset the QP state to send starting from PSN
734 * let the normal send code handle initialization. reset_psn()
757 * let the normal send code handle initialization. reset_psn()
948 /* Post a send completion queue entry if requested. */ do_rc_ack()
1372 * send the earliest so that RDMA reads can be restarted at ipath_rc_rcv_error()
1417 * or the send tasklet is already backed up to send an OP()
1469 * or the send tasklet is already backed up to send an OP()
1493 * Try to send a simple ACK to work around a Mellanox bug
1613 * packet sequence number will be for something in the send work ipath_rc_rcv()
1661 * Note that it is up to the requester to not send a new ipath_rc_rcv()
1860 /* Schedule the send tasklet. */ OP()
1923 /* Schedule the send tasklet. */ OP()
/linux-4.1.27/drivers/media/usb/cx231xx/
H A Dcx231xx-i2c.c160 /* usb send command */ cx231xx_i2c_send_bytes()
184 /* usb send command */ cx231xx_i2c_send_bytes()
257 /* usb send command */ cx231xx_i2c_recv_bytes()
270 /* usb send command */ cx231xx_i2c_recv_bytes()
335 /* usb send command */ cx231xx_i2c_recv_bytes_with_saddr()
362 /* usb send command */ cx231xx_i2c_check_for_device()
/linux-4.1.27/drivers/misc/mei/
H A Dinterrupt.c160 * mei_cl_irq_disconnect_rsp - send disconnection response message
195 * interrupt thread context - send disconnect request
272 * mei_cl_irq_connect - send connect request in irq_thread context
465 /* send disconnect message */ mei_irq_write_handler()
472 /* send flow control message */ mei_irq_write_handler()
486 /* send disconnect resp */ mei_irq_write_handler()
/linux-4.1.27/fs/dlm/
H A Drcom.c455 /* We send back the same rcom_lock struct we received, but receive_rcom_lock()
465 /* If the lockspace doesn't exist then still send a status message
510 * - send nothing
521 * - send NAMES
531 * - send LOOKUP
535 * - send LOCKS
/linux-4.1.27/net/wimax/
H A Dop-msg.c93 * @msg: pointer to the message data to send
94 * @size: size of the message to send (in bytes), including the header.
101 * Allocates an skb that will contain the message to send to user
118 * to send stuff around or communicate with user space.
271 * to send stuff around or communicate with user space.
294 * @buf: pointer to the message to send.
/linux-4.1.27/drivers/net/plip/
H A Dplip.c70 Each octet is sent as <wait for rx. '0x1?'> <send 0x10+(octet&0x0F)>
71 <wait for rx. '0x0?'> <send 0x00+((octet>>4)&0x0F)>
405 * This is tricky. If we got here from the beginning of send (either plip_bh_timeout_error()
410 * that is). So we have to treat HS_TIMEOUT and ERROR from send plip_bh_timeout_error()
504 write_data (dev, 0x10); /* send ACK */ plip_receive()
521 write_data (dev, 0x00); /* send ACK */ plip_receive()
596 write_data (dev, 0x01); /* send ACK */ plip_receive_packet()
699 /* PLIP_SEND --- send a byte (two nibbles)
744 /* PLIP_SEND_PACKET --- send a packet */
755 printk(KERN_DEBUG "%s: send skb lost\n", dev->name); plip_send_packet()
796 printk(KERN_DEBUG "%s: send start\n", dev->name); plip_send_packet()
849 printk(KERN_DEBUG "%s: send end\n", dev->name); plip_send_packet()
978 printk(KERN_DEBUG "%s: send request\n", dev->name); plip_tx_packet()
/linux-4.1.27/lib/zlib_deflate/
H A Ddeftree.c181 int value, /* value to send */ send_bits()
722 * bl_order of the last bit length code to send.
740 /* Determine the number of bit length codes to send. The pkzip format build_bl_tree()
799 send_bits(s, (STORED_BLOCK<<1)+eof, 3); /* send block type */ zlib_tr_stored_block()
824 * the last real code. In this case we send two empty static blocks instead
885 * in bl_order of the last bit length code to send.
1027 unsigned code; /* the code to send */ compress_block()
1028 int extra; /* number of extra bits to send */ compress_block()
1034 send_code(s, lc, ltree); /* send a literal byte */ compress_block()
1039 send_code(s, code+LITERALS+1, ltree); /* send the length code */ compress_block()
1043 send_bits(s, lc, extra); /* send the extra length bits */ compress_block()
1049 send_code(s, code, dtree); /* send the distance code */ compress_block()
1053 send_bits(s, dist, extra); /* send the extra distance bits */ compress_block()
/linux-4.1.27/firmware/keyspan_pda/
H A Dkeyspan_pda.S42 * empties the ring below the desired threshold, we set a bit that will send
45 * EP2in (interrupt) is used to send both rx chars and rx status messages
52 * When the EP2in buffer returns, the interrupt prompts us to send more
310 send: label
319 djnz r1, send
416 ;; 04 is send break (really "turn break on/off"). TODO
572 ;; is to manually send out all of the data. The consequence of not
854 ;; now send it
874 ;; tx was not running. send the first character, setup the TI int
901 ;; prod rx, which will actually send the message when in2 becomes free
913 ;; tx finished. send another character if we have one
997 ;; special sequence: send a tx unthrottle message
1112 ;; read/send [tx_ring_out+1], then tx_ring_out++
H A Dxircom_pgs.S42 * empties the ring below the desired threshold, we set a bit that will send
45 * EP2in (interrupt) is used to send both rx chars and rx status messages
52 * When the EP2in buffer returns, the interrupt prompts us to send more
348 send: label
357 djnz r1, send
454 ;; 04 is send break (really "turn break on/off"). TODO
610 ;; is to manually send out all of the data. The consequence of not
915 ;; now send it
935 ;; tx was not running. send the first character, setup the TI int
962 ;; prod rx, which will actually send the message when in2 becomes free
974 ;; tx finished. send another character if we have one
1065 ;; special sequence: send a tx unthrottle message
1180 ;; read/send [tx_ring_out+1], then tx_ring_out++
/linux-4.1.27/drivers/md/
H A Ddm-uevent.c128 * dm_send_uevents - send uevents for given list
130 * @events: list of events to send
/linux-4.1.27/drivers/infiniband/core/
H A Dagent.c160 /* Obtain send only MAD agent for SMI QP */ ib_agent_port_open()
171 /* Obtain send only MAD agent for GSI QP */ ib_agent_port_open()
H A Diwpm_msg.c108 skb = NULL; /* skb is freed in the netlink send-op handling */ iwpm_register_pid()
111 err_str = "Unable to send a nlmsg"; iwpm_register_pid()
179 skb = NULL; /* skb is freed in the netlink send-op handling */ iwpm_add_mapping()
181 err_str = "Unable to send a nlmsg"; iwpm_add_mapping()
256 skb = NULL; /* skb is freed in the netlink send-op handling */ iwpm_add_and_query_mapping()
257 err_str = "Unable to send a nlmsg"; iwpm_add_and_query_mapping()
317 skb = NULL; /* skb is freed in the netlink send-op handling */ iwpm_remove_mapping()
319 err_str = "Unable to send a nlmsg"; iwpm_remove_mapping()
/linux-4.1.27/drivers/infiniband/hw/cxgb4/
H A Dt4.h93 struct fw_ri_send_wr send; member in union:t4_wr
111 wqe->send.opcode = (u8)opcode; init_wr_hdr()
112 wqe->send.flags = flags; init_wr_hdr()
113 wqe->send.wrid = wrid; init_wr_hdr()
114 wqe->send.r1[0] = 0; init_wr_hdr()
115 wqe->send.r1[1] = 0; init_wr_hdr()
116 wqe->send.r1[2] = 0; init_wr_hdr()
117 wqe->send.len16 = len16; init_wr_hdr()
/linux-4.1.27/drivers/misc/ti-st/
H A Dst_ll.c76 send_ll_cmd(st_data, LL_WAKE_UP_ACK); /* send wake_ack */ ll_device_want_to_wakeup()
131 /* don't send the duplicate wake_indication */ st_ll_wakeup()
/linux-4.1.27/drivers/net/ethernet/intel/ixgbe/
H A Dixgbe_mbx.h40 #define IXGBE_PFMAILBOX_STS 0x00000001 /* Initiate message send to VF */
61 clear to send requests */
/linux-4.1.27/drivers/net/ethernet/intel/ixgbevf/
H A Dmbx.h52 #define IXGBE_PFMAILBOX_STS 0x00000001 /* Initiate message send to VF */
71 /* Indicates that VF is still clear to send requests */
/linux-4.1.27/drivers/net/wan/
H A Dhdlc_ppp.c332 if (action & SCR) /* send Configure-Request */ ppp_cp_event()
335 if (action & SCA) /* send Configure-Ack */ ppp_cp_event()
337 if (action & SCN) /* send Configure-Nak/Reject */ ppp_cp_event()
339 if (action & STR) /* send Terminate-Request */ ppp_cp_event()
341 if (action & STA) /* send Terminate-Ack */ ppp_cp_event()
343 if (action & SCJ) /* send Code-Reject */ ppp_cp_event()
499 case LCP_ECHO_REQ: /* send Echo-Reply */ ppp_rx()
590 } else { /* send keep-alive packet */ ppp_timer()
/linux-4.1.27/drivers/net/wireless/
H A Dmac80211_hwsim.h59 * @HWSIM_CMD_FRAME: send/receive a broadcasted frame from/to kernel/user
158 * @idx: rate index to attempt to send with
/linux-4.1.27/drivers/net/wireless/rsi/
H A Drsi_91x_pkt.c53 rsi_dbg(ERR_ZONE, "%s: Unable to send pkt\n", __func__); rsi_send_data_pkt()
143 rsi_dbg(ERR_ZONE, "%s: Unable to send pkt\n", __func__); rsi_send_mgmt_pkt()
/linux-4.1.27/drivers/scsi/ibmvscsi/
H A Dviosrp.h25 /* standard. SRP is used on IBM iSeries and pSeries platforms to send SCSI */
135 * client to the server. There is no way for the server to send
/linux-4.1.27/drivers/scsi/libfc/
H A Dfc_elsct.c21 * Provide interface to send ELS/CT FC frames
35 * @lport: The local port to send the frame on
/linux-4.1.27/drivers/staging/vme/devices/
H A Dvme_pio2_gpio.c101 /* Directionality configured at board build - send appropriate response */ pio2_gpio_dir_in()
120 /* Directionality configured at board build - send appropriate response */ pio2_gpio_dir_out()
/linux-4.1.27/drivers/gpio/
H A Dgpio-74x164.c51 * chain. Thus, the first byte send will end up in the last __gen_74x164_write_config()
53 * numbering, send the bytes in reverse order so that the last __gen_74x164_write_config()

Completed in 6198 milliseconds

1234567891011>>