Lines Matching refs:txreq

943 		tx = list_entry(l, struct qib_verbs_txreq, txreq.list);  in __get_txreq()
972 tx = list_entry(l, struct qib_verbs_txreq, txreq.list); in get_txreq()
996 if (tx->txreq.flags & QIB_SDMA_TXREQ_F_FREEBUF) { in qib_put_txreq()
997 tx->txreq.flags &= ~QIB_SDMA_TXREQ_F_FREEBUF; in qib_put_txreq()
999 tx->txreq.addr, tx->hdr_dwords << 2, in qib_put_txreq()
1007 list_add(&tx->txreq.list, &dev->txreq_free); in qib_put_txreq()
1052 if (qp->s_tx->txreq.sg_count > avail) in qib_verbs_sdma_desc_avail()
1054 avail -= qp->s_tx->txreq.sg_count; in qib_verbs_sdma_desc_avail()
1081 container_of(cookie, struct qib_verbs_txreq, txreq); in sdma_complete()
1090 if (tx->txreq.flags & QIB_SDMA_TXREQ_F_FREEBUF) in sdma_complete()
1169 tx->txreq.callback = sdma_complete; in qib_verbs_send_dma()
1171 tx->txreq.flags = QIB_SDMA_TXREQ_F_HEADTOHOST; in qib_verbs_send_dma()
1173 tx->txreq.flags = QIB_SDMA_TXREQ_F_INTREQ; in qib_verbs_send_dma()
1175 tx->txreq.flags |= QIB_SDMA_TXREQ_F_USELARGEBUF; in qib_verbs_send_dma()
1192 tx->txreq.flags |= QIB_SDMA_TXREQ_F_FREEDESC; in qib_verbs_send_dma()
1193 tx->txreq.sg_count = ndesc; in qib_verbs_send_dma()
1194 tx->txreq.addr = dev->pio_hdrs_phys + in qib_verbs_send_dma()
1211 tx->txreq.addr = dma_map_single(&dd->pcidev->dev, phdr, in qib_verbs_send_dma()
1213 if (dma_mapping_error(&dd->pcidev->dev, tx->txreq.addr)) in qib_verbs_send_dma()
1216 tx->txreq.flags |= QIB_SDMA_TXREQ_F_FREEBUF; in qib_verbs_send_dma()
1217 tx->txreq.sg_count = 1; in qib_verbs_send_dma()
2141 list_add(&tx->txreq.list, &dev->txreq_free); in qib_register_ib_device()
2266 tx = list_entry(l, struct qib_verbs_txreq, txreq.list); in qib_register_ib_device()
2320 tx = list_entry(l, struct qib_verbs_txreq, txreq.list); in qib_unregister_ib_device()