vnic_sdma          84 drivers/infiniband/hw/hfi1/vnic_sdma.c 	struct hfi1_vnic_sdma *vnic_sdma = tx->sdma;
vnic_sdma          86 drivers/infiniband/hw/hfi1/vnic_sdma.c 	sdma_txclean(vnic_sdma->dd, txreq);
vnic_sdma          88 drivers/infiniband/hw/hfi1/vnic_sdma.c 	kmem_cache_free(vnic_sdma->dd->vnic.txreq_cache, tx);
vnic_sdma         171 drivers/infiniband/hw/hfi1/vnic_sdma.c 	struct hfi1_vnic_sdma *vnic_sdma = &vinfo->sdma[q_idx];
vnic_sdma         172 drivers/infiniband/hw/hfi1/vnic_sdma.c 	struct sdma_engine *sde = vnic_sdma->sde;
vnic_sdma         176 drivers/infiniband/hw/hfi1/vnic_sdma.c 	if (unlikely(READ_ONCE(vnic_sdma->state) != HFI1_VNIC_SDMA_Q_ACTIVE))
vnic_sdma         188 drivers/infiniband/hw/hfi1/vnic_sdma.c 	tx->sdma = vnic_sdma;
vnic_sdma         196 drivers/infiniband/hw/hfi1/vnic_sdma.c 	ret = sdma_send_txreq(sde, iowait_get_ib_work(&vnic_sdma->wait),
vnic_sdma         197 drivers/infiniband/hw/hfi1/vnic_sdma.c 			      &tx->txreq, vnic_sdma->pkts_sent);
vnic_sdma         203 drivers/infiniband/hw/hfi1/vnic_sdma.c 		vnic_sdma->pkts_sent = true;
vnic_sdma         204 drivers/infiniband/hw/hfi1/vnic_sdma.c 		iowait_starve_clear(vnic_sdma->pkts_sent, &vnic_sdma->wait);
vnic_sdma         215 drivers/infiniband/hw/hfi1/vnic_sdma.c 		vnic_sdma->pkts_sent = false;
vnic_sdma         233 drivers/infiniband/hw/hfi1/vnic_sdma.c 	struct hfi1_vnic_sdma *vnic_sdma =
vnic_sdma         242 drivers/infiniband/hw/hfi1/vnic_sdma.c 	vnic_sdma->state = HFI1_VNIC_SDMA_Q_DEFERRED;
vnic_sdma         243 drivers/infiniband/hw/hfi1/vnic_sdma.c 	if (list_empty(&vnic_sdma->wait.list)) {
vnic_sdma         260 drivers/infiniband/hw/hfi1/vnic_sdma.c 	struct hfi1_vnic_sdma *vnic_sdma =
vnic_sdma         262 drivers/infiniband/hw/hfi1/vnic_sdma.c 	struct hfi1_vnic_vport_info *vinfo = vnic_sdma->vinfo;
vnic_sdma         264 drivers/infiniband/hw/hfi1/vnic_sdma.c 	vnic_sdma->state = HFI1_VNIC_SDMA_Q_ACTIVE;
vnic_sdma         265 drivers/infiniband/hw/hfi1/vnic_sdma.c 	if (__netif_subqueue_stopped(vinfo->netdev, vnic_sdma->q_idx))
vnic_sdma         266 drivers/infiniband/hw/hfi1/vnic_sdma.c 		netif_wake_subqueue(vinfo->netdev, vnic_sdma->q_idx);
vnic_sdma         272 drivers/infiniband/hw/hfi1/vnic_sdma.c 	struct hfi1_vnic_sdma *vnic_sdma = &vinfo->sdma[q_idx];
vnic_sdma         274 drivers/infiniband/hw/hfi1/vnic_sdma.c 	return (READ_ONCE(vnic_sdma->state) == HFI1_VNIC_SDMA_Q_ACTIVE);
vnic_sdma         282 drivers/infiniband/hw/hfi1/vnic_sdma.c 		struct hfi1_vnic_sdma *vnic_sdma = &vinfo->sdma[i];
vnic_sdma         284 drivers/infiniband/hw/hfi1/vnic_sdma.c 		iowait_init(&vnic_sdma->wait, 0, NULL, NULL,
vnic_sdma         287 drivers/infiniband/hw/hfi1/vnic_sdma.c 		vnic_sdma->sde = &vinfo->dd->per_sdma[i];
vnic_sdma         288 drivers/infiniband/hw/hfi1/vnic_sdma.c 		vnic_sdma->dd = vinfo->dd;
vnic_sdma         289 drivers/infiniband/hw/hfi1/vnic_sdma.c 		vnic_sdma->vinfo = vinfo;
vnic_sdma         290 drivers/infiniband/hw/hfi1/vnic_sdma.c 		vnic_sdma->q_idx = i;
vnic_sdma         291 drivers/infiniband/hw/hfi1/vnic_sdma.c 		vnic_sdma->state = HFI1_VNIC_SDMA_Q_ACTIVE;
vnic_sdma         294 drivers/infiniband/hw/hfi1/vnic_sdma.c 		if (vnic_sdma->sde->descq_cnt > HFI1_VNIC_SDMA_DESC_WTRMRK) {
vnic_sdma         297 drivers/infiniband/hw/hfi1/vnic_sdma.c 			INIT_LIST_HEAD(&vnic_sdma->stx.list);
vnic_sdma         298 drivers/infiniband/hw/hfi1/vnic_sdma.c 			vnic_sdma->stx.num_desc = HFI1_VNIC_SDMA_DESC_WTRMRK;
vnic_sdma         299 drivers/infiniband/hw/hfi1/vnic_sdma.c 			work = iowait_get_ib_work(&vnic_sdma->wait);
vnic_sdma         300 drivers/infiniband/hw/hfi1/vnic_sdma.c 			list_add_tail(&vnic_sdma->stx.list, &work->tx_head);