Searched refs:ctrlq (Results 1 – 7 of 7) sorted by relevance
65 schedule_work(&vgdev->ctrlq.dequeue_work); in virtio_gpu_ctrl_ack()83 count += virtqueue_get_vring_size(vgdev->ctrlq.vq); in virtio_gpu_alloc_vbufs()107 count += virtqueue_get_vring_size(vgdev->ctrlq.vq); in virtio_gpu_free_vbufs()232 ctrlq.dequeue_work); in virtio_gpu_dequeue_ctrl_func()239 spin_lock(&vgdev->ctrlq.qlock); in virtio_gpu_dequeue_ctrl_func()241 virtqueue_disable_cb(vgdev->ctrlq.vq); in virtio_gpu_dequeue_ctrl_func()242 reclaim_vbufs(vgdev->ctrlq.vq, &reclaim_list); in virtio_gpu_dequeue_ctrl_func()244 } while (!virtqueue_enable_cb(vgdev->ctrlq.vq)); in virtio_gpu_dequeue_ctrl_func()245 spin_unlock(&vgdev->ctrlq.qlock); in virtio_gpu_dequeue_ctrl_func()267 wake_up(&vgdev->ctrlq.ack_queue); in virtio_gpu_dequeue_ctrl_func()[all …]
159 virtio_gpu_init_vq(&vgdev->ctrlq, virtio_gpu_dequeue_ctrl_func); in virtio_gpu_driver_load()179 vgdev->ctrlq.vq = vqs[0]; in virtio_gpu_driver_load()253 flush_work(&vgdev->ctrlq.dequeue_work); in virtio_gpu_driver_unload()
177 struct virtio_gpu_queue ctrlq; member
1471 ret = ctrl_xmit(&adap->sge.ctrlq[0], skb); in t4_mgmt_tx()1682 return ctrl_xmit(&adap->sge.ctrlq[idx], skb); in ofld_send()2907 for (i = 0; i < ARRAY_SIZE(adap->sge.ctrlq); i++) { in t4_free_sge_resources()2908 struct sge_ctrl_txq *cq = &adap->sge.ctrlq[i]; in t4_free_sge_resources()2964 for (i = 0; i < ARRAY_SIZE(s->ctrlq); i++) { in t4_sge_stop()2965 struct sge_ctrl_txq *cq = &s->ctrlq[i]; in t4_sge_stop()
1125 err = t4_sge_alloc_ctrl_txq(adap, &s->ctrlq[i], adap->port[i], in setup_sge_queues()2289 disable_txq_db(&adap->sge.ctrlq[i].q); in disable_dbs()2301 enable_txq_db(adap, &adap->sge.ctrlq[i].q); in enable_dbs()2371 sync_txq_pidx(adap, &adap->sge.ctrlq[i].q); in recover_all_queues()4380 for (i = 0; i < ARRAY_SIZE(s->ctrlq); i++) in cfg_queues()4381 s->ctrlq[i].q.size = 512; in cfg_queues()
636 struct sge_ctrl_txq ctrlq[MAX_CTRL_QUEUES]; member
2412 const struct sge_ctrl_txq *tx = &adap->sge.ctrlq[ctrl_idx * 4]; in sge_qinfo_show()