cq_idx             95 drivers/infiniband/hw/efa/efa.h 	u16 cq_idx;
cq_idx            466 drivers/infiniband/hw/efa/efa_admin_cmds_defs.h 	u16 cq_idx;
cq_idx            475 drivers/infiniband/hw/efa/efa_admin_cmds_defs.h 	u16 cq_idx;
cq_idx            185 drivers/infiniband/hw/efa/efa_com_cmd.c 	result->cq_idx = cmd_completion.cq_idx;
cq_idx            199 drivers/infiniband/hw/efa/efa_com_cmd.c 	destroy_cmd.cq_idx = params->cq_idx;
cq_idx            211 drivers/infiniband/hw/efa/efa_com_cmd.c 				      params->cq_idx, err);
cq_idx             79 drivers/infiniband/hw/efa/efa_com_cmd.h 	u16 cq_idx;
cq_idx             85 drivers/infiniband/hw/efa/efa_com_cmd.h 	u16 cq_idx;
cq_idx            707 drivers/infiniband/hw/efa/efa_verbs.c 	create_qp_params.send_cq_idx = to_ecq(init_attr->send_cq)->cq_idx;
cq_idx            708 drivers/infiniband/hw/efa/efa_verbs.c 	create_qp_params.recv_cq_idx = to_ecq(init_attr->recv_cq)->cq_idx;
cq_idx            881 drivers/infiniband/hw/efa/efa_verbs.c static int efa_destroy_cq_idx(struct efa_dev *dev, int cq_idx)
cq_idx            883 drivers/infiniband/hw/efa/efa_verbs.c 	struct efa_com_destroy_cq_params params = { .cq_idx = cq_idx };
cq_idx            895 drivers/infiniband/hw/efa/efa_verbs.c 		  cq->cq_idx, cq->cpu_addr, cq->size, &cq->dma_addr);
cq_idx            897 drivers/infiniband/hw/efa/efa_verbs.c 	efa_destroy_cq_idx(dev, cq->cq_idx);
cq_idx           1004 drivers/infiniband/hw/efa/efa_verbs.c 	resp.cq_idx = result.cq_idx;
cq_idx           1005 drivers/infiniband/hw/efa/efa_verbs.c 	cq->cq_idx = result.cq_idx;
cq_idx           1012 drivers/infiniband/hw/efa/efa_verbs.c 			  cq->cq_idx);
cq_idx           1029 drivers/infiniband/hw/efa/efa_verbs.c 		  cq->cq_idx, result.actual_depth, &cq->dma_addr, cq->cpu_addr);
cq_idx           1034 drivers/infiniband/hw/efa/efa_verbs.c 	efa_destroy_cq_idx(dev, cq->cq_idx);
cq_idx            965 drivers/infiniband/hw/ocrdma/ocrdma_hw.c static void ocrdma_qp_cq_handler(struct ocrdma_dev *dev, u16 cq_idx)
cq_idx            970 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	if (cq_idx >= OCRDMA_MAX_CQ)
cq_idx            973 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	cq = dev->cq_tbl[cq_idx];
cq_idx             83 drivers/infiniband/hw/usnic/usnic_abi.h 	u32				cq_idx[USNIC_QP_GRP_MAX_CQS];
cq_idx            155 drivers/infiniband/hw/usnic/usnic_ib_verbs.c 		resp.cq_idx[i] = chunk->res[i]->vnic_idx;
cq_idx            244 drivers/net/ethernet/amazon/ena/ena_admin_defs.h 	u16 cq_idx;
cq_idx            332 drivers/net/ethernet/amazon/ena/ena_admin_defs.h 	u16 cq_idx;
cq_idx            347 drivers/net/ethernet/amazon/ena/ena_admin_defs.h 	u16 cq_idx;
cq_idx            820 drivers/net/ethernet/amazon/ena/ena_admin_defs.h 	u16 cq_idx;
cq_idx           1198 drivers/net/ethernet/amazon/ena/ena_com.c 				struct ena_com_io_sq *io_sq, u16 cq_idx)
cq_idx           1229 drivers/net/ethernet/amazon/ena/ena_com.c 	create_cmd.cq_idx = cq_idx;
cq_idx           1288 drivers/net/ethernet/amazon/ena/ena_com.c 		rss->rss_ind_tbl[i].cq_idx = io_sq->idx;
cq_idx           1397 drivers/net/ethernet/amazon/ena/ena_com.c 	io_cq->idx = cmd_completion.cq_idx;
cq_idx           1477 drivers/net/ethernet/amazon/ena/ena_com.c 	destroy_cmd.cq_idx = io_cq->idx;
cq_idx            123 drivers/net/ethernet/cavium/thunder/nic.h 	u8	cq_idx;		/* Completion queue index */
cq_idx            848 drivers/net/ethernet/cavium/thunder/nicvf_main.c static int nicvf_cq_intr_handler(struct net_device *netdev, u8 cq_idx,
cq_idx            856 drivers/net/ethernet/cavium/thunder/nicvf_main.c 	struct cmp_queue *cq = &qs->cq[cq_idx];
cq_idx            859 drivers/net/ethernet/cavium/thunder/nicvf_main.c 	struct snd_queue *sq = &qs->sq[cq_idx];
cq_idx            860 drivers/net/ethernet/cavium/thunder/nicvf_main.c 	struct rcv_queue *rq = &qs->rq[cq_idx];
cq_idx            867 drivers/net/ethernet/cavium/thunder/nicvf_main.c 	cqe_count = nicvf_queue_reg_read(nic, NIC_QSET_CQ_0_7_STATUS, cq_idx);
cq_idx            873 drivers/net/ethernet/cavium/thunder/nicvf_main.c 	cqe_head = nicvf_queue_reg_read(nic, NIC_QSET_CQ_0_7_HEAD, cq_idx) >> 9;
cq_idx            914 drivers/net/ethernet/cavium/thunder/nicvf_main.c 			      cq_idx, processed_cqe);
cq_idx            924 drivers/net/ethernet/cavium/thunder/nicvf_main.c 	txq_idx = nicvf_netdev_qidx(nic, cq_idx);
cq_idx            928 drivers/net/ethernet/cavium/thunder/nicvf_main.c 			nicvf_xdp_sq_doorbell(nic, sq, cq_idx);
cq_idx            968 drivers/net/ethernet/cavium/thunder/nicvf_main.c 	work_done = nicvf_cq_intr_handler(netdev, cq->cq_idx, napi, budget);
cq_idx            975 drivers/net/ethernet/cavium/thunder/nicvf_main.c 					       cq->cq_idx);
cq_idx            976 drivers/net/ethernet/cavium/thunder/nicvf_main.c 		nicvf_clear_intr(nic, NICVF_INTR_CQ, cq->cq_idx);
cq_idx            978 drivers/net/ethernet/cavium/thunder/nicvf_main.c 				      cq->cq_idx, cq_head);
cq_idx            979 drivers/net/ethernet/cavium/thunder/nicvf_main.c 		nicvf_enable_intr(nic, NICVF_INTR_CQ, cq->cq_idx);
cq_idx           1046 drivers/net/ethernet/cavium/thunder/nicvf_main.c 	int qidx = cq_poll->cq_idx;
cq_idx           1476 drivers/net/ethernet/cavium/thunder/nicvf_main.c 		cq_poll->cq_idx = qidx;
cq_idx            764 drivers/net/ethernet/cavium/thunder/nicvf_queues.c 	rq->cq_idx = qidx;
cq_idx            780 drivers/net/ethernet/cavium/thunder/nicvf_queues.c 			  (rq->cq_idx << 16) | (rq->cont_rbdr_qs << 9) |
cq_idx            877 drivers/net/ethernet/cavium/thunder/nicvf_queues.c 	sq->cq_idx = qidx;
cq_idx            884 drivers/net/ethernet/cavium/thunder/nicvf_queues.c 	mbx.sq.cfg = (sq->cq_qs << 3) | sq->cq_idx;
cq_idx            245 drivers/net/ethernet/cavium/thunder/nicvf_queues.h 	u8		cq_idx; /* CQ index (0 to 7) in the QS */
cq_idx            267 drivers/net/ethernet/cavium/thunder/nicvf_queues.h 	u8		cq_idx; /* CQ index (0 to 7) in the above QS */
cq_idx            234 drivers/net/ethernet/chelsio/cxgb3/t3_cpl.h 	__be16 cq_idx;
cq_idx             90 drivers/net/ethernet/mellanox/mlx4/en_cq.c 			int cq_idx)
cq_idx            128 drivers/net/ethernet/mellanox/mlx4/en_cq.c 		cq_idx = cq_idx % priv->rx_ring_num;
cq_idx            129 drivers/net/ethernet/mellanox/mlx4/en_cq.c 		rx_cq = priv->rx_cq[cq_idx];
cq_idx            694 drivers/net/ethernet/mellanox/mlx4/mlx4_en.h 			int cq_idx);
cq_idx            277 drivers/nvme/host/pci.c 	nvmeq->dbbuf_cq_db = &dev->dbbuf_dbs[cq_idx(qid, dev->db_stride)];
cq_idx            279 drivers/nvme/host/pci.c 	nvmeq->dbbuf_cq_ei = &dev->dbbuf_eis[cq_idx(qid, dev->db_stride)];
cq_idx            637 drivers/scsi/lpfc/lpfc_debugfs.h 	int cq_idx;
cq_idx            639 drivers/scsi/lpfc/lpfc_debugfs.h 	for (cq_idx = 0; cq_idx < phba->cfg_hdw_queue; cq_idx++)
cq_idx            640 drivers/scsi/lpfc/lpfc_debugfs.h 		if (phba->sli4_hba.hdwq[cq_idx].io_cq->queue_id == qid)
cq_idx            643 drivers/scsi/lpfc/lpfc_debugfs.h 	if (cq_idx < phba->cfg_hdw_queue) {
cq_idx            644 drivers/scsi/lpfc/lpfc_debugfs.h 		pr_err("IO CQ[Idx:%d|Qid:%d]\n", cq_idx, qid);
cq_idx            645 drivers/scsi/lpfc/lpfc_debugfs.h 		lpfc_debug_dump_q(phba->sli4_hba.hdwq[cq_idx].io_cq);
cq_idx            605 drivers/scsi/qedf/qedf_io.c 	uint16_t cq_idx = smp_processor_id() % qedf->num_queues;
cq_idx            638 drivers/scsi/qedf/qedf_io.c 	io_req->task_params->cq_rss_number = cq_idx;
cq_idx           2053 drivers/scsi/qedi/qedi_fw.c 	u16 cq_idx;
cq_idx           2081 drivers/scsi/qedi/qedi_fw.c 	cq_idx = smp_processor_id() % qedi->num_queues;
cq_idx           2146 drivers/scsi/qedi/qedi_fw.c 	task_params.cq_rss_number = cq_idx;
cq_idx           1201 drivers/scsi/snic/snic_scsi.c 		     unsigned int cq_idx,
cq_idx           1284 drivers/scsi/snic/snic_scsi.c 	unsigned int cq_idx;
cq_idx           1288 drivers/scsi/snic/snic_scsi.c 	for (cq_idx = snic->wq_count; cq_idx < snic->cq_count; cq_idx++) {
cq_idx           1289 drivers/scsi/snic/snic_scsi.c 		nent_per_cq = vnic_cq_fw_service(&snic->cq[cq_idx],
cq_idx             54 include/uapi/rdma/efa-abi.h 	__u16 cq_idx;