io_cq             387 block/bfq-iosched.c static struct bfq_io_cq *icq_to_bic(struct io_cq *icq)
io_cq            4955 block/bfq-iosched.c static void bfq_exit_icq(struct io_cq *icq)
io_cq             388 block/bfq-iosched.h 	struct io_cq icq; /* must be the first member */
io_cq              34 block/blk-ioc.c 	struct io_cq *icq = container_of(head, struct io_cq, __rcu_head);
io_cq              43 block/blk-ioc.c static void ioc_exit_icq(struct io_cq *icq)
io_cq              60 block/blk-ioc.c static void ioc_destroy_icq(struct io_cq *icq)
io_cq             110 block/blk-ioc.c 		struct io_cq *icq = hlist_entry(ioc->icq_list.first,
io_cq             111 block/blk-ioc.c 						struct io_cq, ioc_node);
io_cq             174 block/blk-ioc.c 	struct io_cq *icq;
io_cq             218 block/blk-ioc.c 		struct io_cq *icq = list_entry(icq_list->next,
io_cq             219 block/blk-ioc.c 						struct io_cq, q_node);
io_cq             332 block/blk-ioc.c struct io_cq *ioc_lookup_icq(struct io_context *ioc, struct request_queue *q)
io_cq             334 block/blk-ioc.c 	struct io_cq *icq;
io_cq             372 block/blk-ioc.c struct io_cq *ioc_create_icq(struct io_context *ioc, struct request_queue *q,
io_cq             376 block/blk-ioc.c 	struct io_cq *icq;
io_cq              39 block/blk-mq-sched.c 	struct io_cq *icq;
io_cq             281 block/blk.h    struct io_cq *ioc_lookup_icq(struct io_context *ioc, struct request_queue *q);
io_cq             282 block/blk.h    struct io_cq *ioc_create_icq(struct io_context *ioc, struct request_queue *q,
io_cq             534 block/elevator.c 		if (WARN_ON(e->icq_size < sizeof(struct io_cq)) ||
io_cq             535 block/elevator.c 		    WARN_ON(e->icq_align < __alignof__(struct io_cq)))
io_cq             419 drivers/net/ethernet/amazon/ena/ena_com.c 			      struct ena_com_io_cq *io_cq)
io_cq             424 drivers/net/ethernet/amazon/ena/ena_com.c 	memset(&io_cq->cdesc_addr, 0x0, sizeof(io_cq->cdesc_addr));
io_cq             427 drivers/net/ethernet/amazon/ena/ena_com.c 	io_cq->cdesc_entry_size_in_bytes =
io_cq             428 drivers/net/ethernet/amazon/ena/ena_com.c 		(io_cq->direction == ENA_COM_IO_QUEUE_DIRECTION_TX) ?
io_cq             432 drivers/net/ethernet/amazon/ena/ena_com.c 	size = io_cq->cdesc_entry_size_in_bytes * io_cq->q_depth;
io_cq             436 drivers/net/ethernet/amazon/ena/ena_com.c 	io_cq->cdesc_addr.virt_addr =
io_cq             438 drivers/net/ethernet/amazon/ena/ena_com.c 				   &io_cq->cdesc_addr.phys_addr, GFP_KERNEL);
io_cq             440 drivers/net/ethernet/amazon/ena/ena_com.c 	if (!io_cq->cdesc_addr.virt_addr) {
io_cq             441 drivers/net/ethernet/amazon/ena/ena_com.c 		io_cq->cdesc_addr.virt_addr =
io_cq             443 drivers/net/ethernet/amazon/ena/ena_com.c 					   &io_cq->cdesc_addr.phys_addr,
io_cq             447 drivers/net/ethernet/amazon/ena/ena_com.c 	if (!io_cq->cdesc_addr.virt_addr) {
io_cq             452 drivers/net/ethernet/amazon/ena/ena_com.c 	io_cq->phase = 1;
io_cq             453 drivers/net/ethernet/amazon/ena/ena_com.c 	io_cq->head = 0;
io_cq             913 drivers/net/ethernet/amazon/ena/ena_com.c 				  struct ena_com_io_cq *io_cq)
io_cq             917 drivers/net/ethernet/amazon/ena/ena_com.c 	if (io_cq->cdesc_addr.virt_addr) {
io_cq             918 drivers/net/ethernet/amazon/ena/ena_com.c 		size = io_cq->cdesc_entry_size_in_bytes * io_cq->q_depth;
io_cq             921 drivers/net/ethernet/amazon/ena/ena_com.c 				  io_cq->cdesc_addr.virt_addr,
io_cq             922 drivers/net/ethernet/amazon/ena/ena_com.c 				  io_cq->cdesc_addr.phys_addr);
io_cq             924 drivers/net/ethernet/amazon/ena/ena_com.c 		io_cq->cdesc_addr.virt_addr = NULL;
io_cq            1360 drivers/net/ethernet/amazon/ena/ena_com.c 			 struct ena_com_io_cq *io_cq)
io_cq            1371 drivers/net/ethernet/amazon/ena/ena_com.c 	create_cmd.cq_caps_2 |= (io_cq->cdesc_entry_size_in_bytes / 4) &
io_cq            1376 drivers/net/ethernet/amazon/ena/ena_com.c 	create_cmd.msix_vector = io_cq->msix_vector;
io_cq            1377 drivers/net/ethernet/amazon/ena/ena_com.c 	create_cmd.cq_depth = io_cq->q_depth;
io_cq            1381 drivers/net/ethernet/amazon/ena/ena_com.c 				   io_cq->cdesc_addr.phys_addr);
io_cq            1397 drivers/net/ethernet/amazon/ena/ena_com.c 	io_cq->idx = cmd_completion.cq_idx;
io_cq            1399 drivers/net/ethernet/amazon/ena/ena_com.c 	io_cq->unmask_reg = (u32 __iomem *)((uintptr_t)ena_dev->reg_bar +
io_cq            1403 drivers/net/ethernet/amazon/ena/ena_com.c 		io_cq->cq_head_db_reg =
io_cq            1408 drivers/net/ethernet/amazon/ena/ena_com.c 		io_cq->numa_node_cfg_reg =
io_cq            1412 drivers/net/ethernet/amazon/ena/ena_com.c 	pr_debug("created cq[%u], depth[%u]\n", io_cq->idx, io_cq->q_depth);
io_cq            1419 drivers/net/ethernet/amazon/ena/ena_com.c 			    struct ena_com_io_cq **io_cq)
io_cq            1428 drivers/net/ethernet/amazon/ena/ena_com.c 	*io_cq = &ena_dev->io_cq_queues[qid];
io_cq            1468 drivers/net/ethernet/amazon/ena/ena_com.c 			  struct ena_com_io_cq *io_cq)
io_cq            1477 drivers/net/ethernet/amazon/ena/ena_com.c 	destroy_cmd.cq_idx = io_cq->idx;
io_cq            1822 drivers/net/ethernet/amazon/ena/ena_com.c 	struct ena_com_io_cq *io_cq;
io_cq            1832 drivers/net/ethernet/amazon/ena/ena_com.c 	io_cq = &ena_dev->io_cq_queues[ctx->qid];
io_cq            1835 drivers/net/ethernet/amazon/ena/ena_com.c 	memset(io_cq, 0x0, sizeof(*io_cq));
io_cq            1838 drivers/net/ethernet/amazon/ena/ena_com.c 	io_cq->q_depth = ctx->queue_size;
io_cq            1839 drivers/net/ethernet/amazon/ena/ena_com.c 	io_cq->direction = ctx->direction;
io_cq            1840 drivers/net/ethernet/amazon/ena/ena_com.c 	io_cq->qid = ctx->qid;
io_cq            1842 drivers/net/ethernet/amazon/ena/ena_com.c 	io_cq->msix_vector = ctx->msix_vector;
io_cq            1858 drivers/net/ethernet/amazon/ena/ena_com.c 	ret = ena_com_init_io_cq(ena_dev, ctx, io_cq);
io_cq            1862 drivers/net/ethernet/amazon/ena/ena_com.c 	ret = ena_com_create_io_cq(ena_dev, io_cq);
io_cq            1866 drivers/net/ethernet/amazon/ena/ena_com.c 	ret = ena_com_create_io_sq(ena_dev, io_sq, io_cq->idx);
io_cq            1873 drivers/net/ethernet/amazon/ena/ena_com.c 	ena_com_destroy_io_cq(ena_dev, io_cq);
io_cq            1875 drivers/net/ethernet/amazon/ena/ena_com.c 	ena_com_io_queue_free(ena_dev, io_sq, io_cq);
io_cq            1882 drivers/net/ethernet/amazon/ena/ena_com.c 	struct ena_com_io_cq *io_cq;
io_cq            1891 drivers/net/ethernet/amazon/ena/ena_com.c 	io_cq = &ena_dev->io_cq_queues[qid];
io_cq            1894 drivers/net/ethernet/amazon/ena/ena_com.c 	ena_com_destroy_io_cq(ena_dev, io_cq);
io_cq            1896 drivers/net/ethernet/amazon/ena/ena_com.c 	ena_com_io_queue_free(ena_dev, io_sq, io_cq);
io_cq             471 drivers/net/ethernet/amazon/ena/ena_com.h 			    struct ena_com_io_cq **io_cq);
io_cq             855 drivers/net/ethernet/amazon/ena/ena_com.h 			 struct ena_com_io_cq *io_cq);
io_cq             866 drivers/net/ethernet/amazon/ena/ena_com.h 			  struct ena_com_io_cq *io_cq);
io_cq              36 drivers/net/ethernet/amazon/ena/ena_eth_com.c 	struct ena_com_io_cq *io_cq)
io_cq              42 drivers/net/ethernet/amazon/ena/ena_eth_com.c 	head_masked = io_cq->head & (io_cq->q_depth - 1);
io_cq              43 drivers/net/ethernet/amazon/ena/ena_eth_com.c 	expected_phase = io_cq->phase;
io_cq              45 drivers/net/ethernet/amazon/ena/ena_eth_com.c 	cdesc = (struct ena_eth_io_rx_cdesc_base *)(io_cq->cdesc_addr.virt_addr
io_cq              46 drivers/net/ethernet/amazon/ena/ena_eth_com.c 			+ (head_masked * io_cq->cdesc_entry_size_in_bytes));
io_cq             243 drivers/net/ethernet/amazon/ena/ena_eth_com.c 	ena_com_rx_cdesc_idx_to_ptr(struct ena_com_io_cq *io_cq, u16 idx)
io_cq             245 drivers/net/ethernet/amazon/ena/ena_eth_com.c 	idx &= (io_cq->q_depth - 1);
io_cq             247 drivers/net/ethernet/amazon/ena/ena_eth_com.c 		((uintptr_t)io_cq->cdesc_addr.virt_addr +
io_cq             248 drivers/net/ethernet/amazon/ena/ena_eth_com.c 		idx * io_cq->cdesc_entry_size_in_bytes);
io_cq             251 drivers/net/ethernet/amazon/ena/ena_eth_com.c static u16 ena_com_cdesc_rx_pkt_get(struct ena_com_io_cq *io_cq,
io_cq             259 drivers/net/ethernet/amazon/ena/ena_eth_com.c 		cdesc = ena_com_get_next_rx_cdesc(io_cq);
io_cq             263 drivers/net/ethernet/amazon/ena/ena_eth_com.c 		ena_com_cq_inc_head(io_cq);
io_cq             270 drivers/net/ethernet/amazon/ena/ena_eth_com.c 		*first_cdesc_idx = io_cq->cur_rx_pkt_cdesc_start_idx;
io_cq             271 drivers/net/ethernet/amazon/ena/ena_eth_com.c 		count += io_cq->cur_rx_pkt_cdesc_count;
io_cq             273 drivers/net/ethernet/amazon/ena/ena_eth_com.c 		head_masked = io_cq->head & (io_cq->q_depth - 1);
io_cq             275 drivers/net/ethernet/amazon/ena/ena_eth_com.c 		io_cq->cur_rx_pkt_cdesc_count = 0;
io_cq             276 drivers/net/ethernet/amazon/ena/ena_eth_com.c 		io_cq->cur_rx_pkt_cdesc_start_idx = head_masked;
io_cq             279 drivers/net/ethernet/amazon/ena/ena_eth_com.c 			 io_cq->qid, *first_cdesc_idx, count);
io_cq             281 drivers/net/ethernet/amazon/ena/ena_eth_com.c 		io_cq->cur_rx_pkt_cdesc_count += count;
io_cq             514 drivers/net/ethernet/amazon/ena/ena_eth_com.c int ena_com_rx_pkt(struct ena_com_io_cq *io_cq,
io_cq             524 drivers/net/ethernet/amazon/ena/ena_eth_com.c 	WARN(io_cq->direction != ENA_COM_IO_QUEUE_DIRECTION_RX, "wrong Q type");
io_cq             526 drivers/net/ethernet/amazon/ena/ena_eth_com.c 	nb_hw_desc = ena_com_cdesc_rx_pkt_get(io_cq, &cdesc_idx);
io_cq             532 drivers/net/ethernet/amazon/ena/ena_eth_com.c 	pr_debug("fetch rx packet: queue %d completed desc: %d\n", io_cq->qid,
io_cq             542 drivers/net/ethernet/amazon/ena/ena_eth_com.c 		cdesc = ena_com_rx_cdesc_idx_to_ptr(io_cq, cdesc_idx + i);
io_cq             595 drivers/net/ethernet/amazon/ena/ena_eth_com.c bool ena_com_cq_empty(struct ena_com_io_cq *io_cq)
io_cq             599 drivers/net/ethernet/amazon/ena/ena_eth_com.c 	cdesc = ena_com_get_next_rx_cdesc(io_cq);
io_cq              82 drivers/net/ethernet/amazon/ena/ena_eth_com.h int ena_com_rx_pkt(struct ena_com_io_cq *io_cq,
io_cq              90 drivers/net/ethernet/amazon/ena/ena_eth_com.h bool ena_com_cq_empty(struct ena_com_io_cq *io_cq);
io_cq              92 drivers/net/ethernet/amazon/ena/ena_eth_com.h static inline void ena_com_unmask_intr(struct ena_com_io_cq *io_cq,
io_cq              95 drivers/net/ethernet/amazon/ena/ena_eth_com.h 	writel(intr_reg->intr_control, io_cq->unmask_reg);
io_cq             193 drivers/net/ethernet/amazon/ena/ena_eth_com.h static inline int ena_com_update_dev_comp_head(struct ena_com_io_cq *io_cq)
io_cq             198 drivers/net/ethernet/amazon/ena/ena_eth_com.h 	if (unlikely(io_cq->cq_head_db_reg)) {
io_cq             199 drivers/net/ethernet/amazon/ena/ena_eth_com.h 		head = io_cq->head;
io_cq             200 drivers/net/ethernet/amazon/ena/ena_eth_com.h 		unreported_comp = head - io_cq->last_head_update;
io_cq             201 drivers/net/ethernet/amazon/ena/ena_eth_com.h 		need_update = unreported_comp > (io_cq->q_depth / ENA_COMP_HEAD_THRESH);
io_cq             205 drivers/net/ethernet/amazon/ena/ena_eth_com.h 				 io_cq->qid, head);
io_cq             206 drivers/net/ethernet/amazon/ena/ena_eth_com.h 			writel(head, io_cq->cq_head_db_reg);
io_cq             207 drivers/net/ethernet/amazon/ena/ena_eth_com.h 			io_cq->last_head_update = head;
io_cq             214 drivers/net/ethernet/amazon/ena/ena_eth_com.h static inline void ena_com_update_numa_node(struct ena_com_io_cq *io_cq,
io_cq             219 drivers/net/ethernet/amazon/ena/ena_eth_com.h 	if (!io_cq->numa_node_cfg_reg)
io_cq             225 drivers/net/ethernet/amazon/ena/ena_eth_com.h 	writel(numa_cfg.numa_cfg, io_cq->numa_node_cfg_reg);
io_cq             233 drivers/net/ethernet/amazon/ena/ena_eth_com.h static inline void ena_com_cq_inc_head(struct ena_com_io_cq *io_cq)
io_cq             235 drivers/net/ethernet/amazon/ena/ena_eth_com.h 	io_cq->head++;
io_cq             238 drivers/net/ethernet/amazon/ena/ena_eth_com.h 	if (unlikely((io_cq->head & (io_cq->q_depth - 1)) == 0))
io_cq             239 drivers/net/ethernet/amazon/ena/ena_eth_com.h 		io_cq->phase ^= 1;
io_cq             242 drivers/net/ethernet/amazon/ena/ena_eth_com.h static inline int ena_com_tx_comp_req_id_get(struct ena_com_io_cq *io_cq,
io_cq             249 drivers/net/ethernet/amazon/ena/ena_eth_com.h 	masked_head = io_cq->head & (io_cq->q_depth - 1);
io_cq             250 drivers/net/ethernet/amazon/ena/ena_eth_com.h 	expected_phase = io_cq->phase;
io_cq             253 drivers/net/ethernet/amazon/ena/ena_eth_com.h 		((uintptr_t)io_cq->cdesc_addr.virt_addr +
io_cq             254 drivers/net/ethernet/amazon/ena/ena_eth_com.h 		(masked_head * io_cq->cdesc_entry_size_in_bytes));
io_cq             267 drivers/net/ethernet/amazon/ena/ena_eth_com.h 	if (unlikely(*req_id >= io_cq->q_depth)) {
io_cq             272 drivers/net/ethernet/amazon/ena/ena_eth_com.h 	ena_com_cq_inc_head(io_cq);
io_cq            3763 drivers/scsi/lpfc/lpfc_debugfs.c 	qp = phba->sli4_hba.hdwq[eqidx].io_cq;
io_cq            4226 drivers/scsi/lpfc/lpfc_debugfs.c 				qp = phba->sli4_hba.hdwq[qidx].io_cq;
io_cq             463 drivers/scsi/lpfc/lpfc_debugfs.h 		cq = phba->sli4_hba.hdwq[wqidx].io_cq;
io_cq             640 drivers/scsi/lpfc/lpfc_debugfs.h 		if (phba->sli4_hba.hdwq[cq_idx].io_cq->queue_id == qid)
io_cq             645 drivers/scsi/lpfc/lpfc_debugfs.h 		lpfc_debug_dump_q(phba->sli4_hba.hdwq[cq_idx].io_cq);
io_cq            8678 drivers/scsi/lpfc/lpfc_init.c 	phba->sli4_hba.hdwq[idx].io_cq = qdesc;
io_cq            9125 drivers/scsi/lpfc/lpfc_init.c 		lpfc_sli4_queue_free(hdwq[idx].io_cq);
io_cq            9127 drivers/scsi/lpfc/lpfc_init.c 		hdwq[idx].io_cq = NULL;
io_cq            9466 drivers/scsi/lpfc/lpfc_init.c 				       qp[qidx].io_cq,
io_cq            9775 drivers/scsi/lpfc/lpfc_init.c 			lpfc_cq_destroy(phba, qp->io_cq);
io_cq            5567 drivers/scsi/lpfc/lpfc_sli.c 			sli4_hba->sli4_write_cq_db(phba, qp->io_cq, 0,
io_cq             660 drivers/scsi/lpfc/lpfc_sli4.h 	struct lpfc_queue *io_cq;   /* Fast-path FCP & NVME compl queue */
io_cq             184 include/linux/blkdev.h 			struct io_cq		*icq;
io_cq              10 include/linux/elevator.h struct io_cq;
io_cq              51 include/linux/elevator.h 	void (*init_icq)(struct io_cq *);
io_cq              52 include/linux/elevator.h 	void (*exit_icq)(struct io_cq *);
io_cq             116 include/linux/iocontext.h 	struct io_cq __rcu	*icq_hint;