io_cq 387 block/bfq-iosched.c static struct bfq_io_cq *icq_to_bic(struct io_cq *icq) io_cq 4955 block/bfq-iosched.c static void bfq_exit_icq(struct io_cq *icq) io_cq 388 block/bfq-iosched.h struct io_cq icq; /* must be the first member */ io_cq 34 block/blk-ioc.c struct io_cq *icq = container_of(head, struct io_cq, __rcu_head); io_cq 43 block/blk-ioc.c static void ioc_exit_icq(struct io_cq *icq) io_cq 60 block/blk-ioc.c static void ioc_destroy_icq(struct io_cq *icq) io_cq 110 block/blk-ioc.c struct io_cq *icq = hlist_entry(ioc->icq_list.first, io_cq 111 block/blk-ioc.c struct io_cq, ioc_node); io_cq 174 block/blk-ioc.c struct io_cq *icq; io_cq 218 block/blk-ioc.c struct io_cq *icq = list_entry(icq_list->next, io_cq 219 block/blk-ioc.c struct io_cq, q_node); io_cq 332 block/blk-ioc.c struct io_cq *ioc_lookup_icq(struct io_context *ioc, struct request_queue *q) io_cq 334 block/blk-ioc.c struct io_cq *icq; io_cq 372 block/blk-ioc.c struct io_cq *ioc_create_icq(struct io_context *ioc, struct request_queue *q, io_cq 376 block/blk-ioc.c struct io_cq *icq; io_cq 39 block/blk-mq-sched.c struct io_cq *icq; io_cq 281 block/blk.h struct io_cq *ioc_lookup_icq(struct io_context *ioc, struct request_queue *q); io_cq 282 block/blk.h struct io_cq *ioc_create_icq(struct io_context *ioc, struct request_queue *q, io_cq 534 block/elevator.c if (WARN_ON(e->icq_size < sizeof(struct io_cq)) || io_cq 535 block/elevator.c WARN_ON(e->icq_align < __alignof__(struct io_cq))) io_cq 419 drivers/net/ethernet/amazon/ena/ena_com.c struct ena_com_io_cq *io_cq) io_cq 424 drivers/net/ethernet/amazon/ena/ena_com.c memset(&io_cq->cdesc_addr, 0x0, sizeof(io_cq->cdesc_addr)); io_cq 427 drivers/net/ethernet/amazon/ena/ena_com.c io_cq->cdesc_entry_size_in_bytes = io_cq 428 drivers/net/ethernet/amazon/ena/ena_com.c (io_cq->direction == ENA_COM_IO_QUEUE_DIRECTION_TX) ? io_cq 432 drivers/net/ethernet/amazon/ena/ena_com.c size = io_cq->cdesc_entry_size_in_bytes * io_cq->q_depth; io_cq 436 drivers/net/ethernet/amazon/ena/ena_com.c io_cq->cdesc_addr.virt_addr = io_cq 438 drivers/net/ethernet/amazon/ena/ena_com.c &io_cq->cdesc_addr.phys_addr, GFP_KERNEL); io_cq 440 drivers/net/ethernet/amazon/ena/ena_com.c if (!io_cq->cdesc_addr.virt_addr) { io_cq 441 drivers/net/ethernet/amazon/ena/ena_com.c io_cq->cdesc_addr.virt_addr = io_cq 443 drivers/net/ethernet/amazon/ena/ena_com.c &io_cq->cdesc_addr.phys_addr, io_cq 447 drivers/net/ethernet/amazon/ena/ena_com.c if (!io_cq->cdesc_addr.virt_addr) { io_cq 452 drivers/net/ethernet/amazon/ena/ena_com.c io_cq->phase = 1; io_cq 453 drivers/net/ethernet/amazon/ena/ena_com.c io_cq->head = 0; io_cq 913 drivers/net/ethernet/amazon/ena/ena_com.c struct ena_com_io_cq *io_cq) io_cq 917 drivers/net/ethernet/amazon/ena/ena_com.c if (io_cq->cdesc_addr.virt_addr) { io_cq 918 drivers/net/ethernet/amazon/ena/ena_com.c size = io_cq->cdesc_entry_size_in_bytes * io_cq->q_depth; io_cq 921 drivers/net/ethernet/amazon/ena/ena_com.c io_cq->cdesc_addr.virt_addr, io_cq 922 drivers/net/ethernet/amazon/ena/ena_com.c io_cq->cdesc_addr.phys_addr); io_cq 924 drivers/net/ethernet/amazon/ena/ena_com.c io_cq->cdesc_addr.virt_addr = NULL; io_cq 1360 drivers/net/ethernet/amazon/ena/ena_com.c struct ena_com_io_cq *io_cq) io_cq 1371 drivers/net/ethernet/amazon/ena/ena_com.c create_cmd.cq_caps_2 |= (io_cq->cdesc_entry_size_in_bytes / 4) & io_cq 1376 drivers/net/ethernet/amazon/ena/ena_com.c create_cmd.msix_vector = io_cq->msix_vector; io_cq 1377 drivers/net/ethernet/amazon/ena/ena_com.c create_cmd.cq_depth = io_cq->q_depth; io_cq 1381 drivers/net/ethernet/amazon/ena/ena_com.c io_cq->cdesc_addr.phys_addr); io_cq 1397 drivers/net/ethernet/amazon/ena/ena_com.c io_cq->idx = cmd_completion.cq_idx; io_cq 1399 drivers/net/ethernet/amazon/ena/ena_com.c io_cq->unmask_reg = (u32 __iomem *)((uintptr_t)ena_dev->reg_bar + io_cq 1403 drivers/net/ethernet/amazon/ena/ena_com.c io_cq->cq_head_db_reg = io_cq 1408 drivers/net/ethernet/amazon/ena/ena_com.c io_cq->numa_node_cfg_reg = io_cq 1412 drivers/net/ethernet/amazon/ena/ena_com.c pr_debug("created cq[%u], depth[%u]\n", io_cq->idx, io_cq->q_depth); io_cq 1419 drivers/net/ethernet/amazon/ena/ena_com.c struct ena_com_io_cq **io_cq) io_cq 1428 drivers/net/ethernet/amazon/ena/ena_com.c *io_cq = &ena_dev->io_cq_queues[qid]; io_cq 1468 drivers/net/ethernet/amazon/ena/ena_com.c struct ena_com_io_cq *io_cq) io_cq 1477 drivers/net/ethernet/amazon/ena/ena_com.c destroy_cmd.cq_idx = io_cq->idx; io_cq 1822 drivers/net/ethernet/amazon/ena/ena_com.c struct ena_com_io_cq *io_cq; io_cq 1832 drivers/net/ethernet/amazon/ena/ena_com.c io_cq = &ena_dev->io_cq_queues[ctx->qid]; io_cq 1835 drivers/net/ethernet/amazon/ena/ena_com.c memset(io_cq, 0x0, sizeof(*io_cq)); io_cq 1838 drivers/net/ethernet/amazon/ena/ena_com.c io_cq->q_depth = ctx->queue_size; io_cq 1839 drivers/net/ethernet/amazon/ena/ena_com.c io_cq->direction = ctx->direction; io_cq 1840 drivers/net/ethernet/amazon/ena/ena_com.c io_cq->qid = ctx->qid; io_cq 1842 drivers/net/ethernet/amazon/ena/ena_com.c io_cq->msix_vector = ctx->msix_vector; io_cq 1858 drivers/net/ethernet/amazon/ena/ena_com.c ret = ena_com_init_io_cq(ena_dev, ctx, io_cq); io_cq 1862 drivers/net/ethernet/amazon/ena/ena_com.c ret = ena_com_create_io_cq(ena_dev, io_cq); io_cq 1866 drivers/net/ethernet/amazon/ena/ena_com.c ret = ena_com_create_io_sq(ena_dev, io_sq, io_cq->idx); io_cq 1873 drivers/net/ethernet/amazon/ena/ena_com.c ena_com_destroy_io_cq(ena_dev, io_cq); io_cq 1875 drivers/net/ethernet/amazon/ena/ena_com.c ena_com_io_queue_free(ena_dev, io_sq, io_cq); io_cq 1882 drivers/net/ethernet/amazon/ena/ena_com.c struct ena_com_io_cq *io_cq; io_cq 1891 drivers/net/ethernet/amazon/ena/ena_com.c io_cq = &ena_dev->io_cq_queues[qid]; io_cq 1894 drivers/net/ethernet/amazon/ena/ena_com.c ena_com_destroy_io_cq(ena_dev, io_cq); io_cq 1896 drivers/net/ethernet/amazon/ena/ena_com.c ena_com_io_queue_free(ena_dev, io_sq, io_cq); io_cq 471 drivers/net/ethernet/amazon/ena/ena_com.h struct ena_com_io_cq **io_cq); io_cq 855 drivers/net/ethernet/amazon/ena/ena_com.h struct ena_com_io_cq *io_cq); io_cq 866 drivers/net/ethernet/amazon/ena/ena_com.h struct ena_com_io_cq *io_cq); io_cq 36 drivers/net/ethernet/amazon/ena/ena_eth_com.c struct ena_com_io_cq *io_cq) io_cq 42 drivers/net/ethernet/amazon/ena/ena_eth_com.c head_masked = io_cq->head & (io_cq->q_depth - 1); io_cq 43 drivers/net/ethernet/amazon/ena/ena_eth_com.c expected_phase = io_cq->phase; io_cq 45 drivers/net/ethernet/amazon/ena/ena_eth_com.c cdesc = (struct ena_eth_io_rx_cdesc_base *)(io_cq->cdesc_addr.virt_addr io_cq 46 drivers/net/ethernet/amazon/ena/ena_eth_com.c + (head_masked * io_cq->cdesc_entry_size_in_bytes)); io_cq 243 drivers/net/ethernet/amazon/ena/ena_eth_com.c ena_com_rx_cdesc_idx_to_ptr(struct ena_com_io_cq *io_cq, u16 idx) io_cq 245 drivers/net/ethernet/amazon/ena/ena_eth_com.c idx &= (io_cq->q_depth - 1); io_cq 247 drivers/net/ethernet/amazon/ena/ena_eth_com.c ((uintptr_t)io_cq->cdesc_addr.virt_addr + io_cq 248 drivers/net/ethernet/amazon/ena/ena_eth_com.c idx * io_cq->cdesc_entry_size_in_bytes); io_cq 251 drivers/net/ethernet/amazon/ena/ena_eth_com.c static u16 ena_com_cdesc_rx_pkt_get(struct ena_com_io_cq *io_cq, io_cq 259 drivers/net/ethernet/amazon/ena/ena_eth_com.c cdesc = ena_com_get_next_rx_cdesc(io_cq); io_cq 263 drivers/net/ethernet/amazon/ena/ena_eth_com.c ena_com_cq_inc_head(io_cq); io_cq 270 drivers/net/ethernet/amazon/ena/ena_eth_com.c *first_cdesc_idx = io_cq->cur_rx_pkt_cdesc_start_idx; io_cq 271 drivers/net/ethernet/amazon/ena/ena_eth_com.c count += io_cq->cur_rx_pkt_cdesc_count; io_cq 273 drivers/net/ethernet/amazon/ena/ena_eth_com.c head_masked = io_cq->head & (io_cq->q_depth - 1); io_cq 275 drivers/net/ethernet/amazon/ena/ena_eth_com.c io_cq->cur_rx_pkt_cdesc_count = 0; io_cq 276 drivers/net/ethernet/amazon/ena/ena_eth_com.c io_cq->cur_rx_pkt_cdesc_start_idx = head_masked; io_cq 279 drivers/net/ethernet/amazon/ena/ena_eth_com.c io_cq->qid, *first_cdesc_idx, count); io_cq 281 drivers/net/ethernet/amazon/ena/ena_eth_com.c io_cq->cur_rx_pkt_cdesc_count += count; io_cq 514 drivers/net/ethernet/amazon/ena/ena_eth_com.c int ena_com_rx_pkt(struct ena_com_io_cq *io_cq, io_cq 524 drivers/net/ethernet/amazon/ena/ena_eth_com.c WARN(io_cq->direction != ENA_COM_IO_QUEUE_DIRECTION_RX, "wrong Q type"); io_cq 526 drivers/net/ethernet/amazon/ena/ena_eth_com.c nb_hw_desc = ena_com_cdesc_rx_pkt_get(io_cq, &cdesc_idx); io_cq 532 drivers/net/ethernet/amazon/ena/ena_eth_com.c pr_debug("fetch rx packet: queue %d completed desc: %d\n", io_cq->qid, io_cq 542 drivers/net/ethernet/amazon/ena/ena_eth_com.c cdesc = ena_com_rx_cdesc_idx_to_ptr(io_cq, cdesc_idx + i); io_cq 595 drivers/net/ethernet/amazon/ena/ena_eth_com.c bool ena_com_cq_empty(struct ena_com_io_cq *io_cq) io_cq 599 drivers/net/ethernet/amazon/ena/ena_eth_com.c cdesc = ena_com_get_next_rx_cdesc(io_cq); io_cq 82 drivers/net/ethernet/amazon/ena/ena_eth_com.h int ena_com_rx_pkt(struct ena_com_io_cq *io_cq, io_cq 90 drivers/net/ethernet/amazon/ena/ena_eth_com.h bool ena_com_cq_empty(struct ena_com_io_cq *io_cq); io_cq 92 drivers/net/ethernet/amazon/ena/ena_eth_com.h static inline void ena_com_unmask_intr(struct ena_com_io_cq *io_cq, io_cq 95 drivers/net/ethernet/amazon/ena/ena_eth_com.h writel(intr_reg->intr_control, io_cq->unmask_reg); io_cq 193 drivers/net/ethernet/amazon/ena/ena_eth_com.h static inline int ena_com_update_dev_comp_head(struct ena_com_io_cq *io_cq) io_cq 198 drivers/net/ethernet/amazon/ena/ena_eth_com.h if (unlikely(io_cq->cq_head_db_reg)) { io_cq 199 drivers/net/ethernet/amazon/ena/ena_eth_com.h head = io_cq->head; io_cq 200 drivers/net/ethernet/amazon/ena/ena_eth_com.h unreported_comp = head - io_cq->last_head_update; io_cq 201 drivers/net/ethernet/amazon/ena/ena_eth_com.h need_update = unreported_comp > (io_cq->q_depth / ENA_COMP_HEAD_THRESH); io_cq 205 drivers/net/ethernet/amazon/ena/ena_eth_com.h io_cq->qid, head); io_cq 206 drivers/net/ethernet/amazon/ena/ena_eth_com.h writel(head, io_cq->cq_head_db_reg); io_cq 207 drivers/net/ethernet/amazon/ena/ena_eth_com.h io_cq->last_head_update = head; io_cq 214 drivers/net/ethernet/amazon/ena/ena_eth_com.h static inline void ena_com_update_numa_node(struct ena_com_io_cq *io_cq, io_cq 219 drivers/net/ethernet/amazon/ena/ena_eth_com.h if (!io_cq->numa_node_cfg_reg) io_cq 225 drivers/net/ethernet/amazon/ena/ena_eth_com.h writel(numa_cfg.numa_cfg, io_cq->numa_node_cfg_reg); io_cq 233 drivers/net/ethernet/amazon/ena/ena_eth_com.h static inline void ena_com_cq_inc_head(struct ena_com_io_cq *io_cq) io_cq 235 drivers/net/ethernet/amazon/ena/ena_eth_com.h io_cq->head++; io_cq 238 drivers/net/ethernet/amazon/ena/ena_eth_com.h if (unlikely((io_cq->head & (io_cq->q_depth - 1)) == 0)) io_cq 239 drivers/net/ethernet/amazon/ena/ena_eth_com.h io_cq->phase ^= 1; io_cq 242 drivers/net/ethernet/amazon/ena/ena_eth_com.h static inline int ena_com_tx_comp_req_id_get(struct ena_com_io_cq *io_cq, io_cq 249 drivers/net/ethernet/amazon/ena/ena_eth_com.h masked_head = io_cq->head & (io_cq->q_depth - 1); io_cq 250 drivers/net/ethernet/amazon/ena/ena_eth_com.h expected_phase = io_cq->phase; io_cq 253 drivers/net/ethernet/amazon/ena/ena_eth_com.h ((uintptr_t)io_cq->cdesc_addr.virt_addr + io_cq 254 drivers/net/ethernet/amazon/ena/ena_eth_com.h (masked_head * io_cq->cdesc_entry_size_in_bytes)); io_cq 267 drivers/net/ethernet/amazon/ena/ena_eth_com.h if (unlikely(*req_id >= io_cq->q_depth)) { io_cq 272 drivers/net/ethernet/amazon/ena/ena_eth_com.h ena_com_cq_inc_head(io_cq); io_cq 3763 drivers/scsi/lpfc/lpfc_debugfs.c qp = phba->sli4_hba.hdwq[eqidx].io_cq; io_cq 4226 drivers/scsi/lpfc/lpfc_debugfs.c qp = phba->sli4_hba.hdwq[qidx].io_cq; io_cq 463 drivers/scsi/lpfc/lpfc_debugfs.h cq = phba->sli4_hba.hdwq[wqidx].io_cq; io_cq 640 drivers/scsi/lpfc/lpfc_debugfs.h if (phba->sli4_hba.hdwq[cq_idx].io_cq->queue_id == qid) io_cq 645 drivers/scsi/lpfc/lpfc_debugfs.h lpfc_debug_dump_q(phba->sli4_hba.hdwq[cq_idx].io_cq); io_cq 8678 drivers/scsi/lpfc/lpfc_init.c phba->sli4_hba.hdwq[idx].io_cq = qdesc; io_cq 9125 drivers/scsi/lpfc/lpfc_init.c lpfc_sli4_queue_free(hdwq[idx].io_cq); io_cq 9127 drivers/scsi/lpfc/lpfc_init.c hdwq[idx].io_cq = NULL; io_cq 9466 drivers/scsi/lpfc/lpfc_init.c qp[qidx].io_cq, io_cq 9775 drivers/scsi/lpfc/lpfc_init.c lpfc_cq_destroy(phba, qp->io_cq); io_cq 5567 drivers/scsi/lpfc/lpfc_sli.c sli4_hba->sli4_write_cq_db(phba, qp->io_cq, 0, io_cq 660 drivers/scsi/lpfc/lpfc_sli4.h struct lpfc_queue *io_cq; /* Fast-path FCP & NVME compl queue */ io_cq 184 include/linux/blkdev.h struct io_cq *icq; io_cq 10 include/linux/elevator.h struct io_cq; io_cq 51 include/linux/elevator.h void (*init_icq)(struct io_cq *); io_cq 52 include/linux/elevator.h void (*exit_icq)(struct io_cq *); io_cq 116 include/linux/iocontext.h struct io_cq __rcu *icq_hint;