kcq 452 block/kyber-iosched.c static void kyber_ctx_queue_init(struct kyber_ctx_queue *kcq) kcq 456 block/kyber-iosched.c spin_lock_init(&kcq->lock); kcq 458 block/kyber-iosched.c INIT_LIST_HEAD(&kcq->rq_list[i]); kcq 570 block/kyber-iosched.c struct kyber_ctx_queue *kcq = &khd->kcqs[ctx->index_hw[hctx->type]]; kcq 572 block/kyber-iosched.c struct list_head *rq_list = &kcq->rq_list[sched_domain]; kcq 575 block/kyber-iosched.c spin_lock(&kcq->lock); kcq 577 block/kyber-iosched.c spin_unlock(&kcq->lock); kcq 595 block/kyber-iosched.c struct kyber_ctx_queue *kcq = &khd->kcqs[rq->mq_ctx->index_hw[hctx->type]]; kcq 596 block/kyber-iosched.c struct list_head *head = &kcq->rq_list[sched_domain]; kcq 598 block/kyber-iosched.c spin_lock(&kcq->lock); kcq 606 block/kyber-iosched.c spin_unlock(&kcq->lock); kcq 666 block/kyber-iosched.c struct kyber_ctx_queue *kcq = &flush_data->khd->kcqs[bitnr]; kcq 668 block/kyber-iosched.c spin_lock(&kcq->lock); kcq 669 block/kyber-iosched.c list_splice_tail_init(&kcq->rq_list[flush_data->sched_domain], kcq 672 block/kyber-iosched.c spin_unlock(&kcq->lock); kcq 988 drivers/net/ethernet/broadcom/cnic.c struct kcqe **kcq; kcq 997 drivers/net/ethernet/broadcom/cnic.c kcq = (struct kcqe **) info->dma.pg_arr; kcq 998 drivers/net/ethernet/broadcom/cnic.c info->kcq = kcq; kcq 1010 drivers/net/ethernet/broadcom/cnic.c (struct bnx2x_bd_chain_next *) &kcq[i][MAX_KCQE_CNT]; kcq 2899 drivers/net/ethernet/broadcom/cnic.c kcqe = &info->kcq[KCQ_PG(ri)][KCQ_IDX(ri)]; kcq 3032 drivers/net/ethernet/broadcom/cnic.c prefetch(&cp->kcq1.kcq[KCQ_PG(prod)][KCQ_IDX(prod)]); kcq 179 drivers/net/ethernet/broadcom/cnic.h struct kcqe **kcq; kcq 511 drivers/scsi/bnx2fc/bnx2fc.h void bnx2fc_indicate_kcqe(void *context, struct kcqe *kcq[], kcq 1303 drivers/scsi/bnx2fc/bnx2fc_hwi.c void bnx2fc_indicate_kcqe(void *context, struct kcqe *kcq[], kcq 1311 drivers/scsi/bnx2fc/bnx2fc_hwi.c kcqe = (struct fcoe_kcqe *) kcq[i++];