qdma 1132 drivers/crypto/hisilicon/qm.c qp->qdma.size = qm->sqe_size * QM_Q_DEPTH + qdma 1134 drivers/crypto/hisilicon/qm.c qp->qdma.va = dma_alloc_coherent(dev, qp->qdma.size, qdma 1135 drivers/crypto/hisilicon/qm.c &qp->qdma.dma, GFP_KERNEL); qdma 1136 drivers/crypto/hisilicon/qm.c if (!qp->qdma.va) { qdma 1142 drivers/crypto/hisilicon/qm.c qp->qdma.va, &qp->qdma.dma, qp->qdma.size); qdma 1159 drivers/crypto/hisilicon/qm.c dma_free_coherent(dev, qp->qdma.size, qp->qdma.va, qdma 1160 drivers/crypto/hisilicon/qm.c qp->qdma.dma); qdma 1181 drivers/crypto/hisilicon/qm.c struct qm_dma *qdma = &qp->qdma; qdma 1184 drivers/crypto/hisilicon/qm.c if (qm->use_dma_api && qdma->va) qdma 1185 drivers/crypto/hisilicon/qm.c dma_free_coherent(dev, qdma->size, qdma->va, qdma->dma); qdma 1282 drivers/crypto/hisilicon/qm.c (qp)->type = ((qp)->qdma.va + (off)); \ qdma 1283 drivers/crypto/hisilicon/qm.c (qp)->type##_dma = (qp)->qdma.dma + (off); \ qdma 1287 drivers/crypto/hisilicon/qm.c if (!qp->qdma.dma) { qdma 1293 drivers/crypto/hisilicon/qm.c if (qp->qdma.dma & QM_SQE_DATA_ALIGN_MASK) { qdma 1489 drivers/crypto/hisilicon/qm.c if (qm->use_dma_api && qm->qdma.va) { qdma 1491 drivers/crypto/hisilicon/qm.c dma_free_coherent(dev, qm->qdma.size, qdma 1492 drivers/crypto/hisilicon/qm.c qm->qdma.va, qm->qdma.dma); qdma 1493 drivers/crypto/hisilicon/qm.c memset(&qm->qdma, 0, sizeof(qm->qdma)); qdma 1628 drivers/crypto/hisilicon/qm.c (qm)->type = ((qm)->qdma.va + (off)); \ qdma 1629 drivers/crypto/hisilicon/qm.c (qm)->type##_dma = (qm)->qdma.dma + (off); \ qdma 1633 drivers/crypto/hisilicon/qm.c WARN_ON(!qm->qdma.dma); qdma 1711 drivers/crypto/hisilicon/qm.c } else if (!qm->qdma.va) { qdma 1712 drivers/crypto/hisilicon/qm.c qm->qdma.size = QMC_ALIGN(sizeof(struct qm_eqe) * QM_Q_DEPTH) + qdma 1716 drivers/crypto/hisilicon/qm.c qm->qdma.va = dma_alloc_coherent(dev, qm->qdma.size, qdma 1717 drivers/crypto/hisilicon/qm.c &qm->qdma.dma, GFP_KERNEL); qdma 1719 drivers/crypto/hisilicon/qm.c qm->qdma.va, &qm->qdma.dma, qm->qdma.size); qdma 1720 drivers/crypto/hisilicon/qm.c if (!qm->qdma.va) qdma 137 drivers/crypto/hisilicon/qm.h struct qm_dma qdma; qdma 182 drivers/crypto/hisilicon/qm.h struct qm_dma qdma; qdma 168 drivers/dma/fsl-qdma.c struct fsl_qdma_engine *qdma; qdma 271 drivers/dma/fsl-qdma.c static u32 qdma_readl(struct fsl_qdma_engine *qdma, void __iomem *addr) qdma 273 drivers/dma/fsl-qdma.c return FSL_DMA_IN(qdma, addr, 32); qdma 276 drivers/dma/fsl-qdma.c static void qdma_writel(struct fsl_qdma_engine *qdma, u32 val, qdma 279 drivers/dma/fsl-qdma.c FSL_DMA_OUT(qdma, addr, val, 32); qdma 296 drivers/dma/fsl-qdma.c struct fsl_qdma_engine *fsl_qdma = fsl_chan->qdma; qdma 945 drivers/dma/fsl-qdma.c reg = qdma_readl(fsl_chan->qdma, block + FSL_QDMA_BCQSR(fsl_queue->id)); qdma 961 drivers/dma/fsl-qdma.c reg = qdma_readl(fsl_chan->qdma, block + FSL_QDMA_BCQMR(fsl_queue->id)); qdma 963 drivers/dma/fsl-qdma.c qdma_writel(fsl_chan->qdma, reg, block + FSL_QDMA_BCQMR(fsl_queue->id)); qdma 1019 drivers/dma/fsl-qdma.c struct fsl_qdma_engine *fsl_qdma = fsl_chan->qdma; qdma 1167 drivers/dma/fsl-qdma.c fsl_chan->qdma = fsl_qdma;