qdma             1132 drivers/crypto/hisilicon/qm.c 		qp->qdma.size = qm->sqe_size * QM_Q_DEPTH +
qdma             1134 drivers/crypto/hisilicon/qm.c 		qp->qdma.va = dma_alloc_coherent(dev, qp->qdma.size,
qdma             1135 drivers/crypto/hisilicon/qm.c 						 &qp->qdma.dma, GFP_KERNEL);
qdma             1136 drivers/crypto/hisilicon/qm.c 		if (!qp->qdma.va) {
qdma             1142 drivers/crypto/hisilicon/qm.c 			qp->qdma.va, &qp->qdma.dma, qp->qdma.size);
qdma             1159 drivers/crypto/hisilicon/qm.c 		dma_free_coherent(dev, qp->qdma.size, qp->qdma.va,
qdma             1160 drivers/crypto/hisilicon/qm.c 				  qp->qdma.dma);
qdma             1181 drivers/crypto/hisilicon/qm.c 	struct qm_dma *qdma = &qp->qdma;
qdma             1184 drivers/crypto/hisilicon/qm.c 	if (qm->use_dma_api && qdma->va)
qdma             1185 drivers/crypto/hisilicon/qm.c 		dma_free_coherent(dev, qdma->size, qdma->va, qdma->dma);
qdma             1282 drivers/crypto/hisilicon/qm.c 	(qp)->type = ((qp)->qdma.va + (off)); \
qdma             1283 drivers/crypto/hisilicon/qm.c 	(qp)->type##_dma = (qp)->qdma.dma + (off); \
qdma             1287 drivers/crypto/hisilicon/qm.c 	if (!qp->qdma.dma) {
qdma             1293 drivers/crypto/hisilicon/qm.c 	if (qp->qdma.dma & QM_SQE_DATA_ALIGN_MASK) {
qdma             1489 drivers/crypto/hisilicon/qm.c 	if (qm->use_dma_api && qm->qdma.va) {
qdma             1491 drivers/crypto/hisilicon/qm.c 		dma_free_coherent(dev, qm->qdma.size,
qdma             1492 drivers/crypto/hisilicon/qm.c 				  qm->qdma.va, qm->qdma.dma);
qdma             1493 drivers/crypto/hisilicon/qm.c 		memset(&qm->qdma, 0, sizeof(qm->qdma));
qdma             1628 drivers/crypto/hisilicon/qm.c 	(qm)->type = ((qm)->qdma.va + (off)); \
qdma             1629 drivers/crypto/hisilicon/qm.c 	(qm)->type##_dma = (qm)->qdma.dma + (off); \
qdma             1633 drivers/crypto/hisilicon/qm.c 	WARN_ON(!qm->qdma.dma);
qdma             1711 drivers/crypto/hisilicon/qm.c 	} else if (!qm->qdma.va) {
qdma             1712 drivers/crypto/hisilicon/qm.c 		qm->qdma.size = QMC_ALIGN(sizeof(struct qm_eqe) * QM_Q_DEPTH) +
qdma             1716 drivers/crypto/hisilicon/qm.c 		qm->qdma.va = dma_alloc_coherent(dev, qm->qdma.size,
qdma             1717 drivers/crypto/hisilicon/qm.c 						 &qm->qdma.dma, GFP_KERNEL);
qdma             1719 drivers/crypto/hisilicon/qm.c 			qm->qdma.va, &qm->qdma.dma, qm->qdma.size);
qdma             1720 drivers/crypto/hisilicon/qm.c 		if (!qm->qdma.va)
qdma              137 drivers/crypto/hisilicon/qm.h 	struct qm_dma qdma;
qdma              182 drivers/crypto/hisilicon/qm.h 	struct qm_dma qdma;
qdma              168 drivers/dma/fsl-qdma.c 	struct fsl_qdma_engine		*qdma;
qdma              271 drivers/dma/fsl-qdma.c static u32 qdma_readl(struct fsl_qdma_engine *qdma, void __iomem *addr)
qdma              273 drivers/dma/fsl-qdma.c 	return FSL_DMA_IN(qdma, addr, 32);
qdma              276 drivers/dma/fsl-qdma.c static void qdma_writel(struct fsl_qdma_engine *qdma, u32 val,
qdma              279 drivers/dma/fsl-qdma.c 	FSL_DMA_OUT(qdma, addr, val, 32);
qdma              296 drivers/dma/fsl-qdma.c 	struct fsl_qdma_engine *fsl_qdma = fsl_chan->qdma;
qdma              945 drivers/dma/fsl-qdma.c 	reg = qdma_readl(fsl_chan->qdma, block + FSL_QDMA_BCQSR(fsl_queue->id));
qdma              961 drivers/dma/fsl-qdma.c 	reg = qdma_readl(fsl_chan->qdma, block + FSL_QDMA_BCQMR(fsl_queue->id));
qdma              963 drivers/dma/fsl-qdma.c 	qdma_writel(fsl_chan->qdma, reg, block + FSL_QDMA_BCQMR(fsl_queue->id));
qdma             1019 drivers/dma/fsl-qdma.c 	struct fsl_qdma_engine *fsl_qdma = fsl_chan->qdma;
qdma             1167 drivers/dma/fsl-qdma.c 		fsl_chan->qdma = fsl_qdma;