bufers 89 drivers/crypto/qat/qat_common/qat_algs.c struct qat_alg_buf bufers[]; bufers 688 drivers/crypto/qat/qat_common/qat_algs.c dma_unmap_single(dev, bl->bufers[i].addr, bufers 689 drivers/crypto/qat/qat_common/qat_algs.c bl->bufers[i].len, DMA_BIDIRECTIONAL); bufers 698 drivers/crypto/qat/qat_common/qat_algs.c dma_unmap_single(dev, blout->bufers[i].addr, bufers 699 drivers/crypto/qat/qat_common/qat_algs.c blout->bufers[i].len, bufers 720 drivers/crypto/qat/qat_common/qat_algs.c size_t sz_out, sz = struct_size(bufl, bufers, n + 1); bufers 740 drivers/crypto/qat/qat_common/qat_algs.c bufl->bufers[y].addr = dma_map_single(dev, sg_virt(sg), bufers 743 drivers/crypto/qat/qat_common/qat_algs.c bufl->bufers[y].len = sg->length; bufers 744 drivers/crypto/qat/qat_common/qat_algs.c if (unlikely(dma_mapping_error(dev, bufl->bufers[y].addr))) bufers 754 drivers/crypto/qat/qat_common/qat_algs.c struct qat_alg_buf *bufers; bufers 757 drivers/crypto/qat/qat_common/qat_algs.c sz_out = struct_size(buflout, bufers, n + 1); bufers 766 drivers/crypto/qat/qat_common/qat_algs.c bufers = buflout->bufers; bufers 773 drivers/crypto/qat/qat_common/qat_algs.c bufers[y].addr = dma_map_single(dev, sg_virt(sg), bufers 776 drivers/crypto/qat/qat_common/qat_algs.c if (unlikely(dma_mapping_error(dev, bufers[y].addr))) bufers 778 drivers/crypto/qat/qat_common/qat_algs.c bufers[y].len = sg->length; bufers 796 drivers/crypto/qat/qat_common/qat_algs.c if (!dma_mapping_error(dev, buflout->bufers[i].addr)) bufers 797 drivers/crypto/qat/qat_common/qat_algs.c dma_unmap_single(dev, buflout->bufers[i].addr, bufers 798 drivers/crypto/qat/qat_common/qat_algs.c buflout->bufers[i].len, bufers 807 drivers/crypto/qat/qat_common/qat_algs.c if (!dma_mapping_error(dev, bufl->bufers[i].addr)) bufers 808 drivers/crypto/qat/qat_common/qat_algs.c dma_unmap_single(dev, bufl->bufers[i].addr, bufers 809 drivers/crypto/qat/qat_common/qat_algs.c bufl->bufers[i].len,