sg_src            283 arch/arm/crypto/aes-ce-glue.c 	struct scatterlist sg_src[2], sg_dst[2];
sg_src            311 arch/arm/crypto/aes-ce-glue.c 		dst = src = scatterwalk_ffwd(sg_src, req->src, subreq.cryptlen);
sg_src            341 arch/arm/crypto/aes-ce-glue.c 	struct scatterlist sg_src[2], sg_dst[2];
sg_src            369 arch/arm/crypto/aes-ce-glue.c 		dst = src = scatterwalk_ffwd(sg_src, req->src, subreq.cryptlen);
sg_src            460 arch/arm/crypto/aes-ce-glue.c 	struct scatterlist sg_src[2], sg_dst[2];
sg_src            506 arch/arm/crypto/aes-ce-glue.c 	dst = src = scatterwalk_ffwd(sg_src, req->src, req->cryptlen);
sg_src            532 arch/arm/crypto/aes-ce-glue.c 	struct scatterlist sg_src[2], sg_dst[2];
sg_src            578 arch/arm/crypto/aes-ce-glue.c 	dst = src = scatterwalk_ffwd(sg_src, req->src, req->cryptlen);
sg_src            296 arch/arm64/crypto/aes-glue.c 	struct scatterlist sg_src[2], sg_dst[2];
sg_src            323 arch/arm64/crypto/aes-glue.c 		dst = src = scatterwalk_ffwd(sg_src, req->src, subreq.cryptlen);
sg_src            353 arch/arm64/crypto/aes-glue.c 	struct scatterlist sg_src[2], sg_dst[2];
sg_src            380 arch/arm64/crypto/aes-glue.c 		dst = src = scatterwalk_ffwd(sg_src, req->src, subreq.cryptlen);
sg_src            531 arch/arm64/crypto/aes-glue.c 	struct scatterlist sg_src[2], sg_dst[2];
sg_src            577 arch/arm64/crypto/aes-glue.c 	dst = src = scatterwalk_ffwd(sg_src, req->src, req->cryptlen);
sg_src            603 arch/arm64/crypto/aes-glue.c 	struct scatterlist sg_src[2], sg_dst[2];
sg_src            649 arch/arm64/crypto/aes-glue.c 	dst = src = scatterwalk_ffwd(sg_src, req->src, req->cryptlen);
sg_src            324 arch/arm64/crypto/aes-neonbs-glue.c 	struct scatterlist sg_src[2], sg_dst[2];
sg_src            394 arch/arm64/crypto/aes-neonbs-glue.c 	dst = src = scatterwalk_ffwd(sg_src, req->src, req->cryptlen);
sg_src             70 arch/x86/crypto/aegis128-aesni-glue.c 		struct aegis_state *state, struct scatterlist *sg_src,
sg_src             77 arch/x86/crypto/aegis128-aesni-glue.c 	scatterwalk_start(&walk, sg_src);
sg_src            282 crypto/aegis128-core.c 				       struct scatterlist *sg_src,
sg_src            289 crypto/aegis128-core.c 	scatterwalk_start(&walk, sg_src);
sg_src            150 drivers/crypto/nx/nx.c 				struct scatterlist *sg_src,
sg_src            161 drivers/crypto/nx/nx.c 		scatterwalk_start(&walk, sg_src);
sg_src            163 drivers/crypto/nx/nx.c 		if (start < offset + sg_src->length)
sg_src            166 drivers/crypto/nx/nx.c 		offset += sg_src->length;
sg_src            167 drivers/crypto/nx/nx.c 		sg_src = sg_next(sg_src);
sg_src             67 drivers/crypto/rockchip/rk3288_crypto.c static int check_alignment(struct scatterlist *sg_src,
sg_src             73 drivers/crypto/rockchip/rk3288_crypto.c 	in = IS_ALIGNED((uint32_t)sg_src->offset, 4) &&
sg_src             74 drivers/crypto/rockchip/rk3288_crypto.c 	     IS_ALIGNED((uint32_t)sg_src->length, align_mask);
sg_src             81 drivers/crypto/rockchip/rk3288_crypto.c 	return (align && (sg_src->length == sg_dst->length));
sg_src             85 drivers/crypto/rockchip/rk3288_crypto.c 			struct scatterlist *sg_src,
sg_src             91 drivers/crypto/rockchip/rk3288_crypto.c 		check_alignment(sg_src, sg_dst, dev->align_size) :
sg_src             94 drivers/crypto/rockchip/rk3288_crypto.c 		count = min(dev->left_bytes, sg_src->length);
sg_src             97 drivers/crypto/rockchip/rk3288_crypto.c 		if (!dma_map_sg(dev->dev, sg_src, 1, DMA_TO_DEVICE)) {
sg_src            102 drivers/crypto/rockchip/rk3288_crypto.c 		dev->addr_in = sg_dma_address(sg_src);
sg_src            109 drivers/crypto/rockchip/rk3288_crypto.c 				dma_unmap_sg(dev->dev, sg_src, 1,
sg_src            156 drivers/crypto/rockchip/rk3288_crypto.c 	sg_in = dev->aligned ? dev->sg_src : &dev->sg_tmp;
sg_src            202 drivers/crypto/rockchip/rk3288_crypto.h 	struct scatterlist		*sg_src;
sg_src            223 drivers/crypto/rockchip/rk3288_crypto.h 			 struct scatterlist *sg_src,
sg_src            252 drivers/crypto/rockchip/rk3288_crypto_ablkcipher.c 	u8 *src_last_blk = page_address(sg_page(dev->sg_src)) +
sg_src            253 drivers/crypto/rockchip/rk3288_crypto_ablkcipher.c 		dev->sg_src->offset + dev->sg_src->length - ivsize;
sg_src            264 drivers/crypto/rockchip/rk3288_crypto_ablkcipher.c 	err = dev->load_data(dev, dev->sg_src, dev->sg_dst);
sg_src            279 drivers/crypto/rockchip/rk3288_crypto_ablkcipher.c 	dev->sg_src = req->src;
sg_src            358 drivers/crypto/rockchip/rk3288_crypto_ablkcipher.c 			if (sg_is_last(dev->sg_src)) {
sg_src            364 drivers/crypto/rockchip/rk3288_crypto_ablkcipher.c 			dev->sg_src = sg_next(dev->sg_src);
sg_src            187 drivers/crypto/rockchip/rk3288_crypto_ahash.c 	err = dev->load_data(dev, dev->sg_src, NULL);
sg_src            204 drivers/crypto/rockchip/rk3288_crypto_ahash.c 	dev->sg_src = req->src;
sg_src            238 drivers/crypto/rockchip/rk3288_crypto_ahash.c 			if (sg_is_last(dev->sg_src)) {
sg_src            244 drivers/crypto/rockchip/rk3288_crypto_ahash.c 			dev->sg_src = sg_next(dev->sg_src);
sg_src            308 drivers/crypto/s5p-sss.c 	struct scatterlist		*sg_src;
sg_src            513 drivers/crypto/s5p-sss.c 	dma_unmap_sg(dev->dev, dev->sg_src, 1, DMA_TO_DEVICE);
sg_src            563 drivers/crypto/s5p-sss.c 	dev->sg_src = sg;
sg_src            603 drivers/crypto/s5p-sss.c 	if (!sg_is_last(dev->sg_src)) {
sg_src            604 drivers/crypto/s5p-sss.c 		ret = s5p_set_indata(dev, sg_next(dev->sg_src));
sg_src            748 drivers/crypto/s5p-sss.c 			s5p_set_dma_indata(dev, dev->sg_src);
sg_src           1983 drivers/crypto/s5p-sss.c 	s5p_set_dma_indata(dev,  dev->sg_src);
sg_src            215 drivers/crypto/ux500/cryp/cryp.h 	struct scatterlist *sg_src;
sg_src            540 drivers/crypto/ux500/cryp/cryp_core.c 		ctx->device->dma.sg_src = sg;
sg_src            542 drivers/crypto/ux500/cryp/cryp_core.c 						 ctx->device->dma.sg_src,
sg_src            557 drivers/crypto/ux500/cryp/cryp_core.c 				ctx->device->dma.sg_src,
sg_src            617 drivers/crypto/ux500/cryp/cryp_core.c 	dma_unmap_sg(chan->device->dev, ctx->device->dma.sg_src,
sg_src           2111 drivers/dma/ste_dma40.c 		struct scatterlist *sg_src, struct scatterlist *sg_dst,
sg_src           2120 drivers/dma/ste_dma40.c 	ret = d40_log_sg_to_lli(sg_src, sg_len,
sg_src           2139 drivers/dma/ste_dma40.c 		struct scatterlist *sg_src, struct scatterlist *sg_dst,
sg_src           2152 drivers/dma/ste_dma40.c 	ret = d40_phy_sg_to_lli(sg_src, sg_len, src_dev_addr,
sg_src           2209 drivers/dma/ste_dma40.c d40_prep_sg(struct dma_chan *dchan, struct scatterlist *sg_src,
sg_src           2229 drivers/dma/ste_dma40.c 	desc = d40_prep_desc(chan, sg_src, sg_len, dma_flags);
sg_src           2233 drivers/dma/ste_dma40.c 	if (sg_next(&sg_src[sg_len - 1]) == sg_src)
sg_src           2244 drivers/dma/ste_dma40.c 		ret = d40_prep_sg_log(chan, desc, sg_src, sg_dst,
sg_src           2247 drivers/dma/ste_dma40.c 		ret = d40_prep_sg_phy(chan, desc, sg_src, sg_dst,