mapped_nents     3472 drivers/crypto/caam/caamalg_qi2.c 	int src_nents, mapped_nents, qm_sg_bytes, qm_sg_src_index;
mapped_nents     3491 drivers/crypto/caam/caamalg_qi2.c 			mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents,
mapped_nents     3493 drivers/crypto/caam/caamalg_qi2.c 			if (!mapped_nents) {
mapped_nents     3498 drivers/crypto/caam/caamalg_qi2.c 			mapped_nents = 0;
mapped_nents     3511 drivers/crypto/caam/caamalg_qi2.c 		qm_sg_bytes = pad_sg_nents(qm_sg_src_index + mapped_nents) *
mapped_nents     3524 drivers/crypto/caam/caamalg_qi2.c 		if (mapped_nents) {
mapped_nents     3668 drivers/crypto/caam/caamalg_qi2.c 	int src_nents, mapped_nents;
mapped_nents     3681 drivers/crypto/caam/caamalg_qi2.c 		mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents,
mapped_nents     3683 drivers/crypto/caam/caamalg_qi2.c 		if (!mapped_nents) {
mapped_nents     3688 drivers/crypto/caam/caamalg_qi2.c 		mapped_nents = 0;
mapped_nents     3700 drivers/crypto/caam/caamalg_qi2.c 	qm_sg_bytes = pad_sg_nents(qm_sg_src_index + mapped_nents) *
mapped_nents     3761 drivers/crypto/caam/caamalg_qi2.c 	int src_nents, mapped_nents;
mapped_nents     3774 drivers/crypto/caam/caamalg_qi2.c 		mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents,
mapped_nents     3776 drivers/crypto/caam/caamalg_qi2.c 		if (!mapped_nents) {
mapped_nents     3781 drivers/crypto/caam/caamalg_qi2.c 		mapped_nents = 0;
mapped_nents     3794 drivers/crypto/caam/caamalg_qi2.c 	if (mapped_nents > 1) {
mapped_nents     3798 drivers/crypto/caam/caamalg_qi2.c 		qm_sg_bytes = pad_sg_nents(mapped_nents) * sizeof(*sg_table);
mapped_nents     3933 drivers/crypto/caam/caamalg_qi2.c 	int qm_sg_bytes, src_nents, mapped_nents;
mapped_nents     3951 drivers/crypto/caam/caamalg_qi2.c 			mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents,
mapped_nents     3953 drivers/crypto/caam/caamalg_qi2.c 			if (!mapped_nents) {
mapped_nents     3958 drivers/crypto/caam/caamalg_qi2.c 			mapped_nents = 0;
mapped_nents     3970 drivers/crypto/caam/caamalg_qi2.c 		qm_sg_bytes = pad_sg_nents(1 + mapped_nents) *
mapped_nents     4059 drivers/crypto/caam/caamalg_qi2.c 	int qm_sg_bytes, src_nents, mapped_nents;
mapped_nents     4072 drivers/crypto/caam/caamalg_qi2.c 		mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents,
mapped_nents     4074 drivers/crypto/caam/caamalg_qi2.c 		if (!mapped_nents) {
mapped_nents     4079 drivers/crypto/caam/caamalg_qi2.c 		mapped_nents = 0;
mapped_nents     4090 drivers/crypto/caam/caamalg_qi2.c 	qm_sg_bytes = pad_sg_nents(2 + mapped_nents) * sizeof(*sg_table);
mapped_nents     4157 drivers/crypto/caam/caamalg_qi2.c 	int src_nents, mapped_nents;
mapped_nents     4176 drivers/crypto/caam/caamalg_qi2.c 			mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents,
mapped_nents     4178 drivers/crypto/caam/caamalg_qi2.c 			if (!mapped_nents) {
mapped_nents     4183 drivers/crypto/caam/caamalg_qi2.c 			mapped_nents = 0;
mapped_nents     4201 drivers/crypto/caam/caamalg_qi2.c 		if (mapped_nents > 1) {
mapped_nents     4205 drivers/crypto/caam/caamalg_qi2.c 			qm_sg_bytes = pad_sg_nents(mapped_nents) *
mapped_nents      796 drivers/crypto/caam/caamhash.c 	int src_nents, mapped_nents, sec4_sg_bytes, sec4_sg_src_index;
mapped_nents      826 drivers/crypto/caam/caamhash.c 			mapped_nents = dma_map_sg(jrdev, req->src, src_nents,
mapped_nents      828 drivers/crypto/caam/caamhash.c 			if (!mapped_nents) {
mapped_nents      833 drivers/crypto/caam/caamhash.c 			mapped_nents = 0;
mapped_nents      837 drivers/crypto/caam/caamhash.c 		pad_nents = pad_sg_nents(sec4_sg_src_index + mapped_nents);
mapped_nents      863 drivers/crypto/caam/caamhash.c 		if (mapped_nents)
mapped_nents      997 drivers/crypto/caam/caamhash.c 	int src_nents, mapped_nents;
mapped_nents     1009 drivers/crypto/caam/caamhash.c 		mapped_nents = dma_map_sg(jrdev, req->src, src_nents,
mapped_nents     1011 drivers/crypto/caam/caamhash.c 		if (!mapped_nents) {
mapped_nents     1016 drivers/crypto/caam/caamhash.c 		mapped_nents = 0;
mapped_nents     1022 drivers/crypto/caam/caamhash.c 	edesc = ahash_edesc_alloc(ctx, sec4_sg_src_index + mapped_nents,
mapped_nents     1043 drivers/crypto/caam/caamhash.c 	ret = ahash_edesc_add_src(ctx, edesc, req, mapped_nents,
mapped_nents     1076 drivers/crypto/caam/caamhash.c 	int src_nents, mapped_nents;
mapped_nents     1089 drivers/crypto/caam/caamhash.c 		mapped_nents = dma_map_sg(jrdev, req->src, src_nents,
mapped_nents     1091 drivers/crypto/caam/caamhash.c 		if (!mapped_nents) {
mapped_nents     1096 drivers/crypto/caam/caamhash.c 		mapped_nents = 0;
mapped_nents     1100 drivers/crypto/caam/caamhash.c 	edesc = ahash_edesc_alloc(ctx, mapped_nents > 1 ? mapped_nents : 0,
mapped_nents     1110 drivers/crypto/caam/caamhash.c 	ret = ahash_edesc_add_src(ctx, edesc, req, mapped_nents, 0, 0,
mapped_nents     1216 drivers/crypto/caam/caamhash.c 	int sec4_sg_bytes, src_nents, mapped_nents;
mapped_nents     1246 drivers/crypto/caam/caamhash.c 			mapped_nents = dma_map_sg(jrdev, req->src, src_nents,
mapped_nents     1248 drivers/crypto/caam/caamhash.c 			if (!mapped_nents) {
mapped_nents     1253 drivers/crypto/caam/caamhash.c 			mapped_nents = 0;
mapped_nents     1256 drivers/crypto/caam/caamhash.c 		pad_nents = pad_sg_nents(1 + mapped_nents);
mapped_nents     1347 drivers/crypto/caam/caamhash.c 	int sec4_sg_bytes, sec4_sg_src_index, src_nents, mapped_nents;
mapped_nents     1359 drivers/crypto/caam/caamhash.c 		mapped_nents = dma_map_sg(jrdev, req->src, src_nents,
mapped_nents     1361 drivers/crypto/caam/caamhash.c 		if (!mapped_nents) {
mapped_nents     1366 drivers/crypto/caam/caamhash.c 		mapped_nents = 0;
mapped_nents     1370 drivers/crypto/caam/caamhash.c 	sec4_sg_bytes = (sec4_sg_src_index + mapped_nents) *
mapped_nents     1374 drivers/crypto/caam/caamhash.c 	edesc = ahash_edesc_alloc(ctx, sec4_sg_src_index + mapped_nents,
mapped_nents     1391 drivers/crypto/caam/caamhash.c 	ret = ahash_edesc_add_src(ctx, edesc, req, mapped_nents, 1, buflen,
mapped_nents     1436 drivers/crypto/caam/caamhash.c 	int src_nents, mapped_nents;
mapped_nents     1463 drivers/crypto/caam/caamhash.c 			mapped_nents = dma_map_sg(jrdev, req->src, src_nents,
mapped_nents     1465 drivers/crypto/caam/caamhash.c 			if (!mapped_nents) {
mapped_nents     1470 drivers/crypto/caam/caamhash.c 			mapped_nents = 0;
mapped_nents     1477 drivers/crypto/caam/caamhash.c 		edesc = ahash_edesc_alloc(ctx, mapped_nents > 1 ?
mapped_nents     1478 drivers/crypto/caam/caamhash.c 					  mapped_nents : 0,
mapped_nents     1489 drivers/crypto/caam/caamhash.c 		ret = ahash_edesc_add_src(ctx, edesc, req, mapped_nents, 0, 0,
mapped_nents      294 drivers/crypto/ccree/cc_buffer_mgr.c 		     u32 max_sg_nents, u32 *lbytes, u32 *mapped_nents)
mapped_nents      313 drivers/crypto/ccree/cc_buffer_mgr.c 	*mapped_nents = ret;
mapped_nents      410 drivers/crypto/ccree/cc_buffer_mgr.c 	u32 mapped_nents = 0;
mapped_nents      435 drivers/crypto/ccree/cc_buffer_mgr.c 		       LLI_MAX_NUM_OF_DATA_ENTRIES, &dummy, &mapped_nents);
mapped_nents      438 drivers/crypto/ccree/cc_buffer_mgr.c 	if (mapped_nents > 1)
mapped_nents      453 drivers/crypto/ccree/cc_buffer_mgr.c 			       &dummy, &mapped_nents);
mapped_nents      456 drivers/crypto/ccree/cc_buffer_mgr.c 		if (mapped_nents > 1)
mapped_nents      549 drivers/crypto/ccree/cc_buffer_mgr.c 	dma_unmap_sg(dev, req->src, areq_ctx->src.mapped_nents,
mapped_nents      554 drivers/crypto/ccree/cc_buffer_mgr.c 		dma_unmap_sg(dev, req->dst, areq_ctx->dst.mapped_nents,
mapped_nents      633 drivers/crypto/ccree/cc_buffer_mgr.c 	int mapped_nents = 0;
mapped_nents      656 drivers/crypto/ccree/cc_buffer_mgr.c 	mapped_nents = sg_nents_for_len(req->src, size_of_assoc);
mapped_nents      657 drivers/crypto/ccree/cc_buffer_mgr.c 	if (mapped_nents < 0)
mapped_nents      658 drivers/crypto/ccree/cc_buffer_mgr.c 		return mapped_nents;
mapped_nents      660 drivers/crypto/ccree/cc_buffer_mgr.c 	if (mapped_nents > LLI_MAX_NUM_OF_ASSOC_DATA_ENTRIES) {
mapped_nents      662 drivers/crypto/ccree/cc_buffer_mgr.c 			mapped_nents, LLI_MAX_NUM_OF_ASSOC_DATA_ENTRIES);
mapped_nents      665 drivers/crypto/ccree/cc_buffer_mgr.c 	areq_ctx->assoc.nents = mapped_nents;
mapped_nents      671 drivers/crypto/ccree/cc_buffer_mgr.c 		if ((mapped_nents + 1) > LLI_MAX_NUM_OF_ASSOC_DATA_ENTRIES) {
mapped_nents      680 drivers/crypto/ccree/cc_buffer_mgr.c 	if (mapped_nents == 1 && areq_ctx->ccm_hdr_size == ccm_header_size_null)
mapped_nents      907 drivers/crypto/ccree/cc_buffer_mgr.c 			       &areq_ctx->dst.mapped_nents,
mapped_nents     1016 drivers/crypto/ccree/cc_buffer_mgr.c 	u32 mapped_nents = 0;
mapped_nents     1124 drivers/crypto/ccree/cc_buffer_mgr.c 		       &areq_ctx->src.mapped_nents,
mapped_nents     1127 drivers/crypto/ccree/cc_buffer_mgr.c 		       &dummy, &mapped_nents);
mapped_nents     1215 drivers/crypto/ccree/cc_buffer_mgr.c 	u32 mapped_nents = 0;
mapped_nents     1242 drivers/crypto/ccree/cc_buffer_mgr.c 			       &dummy, &mapped_nents);
mapped_nents     1245 drivers/crypto/ccree/cc_buffer_mgr.c 		if (src && mapped_nents == 1 &&
mapped_nents     1301 drivers/crypto/ccree/cc_buffer_mgr.c 	u32 mapped_nents = 0;
mapped_nents     1355 drivers/crypto/ccree/cc_buffer_mgr.c 			       &mapped_nents);
mapped_nents     1358 drivers/crypto/ccree/cc_buffer_mgr.c 		if (mapped_nents == 1 &&
mapped_nents       28 drivers/crypto/ccree/cc_buffer_mgr.h 	unsigned int mapped_nents;
mapped_nents      444 net/smc/smc_ib.c 	int mapped_nents;
mapped_nents      446 net/smc/smc_ib.c 	mapped_nents = ib_dma_map_sg(smcibdev->ibdev,
mapped_nents      450 net/smc/smc_ib.c 	if (!mapped_nents)
mapped_nents      453 net/smc/smc_ib.c 	return mapped_nents;