n_sg              173 drivers/block/skd_main.c 	u32 n_sg;
n_sg              515 drivers/block/skd_main.c 	skreq->n_sg = 0;
n_sg              530 drivers/block/skd_main.c 				   skreq->n_sg *
n_sg              624 drivers/block/skd_main.c 	int n_sg;
n_sg              632 drivers/block/skd_main.c 	n_sg = blk_rq_map_sg(skdev->queue, req, sgl);
n_sg              633 drivers/block/skd_main.c 	if (n_sg <= 0)
n_sg              640 drivers/block/skd_main.c 	n_sg = dma_map_sg(&skdev->pdev->dev, sgl, n_sg, skreq->data_dir);
n_sg              641 drivers/block/skd_main.c 	if (n_sg <= 0)
n_sg              644 drivers/block/skd_main.c 	SKD_ASSERT(n_sg <= skdev->sgs_per_request);
n_sg              646 drivers/block/skd_main.c 	skreq->n_sg = n_sg;
n_sg              648 drivers/block/skd_main.c 	for_each_sg(sgl, sg, n_sg, i) {
n_sg              660 drivers/block/skd_main.c 	skreq->sksg_list[n_sg - 1].next_desc_ptr = 0LL;
n_sg              661 drivers/block/skd_main.c 	skreq->sksg_list[n_sg - 1].control = FIT_SGD_CONTROL_LAST;
n_sg              667 drivers/block/skd_main.c 		for (i = 0; i < n_sg; i++) {
n_sg              687 drivers/block/skd_main.c 	skreq->sksg_list[skreq->n_sg - 1].next_desc_ptr =
n_sg              689 drivers/block/skd_main.c 		((skreq->n_sg) * sizeof(struct fit_sg_descriptor));
n_sg              690 drivers/block/skd_main.c 	dma_unmap_sg(&skdev->pdev->dev, &skreq->sg[0], skreq->n_sg,
n_sg              900 drivers/block/skd_main.c 	skspcl->req.n_sg = 1;
n_sg             1245 drivers/block/skd_main.c 	WARN_ON_ONCE(skspcl->req.n_sg != 1);
n_sg             1262 drivers/block/skd_main.c 		for (i = 0; i < skspcl->req.n_sg; i++) {
n_sg             1553 drivers/block/skd_main.c 		if (skreq->n_sg > 0)
n_sg             1906 drivers/block/skd_main.c 	if (skreq->n_sg > 0)
n_sg             2704 drivers/block/skd_main.c 						  u32 n_sg,
n_sg             2716 drivers/block/skd_main.c 		for (i = 0; i < n_sg - 1; i++) {
n_sg             3589 drivers/block/skd_main.c 		skreq->data_dir, skreq->n_sg);
n_sg              328 drivers/block/sx8.c 	u8 n_sg;
n_sg              611 drivers/block/sx8.c 	ab->n_sg	= 1;
n_sg               83 drivers/crypto/ccp/ccp-crypto-rsa.c 	rctx->cmd.u.rsa.mod = &ctx->u.rsa.n_sg;
n_sg              148 drivers/crypto/ccp/ccp-crypto-rsa.c 	sg_init_one(&ctx->u.rsa.n_sg, ctx->u.rsa.n_buf, ctx->u.rsa.n_len);
n_sg              242 drivers/crypto/ccp/ccp-crypto.h 	struct scatterlist n_sg;
n_sg             1102 drivers/crypto/talitos.c 	int n_sg = elen ? sg_count + 1 : sg_count;
n_sg             1106 drivers/crypto/talitos.c 	while (cryptlen && sg && n_sg--) {
n_sg             1080 drivers/dma/stm32-dma.c 	u32 n_sg = next_sg;
n_sg             1112 drivers/dma/stm32-dma.c 		n_sg++;
n_sg             1113 drivers/dma/stm32-dma.c 		if (n_sg == chan->desc->num_sgs)
n_sg             1114 drivers/dma/stm32-dma.c 			n_sg = 0;
n_sg             1125 drivers/dma/stm32-dma.c 	if (!chan->desc->cyclic || n_sg != 0)
n_sg             1126 drivers/dma/stm32-dma.c 		for (i = n_sg; i < desc->num_sgs; i++)
n_sg              873 drivers/mmc/host/mmc_spi.c 	unsigned		n_sg;
n_sg              894 drivers/mmc/host/mmc_spi.c 	for_each_sg(data->sg, sg, data->sg_len, n_sg) {
n_sg              122 drivers/xen/xen-scsiback.c 	uint16_t n_sg;		/* real length of SG list */
n_sg              279 drivers/xen/xen-scsiback.c 	req->n_sg = 0;
n_sg              408 drivers/xen/xen-scsiback.c 			pending_req->sgl, pending_req->n_sg,
n_sg              485 drivers/xen/xen-scsiback.c 	pending_req->n_sg = 0;
n_sg              530 drivers/xen/xen-scsiback.c 	pending_req->n_sg = nr_segments;