Searched refs:sg_len (Results 1 - 121 of 121) sorted by relevance

/linux-4.1.27/drivers/mmc/host/
H A Dtmio_mmc_dma.c57 for_each_sg(sg, sg_tmp, host->sg_len, i) { tmio_mmc_start_dma_rx()
66 if ((!aligned && (host->sg_len > 1 || sg->length > PAGE_CACHE_SIZE || tmio_mmc_start_dma_rx()
86 ret = dma_map_sg(chan->device->dev, sg, host->sg_len, DMA_FROM_DEVICE); tmio_mmc_start_dma_rx()
99 __func__, host->sg_len, ret, cookie, host->mrq); tmio_mmc_start_dma_rx()
120 desc, cookie, host->sg_len); tmio_mmc_start_dma_rx()
133 for_each_sg(sg, sg_tmp, host->sg_len, i) { tmio_mmc_start_dma_tx()
142 if ((!aligned && (host->sg_len > 1 || sg->length > PAGE_CACHE_SIZE || tmio_mmc_start_dma_tx()
166 ret = dma_map_sg(chan->device->dev, sg, host->sg_len, DMA_TO_DEVICE); tmio_mmc_start_dma_tx()
179 __func__, host->sg_len, ret, cookie, host->mrq); tmio_mmc_start_dma_tx()
248 host->sg_ptr, host->sg_len, tmio_mmc_tasklet_fn()
252 host->sg_ptr, host->sg_len, tmio_mmc_tasklet_fn()
H A Dandroid-goldfish.c127 unsigned int sg_len; member in struct:goldfish_mmc_host
229 dma_unmap_sg(mmc_dev(host->mmc), data->sg, host->sg_len, goldfish_mmc_xfer_done()
234 host->sg_len = 0; goldfish_mmc_xfer_done()
371 unsigned sg_len; goldfish_mmc_prepare_data() local
391 sg_len = (data->blocks == 1) ? 1 : data->sg_len; goldfish_mmc_prepare_data()
398 host->sg_len = dma_map_sg(mmc_dev(host->mmc), data->sg, goldfish_mmc_prepare_data()
399 sg_len, dma_data_dir); goldfish_mmc_prepare_data()
H A Dmxs-mmc.c158 data->sg_len, ssp->dma_dir); mxs_mmc_request_done()
231 unsigned int sg_len; mxs_mmc_prep_dma() local
236 data->sg_len, ssp->dma_dir); mxs_mmc_prep_dma()
238 sg_len = data->sg_len; mxs_mmc_prep_dma()
242 sg_len = SSP_PIO_NUM; mxs_mmc_prep_dma()
246 sgl, sg_len, ssp->slave_dirn, flags); mxs_mmc_prep_dma()
253 data->sg_len, ssp->dma_dir); mxs_mmc_prep_dma()
357 unsigned int sg_len = data->sg_len; mxs_mmc_adtc() local
402 for_each_sg(sgl, sg, sg_len, i) mxs_mmc_adtc()
H A Dau1xmmc.c353 dma_unmap_sg(mmc_dev(host->mmc), data->sg, data->sg_len, host->dma.dir); au1xmmc_data_complete()
396 int sg_len, max, count; au1xmmc_send_pio() local
411 sg_len = data->sg[host->pio.index].length - host->pio.offset; au1xmmc_send_pio()
414 max = (sg_len > host->pio.len) ? host->pio.len : sg_len; au1xmmc_send_pio()
433 if (count == sg_len) { au1xmmc_send_pio()
451 int max, count, sg_len = 0; au1xmmc_receive_pio() local
468 sg_len = sg_dma_len(&data->sg[host->pio.index]) - host->pio.offset; au1xmmc_receive_pio()
471 if (sg_len < max) au1xmmc_receive_pio()
472 max = sg_len; au1xmmc_receive_pio()
510 if (sg_len && count == sg_len) { au1xmmc_receive_pio()
631 data->sg_len, host->dma.dir); au1xmmc_prepare_data()
647 int sg_len = sg->length; au1xmmc_prepare_data() local
649 int len = (datalen > sg_len) ? sg_len : datalen; au1xmmc_prepare_data()
682 dma_unmap_sg(mmc_dev(host->mmc), data->sg, data->sg_len, au1xmmc_prepare_data()
H A Djz4740_mmc.c118 int sg_len; member in struct:jz4740_mmc_host_next
153 int sg_len; member in struct:jz4740_mmc_host
220 dma_unmap_sg(chan->device->dev, data->sg, data->sg_len, dir); jz4740_mmc_dma_unmap()
231 int sg_len; jz4740_mmc_prepare_dma_data() local
245 sg_len = dma_map_sg(chan->device->dev, jz4740_mmc_prepare_dma_data()
247 data->sg_len, jz4740_mmc_prepare_dma_data()
251 sg_len = next_data->sg_len; jz4740_mmc_prepare_dma_data()
252 next_data->sg_len = 0; jz4740_mmc_prepare_dma_data()
255 if (sg_len <= 0) { jz4740_mmc_prepare_dma_data()
262 next->sg_len = sg_len; jz4740_mmc_prepare_dma_data()
265 host->sg_len = sg_len; jz4740_mmc_prepare_dma_data()
302 host->sg_len, jz4740_mmc_start_dma_transfer()
690 sg_miter_start(&host->miter, data->sg, data->sg_len, direction); jz_mmc_prepare_data_transfer()
H A Domap.c149 unsigned int sg_len; member in struct:mmc_omap_host
427 dma_unmap_sg(dev, data->sg, host->sg_len, dma_data_dir); mmc_omap_release_dma()
451 host->sg_len = 0; mmc_omap_xfer_done()
508 host->sg_len = 0; mmc_omap_abort_xfer()
673 BUG_ON(host->sg_idx == host->sg_len); mmc_omap_xfer_data()
951 unsigned sg_len; mmc_omap_prepare_data() local
972 sg_len = (data->blocks == 1) ? 1 : data->sg_len; mmc_omap_prepare_data()
975 for (i = 0; i < sg_len; i++) { mmc_omap_prepare_data()
1034 host->sg_len = dma_map_sg(c->device->dev, data->sg, sg_len, mmc_omap_prepare_data()
1036 if (host->sg_len == 0) mmc_omap_prepare_data()
1039 tx = dmaengine_prep_slave_sg(c, data->sg, host->sg_len, mmc_omap_prepare_data()
1060 host->sg_len = sg_len; mmc_omap_prepare_data()
H A Dtifm_sd.c106 int sg_len; member in struct:tifm_sd
175 if (host->sg_pos == host->sg_len) tifm_sd_transfer_data()
182 if (host->sg_pos == host->sg_len) { tifm_sd_transfer_data()
238 if (host->sg_pos == host->sg_len) tifm_sd_bounce_block()
271 if (host->sg_pos == host->sg_len) tifm_sd_set_dma_data()
279 if (host->sg_pos == host->sg_len) tifm_sd_set_dma_data()
287 if (host->sg_pos == host->sg_len) tifm_sd_set_dma_data()
666 host->sg_len = r_data->sg_len; tifm_sd_request()
680 host->sg_len = tifm_map_sg(sock, r_data->sg, tifm_sd_request()
681 r_data->sg_len, tifm_sd_request()
686 if (host->sg_len < 1) { tifm_sd_request()
767 tifm_unmap_sg(sock, r_data->sg, r_data->sg_len, tifm_sd_end_cmd()
H A Ddavinci_mmc.c214 unsigned int sg_len; member in struct:mmc_davinci_host
438 host->sg_len, mmc_davinci_send_dma_request()
460 host->sg_len, mmc_davinci_send_dma_request()
485 host->sg_len = dma_map_sg(mmc_dev(host->mmc), data->sg, data->sg_len, mmc_davinci_start_dma_transfer()
491 for (i = 0; i < host->sg_len; i++) { mmc_davinci_start_dma_transfer()
494 data->sg, data->sg_len, mmc_davinci_start_dma_transfer()
622 host->sg_len = data->sg_len; mmc_davinci_prepare_data()
829 dma_unmap_sg(mmc_dev(host->mmc), data->sg, data->sg_len, mmc_davinci_xfer_done()
H A Dusdhi6rol0.c317 __func__, host->mrq->cmd->opcode, data->sg_len, usdhi6_blk_bounce()
354 struct scatterlist *sg = data->sg_len > 1 ? host->sg : data->sg; usdhi6_sg_map()
398 struct scatterlist *sg = data->sg_len > 1 ? usdhi6_sg_unmap()
540 data->sg_len, DMA_FROM_DEVICE); usdhi6_dma_stop_unmap()
543 data->sg_len, DMA_TO_DEVICE); usdhi6_dma_stop_unmap()
583 ret = dma_map_sg(chan->device->dev, sg, data->sg_len, data_dir); usdhi6_dma_setup()
597 __func__, data->sg_len, ret, cookie, desc); usdhi6_dma_setup()
627 __func__, data->sg_len, data->blocks, data->blksz); usdhi6_dma_kill()
903 data ? data->blksz : 0, data ? data->sg_len : 0)) usdhi6_request_done()
1044 "Bad SG of %u: %ux%u @ %u\n", data->sg_len, usdhi6_rq_start()
1058 data->sg_len, use_dma ? "DMA" : "PIO", usdhi6_rq_start()
1671 host->offset, data->blocks, data->blksz, data->sg_len, usdhi6_timeout_work()
H A Dmxcmmc.c294 for_each_sg(data->sg, sg, data->sg_len, i) mxcmci_swap_buffers()
323 for_each_sg(data->sg, sg, data->sg_len, i) { mxcmci_setup_data()
341 data->sg_len, host->dma_dir); mxcmci_setup_data()
342 if (nents != data->sg_len) mxcmci_setup_data()
346 data->sg, data->sg_len, slave_dirn, mxcmci_setup_data()
350 dma_unmap_sg(host->dma->device->dev, data->sg, data->sg_len, mxcmci_setup_data()
460 dma_unmap_sg(host->dma->device->dev, data->sg, data->sg_len, mxcmci_finish_data()
625 for_each_sg(data->sg, sg, data->sg_len, i) { mxcmci_transfer_data()
632 for_each_sg(data->sg, sg, data->sg_len, i) { mxcmci_transfer_data()
H A Ddw_mmc.c418 data->sg_len, dw_mci_dma_cleanup()
466 unsigned int sg_len) dw_mci_translate_sglist()
475 for (i = 0; i < sg_len; i++) { dw_mci_translate_sglist()
519 for (i = 0; i < sg_len; i++) { dw_mci_translate_sglist()
563 static void dw_mci_idmac_start_dma(struct dw_mci *host, unsigned int sg_len) dw_mci_idmac_start_dma() argument
567 dw_mci_translate_sglist(host, host->data, sg_len); dw_mci_idmac_start_dma()
675 unsigned int i, sg_len; dw_mci_pre_dma_transfer() local
691 for_each_sg(data->sg, sg, data->sg_len, i) { dw_mci_pre_dma_transfer()
696 sg_len = dma_map_sg(host->dev, dw_mci_pre_dma_transfer()
698 data->sg_len, dw_mci_pre_dma_transfer()
700 if (sg_len == 0) dw_mci_pre_dma_transfer()
704 data->host_cookie = sg_len; dw_mci_pre_dma_transfer()
706 return sg_len; dw_mci_pre_dma_transfer()
741 data->sg_len, dw_mci_post_req()
829 int sg_len; dw_mci_submit_data_dma() local
838 sg_len = dw_mci_pre_dma_transfer(host, data, 0); dw_mci_submit_data_dma()
839 if (sg_len < 0) { dw_mci_submit_data_dma()
841 return sg_len; dw_mci_submit_data_dma()
847 "sd sg_cpu: %#lx sg_dma: %#lx sg_len: %d\n", dw_mci_submit_data_dma()
849 sg_len); dw_mci_submit_data_dma()
871 host->dma_ops->start(host, sg_len); dw_mci_submit_data_dma()
901 sg_miter_start(&host->sg_miter, data->sg, data->sg_len, flags); dw_mci_submit_data()
465 dw_mci_translate_sglist(struct dw_mci *host, struct mmc_data *data, unsigned int sg_len) dw_mci_translate_sglist() argument
H A Dsh_mmcif.c295 ret = dma_map_sg(chan->device->dev, sg, data->sg_len, sh_mmcif_start_dma_rx()
311 __func__, data->sg_len, ret, cookie); sh_mmcif_start_dma_rx()
332 desc, cookie, data->sg_len); sh_mmcif_start_dma_rx()
344 ret = dma_map_sg(chan->device->dev, sg, data->sg_len, sh_mmcif_start_dma_tx()
360 __func__, data->sg_len, ret, cookie); sh_mmcif_start_dma_tx()
573 if (++host->sg_idx < data->sg_len) sh_mmcif_next_block()
579 return host->sg_idx != data->sg_len; sh_mmcif_next_block()
621 if (!data->sg_len || !data->sg->length) sh_mmcif_multi_read()
699 if (!data->sg_len || !data->sg->length) sh_mmcif_multi_write()
1133 data->sg, data->sg_len, sh_mmcif_end_cmd()
1137 data->sg, data->sg_len, sh_mmcif_end_cmd()
H A Datmel-mci.c186 unsigned int sg_len; member in struct:atmel_mci
803 data->sg, data->sg_len, atmci_pdc_cleanup()
825 sg_copy_from_buffer(host->data->sg, host->data->sg_len, atmci_pdc_complete()
842 data->sg, data->sg_len, atmci_dma_cleanup()
908 host->sg_len = data->sg_len; atmci_prepare_data()
944 unsigned int sg_len; atmci_prepare_data_pdc() local
973 sg_len = dma_map_sg(&host->pdev->dev, data->sg, data->sg_len, dir); atmci_prepare_data_pdc()
977 sg_copy_to_buffer(host->data->sg, host->data->sg_len, atmci_prepare_data_pdc()
1022 for_each_sg(data->sg, sg, data->sg_len, i) { atmci_prepare_data_dma()
1050 data->sg_len, direction); atmci_prepare_data_dma()
1065 dma_unmap_sg(chan->device->dev, data->sg, data->sg_len, direction); atmci_prepare_data_dma()
1875 host->sg_len--; atmci_read_data_pio()
1876 if (!sg || !host->sg_len) atmci_read_data_pio()
1889 host->sg_len--; atmci_read_data_pio()
1890 if (!sg || !host->sg_len) atmci_read_data_pio()
1941 host->sg_len--; atmci_write_data_pio()
1942 if (!sg || !host->sg_len) atmci_write_data_pio()
1956 host->sg_len--; atmci_write_data_pio()
1957 if (!sg || !host->sg_len) { atmci_write_data_pio()
H A Dmoxart-mmc.c155 host->num_sg = data->sg_len; moxart_init_sg()
278 data->sg_len, dir_data); moxart_transfer_dma()
303 data->sg, data->sg_len, moxart_transfer_dma()
H A Dwmt-sdmmc.c307 req->data->sg_len, DMA_TO_DEVICE); wmt_complete_data_request()
310 req->data->sg_len, DMA_FROM_DEVICE); wmt_complete_data_request()
624 req->data->sg_len, DMA_TO_DEVICE); wmt_mci_request()
630 req->data->sg_len, DMA_FROM_DEVICE); wmt_mci_request()
H A Dsunxi-mmc.c299 for (i = 0; i < data->sg_len; i++) { sunxi_mmc_init_idma_des()
339 dma_len = dma_map_sg(mmc_dev(host->mmc), data->sg, data->sg_len, sunxi_mmc_map_dma()
346 for_each_sg(data->sg, sg, data->sg_len, i) { sunxi_mmc_map_dma()
490 dma_unmap_sg(mmc_dev(host->mmc), data->sg, data->sg_len, sunxi_mmc_finalize_request()
847 dma_unmap_sg(mmc_dev(mmc), data->sg, data->sg_len, sunxi_mmc_request()
H A Ds3cmci.c328 if (host->pio_sgptr >= host->mrq->data->sg_len) { get_data_buffer()
330 host->pio_sgptr, host->mrq->data->sg_len); get_data_buffer()
341 host->pio_sgptr, host->mrq->data->sg_len); get_data_buffer()
1107 dma_map_sg(mmc_dev(host->mmc), data->sg, data->sg_len, s3cmci_prepare_dma()
1111 desc = dmaengine_prep_slave_sg(host->dma, data->sg, data->sg_len, s3cmci_prepare_dma()
1124 dma_unmap_sg(mmc_dev(host->mmc), data->sg, data->sg_len, s3cmci_prepare_dma()
H A Dbfin_sdh.c158 host->dma_len = dma_map_sg(mmc_dev(host->mmc), data->sg, data->sg_len, host->dma_dir); sdh_setup_data()
302 dma_unmap_sg(mmc_dev(host->mmc), data->sg, data->sg_len, sdh_data_done()
H A Dsdricoh_cs.c313 dev_dbg(dev, "transfer: blksz %i blocks %i sg_len %i " sdricoh_request()
315 data->sg_len, data->sg->length); sdricoh_request()
H A Drtsx_pci_sdmmc.c176 count = rtsx_pci_dma_map_sg(pcr, data->sg, data->sg_len, read); sd_pre_dma_transfer()
219 rtsx_pci_dma_unmap_sg(pcr, data->sg, data->sg_len, read); sdmmc_post_req()
608 sg_copy_from_buffer(data->sg, data->sg_len, buf, data->blksz); sd_normal_rw()
610 sg_copy_to_buffer(data->sg, data->sg_len, buf, data->blksz); sd_normal_rw()
H A Dtmio_mmc.h65 unsigned int sg_len; member in struct:tmio_mmc_host
H A Dmmci.c421 sg_miter_start(&host->sg_miter, data->sg, data->sg_len, flags); mmci_init_sg()
521 dma_unmap_sg(chan->device->dev, data->sg, data->sg_len, dir); mmci_dma_unmap()
606 nr_sg = dma_map_sg(device->dev, data->sg, data->sg_len, buffer_dirn); __mmci_dma_prep_data()
625 dma_unmap_sg(device->dev, data->sg, data->sg_len, buffer_dirn); __mmci_dma_prep_data()
660 data->sg_len, data->blksz, data->blocks, data->flags); mmci_dma_start_data()
H A Domap_hsmmc.c955 host->data->sg, host->data->sg_len, omap_hsmmc_dma_cleanup()
1267 data->sg, data->sg_len, omap_hsmmc_dma_callback()
1302 dma_len = dma_map_sg(chan->device->dev, data->sg, data->sg_len, omap_hsmmc_pre_dma_transfer()
1336 for (i = 0; i < data->sg_len; i++) { omap_hsmmc_setup_dma_transfer()
1368 tx = dmaengine_prep_slave_sg(chan, data->sg, data->sg_len, omap_hsmmc_setup_dma_transfer()
1479 dma_unmap_sg(c->device->dev, data->sg, data->sg_len, omap_hsmmc_post_req()
H A Dvub300.c1450 data->sg_len, 0, GFP_KERNEL); __command_read_data()
1494 sg_copy_from_buffer(data->sg, data->sg_len, buf, __command_read_data()
1519 sg_copy_to_buffer(data->sg, data->sg_len, __command_write_data()
1543 sg_copy_to_buffer(data->sg, data->sg_len, buf, __command_write_data()
1565 sg_copy_to_buffer(data->sg, data->sg_len, buf, sizeof(buf)); __command_write_data()
1568 data->sg_len, 0, GFP_KERNEL); __command_write_data()
H A Dcb710-mmc.c277 sg_miter_start(&miter, data->sg, data->sg_len, SG_MITER_TO_SG); cb710_mmc_receive()
321 sg_miter_start(&miter, data->sg, data->sg_len, SG_MITER_FROM_SG); cb710_mmc_send()
H A Dpxamci.c212 host->dma_len = dma_map_sg(mmc_dev(host->mmc), data->sg, data->sg_len, pxamci_setup_data()
362 dma_unmap_sg(mmc_dev(host->mmc), data->sg, data->sg_len, pxamci_data_done()
H A Dsdhci.c636 data->sg_len, direction); sdhci_adma_table_post()
656 data->sg_len, direction); sdhci_adma_table_post()
798 for_each_sg(data->sg, sg, data->sg_len, i) { sdhci_prepare_data()
833 for_each_sg(data->sg, sg, data->sg_len, i) { sdhci_prepare_data()
909 sg_miter_start(&host->sg_miter, data->sg, data->sg_len, flags); sdhci_prepare_data()
983 data->sg, data->sg_len, sdhci_finish_data()
2153 dma_unmap_sg(mmc_dev(host->mmc), data->sg, data->sg_len, sdhci_post_req()
2172 sg_count = dma_map_sg(mmc_dev(host->mmc), data->sg, data->sg_len, sdhci_pre_dma_transfer()
H A Drtsx_usb_sdmmc.c533 data->sg_len, NULL, 10000); sd_rw_multi()
579 sg_copy_from_buffer(data->sg, data->sg_len, buf, data->blksz); sd_normal_rw()
581 sg_copy_to_buffer(data->sg, data->sg_len, buf, data->blksz); sd_normal_rw()
H A Dtmio_mmc_pio.c73 host->sg_len = data->sg_len; tmio_mmc_init_sg()
83 return --host->sg_len; tmio_mmc_next_sg()
H A Dwbsd.c246 host->num_sg = data->sg_len; wbsd_init_sg()
284 len = data->sg_len; wbsd_sg_to_dma()
301 len = data->sg_len; wbsd_dma_to_sg()
H A Dtoshsd.c491 sg_miter_start(&host->sg_miter, data->sg, data->sg_len, flags); toshsd_start_data()
H A Dvia-sdmmc.c493 count = dma_map_sg(mmc_dev(host->mmc), data->sg, data->sg_len, via_sdc_preparedata()
637 dma_unmap_sg(mmc_dev(host->mmc), data->sg, data->sg_len, via_sdc_finish_data()
H A Dmvsdio.c134 data->sg_len, dma_dir); mvsd_setup_data()
H A Dmmc_spi.c910 for (sg = data->sg, n_sg = data->sg_len; n_sg; n_sg--, sg++) { mmc_spi_data_do()
/linux-4.1.27/drivers/mmc/card/
H A Dqueue.c145 static struct scatterlist *mmc_alloc_sg(int sg_len, int *err) mmc_alloc_sg() argument
149 sg = kmalloc(sizeof(struct scatterlist)*sg_len, GFP_KERNEL); mmc_alloc_sg()
154 sg_init_table(sg, sg_len); mmc_alloc_sg()
457 unsigned int sg_len = 0; mmc_queue_packed_map_sg() local
473 sg_len++; mmc_queue_packed_map_sg()
478 sg_len += blk_rq_map_sg(mq->queue, req, __sg); mmc_queue_packed_map_sg()
479 __sg = sg + (sg_len - 1); mmc_queue_packed_map_sg()
482 sg_mark_end(sg + (sg_len - 1)); mmc_queue_packed_map_sg()
483 return sg_len; mmc_queue_packed_map_sg()
491 unsigned int sg_len; mmc_queue_map_sg() local
510 sg_len = mmc_queue_packed_map_sg(mq, mqrq->packed, mmc_queue_map_sg()
513 sg_len = blk_rq_map_sg(mq->queue, mqrq->req, mqrq->bounce_sg); mmc_queue_map_sg()
515 mqrq->bounce_sg_len = sg_len; mmc_queue_map_sg()
518 for_each_sg(mqrq->bounce_sg, sg, sg_len, i) mmc_queue_map_sg()
H A Dmmc_test.c71 * @sg_len: length of currently mapped scatterlist @sg
82 unsigned int sg_len; member in struct:mmc_test_area
161 unsigned int *sg_len; member in struct:mmc_test_multiple_rw
191 struct mmc_request *mrq, struct scatterlist *sg, unsigned sg_len, mmc_test_prepare_mrq()
222 mrq->data->sg_len = sg_len; mmc_test_prepare_mrq()
397 unsigned int *sg_len, int min_sg_len) mmc_test_map_sg()
407 *sg_len = 0; mmc_test_map_sg()
426 *sg_len += 1; mmc_test_map_sg()
450 unsigned int *sg_len) mmc_test_map_sg_max_scatter()
459 *sg_len = 0; mmc_test_map_sg_max_scatter()
481 *sg_len += 1; mmc_test_map_sg_max_scatter()
586 "%u.%02u IOPS, sg_len %d)\n", mmc_test_print_avg_rate()
591 test->area.sg_len); mmc_test_print_avg_rate()
781 struct scatterlist *sg, unsigned sg_len, mmc_test_nonblock_transfer()
814 mmc_test_prepare_mrq(test, cur_areq->mrq, sg, sg_len, dev_addr, mmc_test_nonblock_transfer()
846 struct scatterlist *sg, unsigned sg_len, unsigned dev_addr, mmc_test_simple_transfer()
858 mmc_test_prepare_mrq(test, &mrq, sg, sg_len, dev_addr, mmc_test_simple_transfer()
903 struct scatterlist *sg, unsigned sg_len, unsigned dev_addr, mmc_test_transfer()
916 sg_copy_from_buffer(sg, sg_len, test->scratch, BUFFER_SIZE); mmc_test_transfer()
923 ret = mmc_test_simple_transfer(test, sg, sg_len, dev_addr, mmc_test_transfer()
963 sg_copy_to_buffer(sg, sg_len, test->scratch, BUFFER_SIZE); mmc_test_transfer()
1435 &t->sg_len); mmc_test_area_map()
1438 t->max_seg_sz, &t->sg_len, min_sg_len); mmc_test_area_map()
1454 return mmc_test_simple_transfer(test, t->sg, t->sg_len, dev_addr, mmc_test_area_transfer()
1494 ret = mmc_test_nonblock_transfer(test, t->sg, t->sg_len, mmc_test_area_io_seq()
2176 rw->sg_len[i]); mmc_test_rw_multiple_sg_len()
2264 unsigned int sg_len[] = {1, 1 << 3, 1 << 4, 1 << 5, 1 << 6, mmc_test_profile_sglen_wr_blocking_perf() local
2267 .sg_len = sg_len, mmc_test_profile_sglen_wr_blocking_perf()
2269 .len = ARRAY_SIZE(sg_len), mmc_test_profile_sglen_wr_blocking_perf()
2283 unsigned int sg_len[] = {1, 1 << 3, 1 << 4, 1 << 5, 1 << 6, mmc_test_profile_sglen_wr_nonblock_perf() local
2286 .sg_len = sg_len, mmc_test_profile_sglen_wr_nonblock_perf()
2288 .len = ARRAY_SIZE(sg_len), mmc_test_profile_sglen_wr_nonblock_perf()
2302 unsigned int sg_len[] = {1, 1 << 3, 1 << 4, 1 << 5, 1 << 6, mmc_test_profile_sglen_r_blocking_perf() local
2305 .sg_len = sg_len, mmc_test_profile_sglen_r_blocking_perf()
2307 .len = ARRAY_SIZE(sg_len), mmc_test_profile_sglen_r_blocking_perf()
2321 unsigned int sg_len[] = {1, 1 << 3, 1 << 4, 1 << 5, 1 << 6, mmc_test_profile_sglen_r_nonblock_perf() local
2324 .sg_len = sg_len, mmc_test_profile_sglen_r_nonblock_perf()
2326 .len = ARRAY_SIZE(sg_len), mmc_test_profile_sglen_r_nonblock_perf()
190 mmc_test_prepare_mrq(struct mmc_test_card *test, struct mmc_request *mrq, struct scatterlist *sg, unsigned sg_len, unsigned dev_addr, unsigned blocks, unsigned blksz, int write) mmc_test_prepare_mrq() argument
394 mmc_test_map_sg(struct mmc_test_mem *mem, unsigned long size, struct scatterlist *sglist, int repeat, unsigned int max_segs, unsigned int max_seg_sz, unsigned int *sg_len, int min_sg_len) mmc_test_map_sg() argument
445 mmc_test_map_sg_max_scatter(struct mmc_test_mem *mem, unsigned long sz, struct scatterlist *sglist, unsigned int max_segs, unsigned int max_seg_sz, unsigned int *sg_len) mmc_test_map_sg_max_scatter() argument
780 mmc_test_nonblock_transfer(struct mmc_test_card *test, struct scatterlist *sg, unsigned sg_len, unsigned dev_addr, unsigned blocks, unsigned blksz, int write, int count) mmc_test_nonblock_transfer() argument
845 mmc_test_simple_transfer(struct mmc_test_card *test, struct scatterlist *sg, unsigned sg_len, unsigned dev_addr, unsigned blocks, unsigned blksz, int write) mmc_test_simple_transfer() argument
902 mmc_test_transfer(struct mmc_test_card *test, struct scatterlist *sg, unsigned sg_len, unsigned dev_addr, unsigned blocks, unsigned blksz, int write) mmc_test_transfer() argument
H A Dblock.c494 data.sg_len = 1; mmc_blk_ioctl_cmd()
706 data.sg_len = 1; mmc_sd_num_wr_blocks()
1494 brq->data.sg_len = mmc_queue_map_sg(mq, mqrq); mmc_blk_rw_rq_prep()
1504 for_each_sg(brq->data.sg, sg, brq->data.sg_len, i) { mmc_blk_rw_rq_prep()
1512 brq->data.sg_len = i; mmc_blk_rw_rq_prep()
1721 brq->data.sg_len = mmc_queue_map_sg(mq, mqrq); mmc_blk_packed_hdr_wrq_prep()
/linux-4.1.27/drivers/dma/sh/
H A Dusb-dmac.c47 * @sg_len: length of sg
58 unsigned int sg_len; member in struct:usb_dmac_desc
266 static int usb_dmac_desc_alloc(struct usb_dmac_chan *chan, unsigned int sg_len, usb_dmac_desc_alloc() argument
272 desc = kzalloc(sizeof(*desc) + sg_len * sizeof(desc->sg[0]), gfp); usb_dmac_desc_alloc()
276 desc->sg_allocated_len = sg_len; usb_dmac_desc_alloc()
302 unsigned int sg_len, gfp_t gfp) usb_dmac_desc_get()
310 if (sg_len <= desc->sg_allocated_len) { usb_dmac_desc_get()
319 if (!usb_dmac_desc_alloc(chan, sg_len, gfp)) { usb_dmac_desc_get()
420 unsigned int sg_len, enum dma_transfer_direction dir, usb_dmac_prep_slave_sg()
428 if (!sg_len) { usb_dmac_prep_slave_sg()
430 "%s: bad parameter: len=%d\n", __func__, sg_len); usb_dmac_prep_slave_sg()
434 desc = usb_dmac_desc_get(uchan, sg_len, GFP_NOWAIT); usb_dmac_prep_slave_sg()
439 desc->sg_len = sg_len; for_each_sg()
440 for_each_sg(sgl, sg, sg_len, i) { for_each_sg()
522 for (i = desc->sg_index + 1; i < desc->sg_len; i++) usb_dmac_chan_get_residue()
586 if (++desc->sg_index < desc->sg_len) { usb_dmac_isr_transfer_end()
301 usb_dmac_desc_get(struct usb_dmac_chan *chan, unsigned int sg_len, gfp_t gfp) usb_dmac_desc_get() argument
419 usb_dmac_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction dir, unsigned long dma_flags, void *context) usb_dmac_prep_slave_sg() argument
H A Dshdma-base.c564 struct scatterlist *sgl, unsigned int sg_len, dma_addr_t *addr, shdma_prep_sg()
574 for_each_sg(sgl, sg, sg_len, i) shdma_prep_sg()
591 for_each_sg(sgl, sg, sg_len, i) { for_each_sg()
665 struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, shdma_prep_slave_sg()
680 if (slave_id < 0 || !sg_len) { shdma_prep_slave_sg()
682 __func__, sg_len, slave_id); shdma_prep_slave_sg()
688 return shdma_prep_sg(schan, sgl, sg_len, &slave_addr, shdma_prep_slave_sg()
703 unsigned int sg_len = buf_len / period_len; shdma_prep_dma_cyclic() local
714 if (sg_len > SHDMA_MAX_SG_LEN) { shdma_prep_dma_cyclic()
716 sg_len, SHDMA_MAX_SG_LEN); shdma_prep_dma_cyclic()
734 sgl = kcalloc(sg_len, sizeof(*sgl), GFP_KERNEL); shdma_prep_dma_cyclic()
738 sg_init_table(sgl, sg_len); shdma_prep_dma_cyclic()
740 for (i = 0; i < sg_len; i++) { shdma_prep_dma_cyclic()
749 desc = shdma_prep_sg(schan, sgl, sg_len, &slave_addr, shdma_prep_dma_cyclic()
563 shdma_prep_sg(struct shdma_chan *schan, struct scatterlist *sgl, unsigned int sg_len, dma_addr_t *addr, enum dma_transfer_direction direction, unsigned long flags, bool cyclic) shdma_prep_sg() argument
664 shdma_prep_slave_sg( struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) shdma_prep_slave_sg() argument
H A Drcar-dmac.c824 unsigned int sg_len, dma_addr_t dev_addr, rcar_dmac_chan_prep_sg()
855 for_each_sg(sgl, sg, sg_len, i) { for_each_sg()
1021 unsigned int sg_len, enum dma_transfer_direction dir, rcar_dmac_prep_slave_sg()
1028 if (rchan->mid_rid < 0 || !sg_len) { rcar_dmac_prep_slave_sg()
1031 __func__, sg_len, rchan->mid_rid); rcar_dmac_prep_slave_sg()
1037 return rcar_dmac_chan_prep_sg(rchan, sgl, sg_len, dev_addr, rcar_dmac_prep_slave_sg()
1052 unsigned int sg_len; rcar_dmac_prep_dma_cyclic() local
1063 sg_len = buf_len / period_len; rcar_dmac_prep_dma_cyclic()
1064 if (sg_len > RCAR_DMAC_MAX_SG_LEN) { rcar_dmac_prep_dma_cyclic()
1067 rchan->index, sg_len, RCAR_DMAC_MAX_SG_LEN); rcar_dmac_prep_dma_cyclic()
1075 sgl = kcalloc(sg_len, sizeof(*sgl), GFP_NOWAIT); rcar_dmac_prep_dma_cyclic()
1079 sg_init_table(sgl, sg_len); rcar_dmac_prep_dma_cyclic()
1081 for (i = 0; i < sg_len; ++i) { rcar_dmac_prep_dma_cyclic()
1092 desc = rcar_dmac_chan_prep_sg(rchan, sgl, sg_len, dev_addr, rcar_dmac_prep_dma_cyclic()
823 rcar_dmac_chan_prep_sg(struct rcar_dmac_chan *chan, struct scatterlist *sgl, unsigned int sg_len, dma_addr_t dev_addr, enum dma_transfer_direction dir, unsigned long dma_flags, bool cyclic) rcar_dmac_chan_prep_sg() argument
1020 rcar_dmac_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction dir, unsigned long flags, void *context) rcar_dmac_prep_slave_sg() argument
/linux-4.1.27/drivers/mmc/core/
H A Dsdio_ops.c166 data.sg_len = nents; mmc_io_rw_extended()
168 for_each_sg(data.sg, sg_ptr, data.sg_len, i) { for_each_sg()
176 data.sg_len = 1;
H A Dsd_ops.c285 data.sg_len = 1; mmc_app_send_scr()
336 data.sg_len = 1; mmc_sd_switch()
381 data.sg_len = 1; mmc_app_sd_status()
H A Dmmc_ops.c315 data.sg_len = 1; mmc_send_cxd_data()
625 data.sg_len = 1; mmc_send_tuning()
704 data.sg_len = 1; mmc_send_bus_test()
H A Dcore.c239 for_each_sg(mrq->data->sg, sg, mrq->data->sg_len, i) mmc_start_request()
/linux-4.1.27/net/rds/
H A Dib_rdma.c59 unsigned int sg_len; member in struct:rds_ib_mr
438 ibmr->sg_len = nents; rds_ib_map_fmr()
474 ibmr->sg, ibmr->sg_len, __rds_ib_teardown_mr()
480 if (ibmr->sg_len) { __rds_ib_teardown_mr()
483 for (i = 0; i < ibmr->sg_len; ++i) { __rds_ib_teardown_mr()
495 ibmr->sg_len = 0; __rds_ib_teardown_mr()
501 unsigned int pinned = ibmr->sg_len; rds_ib_teardown_mr()
641 unpinned += ibmr->sg_len; rds_ib_flush_mr_pool()
700 rdsdebug("RDS/IB: free_mr nents %u\n", ibmr->sg_len); rds_ib_free_mr()
708 atomic_add(ibmr->sg_len, &pool->free_pinned); rds_ib_free_mr()
H A Diw_rdma.c252 struct scatterlist *list, unsigned int sg_len) rds_iw_set_scatterlist()
255 sg->len = sg_len; rds_iw_set_scatterlist()
761 unsigned int sg_len) rds_iw_map_fastreg()
768 rds_iw_set_scatterlist(&mapping->m_sg, sg, sg_len); rds_iw_map_fastreg()
251 rds_iw_set_scatterlist(struct rds_iw_scatterlist *sg, struct scatterlist *list, unsigned int sg_len) rds_iw_set_scatterlist() argument
758 rds_iw_map_fastreg(struct rds_iw_mr_pool *pool, struct rds_iw_mr *ibmr, struct scatterlist *sg, unsigned int sg_len) rds_iw_map_fastreg() argument
/linux-4.1.27/drivers/dma/
H A Dste_dma40_ll.c266 int sg_len, d40_phy_sg_to_lli()
284 for_each_sg(sg, current_sg, sg_len, i) { for_each_sg()
291 if (i == sg_len - 1) for_each_sg()
418 int sg_len, d40_log_sg_to_lli()
433 for_each_sg(sg, current_sg, sg_len, i) { for_each_sg()
265 d40_phy_sg_to_lli(struct scatterlist *sg, int sg_len, dma_addr_t target, struct d40_phy_lli *lli_sg, dma_addr_t lli_phys, u32 reg_cfg, struct stedma40_half_channel_info *info, struct stedma40_half_channel_info *otherinfo, unsigned long flags) d40_phy_sg_to_lli() argument
417 d40_log_sg_to_lli(struct scatterlist *sg, int sg_len, dma_addr_t dev_addr, struct d40_log_lli *lli_sg, u32 lcsp13, u32 data_width1, u32 data_width2) d40_log_sg_to_lli() argument
H A Dfsl-edma.c503 int sg_len) fsl_edma_alloc_desc()
508 fsl_desc = kzalloc(sizeof(*fsl_desc) + sizeof(struct fsl_edma_sw_tcd) * sg_len, fsl_edma_alloc_desc()
514 fsl_desc->n_tcds = sg_len; fsl_edma_alloc_desc()
515 for (i = 0; i < sg_len; i++) { fsl_edma_alloc_desc()
539 int sg_len, i; fsl_edma_prep_dma_cyclic() local
546 sg_len = buf_len / period_len; fsl_edma_prep_dma_cyclic()
547 fsl_desc = fsl_edma_alloc_desc(fsl_chan, sg_len); fsl_edma_prep_dma_cyclic()
556 for (i = 0; i < sg_len; i++) { fsl_edma_prep_dma_cyclic()
561 last_sg = fsl_desc->tcd[(i + 1) % sg_len].ptcd; fsl_edma_prep_dma_cyclic()
586 unsigned int sg_len, enum dma_transfer_direction direction, fsl_edma_prep_slave_sg()
599 fsl_desc = fsl_edma_alloc_desc(fsl_chan, sg_len); fsl_edma_prep_slave_sg()
605 for_each_sg(sgl, sg, sg_len, i) { for_each_sg()
607 last_sg = fsl_desc->tcd[(i + 1) % sg_len].ptcd; for_each_sg()
622 if (i < sg_len - 1) { for_each_sg()
502 fsl_edma_alloc_desc(struct fsl_edma_chan *fsl_chan, int sg_len) fsl_edma_alloc_desc() argument
584 fsl_edma_prep_slave_sg( struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) fsl_edma_prep_slave_sg() argument
H A Dedma.c94 * - sg_len: Tracks the length of the current intermediate transfer,
99 u32 sg_len; member in struct:edma_desc
175 edesc->sg_len = 0; edma_execute()
181 edesc->sg_len += edesc->pset[j].len; edma_execute()
450 unsigned int sg_len, enum dma_transfer_direction direction, edma_prep_slave_sg()
462 if (unlikely(!echan || !sgl || !sg_len)) edma_prep_slave_sg()
483 edesc = kzalloc(sizeof(*edesc) + sg_len * edma_prep_slave_sg()
490 edesc->pset_nr = sg_len; edma_prep_slave_sg()
496 nslots = min_t(unsigned, MAX_NR_SG, sg_len); edma_prep_slave_sg()
513 for_each_sg(sgl, sg, sg_len, i) { for_each_sg()
537 if (i == sg_len - 1) for_each_sg()
754 edesc->residue -= edesc->sg_len; edma_callback()
448 edma_prep_slave_sg( struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long tx_flags, void *context) edma_prep_slave_sg() argument
H A Dmxs-dma.c501 unsigned int sg_len, enum dma_transfer_direction direction, mxs_dma_prep_slave_sg()
516 if (sg_len + (append ? idx : 0) > NUM_CCW) { mxs_dma_prep_slave_sg()
519 sg_len, NUM_CCW); mxs_dma_prep_slave_sg()
545 for (j = 0; j < sg_len;) mxs_dma_prep_slave_sg()
555 ccw->bits |= BF_CCW(sg_len, PIO_NUM); mxs_dma_prep_slave_sg()
558 for_each_sg(sgl, sg, sg_len, i) { for_each_sg()
579 if (i + 1 == sg_len) { for_each_sg()
499 mxs_dma_prep_slave_sg( struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) mxs_dma_prep_slave_sg() argument
H A Ddma-jz4740.c391 unsigned int sg_len, enum dma_transfer_direction direction, jz4740_dma_prep_slave_sg()
399 desc = jz4740_dma_alloc_desc(sg_len); jz4740_dma_prep_slave_sg()
403 for_each_sg(sgl, sg, sg_len, i) { for_each_sg()
408 desc->num_sgs = sg_len;
389 jz4740_dma_prep_slave_sg( struct dma_chan *c, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) jz4740_dma_prep_slave_sg() argument
H A Dmoxart-dma.c270 unsigned int sg_len, enum dma_transfer_direction dir, moxart_prep_slave_sg()
311 d = kzalloc(sizeof(*d) + sg_len * sizeof(d->sg[0]), GFP_ATOMIC); moxart_prep_slave_sg()
319 for_each_sg(sgl, sgent, sg_len, i) { for_each_sg()
324 d->sglen = sg_len;
268 moxart_prep_slave_sg( struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction dir, unsigned long tx_flags, void *context) moxart_prep_slave_sg() argument
H A Dep93xx_dma.c1017 * @sg_len: number of entries in @sgl
1026 unsigned int sg_len, enum dma_transfer_direction dir, ep93xx_dma_prep_slave_sg()
1047 for_each_sg(sgl, sg, sg_len, i) { for_each_sg()
1048 size_t sg_len = sg_dma_len(sg); for_each_sg() local
1050 if (sg_len > DMA_MAX_CHAN_BYTES) { for_each_sg()
1052 sg_len); for_each_sg() local
1069 desc->size = sg_len; for_each_sg()
1025 ep93xx_dma_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction dir, unsigned long flags, void *context) ep93xx_dma_prep_slave_sg() argument
H A Dste_dma40.c1039 static int d40_sg_2_dmalen(struct scatterlist *sgl, int sg_len, d40_sg_2_dmalen() argument
1047 for_each_sg(sgl, sg, sg_len, i) { for_each_sg()
2139 unsigned int sg_len, dma_addr_t src_dev_addr, d40_prep_sg_log()
2147 ret = d40_log_sg_to_lli(sg_src, sg_len, d40_prep_sg_log()
2154 ret = d40_log_sg_to_lli(sg_dst, sg_len, d40_prep_sg_log()
2167 unsigned int sg_len, dma_addr_t src_dev_addr, d40_prep_sg_phy()
2179 ret = d40_phy_sg_to_lli(sg_src, sg_len, src_dev_addr, d40_prep_sg_phy()
2185 ret = d40_phy_sg_to_lli(sg_dst, sg_len, dst_dev_addr, d40_prep_sg_phy()
2199 unsigned int sg_len, unsigned long dma_flags) d40_prep_desc()
2209 desc->lli_len = d40_sg_2_dmalen(sg, sg_len, cfg->src_info.data_width, d40_prep_desc()
2237 struct scatterlist *sg_dst, unsigned int sg_len, d40_prep_sg()
2254 desc = d40_prep_desc(chan, sg_src, sg_len, dma_flags); d40_prep_sg()
2258 if (sg_next(&sg_src[sg_len - 1]) == sg_src) d40_prep_sg()
2268 sg_len, src_dev_addr, dst_dev_addr); d40_prep_sg()
2271 sg_len, src_dev_addr, dst_dev_addr); d40_prep_sg()
2536 unsigned int sg_len, enum dma_transfer_direction direction, d40_prep_slave_sg()
2542 return d40_prep_sg(chan, sgl, sgl, sg_len, direction, dma_flags); d40_prep_slave_sg()
2137 d40_prep_sg_log(struct d40_chan *chan, struct d40_desc *desc, struct scatterlist *sg_src, struct scatterlist *sg_dst, unsigned int sg_len, dma_addr_t src_dev_addr, dma_addr_t dst_dev_addr) d40_prep_sg_log() argument
2165 d40_prep_sg_phy(struct d40_chan *chan, struct d40_desc *desc, struct scatterlist *sg_src, struct scatterlist *sg_dst, unsigned int sg_len, dma_addr_t src_dev_addr, dma_addr_t dst_dev_addr) d40_prep_sg_phy() argument
2198 d40_prep_desc(struct d40_chan *chan, struct scatterlist *sg, unsigned int sg_len, unsigned long dma_flags) d40_prep_desc() argument
2236 d40_prep_sg(struct dma_chan *dchan, struct scatterlist *sg_src, struct scatterlist *sg_dst, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long dma_flags) d40_prep_sg() argument
2535 d40_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long dma_flags, void *context) d40_prep_slave_sg() argument
H A Ddma-jz4780.c296 struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, jz4780_dma_prep_slave_sg()
304 desc = jz4780_dma_desc_alloc(jzchan, sg_len, DMA_SLAVE); jz4780_dma_prep_slave_sg()
308 for (i = 0; i < sg_len; i++) { jz4780_dma_prep_slave_sg()
319 if (i != (sg_len - 1)) { jz4780_dma_prep_slave_sg()
295 jz4780_dma_prep_slave_sg( struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags) jz4780_dma_prep_slave_sg() argument
H A Dtimb_dma.c509 struct scatterlist *sgl, unsigned int sg_len, td_prep_slave_sg()
520 if (!sgl || !sg_len) { td_prep_slave_sg()
540 for_each_sg(sgl, sg, sg_len, i) { for_each_sg()
548 i == (sg_len - 1)); for_each_sg()
508 td_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) td_prep_slave_sg() argument
H A Dste_dma40_ll.h442 int sg_len,
454 int sg_len,
H A Ds3c24xx-dma.c892 unsigned sg_len; s3c24xx_dma_prep_dma_cyclic() local
953 sg_len = size / period; s3c24xx_dma_prep_dma_cyclic()
955 for (i = 0; i < sg_len; i++) { s3c24xx_dma_prep_dma_cyclic()
965 if (i == sg_len - 1) s3c24xx_dma_prep_dma_cyclic()
981 unsigned int sg_len, enum dma_transfer_direction direction, s3c24xx_dma_prep_slave_sg()
1047 for_each_sg(sgl, sg, sg_len, tmp) { for_each_sg()
979 s3c24xx_dma_prep_slave_sg( struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) s3c24xx_dma_prep_slave_sg() argument
H A Dimx-sdma.c1035 unsigned int sg_len, enum dma_transfer_direction direction, sdma_prep_slave_sg()
1053 sg_len, channel); sdma_prep_slave_sg()
1060 if (sg_len > NUM_BD) { sdma_prep_slave_sg()
1062 channel, sg_len, NUM_BD); sdma_prep_slave_sg()
1068 for_each_sg(sgl, sg, sg_len, i) { for_each_sg()
1111 if (i + 1 == sg_len) { for_each_sg()
1125 sdmac->num_bd = sg_len;
1033 sdma_prep_slave_sg( struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) sdma_prep_slave_sg() argument
H A Dpch_dma.c583 struct scatterlist *sgl, unsigned int sg_len, pd_prep_slave_sg()
596 if (unlikely(!sg_len)) { pd_prep_slave_sg()
611 for_each_sg(sgl, sg, sg_len, i) { for_each_sg()
582 pd_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) pd_prep_slave_sg() argument
H A Dqcom_bam_dma.c581 * @sg_len: length of sg
587 struct scatterlist *sgl, unsigned int sg_len, bam_prep_slave_sg()
606 for_each_sg(sgl, sg, sg_len, i) bam_prep_slave_sg()
630 for_each_sg(sgl, sg, sg_len, i) { for_each_sg()
586 bam_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) bam_prep_slave_sg() argument
H A Dcoh901318.c1272 unsigned int sg_len; member in struct:coh901318_desc
1531 desc->sg_len = 0; coh901318_desc_get()
2291 unsigned int sg_len, enum dma_transfer_direction direction, coh901318_prep_slave_sg()
2316 dev_vdbg(COHC_2_DEV(cohc), "[%s] sg_len %d dir %d\n", coh901318_prep_slave_sg()
2317 __func__, sg_len, direction); coh901318_prep_slave_sg() local
2357 for_each_sg(sgl, sg, sg_len, i) { for_each_sg()
2380 ret = coh901318_lli_fill_sg(&cohc->base->pool, lli, sgl, sg_len,
2290 coh901318_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) coh901318_prep_slave_sg() argument
H A Dat_xdmac.c592 unsigned int sg_len, enum dma_transfer_direction direction, at_xdmac_prep_slave_sg()
611 dev_dbg(chan2dev(chan), "%s: sg_len=%d, dir=%s, flags=0x%lx\n", at_xdmac_prep_slave_sg()
612 __func__, sg_len, at_xdmac_prep_slave_sg() local
623 for_each_sg(sgl, sg, sg_len, i) { for_each_sg()
659 | (i == sg_len - 1 ? 0 : AT_XDMAC_MBR_UBC_NDE) /* descriptor fetch */ for_each_sg()
591 at_xdmac_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) at_xdmac_prep_slave_sg() argument
H A Dat_hdmac.c726 * @sg_len: number of entries in @scatterlist
733 unsigned int sg_len, enum dma_transfer_direction direction, atc_prep_slave_sg()
751 sg_len, atc_prep_slave_sg() local
755 if (unlikely(!atslave || !sg_len)) { atc_prep_slave_sg()
773 for_each_sg(sgl, sg, sg_len, i) { for_each_sg()
814 for_each_sg(sgl, sg, sg_len, i) { for_each_sg()
732 atc_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) atc_prep_slave_sg() argument
H A Dcppi41.c493 struct dma_chan *chan, struct scatterlist *sgl, unsigned sg_len, cppi41_dma_prep_slave_sg()
504 for_each_sg(sgl, sg, sg_len, i) { for_each_sg()
492 cppi41_dma_prep_slave_sg( struct dma_chan *chan, struct scatterlist *sgl, unsigned sg_len, enum dma_transfer_direction dir, unsigned long tx_flags, void *context) cppi41_dma_prep_slave_sg() argument
H A Dimg-mdc-dma.c453 unsigned int sg_len, enum dma_transfer_direction dir, mdc_prep_slave_sg()
478 for_each_sg(sgl, sg, sg_len, i) { for_each_sg()
451 mdc_prep_slave_sg( struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction dir, unsigned long flags, void *context) mdc_prep_slave_sg() argument
H A Dtegra20-apb-dma.c922 struct dma_chan *dc, struct scatterlist *sgl, unsigned int sg_len, tegra_dma_prep_slave_sg()
941 if (sg_len < 1) { tegra_dma_prep_slave_sg()
942 dev_err(tdc2dev(tdc), "Invalid segment length %d\n", sg_len); tegra_dma_prep_slave_sg() local
978 for_each_sg(sgl, sg, sg_len, i) { for_each_sg()
921 tegra_dma_prep_slave_sg( struct dma_chan *dc, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) tegra_dma_prep_slave_sg() argument
H A Dimx-dma.c810 unsigned int sg_len, enum dma_transfer_direction direction, imxdma_prep_slave_sg()
824 for_each_sg(sgl, sg, sg_len, i) { for_each_sg()
845 desc->sgcount = sg_len;
808 imxdma_prep_slave_sg( struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) imxdma_prep_slave_sg() argument
H A Dmmp_pdma.c527 unsigned int sg_len, enum dma_transfer_direction dir, mmp_pdma_prep_slave_sg()
537 if ((sgl == NULL) || (sg_len == 0)) mmp_pdma_prep_slave_sg()
542 for_each_sg(sgl, sg, sg_len, i) { for_each_sg()
526 mmp_pdma_prep_slave_sg(struct dma_chan *dchan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction dir, unsigned long flags, void *context) mmp_pdma_prep_slave_sg() argument
H A Dmpc512x_dma.c689 unsigned int sg_len, enum dma_transfer_direction direction, mpc_dma_prep_slave_sg()
704 if (sg_len != 1) mpc_dma_prep_slave_sg()
710 for_each_sg(sgl, sg, sg_len, i) { for_each_sg()
688 mpc_dma_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) mpc_dma_prep_slave_sg() argument
H A Dtxx9dmac.c811 unsigned int sg_len, enum dma_transfer_direction direction, txx9dmac_prep_slave_sg()
829 if (unlikely(!sg_len)) txx9dmac_prep_slave_sg()
834 for_each_sg(sgl, sg, sg_len, i) { for_each_sg()
810 txx9dmac_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) txx9dmac_prep_slave_sg() argument
H A Dnbpfaxi.c998 struct dma_chan *dchan, struct scatterlist *sgl, unsigned int sg_len, nbpf_prep_slave_sg()
1011 return nbpf_prep_sg(chan, sgl, &slave_sg, sg_len, nbpf_prep_slave_sg()
1016 return nbpf_prep_sg(chan, &slave_sg, sgl, sg_len, nbpf_prep_slave_sg()
997 nbpf_prep_slave_sg( struct dma_chan *dchan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) nbpf_prep_slave_sg() argument
H A Dsun6i-dma.c562 unsigned int sg_len, enum dma_transfer_direction dir, sun6i_dma_prep_slave_sg()
586 for_each_sg(sgl, sg, sg_len, i) { for_each_sg()
560 sun6i_dma_prep_slave_sg( struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction dir, unsigned long flags, void *context) sun6i_dma_prep_slave_sg() argument
H A Damba-pl08x.c989 * sg_len == 1 should be true, as there can be two cases here: pl08x_fill_llis_for_desc()
1584 unsigned int sg_len, enum dma_transfer_direction direction, pl08x_prep_slave_sg()
1601 for_each_sg(sgl, sg, sg_len, tmp) { for_each_sg()
1582 pl08x_prep_slave_sg( struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) pl08x_prep_slave_sg() argument
H A Dpl330.c2655 unsigned int sg_len, enum dma_transfer_direction direction, pl330_prep_slave_sg()
2664 if (unlikely(!pch || !sgl || !sg_len)) pl330_prep_slave_sg()
2671 for_each_sg(sgl, sg, sg_len, i) { for_each_sg()
2654 pl330_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flg, void *context) pl330_prep_slave_sg() argument
/linux-4.1.27/drivers/rapidio/devices/
H A Dtsi721_dma.c457 for_each_sg(desc->sg, sg, desc->sg_len, i) { tsi721_submit_sg()
460 i, desc->sg_len, tsi721_submit_sg()
494 desc->sg_len -= i; tsi721_submit_sg()
524 desc->sg_len = 0; tsi721_submit_sg()
599 if (desc->sg_len == 0) { tsi721_dma_tasklet()
757 struct scatterlist *sgl, unsigned int sg_len, tsi721_prep_rio_sg()
767 if (!sgl || !sg_len) { tsi721_prep_rio_sg()
805 desc->sg_len = sg_len; tsi721_prep_rio_sg()
756 tsi721_prep_rio_sg(struct dma_chan *dchan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction dir, unsigned long flags, void *tinfo) tsi721_prep_rio_sg() argument
H A Dtsi721.h657 unsigned int sg_len; member in struct:tsi721_tx_desc
/linux-4.1.27/drivers/crypto/nx/
H A Dnx.c96 unsigned int sg_len = 0; nx_build_sg_list() local
112 * length of data described by that element to sg_len. Once @len bytes nx_build_sg_list()
121 for (sg = sg_head; sg_len < *len; sg++) { nx_build_sg_list()
130 sg_len += sg->len; nx_build_sg_list()
133 is_vmalloc_addr(start_addr + sg_len)) { nx_build_sg_list()
135 start_addr + sg_len)); nx_build_sg_list()
136 end_addr = sg_addr + *len - sg_len; nx_build_sg_list()
146 *len = sg_len; nx_build_sg_list()
/linux-4.1.27/drivers/dma/hsu/
H A Dhsu.c224 unsigned int sg_len, enum dma_transfer_direction direction, hsu_dma_prep_slave_sg()
232 desc = hsu_dma_alloc_desc(sg_len); hsu_dma_prep_slave_sg()
236 for_each_sg(sgl, sg, sg_len, i) { for_each_sg()
241 desc->nents = sg_len;
222 hsu_dma_prep_slave_sg( struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) hsu_dma_prep_slave_sg() argument
/linux-4.1.27/include/linux/dma/
H A Dipu-dma.h150 unsigned int sg_len; /* tx-descriptor. */ member in struct:idmac_tx_desc
/linux-4.1.27/drivers/scsi/
H A Dgdth_ioctl.h38 u32 sg_len; /* length */ member in struct:__anon8913
44 u32 sg_len; /* length */ member in struct:__anon8914
H A Dgdth.c2535 cmdp->u.cache64.sg_lst[i].sg_len = sg_dma_len(sl); scsi_for_each_sg()
2547 cmdp->u.cache.sg_lst[i].sg_len = sg_dma_len(sl); scsi_for_each_sg()
2565 cmdp->u.cache64.sg_lst[0].sg_len));
2574 cmdp->u.cache.sg_lst[0].sg_len));
2710 cmdp->u.raw64.sg_lst[i].sg_len = sg_dma_len(sl); scsi_for_each_sg()
2722 cmdp->u.raw.sg_lst[i].sg_len = sg_dma_len(sl); scsi_for_each_sg()
2738 cmdp->u.raw64.sg_lst[0].sg_len));
2746 cmdp->u.raw.sg_lst[0].sg_len));
4204 gen.command.u.cache64.sg_lst[0].sg_len = gen.data_len; ioc_general()
4205 gen.command.u.cache64.sg_lst[1].sg_len = 0; ioc_general()
4215 gen.command.u.cache.sg_lst[0].sg_len = gen.data_len; ioc_general()
4216 gen.command.u.cache.sg_lst[1].sg_len = 0; ioc_general()
4240 gen.command.u.raw64.sg_lst[0].sg_len = gen.data_len; ioc_general()
4241 gen.command.u.raw64.sg_lst[1].sg_len = 0; ioc_general()
4252 gen.command.u.raw.sg_lst[0].sg_len = gen.data_len; ioc_general()
4253 gen.command.u.raw.sg_lst[1].sg_len = 0; ioc_general()
H A Da100u2w.c646 scb->sg_len = cpu_to_le32(0); orc_device_reset()
861 scb->sg_len = cpu_to_le32(0); inia100_build_scb()
873 scb->sg_len = cpu_to_le32((u32) (count_sg * 8)); scsi_for_each_sg()
880 scb->sg_len = cpu_to_le32(0);
H A Dscsi_lib.c3091 size_t sg_len = 0, len_complete = 0; scsi_kmap_atomic_sg() local
3098 len_complete = sg_len; /* Complete sg-entries */ for_each_sg()
3099 sg_len += sg->length; for_each_sg()
3100 if (sg_len > *offset) for_each_sg()
3107 __func__, sg_len, *offset, sg_count);
3120 sg_len = PAGE_SIZE - *offset;
3121 if (*len > sg_len)
3122 *len = sg_len;
H A Dips.c1825 ++scb->sg_len; ips_fill_scb_sg_single()
1863 scb->sg_len = 0; ips_flash_firmware()
1932 scb->sg_len = 0; ips_usrcmd()
2697 scb->sg_len = 0; ips_next()
3571 if (!scb->sg_len) { ips_send_cmd()
3592 scb->cmd.basic_io.sg_count = scb->sg_len; ips_send_cmd()
3617 if (!scb->sg_len) { ips_send_cmd()
3638 scb->cmd.basic_io.sg_count = scb->sg_len; ips_send_cmd()
3753 if (!scb->sg_len) { ips_send_cmd()
3786 tapeDCDB->sg_count = scb->sg_len; ips_send_cmd()
3793 if (!scb->sg_len) { ips_send_cmd()
3825 scb->dcdb.sg_count = scb->sg_len; ips_send_cmd()
H A Da100u2w.h154 u32 sg_len; /*10 SG list # * 8 */ member in struct:orc_scb
H A Dips.h1097 uint32_t sg_len; member in struct:ips_scb
1123 uint32_t sg_len; member in struct:ips_scb_pt
/linux-4.1.27/drivers/scsi/bnx2fc/
H A Dbnx2fc_io.c18 static int bnx2fc_split_bd(struct bnx2fc_cmd *io_req, u64 addr, int sg_len,
1617 static int bnx2fc_split_bd(struct bnx2fc_cmd *io_req, u64 addr, int sg_len, bnx2fc_split_bd() argument
1624 while (sg_len) { bnx2fc_split_bd()
1625 if (sg_len >= BNX2FC_BD_SPLIT_SZ) bnx2fc_split_bd()
1628 frag_size = sg_len; bnx2fc_split_bd()
1636 sg_len -= frag_size; bnx2fc_split_bd()
1653 unsigned int sg_len; bnx2fc_map_sg() local
1664 sg_len = sg_dma_len(sg); scsi_for_each_sg()
1666 if (sg_len > BNX2FC_MAX_BD_LEN) { scsi_for_each_sg()
1667 sg_frags = bnx2fc_split_bd(io_req, addr, sg_len, scsi_for_each_sg()
1674 bd[bd_count].buf_len = (u16)sg_len; scsi_for_each_sg()
1678 byte_count += sg_len; scsi_for_each_sg()
/linux-4.1.27/drivers/crypto/ux500/hash/
H A Dhash_alg.h305 * @sg_len: Scatterlist length.
314 int sg_len; member in struct:hash_dma
H A Dhash_core.c170 ctx->device->dma.sg_len = dma_map_sg(channel->device->dev, hash_set_dma_transfer()
174 if (!ctx->device->dma.sg_len) { hash_set_dma_transfer()
183 ctx->device->dma.sg, ctx->device->dma.sg_len, hash_set_dma_transfer()
207 ctx->device->dma.sg_len, DMA_TO_DEVICE); hash_dma_done()
/linux-4.1.27/drivers/ata/
H A Dsata_fsl.c467 u32 sg_len = sg_dma_len(sg); sata_fsl_fill_sg() local
469 VPRINTK("SATA FSL : fill_sg, sg_addr = 0x%llx, sg_len = %d\n", sata_fsl_fill_sg()
470 (unsigned long long)sg_addr, sg_len); sata_fsl_fill_sg()
476 if (unlikely(sg_len & 0x03)) sata_fsl_fill_sg()
478 sg_len); sata_fsl_fill_sg()
490 ttl_dwords += sg_len; sata_fsl_fill_sg()
492 prd->ddc_and_ext = cpu_to_le32(data_snoop | (sg_len & ~0x03)); sata_fsl_fill_sg()
500 indirect_ext_segment_sz += sg_len; sata_fsl_fill_sg()
H A Dsata_promise.c604 u32 sg_len; pdc_fill_sg() local
611 sg_len = sg_dma_len(sg); pdc_fill_sg()
613 while (sg_len) { pdc_fill_sg()
615 len = sg_len; pdc_fill_sg()
616 if ((offset + sg_len) > 0x10000) pdc_fill_sg()
624 sg_len -= len; pdc_fill_sg()
H A Dlibata-sff.c2642 u32 sg_len, len; ata_bmdma_fill_sg() local
2649 sg_len = sg_dma_len(sg); ata_bmdma_fill_sg()
2651 while (sg_len) { ata_bmdma_fill_sg()
2653 len = sg_len; ata_bmdma_fill_sg()
2654 if ((offset + sg_len) > 0x10000) ata_bmdma_fill_sg()
2662 sg_len -= len; ata_bmdma_fill_sg()
2693 u32 sg_len, len, blen; ata_bmdma_fill_sg_dumb() local
2700 sg_len = sg_dma_len(sg); ata_bmdma_fill_sg_dumb()
2702 while (sg_len) { ata_bmdma_fill_sg_dumb()
2704 len = sg_len; ata_bmdma_fill_sg_dumb()
2705 if ((offset + sg_len) > 0x10000) ata_bmdma_fill_sg_dumb()
2722 sg_len -= len; ata_bmdma_fill_sg_dumb()
H A Dsata_sil.c322 u32 sg_len = sg_dma_len(sg); sil_fill_sg() local
325 prd->flags_len = cpu_to_le32(sg_len); sil_fill_sg()
326 VPRINTK("PRD[%u] = (0x%X, 0x%X)\n", si, addr, sg_len); sil_fill_sg()
H A Dpata_macio.c529 u32 addr, sg_len, len; pata_macio_qc_prep() local
536 sg_len = sg_dma_len(sg); pata_macio_qc_prep()
538 while (sg_len) { pata_macio_qc_prep()
542 len = (sg_len < MAX_DBDMA_SEG) ? sg_len : MAX_DBDMA_SEG; pata_macio_qc_prep()
550 sg_len -= len; pata_macio_qc_prep()
H A Dacard-ahci.c242 u32 sg_len = sg_dma_len(sg); acard_ahci_fill_sg() local
251 acard_sg[si].size = cpu_to_le32(sg_len); acard_ahci_fill_sg()
H A Dsata_rcar.c536 u32 addr, sg_len; sata_rcar_bmdma_fill_sg() local
543 sg_len = sg_dma_len(sg); sata_rcar_bmdma_fill_sg()
546 prd[si].flags_len = cpu_to_le32(sg_len); sata_rcar_bmdma_fill_sg()
547 VPRINTK("PRD[%u] = (0x%X, 0x%X)\n", si, addr, sg_len); sata_rcar_bmdma_fill_sg()
H A Dsata_nv.c2018 u32 sg_len, len; nv_swncq_fill_sg() local
2021 sg_len = sg_dma_len(sg); nv_swncq_fill_sg()
2023 while (sg_len) { nv_swncq_fill_sg()
2025 len = sg_len; nv_swncq_fill_sg()
2026 if ((offset + sg_len) > 0x10000) nv_swncq_fill_sg()
2033 sg_len -= len; nv_swncq_fill_sg()
H A Dlibahci.c1503 u32 sg_len = sg_dma_len(sg); ahci_fill_sg() local
1507 ahci_sg[si].flags_size = cpu_to_le32(sg_len - 1); ahci_fill_sg()
H A Dsata_mv.c1810 u32 sg_len = sg_dma_len(sg); mv_fill_sg() local
1812 while (sg_len) { mv_fill_sg()
1814 u32 len = sg_len; mv_fill_sg()
1824 sg_len -= len; mv_fill_sg()
/linux-4.1.27/drivers/crypto/amcc/
H A Dcrypto4xx_core.c550 u32 sg_len; crypto4xx_copy_pkt_to_dst() local
559 sg_len = sg->length; crypto4xx_copy_pkt_to_dst()
580 sg_len -= len; crypto4xx_copy_pkt_to_dst()
581 if (sg_len) { crypto4xx_copy_pkt_to_dst()
584 &sg_len, &this_sd, &offset, &nbytes)) crypto4xx_copy_pkt_to_dst()
/linux-4.1.27/include/linux/
H A Ddmaengine.h663 unsigned int sg_len, enum dma_transfer_direction direction,
713 struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, dmaengine_prep_slave_sg()
716 return chan->device->device_prep_slave_sg(chan, sgl, sg_len, dmaengine_prep_slave_sg()
723 struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, dmaengine_prep_rio_sg()
727 return chan->device->device_prep_slave_sg(chan, sgl, sg_len, dmaengine_prep_rio_sg()
712 dmaengine_prep_slave_sg( struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction dir, unsigned long flags) dmaengine_prep_slave_sg() argument
722 dmaengine_prep_rio_sg( struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction dir, unsigned long flags, struct rio_dma_ext *rio_ext) dmaengine_prep_rio_sg() argument
H A Drio.h440 unsigned int sg_len; /* size of scatter list */ member in struct:rio_dma_data
/linux-4.1.27/drivers/scsi/be2iscsi/
H A Dbe_main.c2303 unsigned int sg_len, index; hwi_write_sgl_v2() local
2318 sg_len = sg_dma_len(sg); hwi_write_sgl_v2()
2328 sg_len); hwi_write_sgl_v2()
2329 sge_len = sg_len; hwi_write_sgl_v2()
2333 sg_len = sg_dma_len(sg); hwi_write_sgl_v2()
2343 sg_len); hwi_write_sgl_v2()
2378 sg_len = sg_dma_len(sg); hwi_write_sgl_v2()
2384 AMAP_SET_BITS(struct amap_iscsi_sge, len, psgl, sg_len); hwi_write_sgl_v2()
2387 offset += sg_len; hwi_write_sgl_v2()
2398 unsigned int sg_len, index; hwi_write_sgl() local
2413 sg_len = sg_dma_len(sg); hwi_write_sgl()
2420 sg_len); hwi_write_sgl()
2421 sge_len = sg_len; hwi_write_sgl()
2425 sg_len = sg_dma_len(sg); hwi_write_sgl()
2432 sg_len); hwi_write_sgl()
2466 sg_len = sg_dma_len(sg); hwi_write_sgl()
2472 AMAP_SET_BITS(struct amap_iscsi_sge, len, psgl, sg_len); hwi_write_sgl()
2475 offset += sg_len; hwi_write_sgl()
/linux-4.1.27/include/linux/mmc/
H A Dcore.h123 unsigned int sg_len; /* size of scatter list */ member in struct:mmc_data
H A Ddw_mmc.h213 void (*start)(struct dw_mci *host, unsigned int sg_len);
/linux-4.1.27/drivers/dma/dw/
H A Dcore.c762 unsigned int sg_len, enum dma_transfer_direction direction, dwc_prep_slave_sg()
781 if (unlikely(!is_slave_direction(direction) || !sg_len)) dwc_prep_slave_sg()
802 for_each_sg(sgl, sg, sg_len, i) { for_each_sg()
859 for_each_sg(sgl, sg, sg_len, i) { for_each_sg()
761 dwc_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) dwc_prep_slave_sg() argument
/linux-4.1.27/drivers/dma/ipu/
H A Dipu_idmac.c1323 for_each_sg(desc->sg, sg, desc->sg_len, k) { ipu_gc_tasklet()
1338 struct scatterlist *sgl, unsigned int sg_len, idmac_prep_slave_sg()
1366 desc->sg_len = sg_len; idmac_prep_slave_sg()
1337 idmac_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long tx_flags, void *context) idmac_prep_slave_sg() argument
/linux-4.1.27/drivers/scsi/bfa/
H A Dbfa_fcpim.c2500 sge->sg_len = pgdlen; bfa_ioim_send_ioreq()
2511 sgpge->sg_len = sg_dma_len(sg); bfa_ioim_send_ioreq()
2512 pgcumsz += sgpge->sg_len; bfa_ioim_send_ioreq()
2530 sgpge->sg_len = pgcumsz; bfa_ioim_send_ioreq()
2536 sgpge->sg_len = pgcumsz; bfa_ioim_send_ioreq()
2550 sge->sg_len = pgdlen; bfa_ioim_send_ioreq()
H A Dbfi.h126 sg_len:28; member in struct:bfi_sge_s
128 u32 sg_len:28,
H A Dbfad_bsg.c3212 return sge->sg_len; bfad_fcxp_get_req_sglen_cb()
3234 return sge->sg_len; bfad_fcxp_get_rsp_sglen_cb()
3283 sg_table->sg_len = buf_info->size; bfad_fcxp_map_sg()
H A Dbfa_ioc.h64 u32 sg_len; member in struct:bfa_sge_s
/linux-4.1.27/drivers/scsi/bnx2i/
H A Dbnx2i_iscsi.c151 int sg_len; bnx2i_map_scsi_sg() local
160 sg_len = sg_dma_len(sg); scsi_for_each_sg()
164 bd[bd_count].buffer_length = sg_len; scsi_for_each_sg()
169 byte_count += sg_len; scsi_for_each_sg()
/linux-4.1.27/drivers/scsi/qla2xxx/
H A Dqla_iocb.c881 uint32_t cumulative_partial, sg_len; qla24xx_get_one_block_sg() local
891 sg_len = sg_dma_len(sg); qla24xx_get_one_block_sg()
895 if ((cumulative_partial + (sg_len - sgx->bytes_consumed)) >= blk_sz) { qla24xx_get_one_block_sg()
901 sgx->dma_len = sg_len - sgx->bytes_consumed; qla24xx_get_one_block_sg()
908 if (sg_len == sgx->bytes_consumed) { qla24xx_get_one_block_sg()
/linux-4.1.27/drivers/scsi/csiostor/
H A Dcsio_scsi.c566 uint32_t sg_len = 0; csio_setup_ddp() local
571 sg_len = sg_dma_len(sgel); csio_setup_ddp()
578 sg_addr, sg_len); csio_setup_ddp()
584 ((buf_off + sg_len) & (ddp_pagesz - 1))) { csio_setup_ddp()
587 "(%llx:%d)\n", sg_addr, sg_len); csio_setup_ddp()
/linux-4.1.27/drivers/crypto/ccp/
H A Dccp-ops.c676 struct scatterlist *sg, u64 sg_len, ccp_init_data()
684 ret = ccp_init_sg_workarea(&data->sg_wa, cmd_q->ccp->dev, sg, sg_len, ccp_init_data()
675 ccp_init_data(struct ccp_data *data, struct ccp_cmd_queue *cmd_q, struct scatterlist *sg, u64 sg_len, unsigned int dm_len, enum dma_data_direction dir) ccp_init_data() argument
/linux-4.1.27/drivers/net/wireless/ath/ath6kl/
H A Dsdio.c255 data->sg_len = scat_req->scat_entries; ath6kl_sdio_setup_scat_data()
/linux-4.1.27/drivers/net/wireless/brcm80211/brcmfmac/
H A Dbcmsdh.c612 mmc_dat.sg_len = sg_cnt; brcmf_sdiod_sglist_rw()
/linux-4.1.27/drivers/media/common/saa7146/
H A Dsaa7146_video.c215 DEB_EE("dev:%p, buf:%p, sg_len:%d\n", dev, buf, length); saa7146_pgtable_build()
/linux-4.1.27/fs/exofs/
H A Dore.c992 " dev=%d sg_len=%d\n", _LLU(obj->id), _ore_read_mirror()
/linux-4.1.27/drivers/rapidio/
H A Drio.c1588 return dmaengine_prep_rio_sg(dchan, data->sg, data->sg_len, rio_dma_prep_xfer()

Completed in 3023 milliseconds