/linux-4.1.27/drivers/mmc/host/ |
H A D | tmio_mmc_dma.c | 57 for_each_sg(sg, sg_tmp, host->sg_len, i) { tmio_mmc_start_dma_rx() 66 if ((!aligned && (host->sg_len > 1 || sg->length > PAGE_CACHE_SIZE || tmio_mmc_start_dma_rx() 86 ret = dma_map_sg(chan->device->dev, sg, host->sg_len, DMA_FROM_DEVICE); tmio_mmc_start_dma_rx() 99 __func__, host->sg_len, ret, cookie, host->mrq); tmio_mmc_start_dma_rx() 120 desc, cookie, host->sg_len); tmio_mmc_start_dma_rx() 133 for_each_sg(sg, sg_tmp, host->sg_len, i) { tmio_mmc_start_dma_tx() 142 if ((!aligned && (host->sg_len > 1 || sg->length > PAGE_CACHE_SIZE || tmio_mmc_start_dma_tx() 166 ret = dma_map_sg(chan->device->dev, sg, host->sg_len, DMA_TO_DEVICE); tmio_mmc_start_dma_tx() 179 __func__, host->sg_len, ret, cookie, host->mrq); tmio_mmc_start_dma_tx() 248 host->sg_ptr, host->sg_len, tmio_mmc_tasklet_fn() 252 host->sg_ptr, host->sg_len, tmio_mmc_tasklet_fn()
|
H A D | android-goldfish.c | 127 unsigned int sg_len; member in struct:goldfish_mmc_host 229 dma_unmap_sg(mmc_dev(host->mmc), data->sg, host->sg_len, goldfish_mmc_xfer_done() 234 host->sg_len = 0; goldfish_mmc_xfer_done() 371 unsigned sg_len; goldfish_mmc_prepare_data() local 391 sg_len = (data->blocks == 1) ? 1 : data->sg_len; goldfish_mmc_prepare_data() 398 host->sg_len = dma_map_sg(mmc_dev(host->mmc), data->sg, goldfish_mmc_prepare_data() 399 sg_len, dma_data_dir); goldfish_mmc_prepare_data()
|
H A D | mxs-mmc.c | 158 data->sg_len, ssp->dma_dir); mxs_mmc_request_done() 231 unsigned int sg_len; mxs_mmc_prep_dma() local 236 data->sg_len, ssp->dma_dir); mxs_mmc_prep_dma() 238 sg_len = data->sg_len; mxs_mmc_prep_dma() 242 sg_len = SSP_PIO_NUM; mxs_mmc_prep_dma() 246 sgl, sg_len, ssp->slave_dirn, flags); mxs_mmc_prep_dma() 253 data->sg_len, ssp->dma_dir); mxs_mmc_prep_dma() 357 unsigned int sg_len = data->sg_len; mxs_mmc_adtc() local 402 for_each_sg(sgl, sg, sg_len, i) mxs_mmc_adtc()
|
H A D | au1xmmc.c | 353 dma_unmap_sg(mmc_dev(host->mmc), data->sg, data->sg_len, host->dma.dir); au1xmmc_data_complete() 396 int sg_len, max, count; au1xmmc_send_pio() local 411 sg_len = data->sg[host->pio.index].length - host->pio.offset; au1xmmc_send_pio() 414 max = (sg_len > host->pio.len) ? host->pio.len : sg_len; au1xmmc_send_pio() 433 if (count == sg_len) { au1xmmc_send_pio() 451 int max, count, sg_len = 0; au1xmmc_receive_pio() local 468 sg_len = sg_dma_len(&data->sg[host->pio.index]) - host->pio.offset; au1xmmc_receive_pio() 471 if (sg_len < max) au1xmmc_receive_pio() 472 max = sg_len; au1xmmc_receive_pio() 510 if (sg_len && count == sg_len) { au1xmmc_receive_pio() 631 data->sg_len, host->dma.dir); au1xmmc_prepare_data() 647 int sg_len = sg->length; au1xmmc_prepare_data() local 649 int len = (datalen > sg_len) ? sg_len : datalen; au1xmmc_prepare_data() 682 dma_unmap_sg(mmc_dev(host->mmc), data->sg, data->sg_len, au1xmmc_prepare_data()
|
H A D | jz4740_mmc.c | 118 int sg_len; member in struct:jz4740_mmc_host_next 153 int sg_len; member in struct:jz4740_mmc_host 220 dma_unmap_sg(chan->device->dev, data->sg, data->sg_len, dir); jz4740_mmc_dma_unmap() 231 int sg_len; jz4740_mmc_prepare_dma_data() local 245 sg_len = dma_map_sg(chan->device->dev, jz4740_mmc_prepare_dma_data() 247 data->sg_len, jz4740_mmc_prepare_dma_data() 251 sg_len = next_data->sg_len; jz4740_mmc_prepare_dma_data() 252 next_data->sg_len = 0; jz4740_mmc_prepare_dma_data() 255 if (sg_len <= 0) { jz4740_mmc_prepare_dma_data() 262 next->sg_len = sg_len; jz4740_mmc_prepare_dma_data() 265 host->sg_len = sg_len; jz4740_mmc_prepare_dma_data() 302 host->sg_len, jz4740_mmc_start_dma_transfer() 690 sg_miter_start(&host->miter, data->sg, data->sg_len, direction); jz_mmc_prepare_data_transfer()
|
H A D | omap.c | 149 unsigned int sg_len; member in struct:mmc_omap_host 427 dma_unmap_sg(dev, data->sg, host->sg_len, dma_data_dir); mmc_omap_release_dma() 451 host->sg_len = 0; mmc_omap_xfer_done() 508 host->sg_len = 0; mmc_omap_abort_xfer() 673 BUG_ON(host->sg_idx == host->sg_len); mmc_omap_xfer_data() 951 unsigned sg_len; mmc_omap_prepare_data() local 972 sg_len = (data->blocks == 1) ? 1 : data->sg_len; mmc_omap_prepare_data() 975 for (i = 0; i < sg_len; i++) { mmc_omap_prepare_data() 1034 host->sg_len = dma_map_sg(c->device->dev, data->sg, sg_len, mmc_omap_prepare_data() 1036 if (host->sg_len == 0) mmc_omap_prepare_data() 1039 tx = dmaengine_prep_slave_sg(c, data->sg, host->sg_len, mmc_omap_prepare_data() 1060 host->sg_len = sg_len; mmc_omap_prepare_data()
|
H A D | tifm_sd.c | 106 int sg_len; member in struct:tifm_sd 175 if (host->sg_pos == host->sg_len) tifm_sd_transfer_data() 182 if (host->sg_pos == host->sg_len) { tifm_sd_transfer_data() 238 if (host->sg_pos == host->sg_len) tifm_sd_bounce_block() 271 if (host->sg_pos == host->sg_len) tifm_sd_set_dma_data() 279 if (host->sg_pos == host->sg_len) tifm_sd_set_dma_data() 287 if (host->sg_pos == host->sg_len) tifm_sd_set_dma_data() 666 host->sg_len = r_data->sg_len; tifm_sd_request() 680 host->sg_len = tifm_map_sg(sock, r_data->sg, tifm_sd_request() 681 r_data->sg_len, tifm_sd_request() 686 if (host->sg_len < 1) { tifm_sd_request() 767 tifm_unmap_sg(sock, r_data->sg, r_data->sg_len, tifm_sd_end_cmd()
|
H A D | davinci_mmc.c | 214 unsigned int sg_len; member in struct:mmc_davinci_host 438 host->sg_len, mmc_davinci_send_dma_request() 460 host->sg_len, mmc_davinci_send_dma_request() 485 host->sg_len = dma_map_sg(mmc_dev(host->mmc), data->sg, data->sg_len, mmc_davinci_start_dma_transfer() 491 for (i = 0; i < host->sg_len; i++) { mmc_davinci_start_dma_transfer() 494 data->sg, data->sg_len, mmc_davinci_start_dma_transfer() 622 host->sg_len = data->sg_len; mmc_davinci_prepare_data() 829 dma_unmap_sg(mmc_dev(host->mmc), data->sg, data->sg_len, mmc_davinci_xfer_done()
|
H A D | usdhi6rol0.c | 317 __func__, host->mrq->cmd->opcode, data->sg_len, usdhi6_blk_bounce() 354 struct scatterlist *sg = data->sg_len > 1 ? host->sg : data->sg; usdhi6_sg_map() 398 struct scatterlist *sg = data->sg_len > 1 ? usdhi6_sg_unmap() 540 data->sg_len, DMA_FROM_DEVICE); usdhi6_dma_stop_unmap() 543 data->sg_len, DMA_TO_DEVICE); usdhi6_dma_stop_unmap() 583 ret = dma_map_sg(chan->device->dev, sg, data->sg_len, data_dir); usdhi6_dma_setup() 597 __func__, data->sg_len, ret, cookie, desc); usdhi6_dma_setup() 627 __func__, data->sg_len, data->blocks, data->blksz); usdhi6_dma_kill() 903 data ? data->blksz : 0, data ? data->sg_len : 0)) usdhi6_request_done() 1044 "Bad SG of %u: %ux%u @ %u\n", data->sg_len, usdhi6_rq_start() 1058 data->sg_len, use_dma ? "DMA" : "PIO", usdhi6_rq_start() 1671 host->offset, data->blocks, data->blksz, data->sg_len, usdhi6_timeout_work()
|
H A D | mxcmmc.c | 294 for_each_sg(data->sg, sg, data->sg_len, i) mxcmci_swap_buffers() 323 for_each_sg(data->sg, sg, data->sg_len, i) { mxcmci_setup_data() 341 data->sg_len, host->dma_dir); mxcmci_setup_data() 342 if (nents != data->sg_len) mxcmci_setup_data() 346 data->sg, data->sg_len, slave_dirn, mxcmci_setup_data() 350 dma_unmap_sg(host->dma->device->dev, data->sg, data->sg_len, mxcmci_setup_data() 460 dma_unmap_sg(host->dma->device->dev, data->sg, data->sg_len, mxcmci_finish_data() 625 for_each_sg(data->sg, sg, data->sg_len, i) { mxcmci_transfer_data() 632 for_each_sg(data->sg, sg, data->sg_len, i) { mxcmci_transfer_data()
|
H A D | dw_mmc.c | 418 data->sg_len, dw_mci_dma_cleanup() 466 unsigned int sg_len) dw_mci_translate_sglist() 475 for (i = 0; i < sg_len; i++) { dw_mci_translate_sglist() 519 for (i = 0; i < sg_len; i++) { dw_mci_translate_sglist() 563 static void dw_mci_idmac_start_dma(struct dw_mci *host, unsigned int sg_len) dw_mci_idmac_start_dma() argument 567 dw_mci_translate_sglist(host, host->data, sg_len); dw_mci_idmac_start_dma() 675 unsigned int i, sg_len; dw_mci_pre_dma_transfer() local 691 for_each_sg(data->sg, sg, data->sg_len, i) { dw_mci_pre_dma_transfer() 696 sg_len = dma_map_sg(host->dev, dw_mci_pre_dma_transfer() 698 data->sg_len, dw_mci_pre_dma_transfer() 700 if (sg_len == 0) dw_mci_pre_dma_transfer() 704 data->host_cookie = sg_len; dw_mci_pre_dma_transfer() 706 return sg_len; dw_mci_pre_dma_transfer() 741 data->sg_len, dw_mci_post_req() 829 int sg_len; dw_mci_submit_data_dma() local 838 sg_len = dw_mci_pre_dma_transfer(host, data, 0); dw_mci_submit_data_dma() 839 if (sg_len < 0) { dw_mci_submit_data_dma() 841 return sg_len; dw_mci_submit_data_dma() 847 "sd sg_cpu: %#lx sg_dma: %#lx sg_len: %d\n", dw_mci_submit_data_dma() 849 sg_len); dw_mci_submit_data_dma() 871 host->dma_ops->start(host, sg_len); dw_mci_submit_data_dma() 901 sg_miter_start(&host->sg_miter, data->sg, data->sg_len, flags); dw_mci_submit_data() 465 dw_mci_translate_sglist(struct dw_mci *host, struct mmc_data *data, unsigned int sg_len) dw_mci_translate_sglist() argument
|
H A D | sh_mmcif.c | 295 ret = dma_map_sg(chan->device->dev, sg, data->sg_len, sh_mmcif_start_dma_rx() 311 __func__, data->sg_len, ret, cookie); sh_mmcif_start_dma_rx() 332 desc, cookie, data->sg_len); sh_mmcif_start_dma_rx() 344 ret = dma_map_sg(chan->device->dev, sg, data->sg_len, sh_mmcif_start_dma_tx() 360 __func__, data->sg_len, ret, cookie); sh_mmcif_start_dma_tx() 573 if (++host->sg_idx < data->sg_len) sh_mmcif_next_block() 579 return host->sg_idx != data->sg_len; sh_mmcif_next_block() 621 if (!data->sg_len || !data->sg->length) sh_mmcif_multi_read() 699 if (!data->sg_len || !data->sg->length) sh_mmcif_multi_write() 1133 data->sg, data->sg_len, sh_mmcif_end_cmd() 1137 data->sg, data->sg_len, sh_mmcif_end_cmd()
|
H A D | atmel-mci.c | 186 unsigned int sg_len; member in struct:atmel_mci 803 data->sg, data->sg_len, atmci_pdc_cleanup() 825 sg_copy_from_buffer(host->data->sg, host->data->sg_len, atmci_pdc_complete() 842 data->sg, data->sg_len, atmci_dma_cleanup() 908 host->sg_len = data->sg_len; atmci_prepare_data() 944 unsigned int sg_len; atmci_prepare_data_pdc() local 973 sg_len = dma_map_sg(&host->pdev->dev, data->sg, data->sg_len, dir); atmci_prepare_data_pdc() 977 sg_copy_to_buffer(host->data->sg, host->data->sg_len, atmci_prepare_data_pdc() 1022 for_each_sg(data->sg, sg, data->sg_len, i) { atmci_prepare_data_dma() 1050 data->sg_len, direction); atmci_prepare_data_dma() 1065 dma_unmap_sg(chan->device->dev, data->sg, data->sg_len, direction); atmci_prepare_data_dma() 1875 host->sg_len--; atmci_read_data_pio() 1876 if (!sg || !host->sg_len) atmci_read_data_pio() 1889 host->sg_len--; atmci_read_data_pio() 1890 if (!sg || !host->sg_len) atmci_read_data_pio() 1941 host->sg_len--; atmci_write_data_pio() 1942 if (!sg || !host->sg_len) atmci_write_data_pio() 1956 host->sg_len--; atmci_write_data_pio() 1957 if (!sg || !host->sg_len) { atmci_write_data_pio()
|
H A D | moxart-mmc.c | 155 host->num_sg = data->sg_len; moxart_init_sg() 278 data->sg_len, dir_data); moxart_transfer_dma() 303 data->sg, data->sg_len, moxart_transfer_dma()
|
H A D | wmt-sdmmc.c | 307 req->data->sg_len, DMA_TO_DEVICE); wmt_complete_data_request() 310 req->data->sg_len, DMA_FROM_DEVICE); wmt_complete_data_request() 624 req->data->sg_len, DMA_TO_DEVICE); wmt_mci_request() 630 req->data->sg_len, DMA_FROM_DEVICE); wmt_mci_request()
|
H A D | sunxi-mmc.c | 299 for (i = 0; i < data->sg_len; i++) { sunxi_mmc_init_idma_des() 339 dma_len = dma_map_sg(mmc_dev(host->mmc), data->sg, data->sg_len, sunxi_mmc_map_dma() 346 for_each_sg(data->sg, sg, data->sg_len, i) { sunxi_mmc_map_dma() 490 dma_unmap_sg(mmc_dev(host->mmc), data->sg, data->sg_len, sunxi_mmc_finalize_request() 847 dma_unmap_sg(mmc_dev(mmc), data->sg, data->sg_len, sunxi_mmc_request()
|
H A D | s3cmci.c | 328 if (host->pio_sgptr >= host->mrq->data->sg_len) { get_data_buffer() 330 host->pio_sgptr, host->mrq->data->sg_len); get_data_buffer() 341 host->pio_sgptr, host->mrq->data->sg_len); get_data_buffer() 1107 dma_map_sg(mmc_dev(host->mmc), data->sg, data->sg_len, s3cmci_prepare_dma() 1111 desc = dmaengine_prep_slave_sg(host->dma, data->sg, data->sg_len, s3cmci_prepare_dma() 1124 dma_unmap_sg(mmc_dev(host->mmc), data->sg, data->sg_len, s3cmci_prepare_dma()
|
H A D | bfin_sdh.c | 158 host->dma_len = dma_map_sg(mmc_dev(host->mmc), data->sg, data->sg_len, host->dma_dir); sdh_setup_data() 302 dma_unmap_sg(mmc_dev(host->mmc), data->sg, data->sg_len, sdh_data_done()
|
H A D | sdricoh_cs.c | 313 dev_dbg(dev, "transfer: blksz %i blocks %i sg_len %i " sdricoh_request() 315 data->sg_len, data->sg->length); sdricoh_request()
|
H A D | rtsx_pci_sdmmc.c | 176 count = rtsx_pci_dma_map_sg(pcr, data->sg, data->sg_len, read); sd_pre_dma_transfer() 219 rtsx_pci_dma_unmap_sg(pcr, data->sg, data->sg_len, read); sdmmc_post_req() 608 sg_copy_from_buffer(data->sg, data->sg_len, buf, data->blksz); sd_normal_rw() 610 sg_copy_to_buffer(data->sg, data->sg_len, buf, data->blksz); sd_normal_rw()
|
H A D | tmio_mmc.h | 65 unsigned int sg_len; member in struct:tmio_mmc_host
|
H A D | mmci.c | 421 sg_miter_start(&host->sg_miter, data->sg, data->sg_len, flags); mmci_init_sg() 521 dma_unmap_sg(chan->device->dev, data->sg, data->sg_len, dir); mmci_dma_unmap() 606 nr_sg = dma_map_sg(device->dev, data->sg, data->sg_len, buffer_dirn); __mmci_dma_prep_data() 625 dma_unmap_sg(device->dev, data->sg, data->sg_len, buffer_dirn); __mmci_dma_prep_data() 660 data->sg_len, data->blksz, data->blocks, data->flags); mmci_dma_start_data()
|
H A D | omap_hsmmc.c | 955 host->data->sg, host->data->sg_len, omap_hsmmc_dma_cleanup() 1267 data->sg, data->sg_len, omap_hsmmc_dma_callback() 1302 dma_len = dma_map_sg(chan->device->dev, data->sg, data->sg_len, omap_hsmmc_pre_dma_transfer() 1336 for (i = 0; i < data->sg_len; i++) { omap_hsmmc_setup_dma_transfer() 1368 tx = dmaengine_prep_slave_sg(chan, data->sg, data->sg_len, omap_hsmmc_setup_dma_transfer() 1479 dma_unmap_sg(c->device->dev, data->sg, data->sg_len, omap_hsmmc_post_req()
|
H A D | vub300.c | 1450 data->sg_len, 0, GFP_KERNEL); __command_read_data() 1494 sg_copy_from_buffer(data->sg, data->sg_len, buf, __command_read_data() 1519 sg_copy_to_buffer(data->sg, data->sg_len, __command_write_data() 1543 sg_copy_to_buffer(data->sg, data->sg_len, buf, __command_write_data() 1565 sg_copy_to_buffer(data->sg, data->sg_len, buf, sizeof(buf)); __command_write_data() 1568 data->sg_len, 0, GFP_KERNEL); __command_write_data()
|
H A D | cb710-mmc.c | 277 sg_miter_start(&miter, data->sg, data->sg_len, SG_MITER_TO_SG); cb710_mmc_receive() 321 sg_miter_start(&miter, data->sg, data->sg_len, SG_MITER_FROM_SG); cb710_mmc_send()
|
H A D | pxamci.c | 212 host->dma_len = dma_map_sg(mmc_dev(host->mmc), data->sg, data->sg_len, pxamci_setup_data() 362 dma_unmap_sg(mmc_dev(host->mmc), data->sg, data->sg_len, pxamci_data_done()
|
H A D | sdhci.c | 636 data->sg_len, direction); sdhci_adma_table_post() 656 data->sg_len, direction); sdhci_adma_table_post() 798 for_each_sg(data->sg, sg, data->sg_len, i) { sdhci_prepare_data() 833 for_each_sg(data->sg, sg, data->sg_len, i) { sdhci_prepare_data() 909 sg_miter_start(&host->sg_miter, data->sg, data->sg_len, flags); sdhci_prepare_data() 983 data->sg, data->sg_len, sdhci_finish_data() 2153 dma_unmap_sg(mmc_dev(host->mmc), data->sg, data->sg_len, sdhci_post_req() 2172 sg_count = dma_map_sg(mmc_dev(host->mmc), data->sg, data->sg_len, sdhci_pre_dma_transfer()
|
H A D | rtsx_usb_sdmmc.c | 533 data->sg_len, NULL, 10000); sd_rw_multi() 579 sg_copy_from_buffer(data->sg, data->sg_len, buf, data->blksz); sd_normal_rw() 581 sg_copy_to_buffer(data->sg, data->sg_len, buf, data->blksz); sd_normal_rw()
|
H A D | tmio_mmc_pio.c | 73 host->sg_len = data->sg_len; tmio_mmc_init_sg() 83 return --host->sg_len; tmio_mmc_next_sg()
|
H A D | wbsd.c | 246 host->num_sg = data->sg_len; wbsd_init_sg() 284 len = data->sg_len; wbsd_sg_to_dma() 301 len = data->sg_len; wbsd_dma_to_sg()
|
H A D | toshsd.c | 491 sg_miter_start(&host->sg_miter, data->sg, data->sg_len, flags); toshsd_start_data()
|
H A D | via-sdmmc.c | 493 count = dma_map_sg(mmc_dev(host->mmc), data->sg, data->sg_len, via_sdc_preparedata() 637 dma_unmap_sg(mmc_dev(host->mmc), data->sg, data->sg_len, via_sdc_finish_data()
|
H A D | mvsdio.c | 134 data->sg_len, dma_dir); mvsd_setup_data()
|
H A D | mmc_spi.c | 910 for (sg = data->sg, n_sg = data->sg_len; n_sg; n_sg--, sg++) { mmc_spi_data_do()
|
/linux-4.1.27/drivers/mmc/card/ |
H A D | queue.c | 145 static struct scatterlist *mmc_alloc_sg(int sg_len, int *err) mmc_alloc_sg() argument 149 sg = kmalloc(sizeof(struct scatterlist)*sg_len, GFP_KERNEL); mmc_alloc_sg() 154 sg_init_table(sg, sg_len); mmc_alloc_sg() 457 unsigned int sg_len = 0; mmc_queue_packed_map_sg() local 473 sg_len++; mmc_queue_packed_map_sg() 478 sg_len += blk_rq_map_sg(mq->queue, req, __sg); mmc_queue_packed_map_sg() 479 __sg = sg + (sg_len - 1); mmc_queue_packed_map_sg() 482 sg_mark_end(sg + (sg_len - 1)); mmc_queue_packed_map_sg() 483 return sg_len; mmc_queue_packed_map_sg() 491 unsigned int sg_len; mmc_queue_map_sg() local 510 sg_len = mmc_queue_packed_map_sg(mq, mqrq->packed, mmc_queue_map_sg() 513 sg_len = blk_rq_map_sg(mq->queue, mqrq->req, mqrq->bounce_sg); mmc_queue_map_sg() 515 mqrq->bounce_sg_len = sg_len; mmc_queue_map_sg() 518 for_each_sg(mqrq->bounce_sg, sg, sg_len, i) mmc_queue_map_sg()
|
H A D | mmc_test.c | 71 * @sg_len: length of currently mapped scatterlist @sg 82 unsigned int sg_len; member in struct:mmc_test_area 161 unsigned int *sg_len; member in struct:mmc_test_multiple_rw 191 struct mmc_request *mrq, struct scatterlist *sg, unsigned sg_len, mmc_test_prepare_mrq() 222 mrq->data->sg_len = sg_len; mmc_test_prepare_mrq() 397 unsigned int *sg_len, int min_sg_len) mmc_test_map_sg() 407 *sg_len = 0; mmc_test_map_sg() 426 *sg_len += 1; mmc_test_map_sg() 450 unsigned int *sg_len) mmc_test_map_sg_max_scatter() 459 *sg_len = 0; mmc_test_map_sg_max_scatter() 481 *sg_len += 1; mmc_test_map_sg_max_scatter() 586 "%u.%02u IOPS, sg_len %d)\n", mmc_test_print_avg_rate() 591 test->area.sg_len); mmc_test_print_avg_rate() 781 struct scatterlist *sg, unsigned sg_len, mmc_test_nonblock_transfer() 814 mmc_test_prepare_mrq(test, cur_areq->mrq, sg, sg_len, dev_addr, mmc_test_nonblock_transfer() 846 struct scatterlist *sg, unsigned sg_len, unsigned dev_addr, mmc_test_simple_transfer() 858 mmc_test_prepare_mrq(test, &mrq, sg, sg_len, dev_addr, mmc_test_simple_transfer() 903 struct scatterlist *sg, unsigned sg_len, unsigned dev_addr, mmc_test_transfer() 916 sg_copy_from_buffer(sg, sg_len, test->scratch, BUFFER_SIZE); mmc_test_transfer() 923 ret = mmc_test_simple_transfer(test, sg, sg_len, dev_addr, mmc_test_transfer() 963 sg_copy_to_buffer(sg, sg_len, test->scratch, BUFFER_SIZE); mmc_test_transfer() 1435 &t->sg_len); mmc_test_area_map() 1438 t->max_seg_sz, &t->sg_len, min_sg_len); mmc_test_area_map() 1454 return mmc_test_simple_transfer(test, t->sg, t->sg_len, dev_addr, mmc_test_area_transfer() 1494 ret = mmc_test_nonblock_transfer(test, t->sg, t->sg_len, mmc_test_area_io_seq() 2176 rw->sg_len[i]); mmc_test_rw_multiple_sg_len() 2264 unsigned int sg_len[] = {1, 1 << 3, 1 << 4, 1 << 5, 1 << 6, mmc_test_profile_sglen_wr_blocking_perf() local 2267 .sg_len = sg_len, mmc_test_profile_sglen_wr_blocking_perf() 2269 .len = ARRAY_SIZE(sg_len), mmc_test_profile_sglen_wr_blocking_perf() 2283 unsigned int sg_len[] = {1, 1 << 3, 1 << 4, 1 << 5, 1 << 6, mmc_test_profile_sglen_wr_nonblock_perf() local 2286 .sg_len = sg_len, mmc_test_profile_sglen_wr_nonblock_perf() 2288 .len = ARRAY_SIZE(sg_len), mmc_test_profile_sglen_wr_nonblock_perf() 2302 unsigned int sg_len[] = {1, 1 << 3, 1 << 4, 1 << 5, 1 << 6, mmc_test_profile_sglen_r_blocking_perf() local 2305 .sg_len = sg_len, mmc_test_profile_sglen_r_blocking_perf() 2307 .len = ARRAY_SIZE(sg_len), mmc_test_profile_sglen_r_blocking_perf() 2321 unsigned int sg_len[] = {1, 1 << 3, 1 << 4, 1 << 5, 1 << 6, mmc_test_profile_sglen_r_nonblock_perf() local 2324 .sg_len = sg_len, mmc_test_profile_sglen_r_nonblock_perf() 2326 .len = ARRAY_SIZE(sg_len), mmc_test_profile_sglen_r_nonblock_perf() 190 mmc_test_prepare_mrq(struct mmc_test_card *test, struct mmc_request *mrq, struct scatterlist *sg, unsigned sg_len, unsigned dev_addr, unsigned blocks, unsigned blksz, int write) mmc_test_prepare_mrq() argument 394 mmc_test_map_sg(struct mmc_test_mem *mem, unsigned long size, struct scatterlist *sglist, int repeat, unsigned int max_segs, unsigned int max_seg_sz, unsigned int *sg_len, int min_sg_len) mmc_test_map_sg() argument 445 mmc_test_map_sg_max_scatter(struct mmc_test_mem *mem, unsigned long sz, struct scatterlist *sglist, unsigned int max_segs, unsigned int max_seg_sz, unsigned int *sg_len) mmc_test_map_sg_max_scatter() argument 780 mmc_test_nonblock_transfer(struct mmc_test_card *test, struct scatterlist *sg, unsigned sg_len, unsigned dev_addr, unsigned blocks, unsigned blksz, int write, int count) mmc_test_nonblock_transfer() argument 845 mmc_test_simple_transfer(struct mmc_test_card *test, struct scatterlist *sg, unsigned sg_len, unsigned dev_addr, unsigned blocks, unsigned blksz, int write) mmc_test_simple_transfer() argument 902 mmc_test_transfer(struct mmc_test_card *test, struct scatterlist *sg, unsigned sg_len, unsigned dev_addr, unsigned blocks, unsigned blksz, int write) mmc_test_transfer() argument
|
H A D | block.c | 494 data.sg_len = 1; mmc_blk_ioctl_cmd() 706 data.sg_len = 1; mmc_sd_num_wr_blocks() 1494 brq->data.sg_len = mmc_queue_map_sg(mq, mqrq); mmc_blk_rw_rq_prep() 1504 for_each_sg(brq->data.sg, sg, brq->data.sg_len, i) { mmc_blk_rw_rq_prep() 1512 brq->data.sg_len = i; mmc_blk_rw_rq_prep() 1721 brq->data.sg_len = mmc_queue_map_sg(mq, mqrq); mmc_blk_packed_hdr_wrq_prep()
|
/linux-4.1.27/drivers/dma/sh/ |
H A D | usb-dmac.c | 47 * @sg_len: length of sg 58 unsigned int sg_len; member in struct:usb_dmac_desc 266 static int usb_dmac_desc_alloc(struct usb_dmac_chan *chan, unsigned int sg_len, usb_dmac_desc_alloc() argument 272 desc = kzalloc(sizeof(*desc) + sg_len * sizeof(desc->sg[0]), gfp); usb_dmac_desc_alloc() 276 desc->sg_allocated_len = sg_len; usb_dmac_desc_alloc() 302 unsigned int sg_len, gfp_t gfp) usb_dmac_desc_get() 310 if (sg_len <= desc->sg_allocated_len) { usb_dmac_desc_get() 319 if (!usb_dmac_desc_alloc(chan, sg_len, gfp)) { usb_dmac_desc_get() 420 unsigned int sg_len, enum dma_transfer_direction dir, usb_dmac_prep_slave_sg() 428 if (!sg_len) { usb_dmac_prep_slave_sg() 430 "%s: bad parameter: len=%d\n", __func__, sg_len); usb_dmac_prep_slave_sg() 434 desc = usb_dmac_desc_get(uchan, sg_len, GFP_NOWAIT); usb_dmac_prep_slave_sg() 439 desc->sg_len = sg_len; for_each_sg() 440 for_each_sg(sgl, sg, sg_len, i) { for_each_sg() 522 for (i = desc->sg_index + 1; i < desc->sg_len; i++) usb_dmac_chan_get_residue() 586 if (++desc->sg_index < desc->sg_len) { usb_dmac_isr_transfer_end() 301 usb_dmac_desc_get(struct usb_dmac_chan *chan, unsigned int sg_len, gfp_t gfp) usb_dmac_desc_get() argument 419 usb_dmac_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction dir, unsigned long dma_flags, void *context) usb_dmac_prep_slave_sg() argument
|
H A D | shdma-base.c | 564 struct scatterlist *sgl, unsigned int sg_len, dma_addr_t *addr, shdma_prep_sg() 574 for_each_sg(sgl, sg, sg_len, i) shdma_prep_sg() 591 for_each_sg(sgl, sg, sg_len, i) { for_each_sg() 665 struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, shdma_prep_slave_sg() 680 if (slave_id < 0 || !sg_len) { shdma_prep_slave_sg() 682 __func__, sg_len, slave_id); shdma_prep_slave_sg() 688 return shdma_prep_sg(schan, sgl, sg_len, &slave_addr, shdma_prep_slave_sg() 703 unsigned int sg_len = buf_len / period_len; shdma_prep_dma_cyclic() local 714 if (sg_len > SHDMA_MAX_SG_LEN) { shdma_prep_dma_cyclic() 716 sg_len, SHDMA_MAX_SG_LEN); shdma_prep_dma_cyclic() 734 sgl = kcalloc(sg_len, sizeof(*sgl), GFP_KERNEL); shdma_prep_dma_cyclic() 738 sg_init_table(sgl, sg_len); shdma_prep_dma_cyclic() 740 for (i = 0; i < sg_len; i++) { shdma_prep_dma_cyclic() 749 desc = shdma_prep_sg(schan, sgl, sg_len, &slave_addr, shdma_prep_dma_cyclic() 563 shdma_prep_sg(struct shdma_chan *schan, struct scatterlist *sgl, unsigned int sg_len, dma_addr_t *addr, enum dma_transfer_direction direction, unsigned long flags, bool cyclic) shdma_prep_sg() argument 664 shdma_prep_slave_sg( struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) shdma_prep_slave_sg() argument
|
H A D | rcar-dmac.c | 824 unsigned int sg_len, dma_addr_t dev_addr, rcar_dmac_chan_prep_sg() 855 for_each_sg(sgl, sg, sg_len, i) { for_each_sg() 1021 unsigned int sg_len, enum dma_transfer_direction dir, rcar_dmac_prep_slave_sg() 1028 if (rchan->mid_rid < 0 || !sg_len) { rcar_dmac_prep_slave_sg() 1031 __func__, sg_len, rchan->mid_rid); rcar_dmac_prep_slave_sg() 1037 return rcar_dmac_chan_prep_sg(rchan, sgl, sg_len, dev_addr, rcar_dmac_prep_slave_sg() 1052 unsigned int sg_len; rcar_dmac_prep_dma_cyclic() local 1063 sg_len = buf_len / period_len; rcar_dmac_prep_dma_cyclic() 1064 if (sg_len > RCAR_DMAC_MAX_SG_LEN) { rcar_dmac_prep_dma_cyclic() 1067 rchan->index, sg_len, RCAR_DMAC_MAX_SG_LEN); rcar_dmac_prep_dma_cyclic() 1075 sgl = kcalloc(sg_len, sizeof(*sgl), GFP_NOWAIT); rcar_dmac_prep_dma_cyclic() 1079 sg_init_table(sgl, sg_len); rcar_dmac_prep_dma_cyclic() 1081 for (i = 0; i < sg_len; ++i) { rcar_dmac_prep_dma_cyclic() 1092 desc = rcar_dmac_chan_prep_sg(rchan, sgl, sg_len, dev_addr, rcar_dmac_prep_dma_cyclic() 823 rcar_dmac_chan_prep_sg(struct rcar_dmac_chan *chan, struct scatterlist *sgl, unsigned int sg_len, dma_addr_t dev_addr, enum dma_transfer_direction dir, unsigned long dma_flags, bool cyclic) rcar_dmac_chan_prep_sg() argument 1020 rcar_dmac_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction dir, unsigned long flags, void *context) rcar_dmac_prep_slave_sg() argument
|
/linux-4.1.27/drivers/mmc/core/ |
H A D | sdio_ops.c | 166 data.sg_len = nents; mmc_io_rw_extended() 168 for_each_sg(data.sg, sg_ptr, data.sg_len, i) { for_each_sg() 176 data.sg_len = 1;
|
H A D | sd_ops.c | 285 data.sg_len = 1; mmc_app_send_scr() 336 data.sg_len = 1; mmc_sd_switch() 381 data.sg_len = 1; mmc_app_sd_status()
|
H A D | mmc_ops.c | 315 data.sg_len = 1; mmc_send_cxd_data() 625 data.sg_len = 1; mmc_send_tuning() 704 data.sg_len = 1; mmc_send_bus_test()
|
H A D | core.c | 239 for_each_sg(mrq->data->sg, sg, mrq->data->sg_len, i) mmc_start_request()
|
/linux-4.1.27/net/rds/ |
H A D | ib_rdma.c | 59 unsigned int sg_len; member in struct:rds_ib_mr 438 ibmr->sg_len = nents; rds_ib_map_fmr() 474 ibmr->sg, ibmr->sg_len, __rds_ib_teardown_mr() 480 if (ibmr->sg_len) { __rds_ib_teardown_mr() 483 for (i = 0; i < ibmr->sg_len; ++i) { __rds_ib_teardown_mr() 495 ibmr->sg_len = 0; __rds_ib_teardown_mr() 501 unsigned int pinned = ibmr->sg_len; rds_ib_teardown_mr() 641 unpinned += ibmr->sg_len; rds_ib_flush_mr_pool() 700 rdsdebug("RDS/IB: free_mr nents %u\n", ibmr->sg_len); rds_ib_free_mr() 708 atomic_add(ibmr->sg_len, &pool->free_pinned); rds_ib_free_mr()
|
H A D | iw_rdma.c | 252 struct scatterlist *list, unsigned int sg_len) rds_iw_set_scatterlist() 255 sg->len = sg_len; rds_iw_set_scatterlist() 761 unsigned int sg_len) rds_iw_map_fastreg() 768 rds_iw_set_scatterlist(&mapping->m_sg, sg, sg_len); rds_iw_map_fastreg() 251 rds_iw_set_scatterlist(struct rds_iw_scatterlist *sg, struct scatterlist *list, unsigned int sg_len) rds_iw_set_scatterlist() argument 758 rds_iw_map_fastreg(struct rds_iw_mr_pool *pool, struct rds_iw_mr *ibmr, struct scatterlist *sg, unsigned int sg_len) rds_iw_map_fastreg() argument
|
/linux-4.1.27/drivers/dma/ |
H A D | ste_dma40_ll.c | 266 int sg_len, d40_phy_sg_to_lli() 284 for_each_sg(sg, current_sg, sg_len, i) { for_each_sg() 291 if (i == sg_len - 1) for_each_sg() 418 int sg_len, d40_log_sg_to_lli() 433 for_each_sg(sg, current_sg, sg_len, i) { for_each_sg() 265 d40_phy_sg_to_lli(struct scatterlist *sg, int sg_len, dma_addr_t target, struct d40_phy_lli *lli_sg, dma_addr_t lli_phys, u32 reg_cfg, struct stedma40_half_channel_info *info, struct stedma40_half_channel_info *otherinfo, unsigned long flags) d40_phy_sg_to_lli() argument 417 d40_log_sg_to_lli(struct scatterlist *sg, int sg_len, dma_addr_t dev_addr, struct d40_log_lli *lli_sg, u32 lcsp13, u32 data_width1, u32 data_width2) d40_log_sg_to_lli() argument
|
H A D | fsl-edma.c | 503 int sg_len) fsl_edma_alloc_desc() 508 fsl_desc = kzalloc(sizeof(*fsl_desc) + sizeof(struct fsl_edma_sw_tcd) * sg_len, fsl_edma_alloc_desc() 514 fsl_desc->n_tcds = sg_len; fsl_edma_alloc_desc() 515 for (i = 0; i < sg_len; i++) { fsl_edma_alloc_desc() 539 int sg_len, i; fsl_edma_prep_dma_cyclic() local 546 sg_len = buf_len / period_len; fsl_edma_prep_dma_cyclic() 547 fsl_desc = fsl_edma_alloc_desc(fsl_chan, sg_len); fsl_edma_prep_dma_cyclic() 556 for (i = 0; i < sg_len; i++) { fsl_edma_prep_dma_cyclic() 561 last_sg = fsl_desc->tcd[(i + 1) % sg_len].ptcd; fsl_edma_prep_dma_cyclic() 586 unsigned int sg_len, enum dma_transfer_direction direction, fsl_edma_prep_slave_sg() 599 fsl_desc = fsl_edma_alloc_desc(fsl_chan, sg_len); fsl_edma_prep_slave_sg() 605 for_each_sg(sgl, sg, sg_len, i) { for_each_sg() 607 last_sg = fsl_desc->tcd[(i + 1) % sg_len].ptcd; for_each_sg() 622 if (i < sg_len - 1) { for_each_sg() 502 fsl_edma_alloc_desc(struct fsl_edma_chan *fsl_chan, int sg_len) fsl_edma_alloc_desc() argument 584 fsl_edma_prep_slave_sg( struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) fsl_edma_prep_slave_sg() argument
|
H A D | edma.c | 94 * - sg_len: Tracks the length of the current intermediate transfer, 99 u32 sg_len; member in struct:edma_desc 175 edesc->sg_len = 0; edma_execute() 181 edesc->sg_len += edesc->pset[j].len; edma_execute() 450 unsigned int sg_len, enum dma_transfer_direction direction, edma_prep_slave_sg() 462 if (unlikely(!echan || !sgl || !sg_len)) edma_prep_slave_sg() 483 edesc = kzalloc(sizeof(*edesc) + sg_len * edma_prep_slave_sg() 490 edesc->pset_nr = sg_len; edma_prep_slave_sg() 496 nslots = min_t(unsigned, MAX_NR_SG, sg_len); edma_prep_slave_sg() 513 for_each_sg(sgl, sg, sg_len, i) { for_each_sg() 537 if (i == sg_len - 1) for_each_sg() 754 edesc->residue -= edesc->sg_len; edma_callback() 448 edma_prep_slave_sg( struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long tx_flags, void *context) edma_prep_slave_sg() argument
|
H A D | mxs-dma.c | 501 unsigned int sg_len, enum dma_transfer_direction direction, mxs_dma_prep_slave_sg() 516 if (sg_len + (append ? idx : 0) > NUM_CCW) { mxs_dma_prep_slave_sg() 519 sg_len, NUM_CCW); mxs_dma_prep_slave_sg() 545 for (j = 0; j < sg_len;) mxs_dma_prep_slave_sg() 555 ccw->bits |= BF_CCW(sg_len, PIO_NUM); mxs_dma_prep_slave_sg() 558 for_each_sg(sgl, sg, sg_len, i) { for_each_sg() 579 if (i + 1 == sg_len) { for_each_sg() 499 mxs_dma_prep_slave_sg( struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) mxs_dma_prep_slave_sg() argument
|
H A D | dma-jz4740.c | 391 unsigned int sg_len, enum dma_transfer_direction direction, jz4740_dma_prep_slave_sg() 399 desc = jz4740_dma_alloc_desc(sg_len); jz4740_dma_prep_slave_sg() 403 for_each_sg(sgl, sg, sg_len, i) { for_each_sg() 408 desc->num_sgs = sg_len; 389 jz4740_dma_prep_slave_sg( struct dma_chan *c, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) jz4740_dma_prep_slave_sg() argument
|
H A D | moxart-dma.c | 270 unsigned int sg_len, enum dma_transfer_direction dir, moxart_prep_slave_sg() 311 d = kzalloc(sizeof(*d) + sg_len * sizeof(d->sg[0]), GFP_ATOMIC); moxart_prep_slave_sg() 319 for_each_sg(sgl, sgent, sg_len, i) { for_each_sg() 324 d->sglen = sg_len; 268 moxart_prep_slave_sg( struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction dir, unsigned long tx_flags, void *context) moxart_prep_slave_sg() argument
|
H A D | ep93xx_dma.c | 1017 * @sg_len: number of entries in @sgl 1026 unsigned int sg_len, enum dma_transfer_direction dir, ep93xx_dma_prep_slave_sg() 1047 for_each_sg(sgl, sg, sg_len, i) { for_each_sg() 1048 size_t sg_len = sg_dma_len(sg); for_each_sg() local 1050 if (sg_len > DMA_MAX_CHAN_BYTES) { for_each_sg() 1052 sg_len); for_each_sg() local 1069 desc->size = sg_len; for_each_sg() 1025 ep93xx_dma_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction dir, unsigned long flags, void *context) ep93xx_dma_prep_slave_sg() argument
|
H A D | ste_dma40.c | 1039 static int d40_sg_2_dmalen(struct scatterlist *sgl, int sg_len, d40_sg_2_dmalen() argument 1047 for_each_sg(sgl, sg, sg_len, i) { for_each_sg() 2139 unsigned int sg_len, dma_addr_t src_dev_addr, d40_prep_sg_log() 2147 ret = d40_log_sg_to_lli(sg_src, sg_len, d40_prep_sg_log() 2154 ret = d40_log_sg_to_lli(sg_dst, sg_len, d40_prep_sg_log() 2167 unsigned int sg_len, dma_addr_t src_dev_addr, d40_prep_sg_phy() 2179 ret = d40_phy_sg_to_lli(sg_src, sg_len, src_dev_addr, d40_prep_sg_phy() 2185 ret = d40_phy_sg_to_lli(sg_dst, sg_len, dst_dev_addr, d40_prep_sg_phy() 2199 unsigned int sg_len, unsigned long dma_flags) d40_prep_desc() 2209 desc->lli_len = d40_sg_2_dmalen(sg, sg_len, cfg->src_info.data_width, d40_prep_desc() 2237 struct scatterlist *sg_dst, unsigned int sg_len, d40_prep_sg() 2254 desc = d40_prep_desc(chan, sg_src, sg_len, dma_flags); d40_prep_sg() 2258 if (sg_next(&sg_src[sg_len - 1]) == sg_src) d40_prep_sg() 2268 sg_len, src_dev_addr, dst_dev_addr); d40_prep_sg() 2271 sg_len, src_dev_addr, dst_dev_addr); d40_prep_sg() 2536 unsigned int sg_len, enum dma_transfer_direction direction, d40_prep_slave_sg() 2542 return d40_prep_sg(chan, sgl, sgl, sg_len, direction, dma_flags); d40_prep_slave_sg() 2137 d40_prep_sg_log(struct d40_chan *chan, struct d40_desc *desc, struct scatterlist *sg_src, struct scatterlist *sg_dst, unsigned int sg_len, dma_addr_t src_dev_addr, dma_addr_t dst_dev_addr) d40_prep_sg_log() argument 2165 d40_prep_sg_phy(struct d40_chan *chan, struct d40_desc *desc, struct scatterlist *sg_src, struct scatterlist *sg_dst, unsigned int sg_len, dma_addr_t src_dev_addr, dma_addr_t dst_dev_addr) d40_prep_sg_phy() argument 2198 d40_prep_desc(struct d40_chan *chan, struct scatterlist *sg, unsigned int sg_len, unsigned long dma_flags) d40_prep_desc() argument 2236 d40_prep_sg(struct dma_chan *dchan, struct scatterlist *sg_src, struct scatterlist *sg_dst, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long dma_flags) d40_prep_sg() argument 2535 d40_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long dma_flags, void *context) d40_prep_slave_sg() argument
|
H A D | dma-jz4780.c | 296 struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, jz4780_dma_prep_slave_sg() 304 desc = jz4780_dma_desc_alloc(jzchan, sg_len, DMA_SLAVE); jz4780_dma_prep_slave_sg() 308 for (i = 0; i < sg_len; i++) { jz4780_dma_prep_slave_sg() 319 if (i != (sg_len - 1)) { jz4780_dma_prep_slave_sg() 295 jz4780_dma_prep_slave_sg( struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags) jz4780_dma_prep_slave_sg() argument
|
H A D | timb_dma.c | 509 struct scatterlist *sgl, unsigned int sg_len, td_prep_slave_sg() 520 if (!sgl || !sg_len) { td_prep_slave_sg() 540 for_each_sg(sgl, sg, sg_len, i) { for_each_sg() 548 i == (sg_len - 1)); for_each_sg() 508 td_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) td_prep_slave_sg() argument
|
H A D | ste_dma40_ll.h | 442 int sg_len, 454 int sg_len,
|
H A D | s3c24xx-dma.c | 892 unsigned sg_len; s3c24xx_dma_prep_dma_cyclic() local 953 sg_len = size / period; s3c24xx_dma_prep_dma_cyclic() 955 for (i = 0; i < sg_len; i++) { s3c24xx_dma_prep_dma_cyclic() 965 if (i == sg_len - 1) s3c24xx_dma_prep_dma_cyclic() 981 unsigned int sg_len, enum dma_transfer_direction direction, s3c24xx_dma_prep_slave_sg() 1047 for_each_sg(sgl, sg, sg_len, tmp) { for_each_sg() 979 s3c24xx_dma_prep_slave_sg( struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) s3c24xx_dma_prep_slave_sg() argument
|
H A D | imx-sdma.c | 1035 unsigned int sg_len, enum dma_transfer_direction direction, sdma_prep_slave_sg() 1053 sg_len, channel); sdma_prep_slave_sg() 1060 if (sg_len > NUM_BD) { sdma_prep_slave_sg() 1062 channel, sg_len, NUM_BD); sdma_prep_slave_sg() 1068 for_each_sg(sgl, sg, sg_len, i) { for_each_sg() 1111 if (i + 1 == sg_len) { for_each_sg() 1125 sdmac->num_bd = sg_len; 1033 sdma_prep_slave_sg( struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) sdma_prep_slave_sg() argument
|
H A D | pch_dma.c | 583 struct scatterlist *sgl, unsigned int sg_len, pd_prep_slave_sg() 596 if (unlikely(!sg_len)) { pd_prep_slave_sg() 611 for_each_sg(sgl, sg, sg_len, i) { for_each_sg() 582 pd_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) pd_prep_slave_sg() argument
|
H A D | qcom_bam_dma.c | 581 * @sg_len: length of sg 587 struct scatterlist *sgl, unsigned int sg_len, bam_prep_slave_sg() 606 for_each_sg(sgl, sg, sg_len, i) bam_prep_slave_sg() 630 for_each_sg(sgl, sg, sg_len, i) { for_each_sg() 586 bam_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) bam_prep_slave_sg() argument
|
H A D | coh901318.c | 1272 unsigned int sg_len; member in struct:coh901318_desc 1531 desc->sg_len = 0; coh901318_desc_get() 2291 unsigned int sg_len, enum dma_transfer_direction direction, coh901318_prep_slave_sg() 2316 dev_vdbg(COHC_2_DEV(cohc), "[%s] sg_len %d dir %d\n", coh901318_prep_slave_sg() 2317 __func__, sg_len, direction); coh901318_prep_slave_sg() local 2357 for_each_sg(sgl, sg, sg_len, i) { for_each_sg() 2380 ret = coh901318_lli_fill_sg(&cohc->base->pool, lli, sgl, sg_len, 2290 coh901318_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) coh901318_prep_slave_sg() argument
|
H A D | at_xdmac.c | 592 unsigned int sg_len, enum dma_transfer_direction direction, at_xdmac_prep_slave_sg() 611 dev_dbg(chan2dev(chan), "%s: sg_len=%d, dir=%s, flags=0x%lx\n", at_xdmac_prep_slave_sg() 612 __func__, sg_len, at_xdmac_prep_slave_sg() local 623 for_each_sg(sgl, sg, sg_len, i) { for_each_sg() 659 | (i == sg_len - 1 ? 0 : AT_XDMAC_MBR_UBC_NDE) /* descriptor fetch */ for_each_sg() 591 at_xdmac_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) at_xdmac_prep_slave_sg() argument
|
H A D | at_hdmac.c | 726 * @sg_len: number of entries in @scatterlist 733 unsigned int sg_len, enum dma_transfer_direction direction, atc_prep_slave_sg() 751 sg_len, atc_prep_slave_sg() local 755 if (unlikely(!atslave || !sg_len)) { atc_prep_slave_sg() 773 for_each_sg(sgl, sg, sg_len, i) { for_each_sg() 814 for_each_sg(sgl, sg, sg_len, i) { for_each_sg() 732 atc_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) atc_prep_slave_sg() argument
|
H A D | cppi41.c | 493 struct dma_chan *chan, struct scatterlist *sgl, unsigned sg_len, cppi41_dma_prep_slave_sg() 504 for_each_sg(sgl, sg, sg_len, i) { for_each_sg() 492 cppi41_dma_prep_slave_sg( struct dma_chan *chan, struct scatterlist *sgl, unsigned sg_len, enum dma_transfer_direction dir, unsigned long tx_flags, void *context) cppi41_dma_prep_slave_sg() argument
|
H A D | img-mdc-dma.c | 453 unsigned int sg_len, enum dma_transfer_direction dir, mdc_prep_slave_sg() 478 for_each_sg(sgl, sg, sg_len, i) { for_each_sg() 451 mdc_prep_slave_sg( struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction dir, unsigned long flags, void *context) mdc_prep_slave_sg() argument
|
H A D | tegra20-apb-dma.c | 922 struct dma_chan *dc, struct scatterlist *sgl, unsigned int sg_len, tegra_dma_prep_slave_sg() 941 if (sg_len < 1) { tegra_dma_prep_slave_sg() 942 dev_err(tdc2dev(tdc), "Invalid segment length %d\n", sg_len); tegra_dma_prep_slave_sg() local 978 for_each_sg(sgl, sg, sg_len, i) { for_each_sg() 921 tegra_dma_prep_slave_sg( struct dma_chan *dc, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) tegra_dma_prep_slave_sg() argument
|
H A D | imx-dma.c | 810 unsigned int sg_len, enum dma_transfer_direction direction, imxdma_prep_slave_sg() 824 for_each_sg(sgl, sg, sg_len, i) { for_each_sg() 845 desc->sgcount = sg_len; 808 imxdma_prep_slave_sg( struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) imxdma_prep_slave_sg() argument
|
H A D | mmp_pdma.c | 527 unsigned int sg_len, enum dma_transfer_direction dir, mmp_pdma_prep_slave_sg() 537 if ((sgl == NULL) || (sg_len == 0)) mmp_pdma_prep_slave_sg() 542 for_each_sg(sgl, sg, sg_len, i) { for_each_sg() 526 mmp_pdma_prep_slave_sg(struct dma_chan *dchan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction dir, unsigned long flags, void *context) mmp_pdma_prep_slave_sg() argument
|
H A D | mpc512x_dma.c | 689 unsigned int sg_len, enum dma_transfer_direction direction, mpc_dma_prep_slave_sg() 704 if (sg_len != 1) mpc_dma_prep_slave_sg() 710 for_each_sg(sgl, sg, sg_len, i) { for_each_sg() 688 mpc_dma_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) mpc_dma_prep_slave_sg() argument
|
H A D | txx9dmac.c | 811 unsigned int sg_len, enum dma_transfer_direction direction, txx9dmac_prep_slave_sg() 829 if (unlikely(!sg_len)) txx9dmac_prep_slave_sg() 834 for_each_sg(sgl, sg, sg_len, i) { for_each_sg() 810 txx9dmac_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) txx9dmac_prep_slave_sg() argument
|
H A D | nbpfaxi.c | 998 struct dma_chan *dchan, struct scatterlist *sgl, unsigned int sg_len, nbpf_prep_slave_sg() 1011 return nbpf_prep_sg(chan, sgl, &slave_sg, sg_len, nbpf_prep_slave_sg() 1016 return nbpf_prep_sg(chan, &slave_sg, sgl, sg_len, nbpf_prep_slave_sg() 997 nbpf_prep_slave_sg( struct dma_chan *dchan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) nbpf_prep_slave_sg() argument
|
H A D | sun6i-dma.c | 562 unsigned int sg_len, enum dma_transfer_direction dir, sun6i_dma_prep_slave_sg() 586 for_each_sg(sgl, sg, sg_len, i) { for_each_sg() 560 sun6i_dma_prep_slave_sg( struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction dir, unsigned long flags, void *context) sun6i_dma_prep_slave_sg() argument
|
H A D | amba-pl08x.c | 989 * sg_len == 1 should be true, as there can be two cases here: pl08x_fill_llis_for_desc() 1584 unsigned int sg_len, enum dma_transfer_direction direction, pl08x_prep_slave_sg() 1601 for_each_sg(sgl, sg, sg_len, tmp) { for_each_sg() 1582 pl08x_prep_slave_sg( struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) pl08x_prep_slave_sg() argument
|
H A D | pl330.c | 2655 unsigned int sg_len, enum dma_transfer_direction direction, pl330_prep_slave_sg() 2664 if (unlikely(!pch || !sgl || !sg_len)) pl330_prep_slave_sg() 2671 for_each_sg(sgl, sg, sg_len, i) { for_each_sg() 2654 pl330_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flg, void *context) pl330_prep_slave_sg() argument
|
/linux-4.1.27/drivers/rapidio/devices/ |
H A D | tsi721_dma.c | 457 for_each_sg(desc->sg, sg, desc->sg_len, i) { tsi721_submit_sg() 460 i, desc->sg_len, tsi721_submit_sg() 494 desc->sg_len -= i; tsi721_submit_sg() 524 desc->sg_len = 0; tsi721_submit_sg() 599 if (desc->sg_len == 0) { tsi721_dma_tasklet() 757 struct scatterlist *sgl, unsigned int sg_len, tsi721_prep_rio_sg() 767 if (!sgl || !sg_len) { tsi721_prep_rio_sg() 805 desc->sg_len = sg_len; tsi721_prep_rio_sg() 756 tsi721_prep_rio_sg(struct dma_chan *dchan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction dir, unsigned long flags, void *tinfo) tsi721_prep_rio_sg() argument
|
H A D | tsi721.h | 657 unsigned int sg_len; member in struct:tsi721_tx_desc
|
/linux-4.1.27/drivers/crypto/nx/ |
H A D | nx.c | 96 unsigned int sg_len = 0; nx_build_sg_list() local 112 * length of data described by that element to sg_len. Once @len bytes nx_build_sg_list() 121 for (sg = sg_head; sg_len < *len; sg++) { nx_build_sg_list() 130 sg_len += sg->len; nx_build_sg_list() 133 is_vmalloc_addr(start_addr + sg_len)) { nx_build_sg_list() 135 start_addr + sg_len)); nx_build_sg_list() 136 end_addr = sg_addr + *len - sg_len; nx_build_sg_list() 146 *len = sg_len; nx_build_sg_list()
|
/linux-4.1.27/drivers/dma/hsu/ |
H A D | hsu.c | 224 unsigned int sg_len, enum dma_transfer_direction direction, hsu_dma_prep_slave_sg() 232 desc = hsu_dma_alloc_desc(sg_len); hsu_dma_prep_slave_sg() 236 for_each_sg(sgl, sg, sg_len, i) { for_each_sg() 241 desc->nents = sg_len; 222 hsu_dma_prep_slave_sg( struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) hsu_dma_prep_slave_sg() argument
|
/linux-4.1.27/include/linux/dma/ |
H A D | ipu-dma.h | 150 unsigned int sg_len; /* tx-descriptor. */ member in struct:idmac_tx_desc
|
/linux-4.1.27/drivers/scsi/ |
H A D | gdth_ioctl.h | 38 u32 sg_len; /* length */ member in struct:__anon8913 44 u32 sg_len; /* length */ member in struct:__anon8914
|
H A D | gdth.c | 2535 cmdp->u.cache64.sg_lst[i].sg_len = sg_dma_len(sl); scsi_for_each_sg() 2547 cmdp->u.cache.sg_lst[i].sg_len = sg_dma_len(sl); scsi_for_each_sg() 2565 cmdp->u.cache64.sg_lst[0].sg_len)); 2574 cmdp->u.cache.sg_lst[0].sg_len)); 2710 cmdp->u.raw64.sg_lst[i].sg_len = sg_dma_len(sl); scsi_for_each_sg() 2722 cmdp->u.raw.sg_lst[i].sg_len = sg_dma_len(sl); scsi_for_each_sg() 2738 cmdp->u.raw64.sg_lst[0].sg_len)); 2746 cmdp->u.raw.sg_lst[0].sg_len)); 4204 gen.command.u.cache64.sg_lst[0].sg_len = gen.data_len; ioc_general() 4205 gen.command.u.cache64.sg_lst[1].sg_len = 0; ioc_general() 4215 gen.command.u.cache.sg_lst[0].sg_len = gen.data_len; ioc_general() 4216 gen.command.u.cache.sg_lst[1].sg_len = 0; ioc_general() 4240 gen.command.u.raw64.sg_lst[0].sg_len = gen.data_len; ioc_general() 4241 gen.command.u.raw64.sg_lst[1].sg_len = 0; ioc_general() 4252 gen.command.u.raw.sg_lst[0].sg_len = gen.data_len; ioc_general() 4253 gen.command.u.raw.sg_lst[1].sg_len = 0; ioc_general()
|
H A D | a100u2w.c | 646 scb->sg_len = cpu_to_le32(0); orc_device_reset() 861 scb->sg_len = cpu_to_le32(0); inia100_build_scb() 873 scb->sg_len = cpu_to_le32((u32) (count_sg * 8)); scsi_for_each_sg() 880 scb->sg_len = cpu_to_le32(0);
|
H A D | scsi_lib.c | 3091 size_t sg_len = 0, len_complete = 0; scsi_kmap_atomic_sg() local 3098 len_complete = sg_len; /* Complete sg-entries */ for_each_sg() 3099 sg_len += sg->length; for_each_sg() 3100 if (sg_len > *offset) for_each_sg() 3107 __func__, sg_len, *offset, sg_count); 3120 sg_len = PAGE_SIZE - *offset; 3121 if (*len > sg_len) 3122 *len = sg_len;
|
H A D | ips.c | 1825 ++scb->sg_len; ips_fill_scb_sg_single() 1863 scb->sg_len = 0; ips_flash_firmware() 1932 scb->sg_len = 0; ips_usrcmd() 2697 scb->sg_len = 0; ips_next() 3571 if (!scb->sg_len) { ips_send_cmd() 3592 scb->cmd.basic_io.sg_count = scb->sg_len; ips_send_cmd() 3617 if (!scb->sg_len) { ips_send_cmd() 3638 scb->cmd.basic_io.sg_count = scb->sg_len; ips_send_cmd() 3753 if (!scb->sg_len) { ips_send_cmd() 3786 tapeDCDB->sg_count = scb->sg_len; ips_send_cmd() 3793 if (!scb->sg_len) { ips_send_cmd() 3825 scb->dcdb.sg_count = scb->sg_len; ips_send_cmd()
|
H A D | a100u2w.h | 154 u32 sg_len; /*10 SG list # * 8 */ member in struct:orc_scb
|
H A D | ips.h | 1097 uint32_t sg_len; member in struct:ips_scb 1123 uint32_t sg_len; member in struct:ips_scb_pt
|
/linux-4.1.27/drivers/scsi/bnx2fc/ |
H A D | bnx2fc_io.c | 18 static int bnx2fc_split_bd(struct bnx2fc_cmd *io_req, u64 addr, int sg_len, 1617 static int bnx2fc_split_bd(struct bnx2fc_cmd *io_req, u64 addr, int sg_len, bnx2fc_split_bd() argument 1624 while (sg_len) { bnx2fc_split_bd() 1625 if (sg_len >= BNX2FC_BD_SPLIT_SZ) bnx2fc_split_bd() 1628 frag_size = sg_len; bnx2fc_split_bd() 1636 sg_len -= frag_size; bnx2fc_split_bd() 1653 unsigned int sg_len; bnx2fc_map_sg() local 1664 sg_len = sg_dma_len(sg); scsi_for_each_sg() 1666 if (sg_len > BNX2FC_MAX_BD_LEN) { scsi_for_each_sg() 1667 sg_frags = bnx2fc_split_bd(io_req, addr, sg_len, scsi_for_each_sg() 1674 bd[bd_count].buf_len = (u16)sg_len; scsi_for_each_sg() 1678 byte_count += sg_len; scsi_for_each_sg()
|
/linux-4.1.27/drivers/crypto/ux500/hash/ |
H A D | hash_alg.h | 305 * @sg_len: Scatterlist length. 314 int sg_len; member in struct:hash_dma
|
H A D | hash_core.c | 170 ctx->device->dma.sg_len = dma_map_sg(channel->device->dev, hash_set_dma_transfer() 174 if (!ctx->device->dma.sg_len) { hash_set_dma_transfer() 183 ctx->device->dma.sg, ctx->device->dma.sg_len, hash_set_dma_transfer() 207 ctx->device->dma.sg_len, DMA_TO_DEVICE); hash_dma_done()
|
/linux-4.1.27/drivers/ata/ |
H A D | sata_fsl.c | 467 u32 sg_len = sg_dma_len(sg); sata_fsl_fill_sg() local 469 VPRINTK("SATA FSL : fill_sg, sg_addr = 0x%llx, sg_len = %d\n", sata_fsl_fill_sg() 470 (unsigned long long)sg_addr, sg_len); sata_fsl_fill_sg() 476 if (unlikely(sg_len & 0x03)) sata_fsl_fill_sg() 478 sg_len); sata_fsl_fill_sg() 490 ttl_dwords += sg_len; sata_fsl_fill_sg() 492 prd->ddc_and_ext = cpu_to_le32(data_snoop | (sg_len & ~0x03)); sata_fsl_fill_sg() 500 indirect_ext_segment_sz += sg_len; sata_fsl_fill_sg()
|
H A D | sata_promise.c | 604 u32 sg_len; pdc_fill_sg() local 611 sg_len = sg_dma_len(sg); pdc_fill_sg() 613 while (sg_len) { pdc_fill_sg() 615 len = sg_len; pdc_fill_sg() 616 if ((offset + sg_len) > 0x10000) pdc_fill_sg() 624 sg_len -= len; pdc_fill_sg()
|
H A D | libata-sff.c | 2642 u32 sg_len, len; ata_bmdma_fill_sg() local 2649 sg_len = sg_dma_len(sg); ata_bmdma_fill_sg() 2651 while (sg_len) { ata_bmdma_fill_sg() 2653 len = sg_len; ata_bmdma_fill_sg() 2654 if ((offset + sg_len) > 0x10000) ata_bmdma_fill_sg() 2662 sg_len -= len; ata_bmdma_fill_sg() 2693 u32 sg_len, len, blen; ata_bmdma_fill_sg_dumb() local 2700 sg_len = sg_dma_len(sg); ata_bmdma_fill_sg_dumb() 2702 while (sg_len) { ata_bmdma_fill_sg_dumb() 2704 len = sg_len; ata_bmdma_fill_sg_dumb() 2705 if ((offset + sg_len) > 0x10000) ata_bmdma_fill_sg_dumb() 2722 sg_len -= len; ata_bmdma_fill_sg_dumb()
|
H A D | sata_sil.c | 322 u32 sg_len = sg_dma_len(sg); sil_fill_sg() local 325 prd->flags_len = cpu_to_le32(sg_len); sil_fill_sg() 326 VPRINTK("PRD[%u] = (0x%X, 0x%X)\n", si, addr, sg_len); sil_fill_sg()
|
H A D | pata_macio.c | 529 u32 addr, sg_len, len; pata_macio_qc_prep() local 536 sg_len = sg_dma_len(sg); pata_macio_qc_prep() 538 while (sg_len) { pata_macio_qc_prep() 542 len = (sg_len < MAX_DBDMA_SEG) ? sg_len : MAX_DBDMA_SEG; pata_macio_qc_prep() 550 sg_len -= len; pata_macio_qc_prep()
|
H A D | acard-ahci.c | 242 u32 sg_len = sg_dma_len(sg); acard_ahci_fill_sg() local 251 acard_sg[si].size = cpu_to_le32(sg_len); acard_ahci_fill_sg()
|
H A D | sata_rcar.c | 536 u32 addr, sg_len; sata_rcar_bmdma_fill_sg() local 543 sg_len = sg_dma_len(sg); sata_rcar_bmdma_fill_sg() 546 prd[si].flags_len = cpu_to_le32(sg_len); sata_rcar_bmdma_fill_sg() 547 VPRINTK("PRD[%u] = (0x%X, 0x%X)\n", si, addr, sg_len); sata_rcar_bmdma_fill_sg()
|
H A D | sata_nv.c | 2018 u32 sg_len, len; nv_swncq_fill_sg() local 2021 sg_len = sg_dma_len(sg); nv_swncq_fill_sg() 2023 while (sg_len) { nv_swncq_fill_sg() 2025 len = sg_len; nv_swncq_fill_sg() 2026 if ((offset + sg_len) > 0x10000) nv_swncq_fill_sg() 2033 sg_len -= len; nv_swncq_fill_sg()
|
H A D | libahci.c | 1503 u32 sg_len = sg_dma_len(sg); ahci_fill_sg() local 1507 ahci_sg[si].flags_size = cpu_to_le32(sg_len - 1); ahci_fill_sg()
|
H A D | sata_mv.c | 1810 u32 sg_len = sg_dma_len(sg); mv_fill_sg() local 1812 while (sg_len) { mv_fill_sg() 1814 u32 len = sg_len; mv_fill_sg() 1824 sg_len -= len; mv_fill_sg()
|
/linux-4.1.27/drivers/crypto/amcc/ |
H A D | crypto4xx_core.c | 550 u32 sg_len; crypto4xx_copy_pkt_to_dst() local 559 sg_len = sg->length; crypto4xx_copy_pkt_to_dst() 580 sg_len -= len; crypto4xx_copy_pkt_to_dst() 581 if (sg_len) { crypto4xx_copy_pkt_to_dst() 584 &sg_len, &this_sd, &offset, &nbytes)) crypto4xx_copy_pkt_to_dst()
|
/linux-4.1.27/include/linux/ |
H A D | dmaengine.h | 663 unsigned int sg_len, enum dma_transfer_direction direction, 713 struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, dmaengine_prep_slave_sg() 716 return chan->device->device_prep_slave_sg(chan, sgl, sg_len, dmaengine_prep_slave_sg() 723 struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, dmaengine_prep_rio_sg() 727 return chan->device->device_prep_slave_sg(chan, sgl, sg_len, dmaengine_prep_rio_sg() 712 dmaengine_prep_slave_sg( struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction dir, unsigned long flags) dmaengine_prep_slave_sg() argument 722 dmaengine_prep_rio_sg( struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction dir, unsigned long flags, struct rio_dma_ext *rio_ext) dmaengine_prep_rio_sg() argument
|
H A D | rio.h | 440 unsigned int sg_len; /* size of scatter list */ member in struct:rio_dma_data
|
/linux-4.1.27/drivers/scsi/be2iscsi/ |
H A D | be_main.c | 2303 unsigned int sg_len, index; hwi_write_sgl_v2() local 2318 sg_len = sg_dma_len(sg); hwi_write_sgl_v2() 2328 sg_len); hwi_write_sgl_v2() 2329 sge_len = sg_len; hwi_write_sgl_v2() 2333 sg_len = sg_dma_len(sg); hwi_write_sgl_v2() 2343 sg_len); hwi_write_sgl_v2() 2378 sg_len = sg_dma_len(sg); hwi_write_sgl_v2() 2384 AMAP_SET_BITS(struct amap_iscsi_sge, len, psgl, sg_len); hwi_write_sgl_v2() 2387 offset += sg_len; hwi_write_sgl_v2() 2398 unsigned int sg_len, index; hwi_write_sgl() local 2413 sg_len = sg_dma_len(sg); hwi_write_sgl() 2420 sg_len); hwi_write_sgl() 2421 sge_len = sg_len; hwi_write_sgl() 2425 sg_len = sg_dma_len(sg); hwi_write_sgl() 2432 sg_len); hwi_write_sgl() 2466 sg_len = sg_dma_len(sg); hwi_write_sgl() 2472 AMAP_SET_BITS(struct amap_iscsi_sge, len, psgl, sg_len); hwi_write_sgl() 2475 offset += sg_len; hwi_write_sgl()
|
/linux-4.1.27/include/linux/mmc/ |
H A D | core.h | 123 unsigned int sg_len; /* size of scatter list */ member in struct:mmc_data
|
H A D | dw_mmc.h | 213 void (*start)(struct dw_mci *host, unsigned int sg_len);
|
/linux-4.1.27/drivers/dma/dw/ |
H A D | core.c | 762 unsigned int sg_len, enum dma_transfer_direction direction, dwc_prep_slave_sg() 781 if (unlikely(!is_slave_direction(direction) || !sg_len)) dwc_prep_slave_sg() 802 for_each_sg(sgl, sg, sg_len, i) { for_each_sg() 859 for_each_sg(sgl, sg, sg_len, i) { for_each_sg() 761 dwc_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) dwc_prep_slave_sg() argument
|
/linux-4.1.27/drivers/dma/ipu/ |
H A D | ipu_idmac.c | 1323 for_each_sg(desc->sg, sg, desc->sg_len, k) { ipu_gc_tasklet() 1338 struct scatterlist *sgl, unsigned int sg_len, idmac_prep_slave_sg() 1366 desc->sg_len = sg_len; idmac_prep_slave_sg() 1337 idmac_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long tx_flags, void *context) idmac_prep_slave_sg() argument
|
/linux-4.1.27/drivers/scsi/bfa/ |
H A D | bfa_fcpim.c | 2500 sge->sg_len = pgdlen; bfa_ioim_send_ioreq() 2511 sgpge->sg_len = sg_dma_len(sg); bfa_ioim_send_ioreq() 2512 pgcumsz += sgpge->sg_len; bfa_ioim_send_ioreq() 2530 sgpge->sg_len = pgcumsz; bfa_ioim_send_ioreq() 2536 sgpge->sg_len = pgcumsz; bfa_ioim_send_ioreq() 2550 sge->sg_len = pgdlen; bfa_ioim_send_ioreq()
|
H A D | bfi.h | 126 sg_len:28; member in struct:bfi_sge_s 128 u32 sg_len:28,
|
H A D | bfad_bsg.c | 3212 return sge->sg_len; bfad_fcxp_get_req_sglen_cb() 3234 return sge->sg_len; bfad_fcxp_get_rsp_sglen_cb() 3283 sg_table->sg_len = buf_info->size; bfad_fcxp_map_sg()
|
H A D | bfa_ioc.h | 64 u32 sg_len; member in struct:bfa_sge_s
|
/linux-4.1.27/drivers/scsi/bnx2i/ |
H A D | bnx2i_iscsi.c | 151 int sg_len; bnx2i_map_scsi_sg() local 160 sg_len = sg_dma_len(sg); scsi_for_each_sg() 164 bd[bd_count].buffer_length = sg_len; scsi_for_each_sg() 169 byte_count += sg_len; scsi_for_each_sg()
|
/linux-4.1.27/drivers/scsi/qla2xxx/ |
H A D | qla_iocb.c | 881 uint32_t cumulative_partial, sg_len; qla24xx_get_one_block_sg() local 891 sg_len = sg_dma_len(sg); qla24xx_get_one_block_sg() 895 if ((cumulative_partial + (sg_len - sgx->bytes_consumed)) >= blk_sz) { qla24xx_get_one_block_sg() 901 sgx->dma_len = sg_len - sgx->bytes_consumed; qla24xx_get_one_block_sg() 908 if (sg_len == sgx->bytes_consumed) { qla24xx_get_one_block_sg()
|
/linux-4.1.27/drivers/scsi/csiostor/ |
H A D | csio_scsi.c | 566 uint32_t sg_len = 0; csio_setup_ddp() local 571 sg_len = sg_dma_len(sgel); csio_setup_ddp() 578 sg_addr, sg_len); csio_setup_ddp() 584 ((buf_off + sg_len) & (ddp_pagesz - 1))) { csio_setup_ddp() 587 "(%llx:%d)\n", sg_addr, sg_len); csio_setup_ddp()
|
/linux-4.1.27/drivers/crypto/ccp/ |
H A D | ccp-ops.c | 676 struct scatterlist *sg, u64 sg_len, ccp_init_data() 684 ret = ccp_init_sg_workarea(&data->sg_wa, cmd_q->ccp->dev, sg, sg_len, ccp_init_data() 675 ccp_init_data(struct ccp_data *data, struct ccp_cmd_queue *cmd_q, struct scatterlist *sg, u64 sg_len, unsigned int dm_len, enum dma_data_direction dir) ccp_init_data() argument
|
/linux-4.1.27/drivers/net/wireless/ath/ath6kl/ |
H A D | sdio.c | 255 data->sg_len = scat_req->scat_entries; ath6kl_sdio_setup_scat_data()
|
/linux-4.1.27/drivers/net/wireless/brcm80211/brcmfmac/ |
H A D | bcmsdh.c | 612 mmc_dat.sg_len = sg_cnt; brcmf_sdiod_sglist_rw()
|
/linux-4.1.27/drivers/media/common/saa7146/ |
H A D | saa7146_video.c | 215 DEB_EE("dev:%p, buf:%p, sg_len:%d\n", dev, buf, length); saa7146_pgtable_build()
|
/linux-4.1.27/fs/exofs/ |
H A D | ore.c | 992 " dev=%d sg_len=%d\n", _LLU(obj->id), _ore_read_mirror()
|
/linux-4.1.27/drivers/rapidio/ |
H A D | rio.c | 1588 return dmaengine_prep_rio_sg(dchan, data->sg, data->sg_len, rio_dma_prep_xfer()
|