dma_dev 101 arch/arm/mach-imx/devices/platform-ipu-core.c copied_pdata->dma_dev = &imx_ipu_coredev->dev; dma_dev 122 arch/arm/mach-imx/devices/platform-ipu-core.c pdata->dma_dev = &imx_ipu_coredev->dev; dma_dev 127 arch/mips/alchemy/common/dma.c static const struct dma_dev dma_dev_table_bank2[DMA_NUM_DEV_BANK2] = { dma_dev 168 arch/mips/alchemy/common/dma.c const struct dma_dev *dev; dma_dev 165 arch/powerpc/platforms/512x/mpc512x_lpbfifo.c struct dma_device *dma_dev = NULL; dma_dev 226 arch/powerpc/platforms/512x/mpc512x_lpbfifo.c dma_dev = lpbfifo.chan->device; dma_dev 243 arch/powerpc/platforms/512x/mpc512x_lpbfifo.c if (dma_dev->device_config(lpbfifo.chan, &dma_conf)) { dma_dev 250 arch/powerpc/platforms/512x/mpc512x_lpbfifo.c sg_dma_address(&sg) = dma_map_single(dma_dev->dev, dma_dev 252 arch/powerpc/platforms/512x/mpc512x_lpbfifo.c if (dma_mapping_error(dma_dev->dev, sg_dma_address(&sg))) dma_dev 327 arch/powerpc/platforms/512x/mpc512x_lpbfifo.c dma_unmap_single(dma_dev->dev, sg_dma_address(&sg), dma_dev 501 arch/powerpc/platforms/512x/mpc512x_lpbfifo.c struct dma_device *dma_dev = lpbfifo.chan->device; dma_dev 509 arch/powerpc/platforms/512x/mpc512x_lpbfifo.c dma_dev->device_terminate_all(lpbfifo.chan); dma_dev 366 arch/s390/include/asm/cio.h void *cio_gp_dma_zalloc(struct gen_pool *gp_dma, struct device *dma_dev, dma_dev 369 arch/s390/include/asm/cio.h void cio_gp_dma_destroy(struct gen_pool *gp_dma, struct device *dma_dev); dma_dev 370 arch/s390/include/asm/cio.h struct gen_pool *cio_gp_dma_create(struct device *dma_dev, int nr_pages); dma_dev 1399 drivers/acpi/scan.c struct device *dma_dev = dev; dma_dev 1408 drivers/acpi/scan.c adev = ACPI_COMPANION(dma_dev); dma_dev 1412 drivers/acpi/scan.c dma_dev = dma_dev->parent; dma_dev 1413 drivers/acpi/scan.c } while (dma_dev); dma_dev 1415 drivers/acpi/scan.c if (!dma_dev) dma_dev 1428 drivers/acpi/scan.c dev_warn(dma_dev, "Can't handle multiple windows with different offsets\n"); dma_dev 1442 drivers/acpi/scan.c dev_dbg(dma_dev, "Invalid DMA regions configuration\n"); dma_dev 884 drivers/ata/pata_octeon_cf.c struct platform_device *dma_dev; dma_dev 885 drivers/ata/pata_octeon_cf.c dma_dev = of_find_device_by_node(dma_node); dma_dev 886 drivers/ata/pata_octeon_cf.c if (dma_dev) { dma_dev 889 drivers/ata/pata_octeon_cf.c res_dma = platform_get_resource(dma_dev, IORESOURCE_MEM, 0); dma_dev 902 drivers/ata/pata_octeon_cf.c i = platform_get_irq(dma_dev, 0); dma_dev 207 drivers/ata/sata_dwc_460ex.c if (dws->dma_dev != chan->device->dev) dma_dev 220 drivers/ata/sata_dwc_460ex.c dws->dma_dev = hsdev->dev; dma_dev 248 drivers/bcma/main.c core->dma_dev = bus->dev; dma_dev 253 drivers/bcma/main.c core->dma_dev = bus->dev; dma_dev 256 drivers/bcma/main.c core->dma_dev = &core->dev; dma_dev 132 drivers/bus/fsl-mc/fsl-mc-bus.c struct device *dma_dev = dev; dma_dev 134 drivers/bus/fsl-mc/fsl-mc-bus.c while (dev_is_fsl_mc(dma_dev)) dma_dev 135 drivers/bus/fsl-mc/fsl-mc-bus.c dma_dev = dma_dev->parent; dma_dev 137 drivers/bus/fsl-mc/fsl-mc-bus.c return of_dma_configure(dev, dma_dev->of_node, 0); dma_dev 2378 drivers/crypto/atmel-aes.c if (sl && sl->dma_dev == chan->device->dev) { dma_dev 2610 drivers/crypto/atmel-sha.c if (sl && sl->dma_dev == chan->device->dev) { dma_dev 709 drivers/crypto/atmel-tdes.c if (sl && sl->dma_dev == chan->device->dev) { dma_dev 391 drivers/crypto/ccp/ccp-dev.h struct dma_device dma_dev; dma_dev 472 drivers/crypto/ccp/ccp-dev.h struct device *dma_dev; dma_dev 639 drivers/crypto/ccp/ccp-dmaengine.c struct dma_device *dma_dev = &ccp->dma_dev; dma_dev 685 drivers/crypto/ccp/ccp-dmaengine.c dma_dev->dev = ccp->dev; dma_dev 686 drivers/crypto/ccp/ccp-dmaengine.c dma_dev->src_addr_widths = CCP_DMA_WIDTH(dma_get_mask(ccp->dev)); dma_dev 687 drivers/crypto/ccp/ccp-dmaengine.c dma_dev->dst_addr_widths = CCP_DMA_WIDTH(dma_get_mask(ccp->dev)); dma_dev 688 drivers/crypto/ccp/ccp-dmaengine.c dma_dev->directions = DMA_MEM_TO_MEM; dma_dev 689 drivers/crypto/ccp/ccp-dmaengine.c dma_dev->residue_granularity = DMA_RESIDUE_GRANULARITY_DESCRIPTOR; dma_dev 690 drivers/crypto/ccp/ccp-dmaengine.c dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask); dma_dev 691 drivers/crypto/ccp/ccp-dmaengine.c dma_cap_set(DMA_INTERRUPT, dma_dev->cap_mask); dma_dev 700 drivers/crypto/ccp/ccp-dmaengine.c dma_cap_set(DMA_PRIVATE, dma_dev->cap_mask); dma_dev 702 drivers/crypto/ccp/ccp-dmaengine.c INIT_LIST_HEAD(&dma_dev->channels); dma_dev 718 drivers/crypto/ccp/ccp-dmaengine.c dma_chan->device = dma_dev; dma_dev 721 drivers/crypto/ccp/ccp-dmaengine.c list_add_tail(&dma_chan->device_node, &dma_dev->channels); dma_dev 724 drivers/crypto/ccp/ccp-dmaengine.c dma_dev->device_free_chan_resources = ccp_free_chan_resources; dma_dev 725 drivers/crypto/ccp/ccp-dmaengine.c dma_dev->device_prep_dma_memcpy = ccp_prep_dma_memcpy; dma_dev 726 drivers/crypto/ccp/ccp-dmaengine.c dma_dev->device_prep_dma_interrupt = ccp_prep_dma_interrupt; dma_dev 727 drivers/crypto/ccp/ccp-dmaengine.c dma_dev->device_issue_pending = ccp_issue_pending; dma_dev 728 drivers/crypto/ccp/ccp-dmaengine.c dma_dev->device_tx_status = ccp_tx_status; dma_dev 729 drivers/crypto/ccp/ccp-dmaengine.c dma_dev->device_pause = ccp_pause; dma_dev 730 drivers/crypto/ccp/ccp-dmaengine.c dma_dev->device_resume = ccp_resume; dma_dev 731 drivers/crypto/ccp/ccp-dmaengine.c dma_dev->device_terminate_all = ccp_terminate_all; dma_dev 733 drivers/crypto/ccp/ccp-dmaengine.c ret = dma_async_device_register(dma_dev); dma_dev 750 drivers/crypto/ccp/ccp-dmaengine.c struct dma_device *dma_dev = &ccp->dma_dev; dma_dev 755 drivers/crypto/ccp/ccp-dmaengine.c dma_async_device_unregister(dma_dev); dma_dev 66 drivers/crypto/ccp/ccp-ops.c dma_unmap_sg(wa->dma_dev, wa->dma_sg, wa->nents, wa->dma_dir); dma_dev 95 drivers/crypto/ccp/ccp-ops.c wa->dma_dev = dev; dma_dev 794 drivers/dma/altera-msgdma.c struct dma_device *dma_dev; dma_dev 842 drivers/dma/altera-msgdma.c dma_dev = &mdev->dmadev; dma_dev 845 drivers/dma/altera-msgdma.c dma_cap_zero(dma_dev->cap_mask); dma_dev 846 drivers/dma/altera-msgdma.c dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask); dma_dev 847 drivers/dma/altera-msgdma.c dma_cap_set(DMA_SLAVE, dma_dev->cap_mask); dma_dev 849 drivers/dma/altera-msgdma.c dma_dev->src_addr_widths = BIT(DMA_SLAVE_BUSWIDTH_4_BYTES); dma_dev 850 drivers/dma/altera-msgdma.c dma_dev->dst_addr_widths = BIT(DMA_SLAVE_BUSWIDTH_4_BYTES); dma_dev 851 drivers/dma/altera-msgdma.c dma_dev->directions = BIT(DMA_MEM_TO_DEV) | BIT(DMA_DEV_TO_MEM) | dma_dev 853 drivers/dma/altera-msgdma.c dma_dev->residue_granularity = DMA_RESIDUE_GRANULARITY_DESCRIPTOR; dma_dev 856 drivers/dma/altera-msgdma.c INIT_LIST_HEAD(&dma_dev->channels); dma_dev 859 drivers/dma/altera-msgdma.c dma_dev->device_tx_status = dma_cookie_status; dma_dev 860 drivers/dma/altera-msgdma.c dma_dev->device_issue_pending = msgdma_issue_pending; dma_dev 861 drivers/dma/altera-msgdma.c dma_dev->dev = &pdev->dev; dma_dev 863 drivers/dma/altera-msgdma.c dma_dev->copy_align = DMAENGINE_ALIGN_4_BYTES; dma_dev 864 drivers/dma/altera-msgdma.c dma_dev->device_prep_dma_memcpy = msgdma_prep_memcpy; dma_dev 865 drivers/dma/altera-msgdma.c dma_dev->device_prep_slave_sg = msgdma_prep_slave_sg; dma_dev 866 drivers/dma/altera-msgdma.c dma_dev->device_config = msgdma_dma_config; dma_dev 868 drivers/dma/altera-msgdma.c dma_dev->device_alloc_chan_resources = msgdma_alloc_chan_resources; dma_dev 869 drivers/dma/altera-msgdma.c dma_dev->device_free_chan_resources = msgdma_free_chan_resources; dma_dev 871 drivers/dma/altera-msgdma.c mdev->dmachan.device = dma_dev; dma_dev 872 drivers/dma/altera-msgdma.c list_add_tail(&mdev->dmachan.device_node, &dma_dev->channels); dma_dev 885 drivers/dma/altera-msgdma.c ret = dma_async_device_register(dma_dev); dma_dev 1566 drivers/dma/at_hdmac.c BUG_ON(!atslave->dma_dev || atslave->dma_dev != atdma->dma_common.dev); dma_dev 1648 drivers/dma/at_hdmac.c if (atslave->dma_dev == chan->device->dev) { dma_dev 1702 drivers/dma/at_hdmac.c atslave->dma_dev = &dmac_pdev->dev; dma_dev 151 drivers/dma/bcm-sba-raid.c struct dma_device dma_dev; dma_dev 1554 drivers/dma/bcm-sba-raid.c struct dma_device *dma_dev = &sba->dma_dev; dma_dev 1557 drivers/dma/bcm-sba-raid.c sba->dma_chan.device = dma_dev; dma_dev 1561 drivers/dma/bcm-sba-raid.c dma_cap_zero(dma_dev->cap_mask); dma_dev 1562 drivers/dma/bcm-sba-raid.c dma_cap_set(DMA_INTERRUPT, dma_dev->cap_mask); dma_dev 1563 drivers/dma/bcm-sba-raid.c dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask); dma_dev 1564 drivers/dma/bcm-sba-raid.c dma_cap_set(DMA_XOR, dma_dev->cap_mask); dma_dev 1565 drivers/dma/bcm-sba-raid.c dma_cap_set(DMA_PQ, dma_dev->cap_mask); dma_dev 1572 drivers/dma/bcm-sba-raid.c dma_dev->dev = sba->mbox_dev; dma_dev 1575 drivers/dma/bcm-sba-raid.c dma_dev->device_free_chan_resources = sba_free_chan_resources; dma_dev 1576 drivers/dma/bcm-sba-raid.c dma_dev->device_terminate_all = sba_device_terminate_all; dma_dev 1577 drivers/dma/bcm-sba-raid.c dma_dev->device_issue_pending = sba_issue_pending; dma_dev 1578 drivers/dma/bcm-sba-raid.c dma_dev->device_tx_status = sba_tx_status; dma_dev 1581 drivers/dma/bcm-sba-raid.c if (dma_has_cap(DMA_INTERRUPT, dma_dev->cap_mask)) dma_dev 1582 drivers/dma/bcm-sba-raid.c dma_dev->device_prep_dma_interrupt = sba_prep_dma_interrupt; dma_dev 1585 drivers/dma/bcm-sba-raid.c if (dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask)) dma_dev 1586 drivers/dma/bcm-sba-raid.c dma_dev->device_prep_dma_memcpy = sba_prep_dma_memcpy; dma_dev 1589 drivers/dma/bcm-sba-raid.c if (dma_has_cap(DMA_XOR, dma_dev->cap_mask)) { dma_dev 1590 drivers/dma/bcm-sba-raid.c dma_dev->device_prep_dma_xor = sba_prep_dma_xor; dma_dev 1591 drivers/dma/bcm-sba-raid.c dma_dev->max_xor = sba->max_xor_srcs; dma_dev 1595 drivers/dma/bcm-sba-raid.c if (dma_has_cap(DMA_PQ, dma_dev->cap_mask)) { dma_dev 1596 drivers/dma/bcm-sba-raid.c dma_dev->device_prep_dma_pq = sba_prep_dma_pq; dma_dev 1597 drivers/dma/bcm-sba-raid.c dma_set_maxpq(dma_dev, sba->max_pq_srcs, 0); dma_dev 1601 drivers/dma/bcm-sba-raid.c INIT_LIST_HEAD(&dma_dev->channels); dma_dev 1602 drivers/dma/bcm-sba-raid.c list_add_tail(&sba->dma_chan.device_node, &dma_dev->channels); dma_dev 1605 drivers/dma/bcm-sba-raid.c ret = dma_async_device_register(dma_dev); dma_dev 1613 drivers/dma/bcm-sba-raid.c dma_has_cap(DMA_INTERRUPT, dma_dev->cap_mask) ? "interrupt " : "", dma_dev 1614 drivers/dma/bcm-sba-raid.c dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask) ? "memcpy " : "", dma_dev 1615 drivers/dma/bcm-sba-raid.c dma_has_cap(DMA_XOR, dma_dev->cap_mask) ? "xor " : "", dma_dev 1616 drivers/dma/bcm-sba-raid.c dma_has_cap(DMA_PQ, dma_dev->cap_mask) ? "pq " : ""); dma_dev 1749 drivers/dma/bcm-sba-raid.c dma_async_device_unregister(&sba->dma_dev); dma_dev 135 drivers/dma/dma-axi-dmac.c struct dma_device dma_dev; dma_dev 144 drivers/dma/dma-axi-dmac.c dma_dev); dma_dev 311 drivers/dma/dma-axi-dmac.c dev_dbg(dmac->dma_dev.dev, dma_dev 315 drivers/dma/dma-axi-dmac.c dev_warn(dmac->dma_dev.dev, dma_dev 800 drivers/dma/dma-axi-dmac.c dev_err(dmac->dma_dev.dev, dma_dev 808 drivers/dma/dma-axi-dmac.c dev_err(dmac->dma_dev.dev, dma_dev 830 drivers/dma/dma-axi-dmac.c struct dma_device *dma_dev; dma_dev 874 drivers/dma/dma-axi-dmac.c dma_dev = &dmac->dma_dev; dma_dev 875 drivers/dma/dma-axi-dmac.c dma_cap_set(DMA_SLAVE, dma_dev->cap_mask); dma_dev 876 drivers/dma/dma-axi-dmac.c dma_cap_set(DMA_CYCLIC, dma_dev->cap_mask); dma_dev 877 drivers/dma/dma-axi-dmac.c dma_cap_set(DMA_INTERLEAVE, dma_dev->cap_mask); dma_dev 878 drivers/dma/dma-axi-dmac.c dma_dev->device_free_chan_resources = axi_dmac_free_chan_resources; dma_dev 879 drivers/dma/dma-axi-dmac.c dma_dev->device_tx_status = dma_cookie_status; dma_dev 880 drivers/dma/dma-axi-dmac.c dma_dev->device_issue_pending = axi_dmac_issue_pending; dma_dev 881 drivers/dma/dma-axi-dmac.c dma_dev->device_prep_slave_sg = axi_dmac_prep_slave_sg; dma_dev 882 drivers/dma/dma-axi-dmac.c dma_dev->device_prep_dma_cyclic = axi_dmac_prep_dma_cyclic; dma_dev 883 drivers/dma/dma-axi-dmac.c dma_dev->device_prep_interleaved_dma = axi_dmac_prep_interleaved; dma_dev 884 drivers/dma/dma-axi-dmac.c dma_dev->device_terminate_all = axi_dmac_terminate_all; dma_dev 885 drivers/dma/dma-axi-dmac.c dma_dev->device_synchronize = axi_dmac_synchronize; dma_dev 886 drivers/dma/dma-axi-dmac.c dma_dev->dev = &pdev->dev; dma_dev 887 drivers/dma/dma-axi-dmac.c dma_dev->chancnt = 1; dma_dev 888 drivers/dma/dma-axi-dmac.c dma_dev->src_addr_widths = BIT(dmac->chan.src_width); dma_dev 889 drivers/dma/dma-axi-dmac.c dma_dev->dst_addr_widths = BIT(dmac->chan.dest_width); dma_dev 890 drivers/dma/dma-axi-dmac.c dma_dev->directions = BIT(dmac->chan.direction); dma_dev 891 drivers/dma/dma-axi-dmac.c dma_dev->residue_granularity = DMA_RESIDUE_GRANULARITY_DESCRIPTOR; dma_dev 892 drivers/dma/dma-axi-dmac.c INIT_LIST_HEAD(&dma_dev->channels); dma_dev 895 drivers/dma/dma-axi-dmac.c vchan_init(&dmac->chan.vchan, dma_dev); dma_dev 905 drivers/dma/dma-axi-dmac.c dma_dev->copy_align = (dmac->chan.address_align_mask + 1); dma_dev 909 drivers/dma/dma-axi-dmac.c ret = dma_async_device_register(dma_dev); dma_dev 914 drivers/dma/dma-axi-dmac.c of_dma_xlate_by_chan_id, dma_dev); dma_dev 939 drivers/dma/dma-axi-dmac.c dma_async_device_unregister(&dmac->dma_dev); dma_dev 953 drivers/dma/dma-axi-dmac.c dma_async_device_unregister(&dmac->dma_dev); dma_dev 145 drivers/dma/dmaengine.c ATTRIBUTE_GROUPS(dma_dev); dma_dev 989 drivers/dma/dmatest.c struct dma_device *dma_dev = chan->device; dma_dev 1002 drivers/dma/dmatest.c if (dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask)) { dma_dev 1009 drivers/dma/dmatest.c if (dma_has_cap(DMA_MEMSET, dma_dev->cap_mask)) { dma_dev 1016 drivers/dma/dmatest.c if (dma_has_cap(DMA_XOR, dma_dev->cap_mask)) { dma_dev 1020 drivers/dma/dmatest.c if (dma_has_cap(DMA_PQ, dma_dev->cap_mask)) { dma_dev 13 drivers/dma/dw/acpi.c .dma_dev = dma_spec->dev, dma_dev 777 drivers/dma/dw/core.c if (dws->dma_dev != chan->device->dev) dma_dev 21 drivers/dma/dw/of.c .dma_dev = dw->dma.dev, dma_dev 202 drivers/dma/ep93xx_dma.c struct dma_device dma_dev; dma_dev 1320 drivers/dma/ep93xx_dma.c struct dma_device *dma_dev; dma_dev 1329 drivers/dma/ep93xx_dma.c dma_dev = &edma->dma_dev; dma_dev 1333 drivers/dma/ep93xx_dma.c INIT_LIST_HEAD(&dma_dev->channels); dma_dev 1338 drivers/dma/ep93xx_dma.c edmac->chan.device = dma_dev; dma_dev 1358 drivers/dma/ep93xx_dma.c &dma_dev->channels); dma_dev 1361 drivers/dma/ep93xx_dma.c dma_cap_zero(dma_dev->cap_mask); dma_dev 1362 drivers/dma/ep93xx_dma.c dma_cap_set(DMA_SLAVE, dma_dev->cap_mask); dma_dev 1363 drivers/dma/ep93xx_dma.c dma_cap_set(DMA_CYCLIC, dma_dev->cap_mask); dma_dev 1365 drivers/dma/ep93xx_dma.c dma_dev->dev = &pdev->dev; dma_dev 1366 drivers/dma/ep93xx_dma.c dma_dev->device_alloc_chan_resources = ep93xx_dma_alloc_chan_resources; dma_dev 1367 drivers/dma/ep93xx_dma.c dma_dev->device_free_chan_resources = ep93xx_dma_free_chan_resources; dma_dev 1368 drivers/dma/ep93xx_dma.c dma_dev->device_prep_slave_sg = ep93xx_dma_prep_slave_sg; dma_dev 1369 drivers/dma/ep93xx_dma.c dma_dev->device_prep_dma_cyclic = ep93xx_dma_prep_dma_cyclic; dma_dev 1370 drivers/dma/ep93xx_dma.c dma_dev->device_config = ep93xx_dma_slave_config; dma_dev 1371 drivers/dma/ep93xx_dma.c dma_dev->device_synchronize = ep93xx_dma_synchronize; dma_dev 1372 drivers/dma/ep93xx_dma.c dma_dev->device_terminate_all = ep93xx_dma_terminate_all; dma_dev 1373 drivers/dma/ep93xx_dma.c dma_dev->device_issue_pending = ep93xx_dma_issue_pending; dma_dev 1374 drivers/dma/ep93xx_dma.c dma_dev->device_tx_status = ep93xx_dma_tx_status; dma_dev 1376 drivers/dma/ep93xx_dma.c dma_set_max_seg_size(dma_dev->dev, DMA_MAX_CHAN_BYTES); dma_dev 1379 drivers/dma/ep93xx_dma.c dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask); dma_dev 1380 drivers/dma/ep93xx_dma.c dma_dev->device_prep_dma_memcpy = ep93xx_dma_prep_dma_memcpy; dma_dev 1387 drivers/dma/ep93xx_dma.c dma_cap_set(DMA_PRIVATE, dma_dev->cap_mask); dma_dev 1396 drivers/dma/ep93xx_dma.c ret = dma_async_device_register(dma_dev); dma_dev 1405 drivers/dma/ep93xx_dma.c dev_info(dma_dev->dev, "EP93xx M2%s DMA ready\n", dma_dev 155 drivers/dma/fsl-edma-common.h struct dma_device dma_dev; dma_dev 109 drivers/dma/fsl-edma.c list_for_each_entry_safe(chan, _chan, &fsl_edma->dma_dev.channels, device_node) { dma_dev 336 drivers/dma/fsl-edma.c INIT_LIST_HEAD(&fsl_edma->dma_dev.channels); dma_dev 346 drivers/dma/fsl-edma.c vchan_init(&fsl_chan->vchan, &fsl_edma->dma_dev); dma_dev 357 drivers/dma/fsl-edma.c dma_cap_set(DMA_PRIVATE, fsl_edma->dma_dev.cap_mask); dma_dev 358 drivers/dma/fsl-edma.c dma_cap_set(DMA_SLAVE, fsl_edma->dma_dev.cap_mask); dma_dev 359 drivers/dma/fsl-edma.c dma_cap_set(DMA_CYCLIC, fsl_edma->dma_dev.cap_mask); dma_dev 361 drivers/dma/fsl-edma.c fsl_edma->dma_dev.dev = &pdev->dev; dma_dev 362 drivers/dma/fsl-edma.c fsl_edma->dma_dev.device_alloc_chan_resources dma_dev 364 drivers/dma/fsl-edma.c fsl_edma->dma_dev.device_free_chan_resources dma_dev 366 drivers/dma/fsl-edma.c fsl_edma->dma_dev.device_tx_status = fsl_edma_tx_status; dma_dev 367 drivers/dma/fsl-edma.c fsl_edma->dma_dev.device_prep_slave_sg = fsl_edma_prep_slave_sg; dma_dev 368 drivers/dma/fsl-edma.c fsl_edma->dma_dev.device_prep_dma_cyclic = fsl_edma_prep_dma_cyclic; dma_dev 369 drivers/dma/fsl-edma.c fsl_edma->dma_dev.device_config = fsl_edma_slave_config; dma_dev 370 drivers/dma/fsl-edma.c fsl_edma->dma_dev.device_pause = fsl_edma_pause; dma_dev 371 drivers/dma/fsl-edma.c fsl_edma->dma_dev.device_resume = fsl_edma_resume; dma_dev 372 drivers/dma/fsl-edma.c fsl_edma->dma_dev.device_terminate_all = fsl_edma_terminate_all; dma_dev 373 drivers/dma/fsl-edma.c fsl_edma->dma_dev.device_synchronize = fsl_edma_synchronize; dma_dev 374 drivers/dma/fsl-edma.c fsl_edma->dma_dev.device_issue_pending = fsl_edma_issue_pending; dma_dev 376 drivers/dma/fsl-edma.c fsl_edma->dma_dev.src_addr_widths = FSL_EDMA_BUSWIDTHS; dma_dev 377 drivers/dma/fsl-edma.c fsl_edma->dma_dev.dst_addr_widths = FSL_EDMA_BUSWIDTHS; dma_dev 378 drivers/dma/fsl-edma.c fsl_edma->dma_dev.directions = BIT(DMA_DEV_TO_MEM) | BIT(DMA_MEM_TO_DEV); dma_dev 382 drivers/dma/fsl-edma.c ret = dma_async_device_register(&fsl_edma->dma_dev); dma_dev 394 drivers/dma/fsl-edma.c dma_async_device_unregister(&fsl_edma->dma_dev); dma_dev 411 drivers/dma/fsl-edma.c fsl_edma_cleanup_vchan(&fsl_edma->dma_dev); dma_dev 413 drivers/dma/fsl-edma.c dma_async_device_unregister(&fsl_edma->dma_dev); dma_dev 198 drivers/dma/fsl-qdma.c struct dma_device dma_dev; dma_dev 707 drivers/dma/fsl-qdma.c dev_err(fsl_qdma->dma_dev.dev, "DMA transaction error!\n"); dma_dev 722 drivers/dma/fsl-qdma.c dev_err(fsl_qdma->dma_dev.dev, dma_dev 740 drivers/dma/fsl-qdma.c dev_err(fsl_qdma->dma_dev.dev, "QDMA: status err!\n"); dma_dev 829 drivers/dma/fsl-qdma.c dev_err(fsl_qdma->dma_dev.dev, "DMA halt failed!"); dma_dev 1162 drivers/dma/fsl-qdma.c INIT_LIST_HEAD(&fsl_qdma->dma_dev.channels); dma_dev 1171 drivers/dma/fsl-qdma.c vchan_init(&fsl_chan->vchan, &fsl_qdma->dma_dev); dma_dev 1174 drivers/dma/fsl-qdma.c dma_cap_set(DMA_MEMCPY, fsl_qdma->dma_dev.cap_mask); dma_dev 1176 drivers/dma/fsl-qdma.c fsl_qdma->dma_dev.dev = &pdev->dev; dma_dev 1177 drivers/dma/fsl-qdma.c fsl_qdma->dma_dev.device_free_chan_resources = dma_dev 1179 drivers/dma/fsl-qdma.c fsl_qdma->dma_dev.device_alloc_chan_resources = dma_dev 1181 drivers/dma/fsl-qdma.c fsl_qdma->dma_dev.device_tx_status = dma_cookie_status; dma_dev 1182 drivers/dma/fsl-qdma.c fsl_qdma->dma_dev.device_prep_dma_memcpy = fsl_qdma_prep_memcpy; dma_dev 1183 drivers/dma/fsl-qdma.c fsl_qdma->dma_dev.device_issue_pending = fsl_qdma_issue_pending; dma_dev 1184 drivers/dma/fsl-qdma.c fsl_qdma->dma_dev.device_synchronize = fsl_qdma_synchronize; dma_dev 1185 drivers/dma/fsl-qdma.c fsl_qdma->dma_dev.device_terminate_all = fsl_qdma_terminate_all; dma_dev 1191 drivers/dma/fsl-qdma.c ret = dma_async_device_register(&fsl_qdma->dma_dev); dma_dev 1226 drivers/dma/fsl-qdma.c fsl_qdma_cleanup_vchan(&fsl_qdma->dma_dev); dma_dev 1228 drivers/dma/fsl-qdma.c dma_async_device_unregister(&fsl_qdma->dma_dev); dma_dev 630 drivers/dma/fsl_raid.c struct dma_device *dma_dev; dma_dev 638 drivers/dma/fsl_raid.c dma_dev = &re_priv->dma_dev; dma_dev 684 drivers/dma/fsl_raid.c chan->chan.device = dma_dev; dma_dev 754 drivers/dma/fsl_raid.c struct dma_device *dma_dev; dma_dev 783 drivers/dma/fsl_raid.c dma_dev = &re_priv->dma_dev; dma_dev 784 drivers/dma/fsl_raid.c dma_dev->dev = dev; dma_dev 785 drivers/dma/fsl_raid.c INIT_LIST_HEAD(&dma_dev->channels); dma_dev 788 drivers/dma/fsl_raid.c dma_dev->device_alloc_chan_resources = fsl_re_alloc_chan_resources; dma_dev 789 drivers/dma/fsl_raid.c dma_dev->device_tx_status = fsl_re_tx_status; dma_dev 790 drivers/dma/fsl_raid.c dma_dev->device_issue_pending = fsl_re_issue_pending; dma_dev 792 drivers/dma/fsl_raid.c dma_dev->max_xor = FSL_RE_MAX_XOR_SRCS; dma_dev 793 drivers/dma/fsl_raid.c dma_dev->device_prep_dma_xor = fsl_re_prep_dma_xor; dma_dev 794 drivers/dma/fsl_raid.c dma_cap_set(DMA_XOR, dma_dev->cap_mask); dma_dev 796 drivers/dma/fsl_raid.c dma_dev->max_pq = FSL_RE_MAX_PQ_SRCS; dma_dev 797 drivers/dma/fsl_raid.c dma_dev->device_prep_dma_pq = fsl_re_prep_dma_pq; dma_dev 798 drivers/dma/fsl_raid.c dma_cap_set(DMA_PQ, dma_dev->cap_mask); dma_dev 800 drivers/dma/fsl_raid.c dma_dev->device_prep_dma_memcpy = fsl_re_prep_dma_memcpy; dma_dev 801 drivers/dma/fsl_raid.c dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask); dma_dev 803 drivers/dma/fsl_raid.c dma_dev->device_free_chan_resources = fsl_re_free_chan_resources; dma_dev 845 drivers/dma/fsl_raid.c dma_async_device_register(dma_dev); dma_dev 875 drivers/dma/fsl_raid.c dma_async_device_unregister(&re_priv->dma_dev); dma_dev 258 drivers/dma/fsl_raid.h struct dma_device dma_dev; dma_dev 134 drivers/dma/img-mdc-dma.c struct dma_device dma_dev; dma_dev 183 drivers/dma/img-mdc-dma.c return mdma->dma_dev.dev; dma_dev 817 drivers/dma/img-mdc-dma.c list_for_each_entry(chan, &mdma->dma_dev.channels, device_node) { dma_dev 915 drivers/dma/img-mdc-dma.c dma_cap_zero(mdma->dma_dev.cap_mask); dma_dev 916 drivers/dma/img-mdc-dma.c dma_cap_set(DMA_SLAVE, mdma->dma_dev.cap_mask); dma_dev 917 drivers/dma/img-mdc-dma.c dma_cap_set(DMA_PRIVATE, mdma->dma_dev.cap_mask); dma_dev 918 drivers/dma/img-mdc-dma.c dma_cap_set(DMA_CYCLIC, mdma->dma_dev.cap_mask); dma_dev 919 drivers/dma/img-mdc-dma.c dma_cap_set(DMA_MEMCPY, mdma->dma_dev.cap_mask); dma_dev 949 drivers/dma/img-mdc-dma.c mdma->dma_dev.dev = &pdev->dev; dma_dev 950 drivers/dma/img-mdc-dma.c mdma->dma_dev.device_prep_slave_sg = mdc_prep_slave_sg; dma_dev 951 drivers/dma/img-mdc-dma.c mdma->dma_dev.device_prep_dma_cyclic = mdc_prep_dma_cyclic; dma_dev 952 drivers/dma/img-mdc-dma.c mdma->dma_dev.device_prep_dma_memcpy = mdc_prep_dma_memcpy; dma_dev 953 drivers/dma/img-mdc-dma.c mdma->dma_dev.device_alloc_chan_resources = mdc_alloc_chan_resources; dma_dev 954 drivers/dma/img-mdc-dma.c mdma->dma_dev.device_free_chan_resources = mdc_free_chan_resources; dma_dev 955 drivers/dma/img-mdc-dma.c mdma->dma_dev.device_tx_status = mdc_tx_status; dma_dev 956 drivers/dma/img-mdc-dma.c mdma->dma_dev.device_issue_pending = mdc_issue_pending; dma_dev 957 drivers/dma/img-mdc-dma.c mdma->dma_dev.device_terminate_all = mdc_terminate_all; dma_dev 958 drivers/dma/img-mdc-dma.c mdma->dma_dev.device_synchronize = mdc_synchronize; dma_dev 959 drivers/dma/img-mdc-dma.c mdma->dma_dev.device_config = mdc_slave_config; dma_dev 961 drivers/dma/img-mdc-dma.c mdma->dma_dev.directions = BIT(DMA_DEV_TO_MEM) | BIT(DMA_MEM_TO_DEV); dma_dev 962 drivers/dma/img-mdc-dma.c mdma->dma_dev.residue_granularity = DMA_RESIDUE_GRANULARITY_BURST; dma_dev 964 drivers/dma/img-mdc-dma.c mdma->dma_dev.src_addr_widths |= BIT(i); dma_dev 965 drivers/dma/img-mdc-dma.c mdma->dma_dev.dst_addr_widths |= BIT(i); dma_dev 968 drivers/dma/img-mdc-dma.c INIT_LIST_HEAD(&mdma->dma_dev.channels); dma_dev 985 drivers/dma/img-mdc-dma.c vchan_init(&mchan->vc, &mdma->dma_dev); dma_dev 1001 drivers/dma/img-mdc-dma.c ret = dma_async_device_register(&mdma->dma_dev); dma_dev 1015 drivers/dma/img-mdc-dma.c dma_async_device_unregister(&mdma->dma_dev); dma_dev 1029 drivers/dma/img-mdc-dma.c dma_async_device_unregister(&mdma->dma_dev); dma_dev 1031 drivers/dma/img-mdc-dma.c list_for_each_entry_safe(mchan, next, &mdma->dma_dev.channels, dma_dev 22 drivers/dma/ioat/dma.h #define to_ioatdma_device(dev) container_of(dev, struct ioatdma_device, dma_dev) dma_dev 69 drivers/dma/ioat/dma.h struct dma_device dma_dev; dma_dev 305 drivers/dma/ioat/init.c struct dma_device *dma = &ioat_dma->dma_dev; dma_dev 351 drivers/dma/ioat/init.c tx = ioat_dma->dma_dev.device_prep_dma_memcpy(dma_chan, dma_dest, dma_dev 424 drivers/dma/ioat/init.c msixcnt = ioat_dma->dma_dev.chancnt; dma_dev 496 drivers/dma/ioat/init.c struct dma_device *dma = &ioat_dma->dma_dev; dma_dev 540 drivers/dma/ioat/init.c int err = dma_async_device_register(&ioat_dma->dma_dev); dma_dev 552 drivers/dma/ioat/init.c struct dma_device *dma = &ioat_dma->dma_dev; dma_dev 573 drivers/dma/ioat/init.c struct dma_device *dma = &ioat_dma->dma_dev; dma_dev 769 drivers/dma/ioat/init.c struct dma_device *dma = &ioat_dma->dma_dev; dma_dev 803 drivers/dma/ioat/init.c struct dma_device *dma = &ioat_dma->dma_dev; dma_dev 1070 drivers/dma/ioat/init.c dma = &ioat_dma->dma_dev; dma_dev 1099 drivers/dma/ioat/init.c dma = &ioat_dma->dma_dev; dma_dev 77 drivers/dma/ioat/sysfs.c struct dma_device *dma = &ioat_dma->dma_dev; dma_dev 98 drivers/dma/ioat/sysfs.c struct dma_device *dma = &ioat_dma->dma_dev; dma_dev 1271 drivers/dma/iop-adma.c struct dma_device *dma_dev; dma_dev 1285 drivers/dma/iop-adma.c dma_dev = &adev->common; dma_dev 1307 drivers/dma/iop-adma.c dma_dev->cap_mask = plat_data->cap_mask; dma_dev 1312 drivers/dma/iop-adma.c INIT_LIST_HEAD(&dma_dev->channels); dma_dev 1315 drivers/dma/iop-adma.c dma_dev->device_alloc_chan_resources = iop_adma_alloc_chan_resources; dma_dev 1316 drivers/dma/iop-adma.c dma_dev->device_free_chan_resources = iop_adma_free_chan_resources; dma_dev 1317 drivers/dma/iop-adma.c dma_dev->device_tx_status = iop_adma_status; dma_dev 1318 drivers/dma/iop-adma.c dma_dev->device_issue_pending = iop_adma_issue_pending; dma_dev 1319 drivers/dma/iop-adma.c dma_dev->dev = &pdev->dev; dma_dev 1322 drivers/dma/iop-adma.c if (dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask)) dma_dev 1323 drivers/dma/iop-adma.c dma_dev->device_prep_dma_memcpy = iop_adma_prep_dma_memcpy; dma_dev 1324 drivers/dma/iop-adma.c if (dma_has_cap(DMA_XOR, dma_dev->cap_mask)) { dma_dev 1325 drivers/dma/iop-adma.c dma_dev->max_xor = iop_adma_get_max_xor(); dma_dev 1326 drivers/dma/iop-adma.c dma_dev->device_prep_dma_xor = iop_adma_prep_dma_xor; dma_dev 1328 drivers/dma/iop-adma.c if (dma_has_cap(DMA_XOR_VAL, dma_dev->cap_mask)) dma_dev 1329 drivers/dma/iop-adma.c dma_dev->device_prep_dma_xor_val = dma_dev 1331 drivers/dma/iop-adma.c if (dma_has_cap(DMA_PQ, dma_dev->cap_mask)) { dma_dev 1332 drivers/dma/iop-adma.c dma_set_maxpq(dma_dev, iop_adma_get_max_pq(), 0); dma_dev 1333 drivers/dma/iop-adma.c dma_dev->device_prep_dma_pq = iop_adma_prep_dma_pq; dma_dev 1335 drivers/dma/iop-adma.c if (dma_has_cap(DMA_PQ_VAL, dma_dev->cap_mask)) dma_dev 1336 drivers/dma/iop-adma.c dma_dev->device_prep_dma_pq_val = dma_dev 1338 drivers/dma/iop-adma.c if (dma_has_cap(DMA_INTERRUPT, dma_dev->cap_mask)) dma_dev 1339 drivers/dma/iop-adma.c dma_dev->device_prep_dma_interrupt = dma_dev 1380 drivers/dma/iop-adma.c iop_chan->common.device = dma_dev; dma_dev 1382 drivers/dma/iop-adma.c list_add_tail(&iop_chan->common.device_node, &dma_dev->channels); dma_dev 1384 drivers/dma/iop-adma.c if (dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask)) { dma_dev 1391 drivers/dma/iop-adma.c if (dma_has_cap(DMA_XOR, dma_dev->cap_mask)) { dma_dev 1398 drivers/dma/iop-adma.c if (dma_has_cap(DMA_PQ, dma_dev->cap_mask) && dma_dev 1399 drivers/dma/iop-adma.c dma_has_cap(DMA_PQ_VAL, dma_dev->cap_mask)) { dma_dev 1405 drivers/dma/iop-adma.c dma_cap_clear(DMA_PQ, dma_dev->cap_mask); dma_dev 1406 drivers/dma/iop-adma.c dma_cap_clear(DMA_PQ_VAL, dma_dev->cap_mask); dma_dev 1414 drivers/dma/iop-adma.c dma_has_cap(DMA_PQ, dma_dev->cap_mask) ? "pq " : "", dma_dev 1415 drivers/dma/iop-adma.c dma_has_cap(DMA_PQ_VAL, dma_dev->cap_mask) ? "pq_val " : "", dma_dev 1416 drivers/dma/iop-adma.c dma_has_cap(DMA_XOR, dma_dev->cap_mask) ? "xor " : "", dma_dev 1417 drivers/dma/iop-adma.c dma_has_cap(DMA_XOR_VAL, dma_dev->cap_mask) ? "xor_val " : "", dma_dev 1418 drivers/dma/iop-adma.c dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask) ? "cpy " : "", dma_dev 1419 drivers/dma/iop-adma.c dma_has_cap(DMA_INTERRUPT, dma_dev->cap_mask) ? "intr " : ""); dma_dev 1421 drivers/dma/iop-adma.c dma_async_device_register(dma_dev); dma_dev 215 drivers/dma/mcf-edma.c INIT_LIST_HEAD(&mcf_edma->dma_dev.channels); dma_dev 224 drivers/dma/mcf-edma.c vchan_init(&mcf_chan->vchan, &mcf_edma->dma_dev); dma_dev 235 drivers/dma/mcf-edma.c dma_cap_set(DMA_PRIVATE, mcf_edma->dma_dev.cap_mask); dma_dev 236 drivers/dma/mcf-edma.c dma_cap_set(DMA_SLAVE, mcf_edma->dma_dev.cap_mask); dma_dev 237 drivers/dma/mcf-edma.c dma_cap_set(DMA_CYCLIC, mcf_edma->dma_dev.cap_mask); dma_dev 239 drivers/dma/mcf-edma.c mcf_edma->dma_dev.dev = &pdev->dev; dma_dev 240 drivers/dma/mcf-edma.c mcf_edma->dma_dev.device_alloc_chan_resources = dma_dev 242 drivers/dma/mcf-edma.c mcf_edma->dma_dev.device_free_chan_resources = dma_dev 244 drivers/dma/mcf-edma.c mcf_edma->dma_dev.device_config = fsl_edma_slave_config; dma_dev 245 drivers/dma/mcf-edma.c mcf_edma->dma_dev.device_prep_dma_cyclic = dma_dev 247 drivers/dma/mcf-edma.c mcf_edma->dma_dev.device_prep_slave_sg = fsl_edma_prep_slave_sg; dma_dev 248 drivers/dma/mcf-edma.c mcf_edma->dma_dev.device_tx_status = fsl_edma_tx_status; dma_dev 249 drivers/dma/mcf-edma.c mcf_edma->dma_dev.device_pause = fsl_edma_pause; dma_dev 250 drivers/dma/mcf-edma.c mcf_edma->dma_dev.device_resume = fsl_edma_resume; dma_dev 251 drivers/dma/mcf-edma.c mcf_edma->dma_dev.device_terminate_all = fsl_edma_terminate_all; dma_dev 252 drivers/dma/mcf-edma.c mcf_edma->dma_dev.device_issue_pending = fsl_edma_issue_pending; dma_dev 254 drivers/dma/mcf-edma.c mcf_edma->dma_dev.src_addr_widths = FSL_EDMA_BUSWIDTHS; dma_dev 255 drivers/dma/mcf-edma.c mcf_edma->dma_dev.dst_addr_widths = FSL_EDMA_BUSWIDTHS; dma_dev 256 drivers/dma/mcf-edma.c mcf_edma->dma_dev.directions = dma_dev 259 drivers/dma/mcf-edma.c mcf_edma->dma_dev.filter.fn = mcf_edma_filter_fn; dma_dev 260 drivers/dma/mcf-edma.c mcf_edma->dma_dev.filter.map = pdata->slave_map; dma_dev 261 drivers/dma/mcf-edma.c mcf_edma->dma_dev.filter.mapcnt = pdata->slavecnt; dma_dev 265 drivers/dma/mcf-edma.c ret = dma_async_device_register(&mcf_edma->dma_dev); dma_dev 283 drivers/dma/mcf-edma.c fsl_edma_cleanup_vchan(&mcf_edma->dma_dev); dma_dev 284 drivers/dma/mcf-edma.c dma_async_device_unregister(&mcf_edma->dma_dev); dma_dev 589 drivers/dma/mic_x100_dma.c dma_cap_zero(mic_dma_dev->dma_dev.cap_mask); dma_dev 594 drivers/dma/mic_x100_dma.c dma_cap_set(DMA_MEMCPY, mic_dma_dev->dma_dev.cap_mask); dma_dev 597 drivers/dma/mic_x100_dma.c dma_cap_set(DMA_PRIVATE, mic_dma_dev->dma_dev.cap_mask); dma_dev 598 drivers/dma/mic_x100_dma.c mic_dma_dev->dma_dev.device_alloc_chan_resources = dma_dev 600 drivers/dma/mic_x100_dma.c mic_dma_dev->dma_dev.device_free_chan_resources = dma_dev 602 drivers/dma/mic_x100_dma.c mic_dma_dev->dma_dev.device_tx_status = mic_dma_tx_status; dma_dev 603 drivers/dma/mic_x100_dma.c mic_dma_dev->dma_dev.device_prep_dma_memcpy = mic_dma_prep_memcpy_lock; dma_dev 604 drivers/dma/mic_x100_dma.c mic_dma_dev->dma_dev.device_prep_dma_imm_data = dma_dev 606 drivers/dma/mic_x100_dma.c mic_dma_dev->dma_dev.device_prep_dma_interrupt = dma_dev 608 drivers/dma/mic_x100_dma.c mic_dma_dev->dma_dev.device_issue_pending = mic_dma_issue_pending; dma_dev 609 drivers/dma/mic_x100_dma.c mic_dma_dev->dma_dev.copy_align = MIC_DMA_ALIGN_SHIFT; dma_dev 610 drivers/dma/mic_x100_dma.c INIT_LIST_HEAD(&mic_dma_dev->dma_dev.channels); dma_dev 612 drivers/dma/mic_x100_dma.c mic_dma_dev->mic_ch[i].api_ch.device = &mic_dma_dev->dma_dev; dma_dev 615 drivers/dma/mic_x100_dma.c &mic_dma_dev->dma_dev.channels); dma_dev 617 drivers/dma/mic_x100_dma.c return dmaenginem_async_device_register(&mic_dma_dev->dma_dev); dma_dev 637 drivers/dma/mic_x100_dma.c mic_dma_dev->dma_dev.dev = dev; dma_dev 132 drivers/dma/mic_x100_dma.h struct dma_device dma_dev; dma_dev 164 drivers/dma/mic_x100_dma.h return to_mic_dma_dev(ch)->dma_dev.dev; dma_dev 1040 drivers/dma/mv_xor.c struct dma_device *dma_dev; dma_dev 1053 drivers/dma/mv_xor.c dma_dev = &mv_chan->dmadev; dma_dev 1054 drivers/dma/mv_xor.c dma_dev->dev = &pdev->dev; dma_dev 1062 drivers/dma/mv_xor.c mv_chan->dummy_src_addr = dma_map_single(dma_dev->dev, dma_dev 1064 drivers/dma/mv_xor.c mv_chan->dummy_dst_addr = dma_map_single(dma_dev->dev, dma_dev 1078 drivers/dma/mv_xor.c dma_dev->cap_mask = cap_mask; dma_dev 1080 drivers/dma/mv_xor.c INIT_LIST_HEAD(&dma_dev->channels); dma_dev 1083 drivers/dma/mv_xor.c dma_dev->device_alloc_chan_resources = mv_xor_alloc_chan_resources; dma_dev 1084 drivers/dma/mv_xor.c dma_dev->device_free_chan_resources = mv_xor_free_chan_resources; dma_dev 1085 drivers/dma/mv_xor.c dma_dev->device_tx_status = mv_xor_status; dma_dev 1086 drivers/dma/mv_xor.c dma_dev->device_issue_pending = mv_xor_issue_pending; dma_dev 1089 drivers/dma/mv_xor.c if (dma_has_cap(DMA_INTERRUPT, dma_dev->cap_mask)) dma_dev 1090 drivers/dma/mv_xor.c dma_dev->device_prep_dma_interrupt = mv_xor_prep_dma_interrupt; dma_dev 1091 drivers/dma/mv_xor.c if (dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask)) dma_dev 1092 drivers/dma/mv_xor.c dma_dev->device_prep_dma_memcpy = mv_xor_prep_dma_memcpy; dma_dev 1093 drivers/dma/mv_xor.c if (dma_has_cap(DMA_XOR, dma_dev->cap_mask)) { dma_dev 1094 drivers/dma/mv_xor.c dma_dev->max_xor = 8; dma_dev 1095 drivers/dma/mv_xor.c dma_dev->device_prep_dma_xor = mv_xor_prep_dma_xor; dma_dev 1123 drivers/dma/mv_xor.c mv_chan->dmachan.device = dma_dev; dma_dev 1126 drivers/dma/mv_xor.c list_add_tail(&mv_chan->dmachan.device_node, &dma_dev->channels); dma_dev 1128 drivers/dma/mv_xor.c if (dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask)) { dma_dev 1135 drivers/dma/mv_xor.c if (dma_has_cap(DMA_XOR, dma_dev->cap_mask)) { dma_dev 1144 drivers/dma/mv_xor.c dma_has_cap(DMA_XOR, dma_dev->cap_mask) ? "xor " : "", dma_dev 1145 drivers/dma/mv_xor.c dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask) ? "cpy " : "", dma_dev 1146 drivers/dma/mv_xor.c dma_has_cap(DMA_INTERRUPT, dma_dev->cap_mask) ? "intr " : ""); dma_dev 1148 drivers/dma/mv_xor.c ret = dma_async_device_register(dma_dev); dma_dev 714 drivers/dma/mv_xor_v2.c struct dma_device *dma_dev; dma_dev 828 drivers/dma/mv_xor_v2.c dma_dev = &xor_dev->dmadev; dma_dev 831 drivers/dma/mv_xor_v2.c dma_cap_zero(dma_dev->cap_mask); dma_dev 832 drivers/dma/mv_xor_v2.c dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask); dma_dev 833 drivers/dma/mv_xor_v2.c dma_cap_set(DMA_XOR, dma_dev->cap_mask); dma_dev 834 drivers/dma/mv_xor_v2.c dma_cap_set(DMA_INTERRUPT, dma_dev->cap_mask); dma_dev 837 drivers/dma/mv_xor_v2.c INIT_LIST_HEAD(&dma_dev->channels); dma_dev 840 drivers/dma/mv_xor_v2.c dma_dev->device_tx_status = dma_cookie_status; dma_dev 841 drivers/dma/mv_xor_v2.c dma_dev->device_issue_pending = mv_xor_v2_issue_pending; dma_dev 842 drivers/dma/mv_xor_v2.c dma_dev->dev = &pdev->dev; dma_dev 844 drivers/dma/mv_xor_v2.c dma_dev->device_prep_dma_memcpy = mv_xor_v2_prep_dma_memcpy; dma_dev 845 drivers/dma/mv_xor_v2.c dma_dev->device_prep_dma_interrupt = mv_xor_v2_prep_dma_interrupt; dma_dev 846 drivers/dma/mv_xor_v2.c dma_dev->max_xor = 8; dma_dev 847 drivers/dma/mv_xor_v2.c dma_dev->device_prep_dma_xor = mv_xor_v2_prep_dma_xor; dma_dev 849 drivers/dma/mv_xor_v2.c xor_dev->dmachan.device = dma_dev; dma_dev 852 drivers/dma/mv_xor_v2.c &dma_dev->channels); dma_dev 858 drivers/dma/mv_xor_v2.c ret = dma_async_device_register(dma_dev); dma_dev 223 drivers/dma/nbpfaxi.c struct dma_device dma_dev; dma_dev 319 drivers/dma/nbpfaxi.c dev_dbg(nbpf->dma_dev.dev, "%s(0x%p + 0x%x) = 0x%x\n", dma_dev 328 drivers/dma/nbpfaxi.c dev_dbg(nbpf->dma_dev.dev, "%s(0x%p + 0x%x) = 0x%x\n", dma_dev 394 drivers/dma/nbpfaxi.c dev_dbg(chan->nbpf->dma_dev.dev, "%s(): next 0x%x, cur 0x%x\n", __func__, dma_dev 850 drivers/dma/nbpfaxi.c dev_dbg(chan->nbpf->dma_dev.dev, "%s(): force-free desc %p cookie %d\n", dma_dev 1091 drivers/dma/nbpfaxi.c dchan = dma_get_any_slave_channel(&nbpf->dma_dev); dma_dev 1222 drivers/dma/nbpfaxi.c dev_warn(nbpf->dma_dev.dev, "DMA error IRQ %u\n", irq); dma_dev 1240 drivers/dma/nbpfaxi.c struct dma_device *dma_dev = &nbpf->dma_dev; dma_dev 1248 drivers/dma/nbpfaxi.c chan->dma_chan.device = dma_dev; dma_dev 1252 drivers/dma/nbpfaxi.c dev_dbg(dma_dev->dev, "%s(): channel %d: -> %p\n", __func__, n, chan->base); dma_dev 1257 drivers/dma/nbpfaxi.c ret = devm_request_irq(dma_dev->dev, chan->irq, dma_dev 1265 drivers/dma/nbpfaxi.c &dma_dev->channels); dma_dev 1289 drivers/dma/nbpfaxi.c struct dma_device *dma_dev; dma_dev 1311 drivers/dma/nbpfaxi.c dma_dev = &nbpf->dma_dev; dma_dev 1312 drivers/dma/nbpfaxi.c dma_dev->dev = dev; dma_dev 1390 drivers/dma/nbpfaxi.c INIT_LIST_HEAD(&dma_dev->channels); dma_dev 1399 drivers/dma/nbpfaxi.c dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask); dma_dev 1400 drivers/dma/nbpfaxi.c dma_cap_set(DMA_SLAVE, dma_dev->cap_mask); dma_dev 1401 drivers/dma/nbpfaxi.c dma_cap_set(DMA_PRIVATE, dma_dev->cap_mask); dma_dev 1404 drivers/dma/nbpfaxi.c dma_dev->device_alloc_chan_resources dma_dev 1406 drivers/dma/nbpfaxi.c dma_dev->device_free_chan_resources = nbpf_free_chan_resources; dma_dev 1407 drivers/dma/nbpfaxi.c dma_dev->device_prep_dma_memcpy = nbpf_prep_memcpy; dma_dev 1408 drivers/dma/nbpfaxi.c dma_dev->device_tx_status = nbpf_tx_status; dma_dev 1409 drivers/dma/nbpfaxi.c dma_dev->device_issue_pending = nbpf_issue_pending; dma_dev 1420 drivers/dma/nbpfaxi.c dma_dev->device_prep_slave_sg = nbpf_prep_slave_sg; dma_dev 1421 drivers/dma/nbpfaxi.c dma_dev->device_config = nbpf_config; dma_dev 1422 drivers/dma/nbpfaxi.c dma_dev->device_pause = nbpf_pause; dma_dev 1423 drivers/dma/nbpfaxi.c dma_dev->device_terminate_all = nbpf_terminate_all; dma_dev 1425 drivers/dma/nbpfaxi.c dma_dev->src_addr_widths = NBPF_DMA_BUSWIDTHS; dma_dev 1426 drivers/dma/nbpfaxi.c dma_dev->dst_addr_widths = NBPF_DMA_BUSWIDTHS; dma_dev 1427 drivers/dma/nbpfaxi.c dma_dev->directions = BIT(DMA_DEV_TO_MEM) | BIT(DMA_MEM_TO_DEV); dma_dev 1437 drivers/dma/nbpfaxi.c ret = dma_async_device_register(dma_dev); dma_dev 1448 drivers/dma/nbpfaxi.c dma_async_device_unregister(dma_dev); dma_dev 1471 drivers/dma/nbpfaxi.c dma_async_device_unregister(&nbpf->dma_dev); dma_dev 37 drivers/dma/sh/shdma-base.c #define to_shdma_dev(d) container_of(d, struct shdma_dev, dma_dev) dma_dev 306 drivers/dma/sh/shdma-base.c dev_warn(sdev->dma_dev.dev, "invalid slave ID passed to dma_request_slave\n"); dma_dev 961 drivers/dma/sh/shdma-base.c schan->dma_chan.device = &sdev->dma_dev; dma_dev 964 drivers/dma/sh/shdma-base.c schan->dev = sdev->dma_dev.dev; dma_dev 978 drivers/dma/sh/shdma-base.c &sdev->dma_dev.channels); dma_dev 992 drivers/dma/sh/shdma-base.c struct dma_device *dma_dev = &sdev->dma_dev; dma_dev 1015 drivers/dma/sh/shdma-base.c INIT_LIST_HEAD(&dma_dev->channels); dma_dev 1018 drivers/dma/sh/shdma-base.c dma_dev->device_alloc_chan_resources dma_dev 1020 drivers/dma/sh/shdma-base.c dma_dev->device_free_chan_resources = shdma_free_chan_resources; dma_dev 1021 drivers/dma/sh/shdma-base.c dma_dev->device_prep_dma_memcpy = shdma_prep_memcpy; dma_dev 1022 drivers/dma/sh/shdma-base.c dma_dev->device_tx_status = shdma_tx_status; dma_dev 1023 drivers/dma/sh/shdma-base.c dma_dev->device_issue_pending = shdma_issue_pending; dma_dev 1026 drivers/dma/sh/shdma-base.c dma_dev->device_prep_slave_sg = shdma_prep_slave_sg; dma_dev 1027 drivers/dma/sh/shdma-base.c dma_dev->device_prep_dma_cyclic = shdma_prep_dma_cyclic; dma_dev 1028 drivers/dma/sh/shdma-base.c dma_dev->device_config = shdma_config; dma_dev 1029 drivers/dma/sh/shdma-base.c dma_dev->device_terminate_all = shdma_terminate_all; dma_dev 1031 drivers/dma/sh/shdma-base.c dma_dev->dev = dev; dma_dev 59 drivers/dma/sh/shdma.h struct sh_dmae_device, shdma_dev.dma_dev) dma_dev 165 drivers/dma/sh/shdmac.c dev_warn(shdev->shdma_dev.dma_dev.dev, "Can't initialize DMAOR.\n"); dma_dev 169 drivers/dma/sh/shdmac.c dev_warn(shdev->shdma_dev.dma_dev.dev, dma_dev 522 drivers/dma/sh/shdmac.c struct platform_device *pdev = to_platform_device(sdev->dma_dev.dev); dma_dev 527 drivers/dma/sh/shdmac.c sh_chan = devm_kzalloc(sdev->dma_dev.dev, sizeof(struct sh_dmae_chan), dma_dev 549 drivers/dma/sh/shdmac.c dev_err(sdev->dma_dev.dev, dma_dev 680 drivers/dma/sh/shdmac.c struct dma_device *dma_dev; dma_dev 720 drivers/dma/sh/shdmac.c dma_dev = &shdev->shdma_dev.dma_dev; dma_dev 731 drivers/dma/sh/shdmac.c dma_dev->src_addr_widths = widths; dma_dev 732 drivers/dma/sh/shdmac.c dma_dev->dst_addr_widths = widths; dma_dev 733 drivers/dma/sh/shdmac.c dma_dev->directions = BIT(DMA_MEM_TO_DEV) | BIT(DMA_DEV_TO_MEM); dma_dev 734 drivers/dma/sh/shdmac.c dma_dev->residue_granularity = DMA_RESIDUE_GRANULARITY_DESCRIPTOR; dma_dev 737 drivers/dma/sh/shdmac.c dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask); dma_dev 739 drivers/dma/sh/shdmac.c dma_cap_set(DMA_SLAVE, dma_dev->cap_mask); dma_dev 742 drivers/dma/sh/shdmac.c dma_dev->copy_align = LOG2_DEFAULT_XFER_SIZE; dma_dev 859 drivers/dma/sh/shdmac.c err = dma_async_device_register(&shdev->shdma_dev.dma_dev); dma_dev 890 drivers/dma/sh/shdmac.c struct dma_device *dma_dev = &shdev->shdma_dev.dma_dev; dma_dev 892 drivers/dma/sh/shdmac.c dma_async_device_unregister(dma_dev); dma_dev 206 drivers/dma/sprd-dma.c struct dma_device dma_dev; dma_dev 355 drivers/dma/sprd-dma.c dev_warn(sdev->dma_dev.dev, dma_dev 419 drivers/dma/sprd-dma.c dev_warn(sdev->dma_dev.dev, "incorrect dma interrupt type\n"); dma_dev 478 drivers/dma/sprd-dma.c dev_err(sdev->dma_dev.dev, "invalid channel mode setting %d\n", dma_dev 737 drivers/dma/sprd-dma.c dev_err(sdev->dma_dev.dev, "invalid source step\n"); dma_dev 753 drivers/dma/sprd-dma.c dev_err(sdev->dma_dev.dev, "invalid destination step\n"); dma_dev 761 drivers/dma/sprd-dma.c dev_err(sdev->dma_dev.dev, "invalid source datawidth\n"); dma_dev 767 drivers/dma/sprd-dma.c dev_err(sdev->dma_dev.dev, "invalid destination datawidth\n"); dma_dev 1134 drivers/dma/sprd-dma.c dma_cap_set(DMA_MEMCPY, sdev->dma_dev.cap_mask); dma_dev 1136 drivers/dma/sprd-dma.c sdev->dma_dev.chancnt = chn_count; dma_dev 1137 drivers/dma/sprd-dma.c INIT_LIST_HEAD(&sdev->dma_dev.channels); dma_dev 1138 drivers/dma/sprd-dma.c INIT_LIST_HEAD(&sdev->dma_dev.global_node); dma_dev 1139 drivers/dma/sprd-dma.c sdev->dma_dev.dev = &pdev->dev; dma_dev 1140 drivers/dma/sprd-dma.c sdev->dma_dev.device_alloc_chan_resources = sprd_dma_alloc_chan_resources; dma_dev 1141 drivers/dma/sprd-dma.c sdev->dma_dev.device_free_chan_resources = sprd_dma_free_chan_resources; dma_dev 1142 drivers/dma/sprd-dma.c sdev->dma_dev.device_tx_status = sprd_dma_tx_status; dma_dev 1143 drivers/dma/sprd-dma.c sdev->dma_dev.device_issue_pending = sprd_dma_issue_pending; dma_dev 1144 drivers/dma/sprd-dma.c sdev->dma_dev.device_prep_dma_memcpy = sprd_dma_prep_dma_memcpy; dma_dev 1145 drivers/dma/sprd-dma.c sdev->dma_dev.device_prep_slave_sg = sprd_dma_prep_slave_sg; dma_dev 1146 drivers/dma/sprd-dma.c sdev->dma_dev.device_config = sprd_dma_slave_config; dma_dev 1147 drivers/dma/sprd-dma.c sdev->dma_dev.device_pause = sprd_dma_pause; dma_dev 1148 drivers/dma/sprd-dma.c sdev->dma_dev.device_resume = sprd_dma_resume; dma_dev 1149 drivers/dma/sprd-dma.c sdev->dma_dev.device_terminate_all = sprd_dma_terminate_all; dma_dev 1160 drivers/dma/sprd-dma.c vchan_init(&dma_chn->vc, &sdev->dma_dev); dma_dev 1175 drivers/dma/sprd-dma.c ret = dma_async_device_register(&sdev->dma_dev); dma_dev 1181 drivers/dma/sprd-dma.c sprd_dma_info.dma_cap = sdev->dma_dev.cap_mask; dma_dev 1191 drivers/dma/sprd-dma.c dma_async_device_unregister(&sdev->dma_dev); dma_dev 1214 drivers/dma/sprd-dma.c list_for_each_entry_safe(c, cn, &sdev->dma_dev.channels, dma_dev 1221 drivers/dma/sprd-dma.c dma_async_device_unregister(&sdev->dma_dev); dma_dev 1249 drivers/dma/sprd-dma.c dev_err(sdev->dma_dev.dev, "enable dma failed\n"); dma_dev 210 drivers/dma/tegra20-apb-dma.c struct dma_device dma_dev; dma_dev 1331 drivers/dma/tegra20-apb-dma.c chan = dma_get_any_slave_channel(&tdma->dma_dev); dma_dev 1443 drivers/dma/tegra20-apb-dma.c INIT_LIST_HEAD(&tdma->dma_dev.channels); dma_dev 1467 drivers/dma/tegra20-apb-dma.c tdc->dma_chan.device = &tdma->dma_dev; dma_dev 1470 drivers/dma/tegra20-apb-dma.c &tdma->dma_dev.channels); dma_dev 1485 drivers/dma/tegra20-apb-dma.c dma_cap_set(DMA_SLAVE, tdma->dma_dev.cap_mask); dma_dev 1486 drivers/dma/tegra20-apb-dma.c dma_cap_set(DMA_PRIVATE, tdma->dma_dev.cap_mask); dma_dev 1487 drivers/dma/tegra20-apb-dma.c dma_cap_set(DMA_CYCLIC, tdma->dma_dev.cap_mask); dma_dev 1490 drivers/dma/tegra20-apb-dma.c tdma->dma_dev.dev = &pdev->dev; dma_dev 1491 drivers/dma/tegra20-apb-dma.c tdma->dma_dev.device_alloc_chan_resources = dma_dev 1493 drivers/dma/tegra20-apb-dma.c tdma->dma_dev.device_free_chan_resources = dma_dev 1495 drivers/dma/tegra20-apb-dma.c tdma->dma_dev.device_prep_slave_sg = tegra_dma_prep_slave_sg; dma_dev 1496 drivers/dma/tegra20-apb-dma.c tdma->dma_dev.device_prep_dma_cyclic = tegra_dma_prep_dma_cyclic; dma_dev 1497 drivers/dma/tegra20-apb-dma.c tdma->dma_dev.src_addr_widths = BIT(DMA_SLAVE_BUSWIDTH_1_BYTE) | dma_dev 1501 drivers/dma/tegra20-apb-dma.c tdma->dma_dev.dst_addr_widths = BIT(DMA_SLAVE_BUSWIDTH_1_BYTE) | dma_dev 1505 drivers/dma/tegra20-apb-dma.c tdma->dma_dev.directions = BIT(DMA_DEV_TO_MEM) | BIT(DMA_MEM_TO_DEV); dma_dev 1506 drivers/dma/tegra20-apb-dma.c tdma->dma_dev.residue_granularity = DMA_RESIDUE_GRANULARITY_BURST; dma_dev 1507 drivers/dma/tegra20-apb-dma.c tdma->dma_dev.device_config = tegra_dma_slave_config; dma_dev 1508 drivers/dma/tegra20-apb-dma.c tdma->dma_dev.device_terminate_all = tegra_dma_terminate_all; dma_dev 1509 drivers/dma/tegra20-apb-dma.c tdma->dma_dev.device_tx_status = tegra_dma_tx_status; dma_dev 1510 drivers/dma/tegra20-apb-dma.c tdma->dma_dev.device_issue_pending = tegra_dma_issue_pending; dma_dev 1512 drivers/dma/tegra20-apb-dma.c ret = dma_async_device_register(&tdma->dma_dev); dma_dev 1532 drivers/dma/tegra20-apb-dma.c dma_async_device_unregister(&tdma->dma_dev); dma_dev 1553 drivers/dma/tegra20-apb-dma.c dma_async_device_unregister(&tdma->dma_dev); dma_dev 153 drivers/dma/tegra210-adma.c struct dma_device dma_dev; dma_dev 702 drivers/dma/tegra210-adma.c chan = dma_get_any_slave_channel(&tdma->dma_dev); dma_dev 851 drivers/dma/tegra210-adma.c INIT_LIST_HEAD(&tdma->dma_dev.channels); dma_dev 864 drivers/dma/tegra210-adma.c vchan_init(&tdc->vc, &tdma->dma_dev); dma_dev 879 drivers/dma/tegra210-adma.c dma_cap_set(DMA_SLAVE, tdma->dma_dev.cap_mask); dma_dev 880 drivers/dma/tegra210-adma.c dma_cap_set(DMA_PRIVATE, tdma->dma_dev.cap_mask); dma_dev 881 drivers/dma/tegra210-adma.c dma_cap_set(DMA_CYCLIC, tdma->dma_dev.cap_mask); dma_dev 883 drivers/dma/tegra210-adma.c tdma->dma_dev.dev = &pdev->dev; dma_dev 884 drivers/dma/tegra210-adma.c tdma->dma_dev.device_alloc_chan_resources = dma_dev 886 drivers/dma/tegra210-adma.c tdma->dma_dev.device_free_chan_resources = dma_dev 888 drivers/dma/tegra210-adma.c tdma->dma_dev.device_issue_pending = tegra_adma_issue_pending; dma_dev 889 drivers/dma/tegra210-adma.c tdma->dma_dev.device_prep_dma_cyclic = tegra_adma_prep_dma_cyclic; dma_dev 890 drivers/dma/tegra210-adma.c tdma->dma_dev.device_config = tegra_adma_slave_config; dma_dev 891 drivers/dma/tegra210-adma.c tdma->dma_dev.device_tx_status = tegra_adma_tx_status; dma_dev 892 drivers/dma/tegra210-adma.c tdma->dma_dev.device_terminate_all = tegra_adma_terminate_all; dma_dev 893 drivers/dma/tegra210-adma.c tdma->dma_dev.src_addr_widths = BIT(DMA_SLAVE_BUSWIDTH_4_BYTES); dma_dev 894 drivers/dma/tegra210-adma.c tdma->dma_dev.dst_addr_widths = BIT(DMA_SLAVE_BUSWIDTH_4_BYTES); dma_dev 895 drivers/dma/tegra210-adma.c tdma->dma_dev.directions = BIT(DMA_DEV_TO_MEM) | BIT(DMA_MEM_TO_DEV); dma_dev 896 drivers/dma/tegra210-adma.c tdma->dma_dev.residue_granularity = DMA_RESIDUE_GRANULARITY_SEGMENT; dma_dev 897 drivers/dma/tegra210-adma.c tdma->dma_dev.device_pause = tegra_adma_pause; dma_dev 898 drivers/dma/tegra210-adma.c tdma->dma_dev.device_resume = tegra_adma_resume; dma_dev 900 drivers/dma/tegra210-adma.c ret = dma_async_device_register(&tdma->dma_dev); dma_dev 921 drivers/dma/tegra210-adma.c dma_async_device_unregister(&tdma->dma_dev); dma_dev 939 drivers/dma/tegra210-adma.c dma_async_device_unregister(&tdma->dma_dev); dma_dev 308 drivers/dma/xgene-dma.c struct dma_device dma_dev[XGENE_DMA_MAX_CHANNEL]; dma_dev 1483 drivers/dma/xgene-dma.c struct dma_device *dma_dev) dma_dev 1486 drivers/dma/xgene-dma.c dma_cap_zero(dma_dev->cap_mask); dma_dev 1501 drivers/dma/xgene-dma.c dma_cap_set(DMA_PQ, dma_dev->cap_mask); dma_dev 1502 drivers/dma/xgene-dma.c dma_cap_set(DMA_XOR, dma_dev->cap_mask); dma_dev 1505 drivers/dma/xgene-dma.c dma_cap_set(DMA_XOR, dma_dev->cap_mask); dma_dev 1509 drivers/dma/xgene-dma.c dma_dev->dev = chan->dev; dma_dev 1510 drivers/dma/xgene-dma.c dma_dev->device_alloc_chan_resources = xgene_dma_alloc_chan_resources; dma_dev 1511 drivers/dma/xgene-dma.c dma_dev->device_free_chan_resources = xgene_dma_free_chan_resources; dma_dev 1512 drivers/dma/xgene-dma.c dma_dev->device_issue_pending = xgene_dma_issue_pending; dma_dev 1513 drivers/dma/xgene-dma.c dma_dev->device_tx_status = xgene_dma_tx_status; dma_dev 1515 drivers/dma/xgene-dma.c if (dma_has_cap(DMA_XOR, dma_dev->cap_mask)) { dma_dev 1516 drivers/dma/xgene-dma.c dma_dev->device_prep_dma_xor = xgene_dma_prep_xor; dma_dev 1517 drivers/dma/xgene-dma.c dma_dev->max_xor = XGENE_DMA_MAX_XOR_SRC; dma_dev 1518 drivers/dma/xgene-dma.c dma_dev->xor_align = DMAENGINE_ALIGN_64_BYTES; dma_dev 1521 drivers/dma/xgene-dma.c if (dma_has_cap(DMA_PQ, dma_dev->cap_mask)) { dma_dev 1522 drivers/dma/xgene-dma.c dma_dev->device_prep_dma_pq = xgene_dma_prep_pq; dma_dev 1523 drivers/dma/xgene-dma.c dma_dev->max_pq = XGENE_DMA_MAX_XOR_SRC; dma_dev 1524 drivers/dma/xgene-dma.c dma_dev->pq_align = DMAENGINE_ALIGN_64_BYTES; dma_dev 1531 drivers/dma/xgene-dma.c struct dma_device *dma_dev = &pdma->dma_dev[id]; dma_dev 1534 drivers/dma/xgene-dma.c chan->dma_chan.device = dma_dev; dma_dev 1548 drivers/dma/xgene-dma.c xgene_dma_set_caps(chan, dma_dev); dma_dev 1551 drivers/dma/xgene-dma.c INIT_LIST_HEAD(&dma_dev->channels); dma_dev 1552 drivers/dma/xgene-dma.c list_add_tail(&chan->dma_chan.device_node, &dma_dev->channels); dma_dev 1555 drivers/dma/xgene-dma.c ret = dma_async_device_register(dma_dev); dma_dev 1566 drivers/dma/xgene-dma.c dma_has_cap(DMA_XOR, dma_dev->cap_mask) ? "XOR " : "", dma_dev 1567 drivers/dma/xgene-dma.c dma_has_cap(DMA_PQ, dma_dev->cap_mask) ? "PQ " : ""); dma_dev 1580 drivers/dma/xgene-dma.c dma_async_device_unregister(&pdma->dma_dev[j]); dma_dev 1596 drivers/dma/xgene-dma.c dma_async_device_unregister(&pdma->dma_dev[i]); dma_dev 66 drivers/gpu/drm/exynos/exynos_drm_dma.c if (get_dma_ops(priv->dma_dev) != get_dma_ops(subdrv_dev)) { dma_dev 126 drivers/gpu/drm/exynos/exynos_drm_dma.c if (!priv->dma_dev) { dma_dev 127 drivers/gpu/drm/exynos/exynos_drm_dma.c priv->dma_dev = dev; dma_dev 142 drivers/gpu/drm/exynos/exynos_drm_dma.c mapping = iommu_get_domain_for_dev(priv->dma_dev); dma_dev 168 drivers/gpu/drm/exynos/exynos_drm_dma.c priv->dma_dev = NULL; dma_dev 203 drivers/gpu/drm/exynos/exynos_drm_drv.h struct device *dma_dev; dma_dev 216 drivers/gpu/drm/exynos/exynos_drm_drv.h return priv->dma_dev; dma_dev 211 drivers/gpu/drm/mediatek/mtk_drm_drv.c struct device *dma_dev; dma_dev 274 drivers/gpu/drm/mediatek/mtk_drm_drv.c dma_dev = &pdev->dev; dma_dev 275 drivers/gpu/drm/mediatek/mtk_drm_drv.c private->dma_dev = dma_dev; dma_dev 281 drivers/gpu/drm/mediatek/mtk_drm_drv.c if (!dma_dev->dma_parms) { dma_dev 283 drivers/gpu/drm/mediatek/mtk_drm_drv.c dma_dev->dma_parms = dma_dev 284 drivers/gpu/drm/mediatek/mtk_drm_drv.c devm_kzalloc(drm->dev, sizeof(*dma_dev->dma_parms), dma_dev 287 drivers/gpu/drm/mediatek/mtk_drm_drv.c if (!dma_dev->dma_parms) { dma_dev 292 drivers/gpu/drm/mediatek/mtk_drm_drv.c ret = dma_set_max_seg_size(dma_dev, (unsigned int)DMA_BIT_MASK(32)); dma_dev 294 drivers/gpu/drm/mediatek/mtk_drm_drv.c dev_err(dma_dev, "Failed to set DMA segment size\n"); dma_dev 315 drivers/gpu/drm/mediatek/mtk_drm_drv.c dma_dev->dma_parms = NULL; dma_dev 332 drivers/gpu/drm/mediatek/mtk_drm_drv.c private->dma_dev->dma_parms = NULL; dma_dev 358 drivers/gpu/drm/mediatek/mtk_drm_drv.c return drm_gem_prime_import_dev(dev, dma_buf, private->dma_dev); dma_dev 36 drivers/gpu/drm/mediatek/mtk_drm_drv.h struct device *dma_dev; dma_dev 57 drivers/gpu/drm/mediatek/mtk_drm_gem.c mtk_gem->cookie = dma_alloc_attrs(priv->dma_dev, obj->size, dma_dev 89 drivers/gpu/drm/mediatek/mtk_drm_gem.c dma_free_attrs(priv->dma_dev, obj->size, mtk_gem->cookie, dma_dev 143 drivers/gpu/drm/mediatek/mtk_drm_gem.c ret = dma_mmap_attrs(priv->dma_dev, vma, mtk_gem->cookie, dma_dev 199 drivers/gpu/drm/mediatek/mtk_drm_gem.c ret = dma_get_sgtable_attrs(priv->dma_dev, sgt, mtk_gem->cookie, dma_dev 85 drivers/gpu/drm/omapdrm/omap_dmm_tiler.c struct dma_device *dma_dev = dmm->wa_dma_chan->device; dma_dev 90 drivers/gpu/drm/omapdrm/omap_dmm_tiler.c tx = dma_dev->device_prep_dma_memcpy(dmm->wa_dma_chan, dst, src, 4, 0); dma_dev 178 drivers/iommu/exynos-iommu.c static struct device *dma_dev; dma_dev 659 drivers/iommu/exynos-iommu.c if (!dma_dev) dma_dev 660 drivers/iommu/exynos-iommu.c dma_dev = &pdev->dev; dma_dev 726 drivers/iommu/exynos-iommu.c dma_sync_single_for_cpu(dma_dev, virt_to_phys(ent), sizeof(*ent), dma_dev 729 drivers/iommu/exynos-iommu.c dma_sync_single_for_device(dma_dev, virt_to_phys(ent), sizeof(*ent), dma_dev 740 drivers/iommu/exynos-iommu.c BUG_ON(PG_ENT_SHIFT < 0 || !dma_dev); dma_dev 765 drivers/iommu/exynos-iommu.c handle = dma_map_single(dma_dev, domain->pgtable, LV1TABLE_SIZE, dma_dev 769 drivers/iommu/exynos-iommu.c if (dma_mapping_error(dma_dev, handle)) dma_dev 819 drivers/iommu/exynos-iommu.c dma_unmap_single(dma_dev, virt_to_phys(domain->pgtable), LV1TABLE_SIZE, dma_dev 826 drivers/iommu/exynos-iommu.c dma_unmap_single(dma_dev, base, LV2TABLE_SIZE, dma_dev 939 drivers/iommu/exynos-iommu.c handle = dma_map_single(dma_dev, pent, LV2TABLE_SIZE, dma_dev 941 drivers/iommu/exynos-iommu.c if (dma_mapping_error(dma_dev, handle)) { dma_dev 1027 drivers/iommu/exynos-iommu.c dma_sync_single_for_cpu(dma_dev, pent_base, dma_dev 1039 drivers/iommu/exynos-iommu.c dma_sync_single_for_device(dma_dev, pent_base, dma_dev 1186 drivers/iommu/exynos-iommu.c dma_sync_single_for_cpu(dma_dev, virt_to_phys(ent), dma_dev 1190 drivers/iommu/exynos-iommu.c dma_sync_single_for_device(dma_dev, virt_to_phys(ent), dma_dev 118 drivers/iommu/rockchip-iommu.c static struct device *dma_dev; dma_dev 125 drivers/iommu/rockchip-iommu.c dma_sync_single_for_device(dma_dev, dma, size, DMA_TO_DEVICE); dma_dev 675 drivers/iommu/rockchip-iommu.c pt_dma = dma_map_single(dma_dev, page_table, SPAGE_SIZE, DMA_TO_DEVICE); dma_dev 676 drivers/iommu/rockchip-iommu.c if (dma_mapping_error(dma_dev, pt_dma)) { dma_dev 677 drivers/iommu/rockchip-iommu.c dev_err(dma_dev, "DMA mapping error while allocating page table\n"); dma_dev 980 drivers/iommu/rockchip-iommu.c if (!dma_dev) dma_dev 1000 drivers/iommu/rockchip-iommu.c rk_domain->dt_dma = dma_map_single(dma_dev, rk_domain->dt, dma_dev 1002 drivers/iommu/rockchip-iommu.c if (dma_mapping_error(dma_dev, rk_domain->dt_dma)) { dma_dev 1003 drivers/iommu/rockchip-iommu.c dev_err(dma_dev, "DMA map error for DT\n"); dma_dev 1042 drivers/iommu/rockchip-iommu.c dma_unmap_single(dma_dev, pt_phys, dma_dev 1048 drivers/iommu/rockchip-iommu.c dma_unmap_single(dma_dev, rk_domain->dt_dma, dma_dev 1108 drivers/iommu/rockchip-iommu.c data = devm_kzalloc(dma_dev, sizeof(*data), GFP_KERNEL); dma_dev 1223 drivers/iommu/rockchip-iommu.c if (!dma_dev) dma_dev 1224 drivers/iommu/rockchip-iommu.c dma_dev = &pdev->dev; dma_dev 233 drivers/mcb/mcb-core.c dev->dma_dev = bus->carrier; dma_dev 129 drivers/mmc/host/mmc_spi.c struct device *dma_dev; dma_dev 168 drivers/mmc/host/mmc_spi.c if (host->dma_dev) dma_dev 169 drivers/mmc/host/mmc_spi.c dma_sync_single_for_device(host->dma_dev, dma_dev 175 drivers/mmc/host/mmc_spi.c if (host->dma_dev) dma_dev 176 drivers/mmc/host/mmc_spi.c dma_sync_single_for_cpu(host->dma_dev, dma_dev 523 drivers/mmc/host/mmc_spi.c if (host->dma_dev) { dma_dev 525 drivers/mmc/host/mmc_spi.c dma_sync_single_for_device(host->dma_dev, dma_dev 531 drivers/mmc/host/mmc_spi.c if (host->dma_dev) dma_dev 532 drivers/mmc/host/mmc_spi.c dma_sync_single_for_cpu(host->dma_dev, dma_dev 665 drivers/mmc/host/mmc_spi.c if (host->dma_dev) dma_dev 666 drivers/mmc/host/mmc_spi.c dma_sync_single_for_device(host->dma_dev, dma_dev 677 drivers/mmc/host/mmc_spi.c if (host->dma_dev) dma_dev 678 drivers/mmc/host/mmc_spi.c dma_sync_single_for_cpu(host->dma_dev, dma_dev 730 drivers/mmc/host/mmc_spi.c if (host->dma_dev) dma_dev 795 drivers/mmc/host/mmc_spi.c if (host->dma_dev) { dma_dev 796 drivers/mmc/host/mmc_spi.c dma_sync_single_for_device(host->dma_dev, dma_dev 799 drivers/mmc/host/mmc_spi.c dma_sync_single_for_device(host->dma_dev, dma_dev 810 drivers/mmc/host/mmc_spi.c if (host->dma_dev) { dma_dev 811 drivers/mmc/host/mmc_spi.c dma_sync_single_for_cpu(host->dma_dev, dma_dev 814 drivers/mmc/host/mmc_spi.c dma_sync_single_for_cpu(host->dma_dev, dma_dev 853 drivers/mmc/host/mmc_spi.c if (host->dma_dev) dma_dev 869 drivers/mmc/host/mmc_spi.c struct device *dma_dev = host->dma_dev; dma_dev 904 drivers/mmc/host/mmc_spi.c if (dma_dev) { dma_dev 910 drivers/mmc/host/mmc_spi.c dma_addr = dma_map_page(dma_dev, sg_page(sg), 0, dma_dev 912 drivers/mmc/host/mmc_spi.c if (dma_mapping_error(dma_dev, dma_addr)) { dma_dev 956 drivers/mmc/host/mmc_spi.c if (dma_dev) dma_dev 957 drivers/mmc/host/mmc_spi.c dma_unmap_page(dma_dev, dma_addr, PAGE_SIZE, dir); dma_dev 997 drivers/mmc/host/mmc_spi.c if (host->dma_dev) dma_dev 998 drivers/mmc/host/mmc_spi.c dma_sync_single_for_device(host->dma_dev, dma_dev 1004 drivers/mmc/host/mmc_spi.c if (host->dma_dev) dma_dev 1005 drivers/mmc/host/mmc_spi.c dma_sync_single_for_cpu(host->dma_dev, dma_dev 1382 drivers/mmc/host/mmc_spi.c host->dma_dev = dev; dma_dev 1392 drivers/mmc/host/mmc_spi.c dma_sync_single_for_cpu(host->dma_dev, dma_dev 1399 drivers/mmc/host/mmc_spi.c host->readback.is_dma_mapped = (host->dma_dev != NULL); dma_dev 1452 drivers/mmc/host/mmc_spi.c host->dma_dev ? "" : ", no DMA", dma_dev 1463 drivers/mmc/host/mmc_spi.c if (host->dma_dev) dma_dev 1464 drivers/mmc/host/mmc_spi.c dma_unmap_single(host->dma_dev, host->data_dma, dma_dev 1467 drivers/mmc/host/mmc_spi.c if (host->dma_dev) dma_dev 1468 drivers/mmc/host/mmc_spi.c dma_unmap_single(host->dma_dev, host->ones_dma, dma_dev 1494 drivers/mmc/host/mmc_spi.c if (host->dma_dev) { dma_dev 1495 drivers/mmc/host/mmc_spi.c dma_unmap_single(host->dma_dev, host->ones_dma, dma_dev 1497 drivers/mmc/host/mmc_spi.c dma_unmap_single(host->dma_dev, host->data_dma, dma_dev 461 drivers/mtd/nand/raw/fsmc_nand.c struct dma_device *dma_dev; dma_dev 476 drivers/mtd/nand/raw/fsmc_nand.c dma_dev = chan->device; dma_dev 477 drivers/mtd/nand/raw/fsmc_nand.c dma_addr = dma_map_single(dma_dev->dev, buffer, len, direction); dma_dev 487 drivers/mtd/nand/raw/fsmc_nand.c tx = dma_dev->device_prep_dma_memcpy(chan, dma_dst, dma_src, dma_dev 520 drivers/mtd/nand/raw/fsmc_nand.c dma_unmap_single(dma_dev->dev, dma_addr, len, direction); dma_dev 153 drivers/net/ethernet/broadcom/b44.c dma_sync_single_for_device(sdev->dma_dev, dma_base + offset, dma_dev 162 drivers/net/ethernet/broadcom/b44.c dma_sync_single_for_cpu(sdev->dma_dev, dma_base + offset, dma_dev 632 drivers/net/ethernet/broadcom/b44.c dma_unmap_single(bp->sdev->dma_dev, dma_dev 677 drivers/net/ethernet/broadcom/b44.c mapping = dma_map_single(bp->sdev->dma_dev, skb->data, dma_dev 683 drivers/net/ethernet/broadcom/b44.c if (dma_mapping_error(bp->sdev->dma_dev, mapping) || dma_dev 686 drivers/net/ethernet/broadcom/b44.c if (!dma_mapping_error(bp->sdev->dma_dev, mapping)) dma_dev 687 drivers/net/ethernet/broadcom/b44.c dma_unmap_single(bp->sdev->dma_dev, mapping, dma_dev 693 drivers/net/ethernet/broadcom/b44.c mapping = dma_map_single(bp->sdev->dma_dev, skb->data, dma_dev 696 drivers/net/ethernet/broadcom/b44.c if (dma_mapping_error(bp->sdev->dma_dev, mapping) || dma_dev 698 drivers/net/ethernet/broadcom/b44.c if (!dma_mapping_error(bp->sdev->dma_dev, mapping)) dma_dev 699 drivers/net/ethernet/broadcom/b44.c dma_unmap_single(bp->sdev->dma_dev, mapping, RX_PKT_BUF_SZ,DMA_FROM_DEVICE); dma_dev 774 drivers/net/ethernet/broadcom/b44.c dma_sync_single_for_device(bp->sdev->dma_dev, dest_map->mapping, dma_dev 796 drivers/net/ethernet/broadcom/b44.c dma_sync_single_for_cpu(bp->sdev->dma_dev, map, dma_dev 830 drivers/net/ethernet/broadcom/b44.c dma_unmap_single(bp->sdev->dma_dev, map, dma_dev 991 drivers/net/ethernet/broadcom/b44.c mapping = dma_map_single(bp->sdev->dma_dev, skb->data, len, DMA_TO_DEVICE); dma_dev 992 drivers/net/ethernet/broadcom/b44.c if (dma_mapping_error(bp->sdev->dma_dev, mapping) || mapping + len > DMA_BIT_MASK(30)) { dma_dev 996 drivers/net/ethernet/broadcom/b44.c if (!dma_mapping_error(bp->sdev->dma_dev, mapping)) dma_dev 997 drivers/net/ethernet/broadcom/b44.c dma_unmap_single(bp->sdev->dma_dev, mapping, len, dma_dev 1004 drivers/net/ethernet/broadcom/b44.c mapping = dma_map_single(bp->sdev->dma_dev, bounce_skb->data, dma_dev 1006 drivers/net/ethernet/broadcom/b44.c if (dma_mapping_error(bp->sdev->dma_dev, mapping) || mapping + len > DMA_BIT_MASK(30)) { dma_dev 1007 drivers/net/ethernet/broadcom/b44.c if (!dma_mapping_error(bp->sdev->dma_dev, mapping)) dma_dev 1008 drivers/net/ethernet/broadcom/b44.c dma_unmap_single(bp->sdev->dma_dev, mapping, dma_dev 1104 drivers/net/ethernet/broadcom/b44.c dma_unmap_single(bp->sdev->dma_dev, rp->mapping, RX_PKT_BUF_SZ, dma_dev 1116 drivers/net/ethernet/broadcom/b44.c dma_unmap_single(bp->sdev->dma_dev, rp->mapping, rp->skb->len, dma_dev 1139 drivers/net/ethernet/broadcom/b44.c dma_sync_single_for_device(bp->sdev->dma_dev, bp->rx_ring_dma, dma_dev 1143 drivers/net/ethernet/broadcom/b44.c dma_sync_single_for_device(bp->sdev->dma_dev, bp->tx_ring_dma, dma_dev 1164 drivers/net/ethernet/broadcom/b44.c dma_unmap_single(bp->sdev->dma_dev, bp->rx_ring_dma, dma_dev 1168 drivers/net/ethernet/broadcom/b44.c dma_free_coherent(bp->sdev->dma_dev, DMA_TABLE_BYTES, dma_dev 1175 drivers/net/ethernet/broadcom/b44.c dma_unmap_single(bp->sdev->dma_dev, bp->tx_ring_dma, dma_dev 1179 drivers/net/ethernet/broadcom/b44.c dma_free_coherent(bp->sdev->dma_dev, DMA_TABLE_BYTES, dma_dev 1205 drivers/net/ethernet/broadcom/b44.c bp->rx_ring = dma_alloc_coherent(bp->sdev->dma_dev, size, dma_dev 1218 drivers/net/ethernet/broadcom/b44.c rx_ring_dma = dma_map_single(bp->sdev->dma_dev, rx_ring, dma_dev 1222 drivers/net/ethernet/broadcom/b44.c if (dma_mapping_error(bp->sdev->dma_dev, rx_ring_dma) || dma_dev 1233 drivers/net/ethernet/broadcom/b44.c bp->tx_ring = dma_alloc_coherent(bp->sdev->dma_dev, size, dma_dev 1246 drivers/net/ethernet/broadcom/b44.c tx_ring_dma = dma_map_single(bp->sdev->dma_dev, tx_ring, dma_dev 1250 drivers/net/ethernet/broadcom/b44.c if (dma_mapping_error(bp->sdev->dma_dev, tx_ring_dma) || dma_dev 2394 drivers/net/ethernet/broadcom/b44.c if (dma_set_mask_and_coherent(sdev->dma_dev, DMA_BIT_MASK(30))) { dma_dev 126 drivers/net/ethernet/broadcom/bgmac-bcma.c bgmac->dma_dev = core->dma_dev; dma_dev 193 drivers/net/ethernet/broadcom/bgmac-platform.c bgmac->dma_dev = &pdev->dev; dma_dev 135 drivers/net/ethernet/broadcom/bgmac.c struct device *dma_dev = bgmac->dma_dev; dma_dev 162 drivers/net/ethernet/broadcom/bgmac.c slot->dma_addr = dma_map_single(dma_dev, skb->data, skb_headlen(skb), dma_dev 164 drivers/net/ethernet/broadcom/bgmac.c if (unlikely(dma_mapping_error(dma_dev, slot->dma_addr))) dma_dev 180 drivers/net/ethernet/broadcom/bgmac.c slot->dma_addr = skb_frag_dma_map(dma_dev, frag, 0, dma_dev 182 drivers/net/ethernet/broadcom/bgmac.c if (unlikely(dma_mapping_error(dma_dev, slot->dma_addr))) dma_dev 211 drivers/net/ethernet/broadcom/bgmac.c dma_unmap_single(dma_dev, slot->dma_addr, skb_headlen(skb), dma_dev 220 drivers/net/ethernet/broadcom/bgmac.c dma_unmap_page(dma_dev, slot->dma_addr, len, DMA_TO_DEVICE); dma_dev 237 drivers/net/ethernet/broadcom/bgmac.c struct device *dma_dev = bgmac->dma_dev; dma_dev 262 drivers/net/ethernet/broadcom/bgmac.c dma_unmap_single(dma_dev, slot->dma_addr, len, dma_dev 265 drivers/net/ethernet/broadcom/bgmac.c dma_unmap_page(dma_dev, slot->dma_addr, len, dma_dev 336 drivers/net/ethernet/broadcom/bgmac.c struct device *dma_dev = bgmac->dma_dev; dma_dev 352 drivers/net/ethernet/broadcom/bgmac.c dma_addr = dma_map_single(dma_dev, buf + BGMAC_RX_BUF_OFFSET, dma_dev 354 drivers/net/ethernet/broadcom/bgmac.c if (dma_mapping_error(dma_dev, dma_addr)) { dma_dev 399 drivers/net/ethernet/broadcom/bgmac.c static void bgmac_dma_rx_poison_buf(struct device *dma_dev, dma_dev 404 drivers/net/ethernet/broadcom/bgmac.c dma_sync_single_for_cpu(dma_dev, slot->dma_addr, BGMAC_RX_BUF_SIZE, dma_dev 408 drivers/net/ethernet/broadcom/bgmac.c dma_sync_single_for_device(dma_dev, slot->dma_addr, BGMAC_RX_BUF_SIZE, dma_dev 425 drivers/net/ethernet/broadcom/bgmac.c struct device *dma_dev = bgmac->dma_dev; dma_dev 436 drivers/net/ethernet/broadcom/bgmac.c bgmac_dma_rx_poison_buf(dma_dev, slot); dma_dev 441 drivers/net/ethernet/broadcom/bgmac.c dma_unmap_single(dma_dev, dma_addr, dma_dev 528 drivers/net/ethernet/broadcom/bgmac.c struct device *dma_dev = bgmac->dma_dev; dma_dev 544 drivers/net/ethernet/broadcom/bgmac.c dma_unmap_single(dma_dev, slot->dma_addr, dma_dev 547 drivers/net/ethernet/broadcom/bgmac.c dma_unmap_page(dma_dev, slot->dma_addr, dma_dev 555 drivers/net/ethernet/broadcom/bgmac.c struct device *dma_dev = bgmac->dma_dev; dma_dev 564 drivers/net/ethernet/broadcom/bgmac.c dma_unmap_single(dma_dev, slot->dma_addr, dma_dev 576 drivers/net/ethernet/broadcom/bgmac.c struct device *dma_dev = bgmac->dma_dev; dma_dev 584 drivers/net/ethernet/broadcom/bgmac.c dma_free_coherent(dma_dev, size, ring->cpu_base, dma_dev 614 drivers/net/ethernet/broadcom/bgmac.c struct device *dma_dev = bgmac->dma_dev; dma_dev 637 drivers/net/ethernet/broadcom/bgmac.c ring->cpu_base = dma_alloc_coherent(dma_dev, size, dma_dev 662 drivers/net/ethernet/broadcom/bgmac.c ring->cpu_base = dma_alloc_coherent(dma_dev, size, dma_dev 494 drivers/net/ethernet/broadcom/bgmac.h struct device *dma_dev; dma_dev 57 drivers/net/ethernet/sfc/falcon/tx.c struct device *dma_dev = &tx_queue->efx->pci_dev->dev; dma_dev 60 drivers/net/ethernet/sfc/falcon/tx.c dma_unmap_single(dma_dev, unmap_addr, buffer->unmap_len, dma_dev 63 drivers/net/ethernet/sfc/falcon/tx.c dma_unmap_page(dma_dev, unmap_addr, buffer->unmap_len, dma_dev 209 drivers/net/ethernet/sfc/falcon/tx.c struct device *dma_dev = &efx->pci_dev->dev; dma_dev 220 drivers/net/ethernet/sfc/falcon/tx.c dma_addr = dma_map_single(dma_dev, skb->data, len, DMA_TO_DEVICE); dma_dev 225 drivers/net/ethernet/sfc/falcon/tx.c if (unlikely(dma_mapping_error(dma_dev, dma_addr))) dma_dev 254 drivers/net/ethernet/sfc/falcon/tx.c dma_addr = skb_frag_dma_map(dma_dev, fragment, dma_dev 260 drivers/net/ethernet/sfc/falcon/tx.c if (unlikely(dma_mapping_error(dma_dev, dma_addr))) dma_dev 65 drivers/net/ethernet/sfc/tx.c struct device *dma_dev = &tx_queue->efx->pci_dev->dev; dma_dev 68 drivers/net/ethernet/sfc/tx.c dma_unmap_single(dma_dev, unmap_addr, buffer->unmap_len, dma_dev 71 drivers/net/ethernet/sfc/tx.c dma_unmap_page(dma_dev, unmap_addr, buffer->unmap_len, dma_dev 364 drivers/net/ethernet/sfc/tx.c struct device *dma_dev = &efx->pci_dev->dev; dma_dev 375 drivers/net/ethernet/sfc/tx.c dma_addr = dma_map_single(dma_dev, skb->data, len, DMA_TO_DEVICE); dma_dev 380 drivers/net/ethernet/sfc/tx.c if (unlikely(dma_mapping_error(dma_dev, dma_addr))) dma_dev 424 drivers/net/ethernet/sfc/tx.c dma_addr = skb_frag_dma_map(dma_dev, fragment, dma_dev 430 drivers/net/ethernet/sfc/tx.c if (unlikely(dma_mapping_error(dma_dev, dma_addr))) dma_dev 173 drivers/net/ethernet/sfc/tx_tso.c struct device *dma_dev = &efx->pci_dev->dev; dma_dev 198 drivers/net/ethernet/sfc/tx_tso.c dma_addr = dma_map_single(dma_dev, skb->data, dma_dev 205 drivers/net/ethernet/sfc/tx_tso.c return unlikely(dma_mapping_error(dma_dev, dma_addr)) ? -ENOMEM : 0; dma_dev 89 drivers/net/ethernet/sgi/ioc3-eth.c struct device *dma_dev; dma_dev 150 drivers/net/ethernet/sgi/ioc3-eth.c d = dma_map_single(ip->dma_dev, new_skb->data, dma_dev 153 drivers/net/ethernet/sgi/ioc3-eth.c if (dma_mapping_error(ip->dma_dev, d)) { dma_dev 580 drivers/net/ethernet/sgi/ioc3-eth.c dma_unmap_single(ip->dma_dev, rxr[rx_entry], dma_dev 812 drivers/net/ethernet/sgi/ioc3-eth.c dma_unmap_single(ip->dma_dev, be64_to_cpu(desc->p1), dma_dev 817 drivers/net/ethernet/sgi/ioc3-eth.c dma_unmap_single(ip->dma_dev, be64_to_cpu(desc->p2), dma_dev 851 drivers/net/ethernet/sgi/ioc3-eth.c dma_unmap_single(ip->dma_dev, dma_dev 1218 drivers/net/ethernet/sgi/ioc3-eth.c ip->dma_dev = &pdev->dev; dma_dev 1245 drivers/net/ethernet/sgi/ioc3-eth.c ip->rxr = dma_direct_alloc_pages(ip->dma_dev, RX_RING_SIZE, dma_dev 1254 drivers/net/ethernet/sgi/ioc3-eth.c ip->txr = dma_direct_alloc_pages(ip->dma_dev, TX_RING_SIZE, dma_dev 1316 drivers/net/ethernet/sgi/ioc3-eth.c dma_direct_free_pages(ip->dma_dev, RX_RING_SIZE, ip->rxr, dma_dev 1319 drivers/net/ethernet/sgi/ioc3-eth.c dma_direct_free_pages(ip->dma_dev, TX_RING_SIZE, ip->txr, dma_dev 1338 drivers/net/ethernet/sgi/ioc3-eth.c dma_direct_free_pages(ip->dma_dev, RX_RING_SIZE, ip->rxr, dma_dev 1340 drivers/net/ethernet/sgi/ioc3-eth.c dma_direct_free_pages(ip->dma_dev, TX_RING_SIZE, ip->txr, dma_dev 1449 drivers/net/ethernet/sgi/ioc3-eth.c d1 = dma_map_single(ip->dma_dev, skb->data, s1, DMA_TO_DEVICE); dma_dev 1450 drivers/net/ethernet/sgi/ioc3-eth.c if (dma_mapping_error(ip->dma_dev, d1)) dma_dev 1452 drivers/net/ethernet/sgi/ioc3-eth.c d2 = dma_map_single(ip->dma_dev, (void *)b2, s1, DMA_TO_DEVICE); dma_dev 1453 drivers/net/ethernet/sgi/ioc3-eth.c if (dma_mapping_error(ip->dma_dev, d2)) { dma_dev 1454 drivers/net/ethernet/sgi/ioc3-eth.c dma_unmap_single(ip->dma_dev, d1, len, DMA_TO_DEVICE); dma_dev 1465 drivers/net/ethernet/sgi/ioc3-eth.c d = dma_map_single(ip->dma_dev, skb->data, len, DMA_TO_DEVICE); dma_dev 1466 drivers/net/ethernet/sgi/ioc3-eth.c if (dma_mapping_error(ip->dma_dev, d)) dma_dev 256 drivers/net/ethernet/sun/sunhme.c ((__hp)->dma_map((__hp)->dma_dev, (__ptr), (__size), (__dir))) dma_dev 258 drivers/net/ethernet/sun/sunhme.c ((__hp)->dma_unmap((__hp)->dma_dev, (__addr), (__size), (__dir))) dma_dev 260 drivers/net/ethernet/sun/sunhme.c ((__hp)->dma_sync_for_cpu((__hp)->dma_dev, (__addr), (__size), (__dir))) dma_dev 262 drivers/net/ethernet/sun/sunhme.c ((__hp)->dma_sync_for_device((__hp)->dma_dev, (__addr), (__size), (__dir))) dma_dev 282 drivers/net/ethernet/sun/sunhme.c dma_map_single((__hp)->dma_dev, (__ptr), (__size), (__dir)) dma_dev 284 drivers/net/ethernet/sun/sunhme.c dma_unmap_single((__hp)->dma_dev, (__addr), (__size), (__dir)) dma_dev 286 drivers/net/ethernet/sun/sunhme.c dma_dma_sync_single_for_cpu((__hp)->dma_dev, (__addr), (__size), (__dir)) dma_dev 288 drivers/net/ethernet/sun/sunhme.c dma_dma_sync_single_for_device((__hp)->dma_dev, (__addr), (__size), (__dir)) dma_dev 310 drivers/net/ethernet/sun/sunhme.c pci_map_single((__hp)->dma_dev, (__ptr), (__size), (__dir)) dma_dev 312 drivers/net/ethernet/sun/sunhme.c pci_unmap_single((__hp)->dma_dev, (__addr), (__size), (__dir)) dma_dev 314 drivers/net/ethernet/sun/sunhme.c pci_dma_sync_single_for_cpu((__hp)->dma_dev, (__addr), (__size), (__dir)) dma_dev 316 drivers/net/ethernet/sun/sunhme.c pci_dma_sync_single_for_device((__hp)->dma_dev, (__addr), (__size), (__dir)) dma_dev 1210 drivers/net/ethernet/sun/sunhme.c dma_unmap_single(hp->dma_dev, dma_addr, dma_dev 1230 drivers/net/ethernet/sun/sunhme.c dma_unmap_single(hp->dma_dev, dma_addr, dma_dev 1235 drivers/net/ethernet/sun/sunhme.c dma_unmap_page(hp->dma_dev, dma_addr, dma_dev 1277 drivers/net/ethernet/sun/sunhme.c mapping = dma_map_single(hp->dma_dev, skb->data, RX_BUF_ALLOC_SIZE, dma_dev 1279 drivers/net/ethernet/sun/sunhme.c if (dma_mapping_error(hp->dma_dev, mapping)) { dma_dev 1957 drivers/net/ethernet/sun/sunhme.c dma_unmap_single(hp->dma_dev, dma_addr, dma_len, DMA_TO_DEVICE); dma_dev 1959 drivers/net/ethernet/sun/sunhme.c dma_unmap_page(hp->dma_dev, dma_addr, dma_len, DMA_TO_DEVICE); dma_dev 2039 drivers/net/ethernet/sun/sunhme.c mapping = dma_map_single(hp->dma_dev, new_skb->data, dma_dev 2042 drivers/net/ethernet/sun/sunhme.c if (unlikely(dma_mapping_error(hp->dma_dev, mapping))) { dma_dev 2048 drivers/net/ethernet/sun/sunhme.c dma_unmap_single(hp->dma_dev, dma_addr, RX_BUF_ALLOC_SIZE, DMA_FROM_DEVICE); dma_dev 2067 drivers/net/ethernet/sun/sunhme.c dma_sync_single_for_cpu(hp->dma_dev, dma_addr, len, DMA_FROM_DEVICE); dma_dev 2069 drivers/net/ethernet/sun/sunhme.c dma_sync_single_for_device(hp->dma_dev, dma_addr, len, DMA_FROM_DEVICE); dma_dev 2272 drivers/net/ethernet/sun/sunhme.c dma_unmap_single(hp->dma_dev, first_mapping, first_len, DMA_TO_DEVICE); dma_dev 2282 drivers/net/ethernet/sun/sunhme.c dma_unmap_page(hp->dma_dev, addr, len, DMA_TO_DEVICE); dma_dev 2321 drivers/net/ethernet/sun/sunhme.c mapping = dma_map_single(hp->dma_dev, skb->data, len, DMA_TO_DEVICE); dma_dev 2322 drivers/net/ethernet/sun/sunhme.c if (unlikely(dma_mapping_error(hp->dma_dev, mapping))) dma_dev 2337 drivers/net/ethernet/sun/sunhme.c first_mapping = dma_map_single(hp->dma_dev, skb->data, first_len, dma_dev 2339 drivers/net/ethernet/sun/sunhme.c if (unlikely(dma_mapping_error(hp->dma_dev, first_mapping))) dma_dev 2348 drivers/net/ethernet/sun/sunhme.c mapping = skb_frag_dma_map(hp->dma_dev, this_frag, dma_dev 2350 drivers/net/ethernet/sun/sunhme.c if (unlikely(dma_mapping_error(hp->dma_dev, mapping))) { dma_dev 2743 drivers/net/ethernet/sun/sunhme.c hp->dma_dev = &op->dev; dma_dev 2806 drivers/net/ethernet/sun/sunhme.c hp->happy_block = dma_alloc_coherent(hp->dma_dev, dma_dev 2871 drivers/net/ethernet/sun/sunhme.c dma_free_coherent(hp->dma_dev, dma_dev 3039 drivers/net/ethernet/sun/sunhme.c hp->dma_dev = &pdev->dev; dma_dev 3223 drivers/net/ethernet/sun/sunhme.c dma_free_coherent(hp->dma_dev, PAGE_SIZE, dma_dev 3300 drivers/net/ethernet/sun/sunhme.c dma_free_coherent(hp->dma_dev, dma_dev 413 drivers/net/ethernet/sun/sunhme.h struct device *dma_dev; dma_dev 326 drivers/net/thunderbolt.c struct device *dma_dev = tb_ring_dma_device(ring->ring); dma_dev 346 drivers/net/thunderbolt.c dma_unmap_page(dma_dev, tf->frame.buffer_phy, size, dma_dev 472 drivers/net/thunderbolt.c struct device *dma_dev = tb_ring_dma_device(ring->ring); dma_dev 490 drivers/net/thunderbolt.c dma_addr = dma_map_page(dma_dev, tf->page, 0, dma_dev 492 drivers/net/thunderbolt.c if (dma_mapping_error(dma_dev, dma_addr)) { dma_dev 515 drivers/net/thunderbolt.c struct device *dma_dev = tb_ring_dma_device(ring->ring); dma_dev 527 drivers/net/thunderbolt.c dma_sync_single_for_cpu(dma_dev, tf->frame.buffer_phy, dma_dev 549 drivers/net/thunderbolt.c struct device *dma_dev = tb_ring_dma_device(ring->ring); dma_dev 562 drivers/net/thunderbolt.c dma_addr = dma_map_page(dma_dev, tf->page, 0, TBNET_FRAME_SIZE, dma_dev 564 drivers/net/thunderbolt.c if (dma_mapping_error(dma_dev, dma_addr)) { dma_dev 744 drivers/net/thunderbolt.c struct device *dma_dev = tb_ring_dma_device(net->rx_ring.ring); dma_dev 770 drivers/net/thunderbolt.c dma_unmap_page(dma_dev, frame->buffer_phy, dma_dev 906 drivers/net/thunderbolt.c struct device *dma_dev = tb_ring_dma_device(net->tx_ring.ring); dma_dev 921 drivers/net/thunderbolt.c dma_sync_single_for_device(dma_dev, dma_dev 995 drivers/net/thunderbolt.c dma_sync_single_for_device(dma_dev, frames[i]->frame.buffer_phy, dma_dev 104 drivers/net/wireless/broadcom/b43/bus.c dev->dma_dev = core->dma_dev; dma_dev 203 drivers/net/wireless/broadcom/b43/bus.c dev->dma_dev = sdev->dma_dev; dma_dev 40 drivers/net/wireless/broadcom/b43/bus.h struct device *dma_dev; dma_dev 353 drivers/net/wireless/broadcom/b43/dma.c dmaaddr = dma_map_single(ring->dev->dev->dma_dev, dma_dev 356 drivers/net/wireless/broadcom/b43/dma.c dmaaddr = dma_map_single(ring->dev->dev->dma_dev, dma_dev 368 drivers/net/wireless/broadcom/b43/dma.c dma_unmap_single(ring->dev->dev->dma_dev, dma_dev 371 drivers/net/wireless/broadcom/b43/dma.c dma_unmap_single(ring->dev->dev->dma_dev, dma_dev 381 drivers/net/wireless/broadcom/b43/dma.c dma_sync_single_for_cpu(ring->dev->dev->dma_dev, dma_dev 390 drivers/net/wireless/broadcom/b43/dma.c dma_sync_single_for_device(ring->dev->dev->dma_dev, dma_dev 421 drivers/net/wireless/broadcom/b43/dma.c ring->descbase = dma_alloc_coherent(ring->dev->dev->dma_dev, dma_dev 434 drivers/net/wireless/broadcom/b43/dma.c dma_free_coherent(ring->dev->dev->dma_dev, ring_mem_size, dma_dev 542 drivers/net/wireless/broadcom/b43/dma.c if (unlikely(dma_mapping_error(ring->dev->dev->dma_dev, addr))) dma_dev 898 drivers/net/wireless/broadcom/b43/dma.c dma_test = dma_map_single(dev->dev->dma_dev, dma_dev 913 drivers/net/wireless/broadcom/b43/dma.c dma_test = dma_map_single(dev->dev->dma_dev, dma_dev 927 drivers/net/wireless/broadcom/b43/dma.c dma_unmap_single(dev->dev->dma_dev, dma_dev 1058 drivers/net/wireless/broadcom/b43/dma.c err = dma_set_mask_and_coherent(dev->dev->dma_dev, DMA_BIT_MASK(type)); dma_dev 255 drivers/net/wireless/broadcom/b43legacy/dma.c dmaaddr = dma_map_single(ring->dev->dev->dma_dev, dma_dev 259 drivers/net/wireless/broadcom/b43legacy/dma.c dmaaddr = dma_map_single(ring->dev->dev->dma_dev, dma_dev 273 drivers/net/wireless/broadcom/b43legacy/dma.c dma_unmap_single(ring->dev->dev->dma_dev, dma_dev 277 drivers/net/wireless/broadcom/b43legacy/dma.c dma_unmap_single(ring->dev->dev->dma_dev, dma_dev 289 drivers/net/wireless/broadcom/b43legacy/dma.c dma_sync_single_for_cpu(ring->dev->dev->dma_dev, dma_dev 300 drivers/net/wireless/broadcom/b43legacy/dma.c dma_sync_single_for_device(ring->dev->dev->dma_dev, dma_dev 321 drivers/net/wireless/broadcom/b43legacy/dma.c ring->descbase = dma_alloc_coherent(ring->dev->dev->dma_dev, dma_dev 332 drivers/net/wireless/broadcom/b43legacy/dma.c dma_free_coherent(ring->dev->dev->dma_dev, B43legacy_DMA_RINGMEMSIZE, dma_dev 416 drivers/net/wireless/broadcom/b43legacy/dma.c if (unlikely(dma_mapping_error(ring->dev->dev->dma_dev, addr))) dma_dev 656 drivers/net/wireless/broadcom/b43legacy/dma.c dma_test = dma_map_single(dev->dev->dma_dev, ring->txhdr_cache, dma_dev 670 drivers/net/wireless/broadcom/b43legacy/dma.c dma_test = dma_map_single(dev->dev->dma_dev, dma_dev 680 drivers/net/wireless/broadcom/b43legacy/dma.c dma_unmap_single(dev->dev->dma_dev, dma_test, dma_dev 783 drivers/net/wireless/broadcom/b43legacy/dma.c err = dma_set_mask_and_coherent(dev->dev->dma_dev, DMA_BIT_MASK(type)); dma_dev 590 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c di->dmadev = core->dma_dev; dma_dev 812 drivers/ntb/ntb_transport.c struct device *dma_dev, size_t align) dma_dev 818 drivers/ntb/ntb_transport.c alloc_addr = dma_alloc_coherent(dma_dev, mw->alloc_size, dma_dev 821 drivers/ntb/ntb_transport.c dev_err(dma_dev, "Unable to alloc MW buff of size %zu\n", dma_dev 850 drivers/ntb/ntb_transport.c dma_free_coherent(dma_dev, mw->alloc_size, alloc_addr, dma_addr); dma_dev 783 drivers/ntb/test/ntb_perf.c struct device *dma_dev; dma_dev 791 drivers/ntb/test/ntb_perf.c dma_dev = pthr->dma_chan->device->dev; dma_dev 797 drivers/ntb/test/ntb_perf.c unmap = dmaengine_get_unmap_data(dma_dev, 2, GFP_NOWAIT); dma_dev 802 drivers/ntb/test/ntb_perf.c unmap->addr[0] = dma_map_page(dma_dev, virt_to_page(src), dma_dev 804 drivers/ntb/test/ntb_perf.c if (dma_mapping_error(dma_dev, unmap->addr[0])) { dma_dev 810 drivers/ntb/test/ntb_perf.c unmap->addr[1] = dma_map_page(dma_dev, virt_to_page(dst), dma_dev 812 drivers/ntb/test/ntb_perf.c if (dma_mapping_error(dma_dev, unmap->addr[1])) { dma_dev 1099 drivers/s390/cio/css.c struct gen_pool *cio_gp_dma_create(struct device *dma_dev, int nr_pages) dma_dev 1110 drivers/s390/cio/css.c cpu_addr = dma_alloc_coherent(dma_dev, PAGE_SIZE, &dma_addr, dma_dev 1130 drivers/s390/cio/css.c void cio_gp_dma_destroy(struct gen_pool *gp_dma, struct device *dma_dev) dma_dev 1135 drivers/s390/cio/css.c gen_pool_for_each_chunk(gp_dma, __gp_dma_free_dma, dma_dev); dma_dev 1148 drivers/s390/cio/css.c void *cio_gp_dma_zalloc(struct gen_pool *gp_dma, struct device *dma_dev, dma_dev 1160 drivers/s390/cio/css.c addr = (unsigned long) dma_alloc_coherent(dma_dev, dma_dev 247 drivers/scsi/aha1542.c struct device *dev = cmd->device->host->dma_dev; dma_dev 430 drivers/scsi/aha1542.c acmd->chain_handle = dma_map_single(sh->dma_dev, acmd->chain, dma_dev 432 drivers/scsi/aha1542.c if (dma_mapping_error(sh->dma_dev, acmd->chain_handle)) dma_dev 830 drivers/scsi/aha1542.c struct device *dev = sh->dma_dev; dma_dev 208 drivers/scsi/hosts.c struct device *dma_dev) dma_dev 232 drivers/scsi/hosts.c if (!dma_dev) dma_dev 233 drivers/scsi/hosts.c dma_dev = shost->shost_gendev.parent; dma_dev 235 drivers/scsi/hosts.c shost->dma_dev = dma_dev; dma_dev 260 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c vio_disable_interrupts(vscsi->dma_dev); dma_dev 817 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c rc = vio_enable_interrupts(vscsi->dma_dev); dma_dev 889 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c vio_enable_interrupts(vscsi->dma_dev); dma_dev 1230 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c vio_enable_interrupts(vscsi->dma_dev); dma_dev 1516 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c info = dma_alloc_coherent(&vscsi->dma_dev->dev, sizeof(*info), &token, dma_dev 1595 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c dma_free_coherent(&vscsi->dma_dev->dev, sizeof(*info), info, token); dma_dev 1635 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c cap = dma_alloc_coherent(&vscsi->dma_dev->dev, olen, &token, dma_dev 1647 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c strncpy(cap->name, dev_name(&vscsi->dma_dev->dev), dma_dev 1711 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c dma_free_coherent(&vscsi->dma_dev->dev, olen, cap, token); dma_dev 1944 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c rc = h_send_crq(vscsi->dma_dev->unit_address, dma_dev 2595 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c struct vio_dev *vdev = vscsi->dma_dev; dma_dev 2646 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c vdev = vscsi->dma_dev; dma_dev 2965 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c vio_disable_interrupts(vscsi->dma_dev); dma_dev 3015 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c struct vio_dev *vdev = vscsi->dma_dev; dma_dev 3051 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c dma_unmap_single(&vscsi->dma_dev->dev, vscsi->cmd_q.crq_token, dma_dev 3353 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c vio_enable_interrupts(vscsi->dma_dev); dma_dev 3417 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c vio_enable_interrupts(vscsi->dma_dev); dma_dev 3452 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c vscsi->dma_dev = vdev; dma_dev 3606 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c dev_dbg(&vscsi->dev, "remove (%s)\n", dev_name(&vscsi->dma_dev->dev)); dma_dev 3650 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c return snprintf(buf, PAGE_SIZE, "%x\n", vscsi->dma_dev->unit_address); dma_dev 295 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.h struct vio_dev *dma_dev; dma_dev 1784 drivers/scsi/scsi_lib.c struct device *dev = shost->dma_dev; dma_dev 28 drivers/scsi/scsi_lib_dma.c struct device *dev = cmd->device->host->dma_dev; dma_dev 46 drivers/scsi/scsi_lib_dma.c struct device *dev = cmd->device->host->dma_dev; dma_dev 31 drivers/spi/spi-dw-mid.c if (s->dma_dev != chan->device->dev) dma_dev 40 drivers/spi/spi-dw-mid.c struct pci_dev *dma_dev; dma_dev 49 drivers/spi/spi-dw-mid.c dma_dev = pci_get_device(PCI_VENDOR_ID_INTEL, 0x0827, NULL); dma_dev 50 drivers/spi/spi-dw-mid.c if (!dma_dev) dma_dev 57 drivers/spi/spi-dw-mid.c rx->dma_dev = &dma_dev->dev; dma_dev 64 drivers/spi/spi-dw-mid.c tx->dma_dev = &dma_dev->dev; dma_dev 67 drivers/spi/spi-pxa2xx-pci.c if (dws->dma_dev != chan->device->dev) dma_dev 76 drivers/spi/spi-pxa2xx-pci.c struct pci_dev *dma_dev; dma_dev 81 drivers/spi/spi-pxa2xx-pci.c dma_dev = pci_get_slot(dev->bus, PCI_DEVFN(PCI_SLOT(dev->devfn), 0)); dma_dev 86 drivers/spi/spi-pxa2xx-pci.c slave->dma_dev = &dma_dev->dev; dma_dev 94 drivers/spi/spi-pxa2xx-pci.c slave->dma_dev = &dma_dev->dev; dma_dev 105 drivers/spi/spi-pxa2xx-pci.c struct pci_dev *dma_dev = pci_get_slot(dev->bus, PCI_DEVFN(21, 0)); dma_dev 132 drivers/spi/spi-pxa2xx-pci.c tx->dma_dev = &dma_dev->dev; dma_dev 135 drivers/spi/spi-pxa2xx-pci.c rx->dma_dev = &dma_dev->dev; dma_dev 833 drivers/spi/spi-topcliff-pch.c (param->dma_dev == chan->device->dev)) { dma_dev 845 drivers/spi/spi-topcliff-pch.c struct pci_dev *dma_dev; dma_dev 860 drivers/spi/spi-topcliff-pch.c dma_dev = pci_get_slot(data->board_dat->pdev->bus, dma_dev 865 drivers/spi/spi-topcliff-pch.c param->dma_dev = &dma_dev->dev; dma_dev 880 drivers/spi/spi-topcliff-pch.c param->dma_dev = &dma_dev->dev; dma_dev 499 drivers/ssb/main.c sdev->dma_dev = dev->parent; dma_dev 515 drivers/ssb/main.c sdev->dma_dev = dev; dma_dev 134 drivers/staging/ralink-gdma/ralink-gdma.c void (*init)(struct gdma_dma_dev *dma_dev); dma_dev 155 drivers/staging/ralink-gdma/ralink-gdma.c static inline uint32_t gdma_dma_read(struct gdma_dma_dev *dma_dev, dma_dev 158 drivers/staging/ralink-gdma/ralink-gdma.c return readl(dma_dev->base + reg); dma_dev 161 drivers/staging/ralink-gdma/ralink-gdma.c static inline void gdma_dma_write(struct gdma_dma_dev *dma_dev, dma_dev 164 drivers/staging/ralink-gdma/ralink-gdma.c writel(val, dma_dev->base + reg); dma_dev 185 drivers/staging/ralink-gdma/ralink-gdma.c struct gdma_dma_dev *dma_dev = gdma_dma_chan_get_dev(chan); dma_dev 188 drivers/staging/ralink-gdma/ralink-gdma.c dev_err(dma_dev->ddev.dev, "not support flow controller\n"); dma_dev 195 drivers/staging/ralink-gdma/ralink-gdma.c dev_err(dma_dev->ddev.dev, "only support 4 byte buswidth\n"); dma_dev 204 drivers/staging/ralink-gdma/ralink-gdma.c dev_err(dma_dev->ddev.dev, "only support 4 byte buswidth\n"); dma_dev 212 drivers/staging/ralink-gdma/ralink-gdma.c dev_err(dma_dev->ddev.dev, "direction type %d error\n", dma_dev 223 drivers/staging/ralink-gdma/ralink-gdma.c struct gdma_dma_dev *dma_dev = gdma_dma_chan_get_dev(chan); dma_dev 230 drivers/staging/ralink-gdma/ralink-gdma.c clear_bit(chan->id, &dma_dev->chan_issued); dma_dev 238 drivers/staging/ralink-gdma/ralink-gdma.c while (gdma_dma_read(dma_dev, GDMA_REG_CTRL0(chan->id)) & dma_dev 241 drivers/staging/ralink-gdma/ralink-gdma.c dev_err(dma_dev->ddev.dev, "chan %d wait timeout\n", dma_dev 244 drivers/staging/ralink-gdma/ralink-gdma.c gdma_dma_write(dma_dev, GDMA_REG_CTRL0(chan->id), 0); dma_dev 252 drivers/staging/ralink-gdma/ralink-gdma.c dev_dbg(dma_dev->ddev.dev, "terminate chan %d loops %d\n", dma_dev 258 drivers/staging/ralink-gdma/ralink-gdma.c static void rt305x_dump_reg(struct gdma_dma_dev *dma_dev, int id) dma_dev 260 drivers/staging/ralink-gdma/ralink-gdma.c dev_dbg(dma_dev->ddev.dev, "chan %d, src %08x, dst %08x, ctr0 %08x, ctr1 %08x, intr %08x, signal %08x\n", dma_dev 262 drivers/staging/ralink-gdma/ralink-gdma.c gdma_dma_read(dma_dev, GDMA_REG_SRC_ADDR(id)), dma_dev 263 drivers/staging/ralink-gdma/ralink-gdma.c gdma_dma_read(dma_dev, GDMA_REG_DST_ADDR(id)), dma_dev 264 drivers/staging/ralink-gdma/ralink-gdma.c gdma_dma_read(dma_dev, GDMA_REG_CTRL0(id)), dma_dev 265 drivers/staging/ralink-gdma/ralink-gdma.c gdma_dma_read(dma_dev, GDMA_REG_CTRL1(id)), dma_dev 266 drivers/staging/ralink-gdma/ralink-gdma.c gdma_dma_read(dma_dev, GDMA_RT305X_STATUS_INT), dma_dev 267 drivers/staging/ralink-gdma/ralink-gdma.c gdma_dma_read(dma_dev, GDMA_RT305X_STATUS_SIGNAL)); dma_dev 272 drivers/staging/ralink-gdma/ralink-gdma.c struct gdma_dma_dev *dma_dev = gdma_dma_chan_get_dev(chan); dma_dev 278 drivers/staging/ralink-gdma/ralink-gdma.c ctrl0 = gdma_dma_read(dma_dev, GDMA_REG_CTRL0(chan->id)); dma_dev 280 drivers/staging/ralink-gdma/ralink-gdma.c dev_err(dma_dev->ddev.dev, "chan %d is start(%08x).\n", dma_dev 282 drivers/staging/ralink-gdma/ralink-gdma.c rt305x_dump_reg(dma_dev, chan->id); dma_dev 310 drivers/staging/ralink-gdma/ralink-gdma.c dev_err(dma_dev->ddev.dev, "direction type %d error\n", dma_dev 321 drivers/staging/ralink-gdma/ralink-gdma.c gdma_dma_write(dma_dev, GDMA_REG_SRC_ADDR(chan->id), src_addr); dma_dev 322 drivers/staging/ralink-gdma/ralink-gdma.c gdma_dma_write(dma_dev, GDMA_REG_DST_ADDR(chan->id), dst_addr); dma_dev 323 drivers/staging/ralink-gdma/ralink-gdma.c gdma_dma_write(dma_dev, GDMA_REG_CTRL1(chan->id), ctrl1); dma_dev 327 drivers/staging/ralink-gdma/ralink-gdma.c gdma_dma_write(dma_dev, GDMA_REG_CTRL0(chan->id), ctrl0); dma_dev 332 drivers/staging/ralink-gdma/ralink-gdma.c static void rt3883_dump_reg(struct gdma_dma_dev *dma_dev, int id) dma_dev 334 drivers/staging/ralink-gdma/ralink-gdma.c dev_dbg(dma_dev->ddev.dev, "chan %d, src %08x, dst %08x, ctr0 %08x, ctr1 %08x, unmask %08x, done %08x, req %08x, ack %08x, fin %08x\n", dma_dev 336 drivers/staging/ralink-gdma/ralink-gdma.c gdma_dma_read(dma_dev, GDMA_REG_SRC_ADDR(id)), dma_dev 337 drivers/staging/ralink-gdma/ralink-gdma.c gdma_dma_read(dma_dev, GDMA_REG_DST_ADDR(id)), dma_dev 338 drivers/staging/ralink-gdma/ralink-gdma.c gdma_dma_read(dma_dev, GDMA_REG_CTRL0(id)), dma_dev 339 drivers/staging/ralink-gdma/ralink-gdma.c gdma_dma_read(dma_dev, GDMA_REG_CTRL1(id)), dma_dev 340 drivers/staging/ralink-gdma/ralink-gdma.c gdma_dma_read(dma_dev, GDMA_REG_UNMASK_INT), dma_dev 341 drivers/staging/ralink-gdma/ralink-gdma.c gdma_dma_read(dma_dev, GDMA_REG_DONE_INT), dma_dev 342 drivers/staging/ralink-gdma/ralink-gdma.c gdma_dma_read(dma_dev, GDMA_REG_REQSTS), dma_dev 343 drivers/staging/ralink-gdma/ralink-gdma.c gdma_dma_read(dma_dev, GDMA_REG_ACKSTS), dma_dev 344 drivers/staging/ralink-gdma/ralink-gdma.c gdma_dma_read(dma_dev, GDMA_REG_FINSTS)); dma_dev 349 drivers/staging/ralink-gdma/ralink-gdma.c struct gdma_dma_dev *dma_dev = gdma_dma_chan_get_dev(chan); dma_dev 355 drivers/staging/ralink-gdma/ralink-gdma.c ctrl0 = gdma_dma_read(dma_dev, GDMA_REG_CTRL0(chan->id)); dma_dev 357 drivers/staging/ralink-gdma/ralink-gdma.c dev_err(dma_dev->ddev.dev, "chan %d is start(%08x).\n", dma_dev 359 drivers/staging/ralink-gdma/ralink-gdma.c rt3883_dump_reg(dma_dev, chan->id); dma_dev 385 drivers/staging/ralink-gdma/ralink-gdma.c dev_err(dma_dev->ddev.dev, "direction type %d error\n", dma_dev 396 drivers/staging/ralink-gdma/ralink-gdma.c gdma_dma_write(dma_dev, GDMA_REG_SRC_ADDR(chan->id), src_addr); dma_dev 397 drivers/staging/ralink-gdma/ralink-gdma.c gdma_dma_write(dma_dev, GDMA_REG_DST_ADDR(chan->id), dst_addr); dma_dev 398 drivers/staging/ralink-gdma/ralink-gdma.c gdma_dma_write(dma_dev, GDMA_REG_CTRL1(chan->id), ctrl1); dma_dev 402 drivers/staging/ralink-gdma/ralink-gdma.c gdma_dma_write(dma_dev, GDMA_REG_CTRL0(chan->id), ctrl0); dma_dev 407 drivers/staging/ralink-gdma/ralink-gdma.c static inline int gdma_start_transfer(struct gdma_dma_dev *dma_dev, dma_dev 410 drivers/staging/ralink-gdma/ralink-gdma.c return dma_dev->data->start_transfer(chan); dma_dev 428 drivers/staging/ralink-gdma/ralink-gdma.c static void gdma_dma_chan_irq(struct gdma_dma_dev *dma_dev, dma_dev 455 drivers/staging/ralink-gdma/ralink-gdma.c dev_dbg(dma_dev->ddev.dev, "chan %d no desc to complete\n", dma_dev 459 drivers/staging/ralink-gdma/ralink-gdma.c set_bit(chan->id, &dma_dev->chan_issued); dma_dev 465 drivers/staging/ralink-gdma/ralink-gdma.c struct gdma_dma_dev *dma_dev = devid; dma_dev 469 drivers/staging/ralink-gdma/ralink-gdma.c done_reg = dma_dev->data->done_int_reg; dma_dev 470 drivers/staging/ralink-gdma/ralink-gdma.c done = gdma_dma_read(dma_dev, done_reg); dma_dev 475 drivers/staging/ralink-gdma/ralink-gdma.c gdma_dma_write(dma_dev, done_reg, done); dma_dev 480 drivers/staging/ralink-gdma/ralink-gdma.c gdma_dma_chan_irq(dma_dev, &dma_dev->chan[i]); dma_dev 481 drivers/staging/ralink-gdma/ralink-gdma.c atomic_dec(&dma_dev->cnt); dma_dev 488 drivers/staging/ralink-gdma/ralink-gdma.c if (dma_dev->chan_issued) dma_dev 489 drivers/staging/ralink-gdma/ralink-gdma.c tasklet_schedule(&dma_dev->task); dma_dev 497 drivers/staging/ralink-gdma/ralink-gdma.c struct gdma_dma_dev *dma_dev = gdma_dma_chan_get_dev(chan); dma_dev 503 drivers/staging/ralink-gdma/ralink-gdma.c set_bit(chan->id, &dma_dev->chan_issued); dma_dev 504 drivers/staging/ralink-gdma/ralink-gdma.c tasklet_schedule(&dma_dev->task); dma_dev 506 drivers/staging/ralink-gdma/ralink-gdma.c dev_dbg(dma_dev->ddev.dev, "chan %d no desc to issue\n", dma_dev 706 drivers/staging/ralink-gdma/ralink-gdma.c struct gdma_dma_dev *dma_dev = (struct gdma_dma_dev *)arg; dma_dev 713 drivers/staging/ralink-gdma/ralink-gdma.c chan_mask = dma_dev->data->chancnt - 1; dma_dev 721 drivers/staging/ralink-gdma/ralink-gdma.c if (atomic_read(&dma_dev->cnt) >= 2) { dma_dev 726 drivers/staging/ralink-gdma/ralink-gdma.c if (test_and_clear_bit(i, &dma_dev->chan_issued)) { dma_dev 727 drivers/staging/ralink-gdma/ralink-gdma.c chan = &dma_dev->chan[i]; dma_dev 729 drivers/staging/ralink-gdma/ralink-gdma.c atomic_inc(&dma_dev->cnt); dma_dev 730 drivers/staging/ralink-gdma/ralink-gdma.c gdma_start_transfer(dma_dev, chan); dma_dev 732 drivers/staging/ralink-gdma/ralink-gdma.c dev_dbg(dma_dev->ddev.dev, dma_dev 736 drivers/staging/ralink-gdma/ralink-gdma.c if (!dma_dev->chan_issued) dma_dev 744 drivers/staging/ralink-gdma/ralink-gdma.c static void rt305x_gdma_init(struct gdma_dma_dev *dma_dev) dma_dev 749 drivers/staging/ralink-gdma/ralink-gdma.c gdma_dma_write(dma_dev, GDMA_RT305X_GCT, GDMA_REG_GCT_ARBIT_RR); dma_dev 751 drivers/staging/ralink-gdma/ralink-gdma.c gct = gdma_dma_read(dma_dev, GDMA_RT305X_GCT); dma_dev 752 drivers/staging/ralink-gdma/ralink-gdma.c dev_info(dma_dev->ddev.dev, "revision: %d, channels: %d\n", dma_dev 758 drivers/staging/ralink-gdma/ralink-gdma.c static void rt3883_gdma_init(struct gdma_dma_dev *dma_dev) dma_dev 763 drivers/staging/ralink-gdma/ralink-gdma.c gdma_dma_write(dma_dev, GDMA_REG_GCT, GDMA_REG_GCT_ARBIT_RR); dma_dev 765 drivers/staging/ralink-gdma/ralink-gdma.c gct = gdma_dma_read(dma_dev, GDMA_REG_GCT); dma_dev 766 drivers/staging/ralink-gdma/ralink-gdma.c dev_info(dma_dev->ddev.dev, "revision: %d, channels: %d\n", dma_dev 796 drivers/staging/ralink-gdma/ralink-gdma.c struct gdma_dma_dev *dma_dev; dma_dev 814 drivers/staging/ralink-gdma/ralink-gdma.c dma_dev = devm_kzalloc(&pdev->dev, dma_dev 815 drivers/staging/ralink-gdma/ralink-gdma.c struct_size(dma_dev, chan, data->chancnt), dma_dev 817 drivers/staging/ralink-gdma/ralink-gdma.c if (!dma_dev) dma_dev 819 drivers/staging/ralink-gdma/ralink-gdma.c dma_dev->data = data; dma_dev 825 drivers/staging/ralink-gdma/ralink-gdma.c dma_dev->base = base; dma_dev 826 drivers/staging/ralink-gdma/ralink-gdma.c tasklet_init(&dma_dev->task, gdma_dma_tasklet, (unsigned long)dma_dev); dma_dev 832 drivers/staging/ralink-gdma/ralink-gdma.c 0, dev_name(&pdev->dev), dma_dev); dma_dev 840 drivers/staging/ralink-gdma/ralink-gdma.c dd = &dma_dev->ddev; dma_dev 859 drivers/staging/ralink-gdma/ralink-gdma.c dd->dev->dma_parms = &dma_dev->dma_parms; dma_dev 864 drivers/staging/ralink-gdma/ralink-gdma.c chan = &dma_dev->chan[i]; dma_dev 871 drivers/staging/ralink-gdma/ralink-gdma.c data->init(dma_dev); dma_dev 880 drivers/staging/ralink-gdma/ralink-gdma.c of_dma_xlate_by_chan_id, dma_dev); dma_dev 886 drivers/staging/ralink-gdma/ralink-gdma.c platform_set_drvdata(pdev, dma_dev); dma_dev 897 drivers/staging/ralink-gdma/ralink-gdma.c struct gdma_dma_dev *dma_dev = platform_get_drvdata(pdev); dma_dev 899 drivers/staging/ralink-gdma/ralink-gdma.c tasklet_kill(&dma_dev->task); dma_dev 901 drivers/staging/ralink-gdma/ralink-gdma.c dma_async_device_unregister(&dma_dev->ddev); dma_dev 125 drivers/tty/serial/8250/8250_lpss.c struct pci_dev *dma_dev = pci_get_slot(pdev->bus, dma_devfn); dma_dev 144 drivers/tty/serial/8250/8250_lpss.c param->dma_dev = &dma_dev->dev; dma_dev 197 drivers/tty/serial/8250/8250_lpss.c param->dma_dev = &pdev->dev; dma_dev 210 drivers/tty/serial/8250/8250_lpss.c if (!param->dma_dev) dma_dev 248 drivers/tty/serial/8250/8250_lpss.c if (dws->dma_dev != chan->device->dev) dma_dev 261 drivers/tty/serial/8250/8250_lpss.c if (!lpss->dma_param.dma_dev) dma_dev 106 drivers/tty/serial/8250/8250_men_mcb.c data[i].uart.port.dev = mdev->dma_dev; dma_dev 45 drivers/tty/serial/8250/8250_mid.c struct pci_dev *dma_dev; dma_dev 71 drivers/tty/serial/8250/8250_mid.c mid->dma_dev = pci_get_slot(pdev->bus, dma_dev 85 drivers/tty/serial/8250/8250_mid.c chip = pci_get_drvdata(mid->dma_dev); dma_dev 121 drivers/tty/serial/8250/8250_mid.c mid->dma_dev = pci_get_slot(pdev->bus, PCI_DEVFN(5, 0)); dma_dev 184 drivers/tty/serial/8250/8250_mid.c mid->dma_dev = pdev; dma_dev 192 drivers/tty/serial/8250/8250_mid.c if (!mid->dma_dev) dma_dev 239 drivers/tty/serial/8250/8250_mid.c if (s->dma_dev != chan->device->dev || s->chan_id != chan->chan_id) dma_dev 253 drivers/tty/serial/8250/8250_mid.c if (!mid->dma_dev) dma_dev 270 drivers/tty/serial/8250/8250_mid.c rx_param->dma_dev = &mid->dma_dev->dev; dma_dev 271 drivers/tty/serial/8250/8250_mid.c tx_param->dma_dev = &mid->dma_dev->dev; dma_dev 602 drivers/tty/serial/amba-pl011.c struct dma_device *dma_dev = chan->device; dma_dev 646 drivers/tty/serial/amba-pl011.c if (dma_map_sg(dma_dev->dev, &dmatx->sg, 1, DMA_TO_DEVICE) != 1) { dma_dev 655 drivers/tty/serial/amba-pl011.c dma_unmap_sg(dma_dev->dev, &dmatx->sg, 1, DMA_TO_DEVICE); dma_dev 673 drivers/tty/serial/amba-pl011.c dma_dev->device_issue_pending(chan); dma_dev 703 drivers/tty/serial/pch_uart.c if ((chan->chan_id == param->chan_id) && (param->dma_dev == dma_dev 716 drivers/tty/serial/pch_uart.c struct pci_dev *dma_dev; dma_dev 724 drivers/tty/serial/pch_uart.c dma_dev = pci_get_slot(priv->pdev->bus, dma_dev 729 drivers/tty/serial/pch_uart.c param->dma_dev = &dma_dev->dev; dma_dev 743 drivers/tty/serial/pch_uart.c param->dma_dev = &dma_dev->dev; dma_dev 351 drivers/usb/host/bcma-hcd.c if (dma_set_mask_and_coherent(dev->dma_dev, DMA_BIT_MASK(32))) dma_dev 165 drivers/usb/host/ssb-hcd.c if (dma_set_mask_and_coherent(dev->dma_dev, DMA_BIT_MASK(32))) dma_dev 187 drivers/usb/renesas_usbhs/mod_gadget.c static int usbhsg_dma_map_ctrl(struct device *dma_dev, struct usbhs_pkt *pkt, dma_dev 202 drivers/usb/renesas_usbhs/mod_gadget.c ret = usb_gadget_map_request_by_dev(dma_dev, req, dir); dma_dev 208 drivers/usb/renesas_usbhs/mod_gadget.c usb_gadget_unmap_request_by_dev(dma_dev, req, dir); dma_dev 917 drivers/usb/renesas_usbhs/mod_host.c static int usbhsh_dma_map_ctrl(struct device *dma_dev, struct usbhs_pkt *pkt, dma_dev 672 drivers/usb/renesas_usbhs/pipe.c int (*dma_map_ctrl)(struct device *dma_dev, dma_dev 41 drivers/usb/renesas_usbhs/pipe.h int (*dma_map_ctrl)(struct device *dma_dev, struct usbhs_pkt *pkt, dma_dev 79 drivers/usb/renesas_usbhs/pipe.h int (*dma_map_ctrl)(struct device *dma_dev, dma_dev 1539 drivers/video/fbdev/mx3fb.c mx3fb_pdata->dma_dev == chan->device->dev; dma_dev 36 drivers/xen/gntdev-common.h struct device *dma_dev; dma_dev 73 drivers/xen/gntdev-common.h struct device *dma_dev; dma_dev 501 drivers/xen/gntdev-dmabuf.c args.dev = priv->dma_dev; dma_dev 825 drivers/xen/gntdev-dmabuf.c priv->dma_dev, op.fd, dma_dev 103 drivers/xen/gntdev.c args.dev = map->dma_dev; dma_dev 169 drivers/xen/gntdev.c add->dma_dev = priv->dma_dev; dma_dev 171 drivers/xen/gntdev.c args.dev = priv->dma_dev; dma_dev 625 drivers/xen/gntdev.c priv->dma_dev = gntdev_miscdev.this_device; dma_dev 626 drivers/xen/gntdev.c dma_coerce_mask_and_coherent(priv->dma_dev, DMA_BIT_MASK(64)); dma_dev 272 include/linux/bcma/bcma.h struct device *dma_dev; dma_dev 147 include/linux/iio/buffer-dma.h struct device *dma_dev, const struct iio_dma_buffer_ops *ops); dma_dev 76 include/linux/mcb.h struct device *dma_dev; dma_dev 18 include/linux/pch_dma.h struct device *dma_dev; dma_dev 28 include/linux/platform_data/dma-atmel.h struct device *dma_dev; dma_dev 27 include/linux/platform_data/dma-dw.h struct device *dma_dev; dma_dev 14 include/linux/platform_data/dma-hsu.h struct device *dma_dev; dma_dev 38 include/linux/platform_data/media/camera-mx3.h struct device *dma_dev; dma_dev 43 include/linux/platform_data/video-mx3fb.h struct device *dma_dev; dma_dev 107 include/linux/shdma-base.h struct dma_device dma_dev; dma_dev 114 include/linux/shdma-base.h i < (d)->dma_dev.chancnt; c = (d)->schan[++i]) dma_dev 273 include/linux/ssb/ssb.h struct device *dev, *dma_dev; dma_dev 692 include/scsi/scsi_host.h struct device *dma_dev; dma_dev 137 include/sound/dmaengine_pcm.h struct device *dma_dev; dma_dev 375 sound/pci/trident/trident.h struct snd_dma_device dma_dev; dma_dev 313 sound/soc/intel/baytrail/sst-baytrail-dsp.c ret = dma_coerce_mask_and_coherent(sst->dma_dev, DMA_BIT_MASK(32)); dma_dev 328 sound/soc/intel/baytrail/sst-baytrail-pcm.c pdata->dma_dev, dma_dev 102 sound/soc/intel/common/sst-acpi.c sst_pdata->dma_dev = dev; dma_dev 265 sound/soc/intel/common/sst-dsp-priv.h struct device *dma_dev; dma_dev 203 sound/soc/intel/common/sst-dsp.h struct device *dma_dev; dma_dev 214 sound/soc/intel/common/sst-firmware.c return chan->device->dev == dsp->dma_dev; dma_dev 299 sound/soc/intel/common/sst-firmware.c dma->chip = dw_probe(sst->dma_dev, &mem, sst_pdata->irq); dma_dev 348 sound/soc/intel/common/sst-firmware.c sst_fw->dma_buf = dma_alloc_coherent(dsp->dma_dev, sst_fw->size, dma_dev 385 sound/soc/intel/common/sst-firmware.c dma_free_coherent(dsp->dma_dev, sst_fw->size, dma_dev 456 sound/soc/intel/common/sst-firmware.c dma_free_coherent(dsp->dma_dev, sst_fw->size, sst_fw->dma_buf, dma_dev 936 sound/soc/intel/common/sst-firmware.c context->buffer = dma_alloc_coherent(dsp->dma_dev, dma_dev 1007 sound/soc/intel/common/sst-firmware.c dma_free_coherent(dsp->dma_dev, module->persistent_size, dma_dev 1218 sound/soc/intel/common/sst-firmware.c sst->dma_dev = pdata->dma_dev; dma_dev 618 sound/soc/intel/haswell/sst-haswell-dsp.c dev = sst->dma_dev; dma_dev 2142 sound/soc/intel/haswell/sst-haswell-ipc.c hsw->dx_context = dma_alloc_coherent(hsw->dsp->dma_dev, dma_dev 2200 sound/soc/intel/haswell/sst-haswell-ipc.c dma_free_coherent(hsw->dsp->dma_dev, SST_HSW_DX_CONTEXT_SIZE, dma_dev 2217 sound/soc/intel/haswell/sst-haswell-ipc.c dma_free_coherent(hsw->dsp->dma_dev, SST_HSW_DX_CONTEXT_SIZE, dma_dev 939 sound/soc/intel/haswell/sst-haswell-pcm.c struct device *dev = pdata->dma_dev; dma_dev 1044 sound/soc/intel/haswell/sst-haswell-pcm.c struct device *dma_dev, *dev; dma_dev 1051 sound/soc/intel/haswell/sst-haswell-pcm.c dma_dev = pdata->dma_dev; dma_dev 1064 sound/soc/intel/haswell/sst-haswell-pcm.c ret = snd_dma_alloc_pages(SNDRV_DMA_TYPE_DEV, dma_dev, dma_dev 1073 sound/soc/intel/haswell/sst-haswell-pcm.c ret = snd_dma_alloc_pages(SNDRV_DMA_TYPE_DEV, dma_dev, dma_dev 17 sound/soc/samsung/dma.h struct device *dma_dev); dma_dev 19 sound/soc/samsung/dmaengine.c struct device *dma_dev) dma_dev 29 sound/soc/samsung/dmaengine.c pcm_conf->dma_dev = dma_dev; dma_dev 118 sound/soc/soc-generic-dmaengine-pcm.c struct device *dma_dev = dmaengine_dma_dev(pcm, substream); dma_dev 141 sound/soc/soc-generic-dmaengine-pcm.c hw.period_bytes_max = dma_get_max_seg_size(dma_dev); dma_dev 416 sound/soc/soc-generic-dmaengine-pcm.c !(config && config->dma_dev && config->dma_dev->of_node))) dma_dev 419 sound/soc/soc-generic-dmaengine-pcm.c if (config && config->dma_dev) { dma_dev 427 sound/soc/soc-generic-dmaengine-pcm.c dev_name(config->dma_dev)); dma_dev 428 sound/soc/soc-generic-dmaengine-pcm.c dev = config->dma_dev; dma_dev 56 sound/soc/tegra/tegra_pcm.c config->dma_dev = dev->parent;