dma_dev           101 arch/arm/mach-imx/devices/platform-ipu-core.c 		copied_pdata->dma_dev = &imx_ipu_coredev->dev;
dma_dev           122 arch/arm/mach-imx/devices/platform-ipu-core.c 	pdata->dma_dev = &imx_ipu_coredev->dev;
dma_dev           127 arch/mips/alchemy/common/dma.c static const struct dma_dev dma_dev_table_bank2[DMA_NUM_DEV_BANK2] = {
dma_dev           168 arch/mips/alchemy/common/dma.c 	const struct dma_dev *dev;
dma_dev           165 arch/powerpc/platforms/512x/mpc512x_lpbfifo.c 	struct dma_device *dma_dev = NULL;
dma_dev           226 arch/powerpc/platforms/512x/mpc512x_lpbfifo.c 	dma_dev = lpbfifo.chan->device;
dma_dev           243 arch/powerpc/platforms/512x/mpc512x_lpbfifo.c 	if (dma_dev->device_config(lpbfifo.chan, &dma_conf)) {
dma_dev           250 arch/powerpc/platforms/512x/mpc512x_lpbfifo.c 	sg_dma_address(&sg) = dma_map_single(dma_dev->dev,
dma_dev           252 arch/powerpc/platforms/512x/mpc512x_lpbfifo.c 	if (dma_mapping_error(dma_dev->dev, sg_dma_address(&sg)))
dma_dev           327 arch/powerpc/platforms/512x/mpc512x_lpbfifo.c 	dma_unmap_single(dma_dev->dev, sg_dma_address(&sg),
dma_dev           501 arch/powerpc/platforms/512x/mpc512x_lpbfifo.c 	struct dma_device *dma_dev = lpbfifo.chan->device;
dma_dev           509 arch/powerpc/platforms/512x/mpc512x_lpbfifo.c 	dma_dev->device_terminate_all(lpbfifo.chan);
dma_dev           366 arch/s390/include/asm/cio.h void *cio_gp_dma_zalloc(struct gen_pool *gp_dma, struct device *dma_dev,
dma_dev           369 arch/s390/include/asm/cio.h void cio_gp_dma_destroy(struct gen_pool *gp_dma, struct device *dma_dev);
dma_dev           370 arch/s390/include/asm/cio.h struct gen_pool *cio_gp_dma_create(struct device *dma_dev, int nr_pages);
dma_dev          1399 drivers/acpi/scan.c 	struct device *dma_dev = dev;
dma_dev          1408 drivers/acpi/scan.c 		adev = ACPI_COMPANION(dma_dev);
dma_dev          1412 drivers/acpi/scan.c 		dma_dev = dma_dev->parent;
dma_dev          1413 drivers/acpi/scan.c 	} while (dma_dev);
dma_dev          1415 drivers/acpi/scan.c 	if (!dma_dev)
dma_dev          1428 drivers/acpi/scan.c 				dev_warn(dma_dev, "Can't handle multiple windows with different offsets\n");
dma_dev          1442 drivers/acpi/scan.c 			dev_dbg(dma_dev, "Invalid DMA regions configuration\n");
dma_dev           884 drivers/ata/pata_octeon_cf.c 			struct platform_device *dma_dev;
dma_dev           885 drivers/ata/pata_octeon_cf.c 			dma_dev = of_find_device_by_node(dma_node);
dma_dev           886 drivers/ata/pata_octeon_cf.c 			if (dma_dev) {
dma_dev           889 drivers/ata/pata_octeon_cf.c 				res_dma = platform_get_resource(dma_dev, IORESOURCE_MEM, 0);
dma_dev           902 drivers/ata/pata_octeon_cf.c 				i = platform_get_irq(dma_dev, 0);
dma_dev           207 drivers/ata/sata_dwc_460ex.c 	if (dws->dma_dev != chan->device->dev)
dma_dev           220 drivers/ata/sata_dwc_460ex.c 	dws->dma_dev = hsdev->dev;
dma_dev           248 drivers/bcma/main.c 		core->dma_dev = bus->dev;
dma_dev           253 drivers/bcma/main.c 			core->dma_dev = bus->dev;
dma_dev           256 drivers/bcma/main.c 			core->dma_dev = &core->dev;
dma_dev           132 drivers/bus/fsl-mc/fsl-mc-bus.c 	struct device *dma_dev = dev;
dma_dev           134 drivers/bus/fsl-mc/fsl-mc-bus.c 	while (dev_is_fsl_mc(dma_dev))
dma_dev           135 drivers/bus/fsl-mc/fsl-mc-bus.c 		dma_dev = dma_dev->parent;
dma_dev           137 drivers/bus/fsl-mc/fsl-mc-bus.c 	return of_dma_configure(dev, dma_dev->of_node, 0);
dma_dev          2378 drivers/crypto/atmel-aes.c 	if (sl && sl->dma_dev == chan->device->dev) {
dma_dev          2610 drivers/crypto/atmel-sha.c 	if (sl && sl->dma_dev == chan->device->dev) {
dma_dev           709 drivers/crypto/atmel-tdes.c 	if (sl && sl->dma_dev == chan->device->dev) {
dma_dev           391 drivers/crypto/ccp/ccp-dev.h 	struct dma_device dma_dev;
dma_dev           472 drivers/crypto/ccp/ccp-dev.h 	struct device *dma_dev;
dma_dev           639 drivers/crypto/ccp/ccp-dmaengine.c 	struct dma_device *dma_dev = &ccp->dma_dev;
dma_dev           685 drivers/crypto/ccp/ccp-dmaengine.c 	dma_dev->dev = ccp->dev;
dma_dev           686 drivers/crypto/ccp/ccp-dmaengine.c 	dma_dev->src_addr_widths = CCP_DMA_WIDTH(dma_get_mask(ccp->dev));
dma_dev           687 drivers/crypto/ccp/ccp-dmaengine.c 	dma_dev->dst_addr_widths = CCP_DMA_WIDTH(dma_get_mask(ccp->dev));
dma_dev           688 drivers/crypto/ccp/ccp-dmaengine.c 	dma_dev->directions = DMA_MEM_TO_MEM;
dma_dev           689 drivers/crypto/ccp/ccp-dmaengine.c 	dma_dev->residue_granularity = DMA_RESIDUE_GRANULARITY_DESCRIPTOR;
dma_dev           690 drivers/crypto/ccp/ccp-dmaengine.c 	dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask);
dma_dev           691 drivers/crypto/ccp/ccp-dmaengine.c 	dma_cap_set(DMA_INTERRUPT, dma_dev->cap_mask);
dma_dev           700 drivers/crypto/ccp/ccp-dmaengine.c 		dma_cap_set(DMA_PRIVATE, dma_dev->cap_mask);
dma_dev           702 drivers/crypto/ccp/ccp-dmaengine.c 	INIT_LIST_HEAD(&dma_dev->channels);
dma_dev           718 drivers/crypto/ccp/ccp-dmaengine.c 		dma_chan->device = dma_dev;
dma_dev           721 drivers/crypto/ccp/ccp-dmaengine.c 		list_add_tail(&dma_chan->device_node, &dma_dev->channels);
dma_dev           724 drivers/crypto/ccp/ccp-dmaengine.c 	dma_dev->device_free_chan_resources = ccp_free_chan_resources;
dma_dev           725 drivers/crypto/ccp/ccp-dmaengine.c 	dma_dev->device_prep_dma_memcpy = ccp_prep_dma_memcpy;
dma_dev           726 drivers/crypto/ccp/ccp-dmaengine.c 	dma_dev->device_prep_dma_interrupt = ccp_prep_dma_interrupt;
dma_dev           727 drivers/crypto/ccp/ccp-dmaengine.c 	dma_dev->device_issue_pending = ccp_issue_pending;
dma_dev           728 drivers/crypto/ccp/ccp-dmaengine.c 	dma_dev->device_tx_status = ccp_tx_status;
dma_dev           729 drivers/crypto/ccp/ccp-dmaengine.c 	dma_dev->device_pause = ccp_pause;
dma_dev           730 drivers/crypto/ccp/ccp-dmaengine.c 	dma_dev->device_resume = ccp_resume;
dma_dev           731 drivers/crypto/ccp/ccp-dmaengine.c 	dma_dev->device_terminate_all = ccp_terminate_all;
dma_dev           733 drivers/crypto/ccp/ccp-dmaengine.c 	ret = dma_async_device_register(dma_dev);
dma_dev           750 drivers/crypto/ccp/ccp-dmaengine.c 	struct dma_device *dma_dev = &ccp->dma_dev;
dma_dev           755 drivers/crypto/ccp/ccp-dmaengine.c 	dma_async_device_unregister(dma_dev);
dma_dev            66 drivers/crypto/ccp/ccp-ops.c 		dma_unmap_sg(wa->dma_dev, wa->dma_sg, wa->nents, wa->dma_dir);
dma_dev            95 drivers/crypto/ccp/ccp-ops.c 	wa->dma_dev = dev;
dma_dev           794 drivers/dma/altera-msgdma.c 	struct dma_device *dma_dev;
dma_dev           842 drivers/dma/altera-msgdma.c 	dma_dev = &mdev->dmadev;
dma_dev           845 drivers/dma/altera-msgdma.c 	dma_cap_zero(dma_dev->cap_mask);
dma_dev           846 drivers/dma/altera-msgdma.c 	dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask);
dma_dev           847 drivers/dma/altera-msgdma.c 	dma_cap_set(DMA_SLAVE, dma_dev->cap_mask);
dma_dev           849 drivers/dma/altera-msgdma.c 	dma_dev->src_addr_widths = BIT(DMA_SLAVE_BUSWIDTH_4_BYTES);
dma_dev           850 drivers/dma/altera-msgdma.c 	dma_dev->dst_addr_widths = BIT(DMA_SLAVE_BUSWIDTH_4_BYTES);
dma_dev           851 drivers/dma/altera-msgdma.c 	dma_dev->directions = BIT(DMA_MEM_TO_DEV) | BIT(DMA_DEV_TO_MEM) |
dma_dev           853 drivers/dma/altera-msgdma.c 	dma_dev->residue_granularity = DMA_RESIDUE_GRANULARITY_DESCRIPTOR;
dma_dev           856 drivers/dma/altera-msgdma.c 	INIT_LIST_HEAD(&dma_dev->channels);
dma_dev           859 drivers/dma/altera-msgdma.c 	dma_dev->device_tx_status = dma_cookie_status;
dma_dev           860 drivers/dma/altera-msgdma.c 	dma_dev->device_issue_pending = msgdma_issue_pending;
dma_dev           861 drivers/dma/altera-msgdma.c 	dma_dev->dev = &pdev->dev;
dma_dev           863 drivers/dma/altera-msgdma.c 	dma_dev->copy_align = DMAENGINE_ALIGN_4_BYTES;
dma_dev           864 drivers/dma/altera-msgdma.c 	dma_dev->device_prep_dma_memcpy = msgdma_prep_memcpy;
dma_dev           865 drivers/dma/altera-msgdma.c 	dma_dev->device_prep_slave_sg = msgdma_prep_slave_sg;
dma_dev           866 drivers/dma/altera-msgdma.c 	dma_dev->device_config = msgdma_dma_config;
dma_dev           868 drivers/dma/altera-msgdma.c 	dma_dev->device_alloc_chan_resources = msgdma_alloc_chan_resources;
dma_dev           869 drivers/dma/altera-msgdma.c 	dma_dev->device_free_chan_resources = msgdma_free_chan_resources;
dma_dev           871 drivers/dma/altera-msgdma.c 	mdev->dmachan.device = dma_dev;
dma_dev           872 drivers/dma/altera-msgdma.c 	list_add_tail(&mdev->dmachan.device_node, &dma_dev->channels);
dma_dev           885 drivers/dma/altera-msgdma.c 	ret = dma_async_device_register(dma_dev);
dma_dev          1566 drivers/dma/at_hdmac.c 		BUG_ON(!atslave->dma_dev || atslave->dma_dev != atdma->dma_common.dev);
dma_dev          1648 drivers/dma/at_hdmac.c 	if (atslave->dma_dev == chan->device->dev) {
dma_dev          1702 drivers/dma/at_hdmac.c 	atslave->dma_dev = &dmac_pdev->dev;
dma_dev           151 drivers/dma/bcm-sba-raid.c 	struct dma_device dma_dev;
dma_dev          1554 drivers/dma/bcm-sba-raid.c 	struct dma_device *dma_dev = &sba->dma_dev;
dma_dev          1557 drivers/dma/bcm-sba-raid.c 	sba->dma_chan.device = dma_dev;
dma_dev          1561 drivers/dma/bcm-sba-raid.c 	dma_cap_zero(dma_dev->cap_mask);
dma_dev          1562 drivers/dma/bcm-sba-raid.c 	dma_cap_set(DMA_INTERRUPT, dma_dev->cap_mask);
dma_dev          1563 drivers/dma/bcm-sba-raid.c 	dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask);
dma_dev          1564 drivers/dma/bcm-sba-raid.c 	dma_cap_set(DMA_XOR, dma_dev->cap_mask);
dma_dev          1565 drivers/dma/bcm-sba-raid.c 	dma_cap_set(DMA_PQ, dma_dev->cap_mask);
dma_dev          1572 drivers/dma/bcm-sba-raid.c 	dma_dev->dev = sba->mbox_dev;
dma_dev          1575 drivers/dma/bcm-sba-raid.c 	dma_dev->device_free_chan_resources = sba_free_chan_resources;
dma_dev          1576 drivers/dma/bcm-sba-raid.c 	dma_dev->device_terminate_all = sba_device_terminate_all;
dma_dev          1577 drivers/dma/bcm-sba-raid.c 	dma_dev->device_issue_pending = sba_issue_pending;
dma_dev          1578 drivers/dma/bcm-sba-raid.c 	dma_dev->device_tx_status = sba_tx_status;
dma_dev          1581 drivers/dma/bcm-sba-raid.c 	if (dma_has_cap(DMA_INTERRUPT, dma_dev->cap_mask))
dma_dev          1582 drivers/dma/bcm-sba-raid.c 		dma_dev->device_prep_dma_interrupt = sba_prep_dma_interrupt;
dma_dev          1585 drivers/dma/bcm-sba-raid.c 	if (dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask))
dma_dev          1586 drivers/dma/bcm-sba-raid.c 		dma_dev->device_prep_dma_memcpy = sba_prep_dma_memcpy;
dma_dev          1589 drivers/dma/bcm-sba-raid.c 	if (dma_has_cap(DMA_XOR, dma_dev->cap_mask)) {
dma_dev          1590 drivers/dma/bcm-sba-raid.c 		dma_dev->device_prep_dma_xor = sba_prep_dma_xor;
dma_dev          1591 drivers/dma/bcm-sba-raid.c 		dma_dev->max_xor = sba->max_xor_srcs;
dma_dev          1595 drivers/dma/bcm-sba-raid.c 	if (dma_has_cap(DMA_PQ, dma_dev->cap_mask)) {
dma_dev          1596 drivers/dma/bcm-sba-raid.c 		dma_dev->device_prep_dma_pq = sba_prep_dma_pq;
dma_dev          1597 drivers/dma/bcm-sba-raid.c 		dma_set_maxpq(dma_dev, sba->max_pq_srcs, 0);
dma_dev          1601 drivers/dma/bcm-sba-raid.c 	INIT_LIST_HEAD(&dma_dev->channels);
dma_dev          1602 drivers/dma/bcm-sba-raid.c 	list_add_tail(&sba->dma_chan.device_node, &dma_dev->channels);
dma_dev          1605 drivers/dma/bcm-sba-raid.c 	ret = dma_async_device_register(dma_dev);
dma_dev          1613 drivers/dma/bcm-sba-raid.c 	dma_has_cap(DMA_INTERRUPT, dma_dev->cap_mask) ? "interrupt " : "",
dma_dev          1614 drivers/dma/bcm-sba-raid.c 	dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask) ? "memcpy " : "",
dma_dev          1615 drivers/dma/bcm-sba-raid.c 	dma_has_cap(DMA_XOR, dma_dev->cap_mask) ? "xor " : "",
dma_dev          1616 drivers/dma/bcm-sba-raid.c 	dma_has_cap(DMA_PQ, dma_dev->cap_mask) ? "pq " : "");
dma_dev          1749 drivers/dma/bcm-sba-raid.c 	dma_async_device_unregister(&sba->dma_dev);
dma_dev           135 drivers/dma/dma-axi-dmac.c 	struct dma_device dma_dev;
dma_dev           144 drivers/dma/dma-axi-dmac.c 		dma_dev);
dma_dev           311 drivers/dma/dma-axi-dmac.c 			dev_dbg(dmac->dma_dev.dev,
dma_dev           315 drivers/dma/dma-axi-dmac.c 			dev_warn(dmac->dma_dev.dev,
dma_dev           800 drivers/dma/dma-axi-dmac.c 		dev_err(dmac->dma_dev.dev,
dma_dev           808 drivers/dma/dma-axi-dmac.c 		dev_err(dmac->dma_dev.dev,
dma_dev           830 drivers/dma/dma-axi-dmac.c 	struct dma_device *dma_dev;
dma_dev           874 drivers/dma/dma-axi-dmac.c 	dma_dev = &dmac->dma_dev;
dma_dev           875 drivers/dma/dma-axi-dmac.c 	dma_cap_set(DMA_SLAVE, dma_dev->cap_mask);
dma_dev           876 drivers/dma/dma-axi-dmac.c 	dma_cap_set(DMA_CYCLIC, dma_dev->cap_mask);
dma_dev           877 drivers/dma/dma-axi-dmac.c 	dma_cap_set(DMA_INTERLEAVE, dma_dev->cap_mask);
dma_dev           878 drivers/dma/dma-axi-dmac.c 	dma_dev->device_free_chan_resources = axi_dmac_free_chan_resources;
dma_dev           879 drivers/dma/dma-axi-dmac.c 	dma_dev->device_tx_status = dma_cookie_status;
dma_dev           880 drivers/dma/dma-axi-dmac.c 	dma_dev->device_issue_pending = axi_dmac_issue_pending;
dma_dev           881 drivers/dma/dma-axi-dmac.c 	dma_dev->device_prep_slave_sg = axi_dmac_prep_slave_sg;
dma_dev           882 drivers/dma/dma-axi-dmac.c 	dma_dev->device_prep_dma_cyclic = axi_dmac_prep_dma_cyclic;
dma_dev           883 drivers/dma/dma-axi-dmac.c 	dma_dev->device_prep_interleaved_dma = axi_dmac_prep_interleaved;
dma_dev           884 drivers/dma/dma-axi-dmac.c 	dma_dev->device_terminate_all = axi_dmac_terminate_all;
dma_dev           885 drivers/dma/dma-axi-dmac.c 	dma_dev->device_synchronize = axi_dmac_synchronize;
dma_dev           886 drivers/dma/dma-axi-dmac.c 	dma_dev->dev = &pdev->dev;
dma_dev           887 drivers/dma/dma-axi-dmac.c 	dma_dev->chancnt = 1;
dma_dev           888 drivers/dma/dma-axi-dmac.c 	dma_dev->src_addr_widths = BIT(dmac->chan.src_width);
dma_dev           889 drivers/dma/dma-axi-dmac.c 	dma_dev->dst_addr_widths = BIT(dmac->chan.dest_width);
dma_dev           890 drivers/dma/dma-axi-dmac.c 	dma_dev->directions = BIT(dmac->chan.direction);
dma_dev           891 drivers/dma/dma-axi-dmac.c 	dma_dev->residue_granularity = DMA_RESIDUE_GRANULARITY_DESCRIPTOR;
dma_dev           892 drivers/dma/dma-axi-dmac.c 	INIT_LIST_HEAD(&dma_dev->channels);
dma_dev           895 drivers/dma/dma-axi-dmac.c 	vchan_init(&dmac->chan.vchan, dma_dev);
dma_dev           905 drivers/dma/dma-axi-dmac.c 	dma_dev->copy_align = (dmac->chan.address_align_mask + 1);
dma_dev           909 drivers/dma/dma-axi-dmac.c 	ret = dma_async_device_register(dma_dev);
dma_dev           914 drivers/dma/dma-axi-dmac.c 		of_dma_xlate_by_chan_id, dma_dev);
dma_dev           939 drivers/dma/dma-axi-dmac.c 	dma_async_device_unregister(&dmac->dma_dev);
dma_dev           953 drivers/dma/dma-axi-dmac.c 	dma_async_device_unregister(&dmac->dma_dev);
dma_dev           145 drivers/dma/dmaengine.c ATTRIBUTE_GROUPS(dma_dev);
dma_dev           989 drivers/dma/dmatest.c 	struct dma_device	*dma_dev = chan->device;
dma_dev          1002 drivers/dma/dmatest.c 	if (dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask)) {
dma_dev          1009 drivers/dma/dmatest.c 	if (dma_has_cap(DMA_MEMSET, dma_dev->cap_mask)) {
dma_dev          1016 drivers/dma/dmatest.c 	if (dma_has_cap(DMA_XOR, dma_dev->cap_mask)) {
dma_dev          1020 drivers/dma/dmatest.c 	if (dma_has_cap(DMA_PQ, dma_dev->cap_mask)) {
dma_dev            13 drivers/dma/dw/acpi.c 		.dma_dev = dma_spec->dev,
dma_dev           777 drivers/dma/dw/core.c 	if (dws->dma_dev != chan->device->dev)
dma_dev            21 drivers/dma/dw/of.c 		.dma_dev = dw->dma.dev,
dma_dev           202 drivers/dma/ep93xx_dma.c 	struct dma_device	dma_dev;
dma_dev          1320 drivers/dma/ep93xx_dma.c 	struct dma_device *dma_dev;
dma_dev          1329 drivers/dma/ep93xx_dma.c 	dma_dev = &edma->dma_dev;
dma_dev          1333 drivers/dma/ep93xx_dma.c 	INIT_LIST_HEAD(&dma_dev->channels);
dma_dev          1338 drivers/dma/ep93xx_dma.c 		edmac->chan.device = dma_dev;
dma_dev          1358 drivers/dma/ep93xx_dma.c 			      &dma_dev->channels);
dma_dev          1361 drivers/dma/ep93xx_dma.c 	dma_cap_zero(dma_dev->cap_mask);
dma_dev          1362 drivers/dma/ep93xx_dma.c 	dma_cap_set(DMA_SLAVE, dma_dev->cap_mask);
dma_dev          1363 drivers/dma/ep93xx_dma.c 	dma_cap_set(DMA_CYCLIC, dma_dev->cap_mask);
dma_dev          1365 drivers/dma/ep93xx_dma.c 	dma_dev->dev = &pdev->dev;
dma_dev          1366 drivers/dma/ep93xx_dma.c 	dma_dev->device_alloc_chan_resources = ep93xx_dma_alloc_chan_resources;
dma_dev          1367 drivers/dma/ep93xx_dma.c 	dma_dev->device_free_chan_resources = ep93xx_dma_free_chan_resources;
dma_dev          1368 drivers/dma/ep93xx_dma.c 	dma_dev->device_prep_slave_sg = ep93xx_dma_prep_slave_sg;
dma_dev          1369 drivers/dma/ep93xx_dma.c 	dma_dev->device_prep_dma_cyclic = ep93xx_dma_prep_dma_cyclic;
dma_dev          1370 drivers/dma/ep93xx_dma.c 	dma_dev->device_config = ep93xx_dma_slave_config;
dma_dev          1371 drivers/dma/ep93xx_dma.c 	dma_dev->device_synchronize = ep93xx_dma_synchronize;
dma_dev          1372 drivers/dma/ep93xx_dma.c 	dma_dev->device_terminate_all = ep93xx_dma_terminate_all;
dma_dev          1373 drivers/dma/ep93xx_dma.c 	dma_dev->device_issue_pending = ep93xx_dma_issue_pending;
dma_dev          1374 drivers/dma/ep93xx_dma.c 	dma_dev->device_tx_status = ep93xx_dma_tx_status;
dma_dev          1376 drivers/dma/ep93xx_dma.c 	dma_set_max_seg_size(dma_dev->dev, DMA_MAX_CHAN_BYTES);
dma_dev          1379 drivers/dma/ep93xx_dma.c 		dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask);
dma_dev          1380 drivers/dma/ep93xx_dma.c 		dma_dev->device_prep_dma_memcpy = ep93xx_dma_prep_dma_memcpy;
dma_dev          1387 drivers/dma/ep93xx_dma.c 		dma_cap_set(DMA_PRIVATE, dma_dev->cap_mask);
dma_dev          1396 drivers/dma/ep93xx_dma.c 	ret = dma_async_device_register(dma_dev);
dma_dev          1405 drivers/dma/ep93xx_dma.c 		dev_info(dma_dev->dev, "EP93xx M2%s DMA ready\n",
dma_dev           155 drivers/dma/fsl-edma-common.h 	struct dma_device	dma_dev;
dma_dev           109 drivers/dma/fsl-edma.c 	list_for_each_entry_safe(chan, _chan, &fsl_edma->dma_dev.channels, device_node) {
dma_dev           336 drivers/dma/fsl-edma.c 	INIT_LIST_HEAD(&fsl_edma->dma_dev.channels);
dma_dev           346 drivers/dma/fsl-edma.c 		vchan_init(&fsl_chan->vchan, &fsl_edma->dma_dev);
dma_dev           357 drivers/dma/fsl-edma.c 	dma_cap_set(DMA_PRIVATE, fsl_edma->dma_dev.cap_mask);
dma_dev           358 drivers/dma/fsl-edma.c 	dma_cap_set(DMA_SLAVE, fsl_edma->dma_dev.cap_mask);
dma_dev           359 drivers/dma/fsl-edma.c 	dma_cap_set(DMA_CYCLIC, fsl_edma->dma_dev.cap_mask);
dma_dev           361 drivers/dma/fsl-edma.c 	fsl_edma->dma_dev.dev = &pdev->dev;
dma_dev           362 drivers/dma/fsl-edma.c 	fsl_edma->dma_dev.device_alloc_chan_resources
dma_dev           364 drivers/dma/fsl-edma.c 	fsl_edma->dma_dev.device_free_chan_resources
dma_dev           366 drivers/dma/fsl-edma.c 	fsl_edma->dma_dev.device_tx_status = fsl_edma_tx_status;
dma_dev           367 drivers/dma/fsl-edma.c 	fsl_edma->dma_dev.device_prep_slave_sg = fsl_edma_prep_slave_sg;
dma_dev           368 drivers/dma/fsl-edma.c 	fsl_edma->dma_dev.device_prep_dma_cyclic = fsl_edma_prep_dma_cyclic;
dma_dev           369 drivers/dma/fsl-edma.c 	fsl_edma->dma_dev.device_config = fsl_edma_slave_config;
dma_dev           370 drivers/dma/fsl-edma.c 	fsl_edma->dma_dev.device_pause = fsl_edma_pause;
dma_dev           371 drivers/dma/fsl-edma.c 	fsl_edma->dma_dev.device_resume = fsl_edma_resume;
dma_dev           372 drivers/dma/fsl-edma.c 	fsl_edma->dma_dev.device_terminate_all = fsl_edma_terminate_all;
dma_dev           373 drivers/dma/fsl-edma.c 	fsl_edma->dma_dev.device_synchronize = fsl_edma_synchronize;
dma_dev           374 drivers/dma/fsl-edma.c 	fsl_edma->dma_dev.device_issue_pending = fsl_edma_issue_pending;
dma_dev           376 drivers/dma/fsl-edma.c 	fsl_edma->dma_dev.src_addr_widths = FSL_EDMA_BUSWIDTHS;
dma_dev           377 drivers/dma/fsl-edma.c 	fsl_edma->dma_dev.dst_addr_widths = FSL_EDMA_BUSWIDTHS;
dma_dev           378 drivers/dma/fsl-edma.c 	fsl_edma->dma_dev.directions = BIT(DMA_DEV_TO_MEM) | BIT(DMA_MEM_TO_DEV);
dma_dev           382 drivers/dma/fsl-edma.c 	ret = dma_async_device_register(&fsl_edma->dma_dev);
dma_dev           394 drivers/dma/fsl-edma.c 		dma_async_device_unregister(&fsl_edma->dma_dev);
dma_dev           411 drivers/dma/fsl-edma.c 	fsl_edma_cleanup_vchan(&fsl_edma->dma_dev);
dma_dev           413 drivers/dma/fsl-edma.c 	dma_async_device_unregister(&fsl_edma->dma_dev);
dma_dev           198 drivers/dma/fsl-qdma.c 	struct dma_device	dma_dev;
dma_dev           707 drivers/dma/fsl-qdma.c 		dev_err(fsl_qdma->dma_dev.dev, "DMA transaction error!\n");
dma_dev           722 drivers/dma/fsl-qdma.c 		dev_err(fsl_qdma->dma_dev.dev,
dma_dev           740 drivers/dma/fsl-qdma.c 		dev_err(fsl_qdma->dma_dev.dev, "QDMA: status err!\n");
dma_dev           829 drivers/dma/fsl-qdma.c 		dev_err(fsl_qdma->dma_dev.dev, "DMA halt failed!");
dma_dev          1162 drivers/dma/fsl-qdma.c 	INIT_LIST_HEAD(&fsl_qdma->dma_dev.channels);
dma_dev          1171 drivers/dma/fsl-qdma.c 		vchan_init(&fsl_chan->vchan, &fsl_qdma->dma_dev);
dma_dev          1174 drivers/dma/fsl-qdma.c 	dma_cap_set(DMA_MEMCPY, fsl_qdma->dma_dev.cap_mask);
dma_dev          1176 drivers/dma/fsl-qdma.c 	fsl_qdma->dma_dev.dev = &pdev->dev;
dma_dev          1177 drivers/dma/fsl-qdma.c 	fsl_qdma->dma_dev.device_free_chan_resources =
dma_dev          1179 drivers/dma/fsl-qdma.c 	fsl_qdma->dma_dev.device_alloc_chan_resources =
dma_dev          1181 drivers/dma/fsl-qdma.c 	fsl_qdma->dma_dev.device_tx_status = dma_cookie_status;
dma_dev          1182 drivers/dma/fsl-qdma.c 	fsl_qdma->dma_dev.device_prep_dma_memcpy = fsl_qdma_prep_memcpy;
dma_dev          1183 drivers/dma/fsl-qdma.c 	fsl_qdma->dma_dev.device_issue_pending = fsl_qdma_issue_pending;
dma_dev          1184 drivers/dma/fsl-qdma.c 	fsl_qdma->dma_dev.device_synchronize = fsl_qdma_synchronize;
dma_dev          1185 drivers/dma/fsl-qdma.c 	fsl_qdma->dma_dev.device_terminate_all = fsl_qdma_terminate_all;
dma_dev          1191 drivers/dma/fsl-qdma.c 	ret = dma_async_device_register(&fsl_qdma->dma_dev);
dma_dev          1226 drivers/dma/fsl-qdma.c 	fsl_qdma_cleanup_vchan(&fsl_qdma->dma_dev);
dma_dev          1228 drivers/dma/fsl-qdma.c 	dma_async_device_unregister(&fsl_qdma->dma_dev);
dma_dev           630 drivers/dma/fsl_raid.c 	struct dma_device *dma_dev;
dma_dev           638 drivers/dma/fsl_raid.c 	dma_dev = &re_priv->dma_dev;
dma_dev           684 drivers/dma/fsl_raid.c 	chan->chan.device = dma_dev;
dma_dev           754 drivers/dma/fsl_raid.c 	struct dma_device *dma_dev;
dma_dev           783 drivers/dma/fsl_raid.c 	dma_dev = &re_priv->dma_dev;
dma_dev           784 drivers/dma/fsl_raid.c 	dma_dev->dev = dev;
dma_dev           785 drivers/dma/fsl_raid.c 	INIT_LIST_HEAD(&dma_dev->channels);
dma_dev           788 drivers/dma/fsl_raid.c 	dma_dev->device_alloc_chan_resources = fsl_re_alloc_chan_resources;
dma_dev           789 drivers/dma/fsl_raid.c 	dma_dev->device_tx_status = fsl_re_tx_status;
dma_dev           790 drivers/dma/fsl_raid.c 	dma_dev->device_issue_pending = fsl_re_issue_pending;
dma_dev           792 drivers/dma/fsl_raid.c 	dma_dev->max_xor = FSL_RE_MAX_XOR_SRCS;
dma_dev           793 drivers/dma/fsl_raid.c 	dma_dev->device_prep_dma_xor = fsl_re_prep_dma_xor;
dma_dev           794 drivers/dma/fsl_raid.c 	dma_cap_set(DMA_XOR, dma_dev->cap_mask);
dma_dev           796 drivers/dma/fsl_raid.c 	dma_dev->max_pq = FSL_RE_MAX_PQ_SRCS;
dma_dev           797 drivers/dma/fsl_raid.c 	dma_dev->device_prep_dma_pq = fsl_re_prep_dma_pq;
dma_dev           798 drivers/dma/fsl_raid.c 	dma_cap_set(DMA_PQ, dma_dev->cap_mask);
dma_dev           800 drivers/dma/fsl_raid.c 	dma_dev->device_prep_dma_memcpy = fsl_re_prep_dma_memcpy;
dma_dev           801 drivers/dma/fsl_raid.c 	dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask);
dma_dev           803 drivers/dma/fsl_raid.c 	dma_dev->device_free_chan_resources = fsl_re_free_chan_resources;
dma_dev           845 drivers/dma/fsl_raid.c 	dma_async_device_register(dma_dev);
dma_dev           875 drivers/dma/fsl_raid.c 	dma_async_device_unregister(&re_priv->dma_dev);
dma_dev           258 drivers/dma/fsl_raid.h 	struct dma_device dma_dev;
dma_dev           134 drivers/dma/img-mdc-dma.c 	struct dma_device dma_dev;
dma_dev           183 drivers/dma/img-mdc-dma.c 	return mdma->dma_dev.dev;
dma_dev           817 drivers/dma/img-mdc-dma.c 	list_for_each_entry(chan, &mdma->dma_dev.channels, device_node) {
dma_dev           915 drivers/dma/img-mdc-dma.c 	dma_cap_zero(mdma->dma_dev.cap_mask);
dma_dev           916 drivers/dma/img-mdc-dma.c 	dma_cap_set(DMA_SLAVE, mdma->dma_dev.cap_mask);
dma_dev           917 drivers/dma/img-mdc-dma.c 	dma_cap_set(DMA_PRIVATE, mdma->dma_dev.cap_mask);
dma_dev           918 drivers/dma/img-mdc-dma.c 	dma_cap_set(DMA_CYCLIC, mdma->dma_dev.cap_mask);
dma_dev           919 drivers/dma/img-mdc-dma.c 	dma_cap_set(DMA_MEMCPY, mdma->dma_dev.cap_mask);
dma_dev           949 drivers/dma/img-mdc-dma.c 	mdma->dma_dev.dev = &pdev->dev;
dma_dev           950 drivers/dma/img-mdc-dma.c 	mdma->dma_dev.device_prep_slave_sg = mdc_prep_slave_sg;
dma_dev           951 drivers/dma/img-mdc-dma.c 	mdma->dma_dev.device_prep_dma_cyclic = mdc_prep_dma_cyclic;
dma_dev           952 drivers/dma/img-mdc-dma.c 	mdma->dma_dev.device_prep_dma_memcpy = mdc_prep_dma_memcpy;
dma_dev           953 drivers/dma/img-mdc-dma.c 	mdma->dma_dev.device_alloc_chan_resources = mdc_alloc_chan_resources;
dma_dev           954 drivers/dma/img-mdc-dma.c 	mdma->dma_dev.device_free_chan_resources = mdc_free_chan_resources;
dma_dev           955 drivers/dma/img-mdc-dma.c 	mdma->dma_dev.device_tx_status = mdc_tx_status;
dma_dev           956 drivers/dma/img-mdc-dma.c 	mdma->dma_dev.device_issue_pending = mdc_issue_pending;
dma_dev           957 drivers/dma/img-mdc-dma.c 	mdma->dma_dev.device_terminate_all = mdc_terminate_all;
dma_dev           958 drivers/dma/img-mdc-dma.c 	mdma->dma_dev.device_synchronize = mdc_synchronize;
dma_dev           959 drivers/dma/img-mdc-dma.c 	mdma->dma_dev.device_config = mdc_slave_config;
dma_dev           961 drivers/dma/img-mdc-dma.c 	mdma->dma_dev.directions = BIT(DMA_DEV_TO_MEM) | BIT(DMA_MEM_TO_DEV);
dma_dev           962 drivers/dma/img-mdc-dma.c 	mdma->dma_dev.residue_granularity = DMA_RESIDUE_GRANULARITY_BURST;
dma_dev           964 drivers/dma/img-mdc-dma.c 		mdma->dma_dev.src_addr_widths |= BIT(i);
dma_dev           965 drivers/dma/img-mdc-dma.c 		mdma->dma_dev.dst_addr_widths |= BIT(i);
dma_dev           968 drivers/dma/img-mdc-dma.c 	INIT_LIST_HEAD(&mdma->dma_dev.channels);
dma_dev           985 drivers/dma/img-mdc-dma.c 		vchan_init(&mchan->vc, &mdma->dma_dev);
dma_dev          1001 drivers/dma/img-mdc-dma.c 	ret = dma_async_device_register(&mdma->dma_dev);
dma_dev          1015 drivers/dma/img-mdc-dma.c 	dma_async_device_unregister(&mdma->dma_dev);
dma_dev          1029 drivers/dma/img-mdc-dma.c 	dma_async_device_unregister(&mdma->dma_dev);
dma_dev          1031 drivers/dma/img-mdc-dma.c 	list_for_each_entry_safe(mchan, next, &mdma->dma_dev.channels,
dma_dev            22 drivers/dma/ioat/dma.h #define to_ioatdma_device(dev) container_of(dev, struct ioatdma_device, dma_dev)
dma_dev            69 drivers/dma/ioat/dma.h 	struct dma_device dma_dev;
dma_dev           305 drivers/dma/ioat/init.c 	struct dma_device *dma = &ioat_dma->dma_dev;
dma_dev           351 drivers/dma/ioat/init.c 	tx = ioat_dma->dma_dev.device_prep_dma_memcpy(dma_chan, dma_dest,
dma_dev           424 drivers/dma/ioat/init.c 	msixcnt = ioat_dma->dma_dev.chancnt;
dma_dev           496 drivers/dma/ioat/init.c 	struct dma_device *dma = &ioat_dma->dma_dev;
dma_dev           540 drivers/dma/ioat/init.c 	int err = dma_async_device_register(&ioat_dma->dma_dev);
dma_dev           552 drivers/dma/ioat/init.c 	struct dma_device *dma = &ioat_dma->dma_dev;
dma_dev           573 drivers/dma/ioat/init.c 	struct dma_device *dma = &ioat_dma->dma_dev;
dma_dev           769 drivers/dma/ioat/init.c 	struct dma_device *dma = &ioat_dma->dma_dev;
dma_dev           803 drivers/dma/ioat/init.c 	struct dma_device *dma = &ioat_dma->dma_dev;
dma_dev          1070 drivers/dma/ioat/init.c 	dma = &ioat_dma->dma_dev;
dma_dev          1099 drivers/dma/ioat/init.c 	dma = &ioat_dma->dma_dev;
dma_dev            77 drivers/dma/ioat/sysfs.c 	struct dma_device *dma = &ioat_dma->dma_dev;
dma_dev            98 drivers/dma/ioat/sysfs.c 	struct dma_device *dma = &ioat_dma->dma_dev;
dma_dev          1271 drivers/dma/iop-adma.c 	struct dma_device *dma_dev;
dma_dev          1285 drivers/dma/iop-adma.c 	dma_dev = &adev->common;
dma_dev          1307 drivers/dma/iop-adma.c 	dma_dev->cap_mask = plat_data->cap_mask;
dma_dev          1312 drivers/dma/iop-adma.c 	INIT_LIST_HEAD(&dma_dev->channels);
dma_dev          1315 drivers/dma/iop-adma.c 	dma_dev->device_alloc_chan_resources = iop_adma_alloc_chan_resources;
dma_dev          1316 drivers/dma/iop-adma.c 	dma_dev->device_free_chan_resources = iop_adma_free_chan_resources;
dma_dev          1317 drivers/dma/iop-adma.c 	dma_dev->device_tx_status = iop_adma_status;
dma_dev          1318 drivers/dma/iop-adma.c 	dma_dev->device_issue_pending = iop_adma_issue_pending;
dma_dev          1319 drivers/dma/iop-adma.c 	dma_dev->dev = &pdev->dev;
dma_dev          1322 drivers/dma/iop-adma.c 	if (dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask))
dma_dev          1323 drivers/dma/iop-adma.c 		dma_dev->device_prep_dma_memcpy = iop_adma_prep_dma_memcpy;
dma_dev          1324 drivers/dma/iop-adma.c 	if (dma_has_cap(DMA_XOR, dma_dev->cap_mask)) {
dma_dev          1325 drivers/dma/iop-adma.c 		dma_dev->max_xor = iop_adma_get_max_xor();
dma_dev          1326 drivers/dma/iop-adma.c 		dma_dev->device_prep_dma_xor = iop_adma_prep_dma_xor;
dma_dev          1328 drivers/dma/iop-adma.c 	if (dma_has_cap(DMA_XOR_VAL, dma_dev->cap_mask))
dma_dev          1329 drivers/dma/iop-adma.c 		dma_dev->device_prep_dma_xor_val =
dma_dev          1331 drivers/dma/iop-adma.c 	if (dma_has_cap(DMA_PQ, dma_dev->cap_mask)) {
dma_dev          1332 drivers/dma/iop-adma.c 		dma_set_maxpq(dma_dev, iop_adma_get_max_pq(), 0);
dma_dev          1333 drivers/dma/iop-adma.c 		dma_dev->device_prep_dma_pq = iop_adma_prep_dma_pq;
dma_dev          1335 drivers/dma/iop-adma.c 	if (dma_has_cap(DMA_PQ_VAL, dma_dev->cap_mask))
dma_dev          1336 drivers/dma/iop-adma.c 		dma_dev->device_prep_dma_pq_val =
dma_dev          1338 drivers/dma/iop-adma.c 	if (dma_has_cap(DMA_INTERRUPT, dma_dev->cap_mask))
dma_dev          1339 drivers/dma/iop-adma.c 		dma_dev->device_prep_dma_interrupt =
dma_dev          1380 drivers/dma/iop-adma.c 	iop_chan->common.device = dma_dev;
dma_dev          1382 drivers/dma/iop-adma.c 	list_add_tail(&iop_chan->common.device_node, &dma_dev->channels);
dma_dev          1384 drivers/dma/iop-adma.c 	if (dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask)) {
dma_dev          1391 drivers/dma/iop-adma.c 	if (dma_has_cap(DMA_XOR, dma_dev->cap_mask)) {
dma_dev          1398 drivers/dma/iop-adma.c 	if (dma_has_cap(DMA_PQ, dma_dev->cap_mask) &&
dma_dev          1399 drivers/dma/iop-adma.c 	    dma_has_cap(DMA_PQ_VAL, dma_dev->cap_mask)) {
dma_dev          1405 drivers/dma/iop-adma.c 		dma_cap_clear(DMA_PQ, dma_dev->cap_mask);
dma_dev          1406 drivers/dma/iop-adma.c 		dma_cap_clear(DMA_PQ_VAL, dma_dev->cap_mask);
dma_dev          1414 drivers/dma/iop-adma.c 		 dma_has_cap(DMA_PQ, dma_dev->cap_mask) ? "pq " : "",
dma_dev          1415 drivers/dma/iop-adma.c 		 dma_has_cap(DMA_PQ_VAL, dma_dev->cap_mask) ? "pq_val " : "",
dma_dev          1416 drivers/dma/iop-adma.c 		 dma_has_cap(DMA_XOR, dma_dev->cap_mask) ? "xor " : "",
dma_dev          1417 drivers/dma/iop-adma.c 		 dma_has_cap(DMA_XOR_VAL, dma_dev->cap_mask) ? "xor_val " : "",
dma_dev          1418 drivers/dma/iop-adma.c 		 dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask) ? "cpy " : "",
dma_dev          1419 drivers/dma/iop-adma.c 		 dma_has_cap(DMA_INTERRUPT, dma_dev->cap_mask) ? "intr " : "");
dma_dev          1421 drivers/dma/iop-adma.c 	dma_async_device_register(dma_dev);
dma_dev           215 drivers/dma/mcf-edma.c 	INIT_LIST_HEAD(&mcf_edma->dma_dev.channels);
dma_dev           224 drivers/dma/mcf-edma.c 		vchan_init(&mcf_chan->vchan, &mcf_edma->dma_dev);
dma_dev           235 drivers/dma/mcf-edma.c 	dma_cap_set(DMA_PRIVATE, mcf_edma->dma_dev.cap_mask);
dma_dev           236 drivers/dma/mcf-edma.c 	dma_cap_set(DMA_SLAVE, mcf_edma->dma_dev.cap_mask);
dma_dev           237 drivers/dma/mcf-edma.c 	dma_cap_set(DMA_CYCLIC, mcf_edma->dma_dev.cap_mask);
dma_dev           239 drivers/dma/mcf-edma.c 	mcf_edma->dma_dev.dev = &pdev->dev;
dma_dev           240 drivers/dma/mcf-edma.c 	mcf_edma->dma_dev.device_alloc_chan_resources =
dma_dev           242 drivers/dma/mcf-edma.c 	mcf_edma->dma_dev.device_free_chan_resources =
dma_dev           244 drivers/dma/mcf-edma.c 	mcf_edma->dma_dev.device_config = fsl_edma_slave_config;
dma_dev           245 drivers/dma/mcf-edma.c 	mcf_edma->dma_dev.device_prep_dma_cyclic =
dma_dev           247 drivers/dma/mcf-edma.c 	mcf_edma->dma_dev.device_prep_slave_sg = fsl_edma_prep_slave_sg;
dma_dev           248 drivers/dma/mcf-edma.c 	mcf_edma->dma_dev.device_tx_status = fsl_edma_tx_status;
dma_dev           249 drivers/dma/mcf-edma.c 	mcf_edma->dma_dev.device_pause = fsl_edma_pause;
dma_dev           250 drivers/dma/mcf-edma.c 	mcf_edma->dma_dev.device_resume = fsl_edma_resume;
dma_dev           251 drivers/dma/mcf-edma.c 	mcf_edma->dma_dev.device_terminate_all = fsl_edma_terminate_all;
dma_dev           252 drivers/dma/mcf-edma.c 	mcf_edma->dma_dev.device_issue_pending = fsl_edma_issue_pending;
dma_dev           254 drivers/dma/mcf-edma.c 	mcf_edma->dma_dev.src_addr_widths = FSL_EDMA_BUSWIDTHS;
dma_dev           255 drivers/dma/mcf-edma.c 	mcf_edma->dma_dev.dst_addr_widths = FSL_EDMA_BUSWIDTHS;
dma_dev           256 drivers/dma/mcf-edma.c 	mcf_edma->dma_dev.directions =
dma_dev           259 drivers/dma/mcf-edma.c 	mcf_edma->dma_dev.filter.fn = mcf_edma_filter_fn;
dma_dev           260 drivers/dma/mcf-edma.c 	mcf_edma->dma_dev.filter.map = pdata->slave_map;
dma_dev           261 drivers/dma/mcf-edma.c 	mcf_edma->dma_dev.filter.mapcnt = pdata->slavecnt;
dma_dev           265 drivers/dma/mcf-edma.c 	ret = dma_async_device_register(&mcf_edma->dma_dev);
dma_dev           283 drivers/dma/mcf-edma.c 	fsl_edma_cleanup_vchan(&mcf_edma->dma_dev);
dma_dev           284 drivers/dma/mcf-edma.c 	dma_async_device_unregister(&mcf_edma->dma_dev);
dma_dev           589 drivers/dma/mic_x100_dma.c 	dma_cap_zero(mic_dma_dev->dma_dev.cap_mask);
dma_dev           594 drivers/dma/mic_x100_dma.c 	dma_cap_set(DMA_MEMCPY, mic_dma_dev->dma_dev.cap_mask);
dma_dev           597 drivers/dma/mic_x100_dma.c 		dma_cap_set(DMA_PRIVATE, mic_dma_dev->dma_dev.cap_mask);
dma_dev           598 drivers/dma/mic_x100_dma.c 	mic_dma_dev->dma_dev.device_alloc_chan_resources =
dma_dev           600 drivers/dma/mic_x100_dma.c 	mic_dma_dev->dma_dev.device_free_chan_resources =
dma_dev           602 drivers/dma/mic_x100_dma.c 	mic_dma_dev->dma_dev.device_tx_status = mic_dma_tx_status;
dma_dev           603 drivers/dma/mic_x100_dma.c 	mic_dma_dev->dma_dev.device_prep_dma_memcpy = mic_dma_prep_memcpy_lock;
dma_dev           604 drivers/dma/mic_x100_dma.c 	mic_dma_dev->dma_dev.device_prep_dma_imm_data =
dma_dev           606 drivers/dma/mic_x100_dma.c 	mic_dma_dev->dma_dev.device_prep_dma_interrupt =
dma_dev           608 drivers/dma/mic_x100_dma.c 	mic_dma_dev->dma_dev.device_issue_pending = mic_dma_issue_pending;
dma_dev           609 drivers/dma/mic_x100_dma.c 	mic_dma_dev->dma_dev.copy_align = MIC_DMA_ALIGN_SHIFT;
dma_dev           610 drivers/dma/mic_x100_dma.c 	INIT_LIST_HEAD(&mic_dma_dev->dma_dev.channels);
dma_dev           612 drivers/dma/mic_x100_dma.c 		mic_dma_dev->mic_ch[i].api_ch.device = &mic_dma_dev->dma_dev;
dma_dev           615 drivers/dma/mic_x100_dma.c 			      &mic_dma_dev->dma_dev.channels);
dma_dev           617 drivers/dma/mic_x100_dma.c 	return dmaenginem_async_device_register(&mic_dma_dev->dma_dev);
dma_dev           637 drivers/dma/mic_x100_dma.c 	mic_dma_dev->dma_dev.dev = dev;
dma_dev           132 drivers/dma/mic_x100_dma.h 	struct dma_device dma_dev;
dma_dev           164 drivers/dma/mic_x100_dma.h 	return to_mic_dma_dev(ch)->dma_dev.dev;
dma_dev          1040 drivers/dma/mv_xor.c 	struct dma_device *dma_dev;
dma_dev          1053 drivers/dma/mv_xor.c 	dma_dev = &mv_chan->dmadev;
dma_dev          1054 drivers/dma/mv_xor.c 	dma_dev->dev = &pdev->dev;
dma_dev          1062 drivers/dma/mv_xor.c 	mv_chan->dummy_src_addr = dma_map_single(dma_dev->dev,
dma_dev          1064 drivers/dma/mv_xor.c 	mv_chan->dummy_dst_addr = dma_map_single(dma_dev->dev,
dma_dev          1078 drivers/dma/mv_xor.c 	dma_dev->cap_mask = cap_mask;
dma_dev          1080 drivers/dma/mv_xor.c 	INIT_LIST_HEAD(&dma_dev->channels);
dma_dev          1083 drivers/dma/mv_xor.c 	dma_dev->device_alloc_chan_resources = mv_xor_alloc_chan_resources;
dma_dev          1084 drivers/dma/mv_xor.c 	dma_dev->device_free_chan_resources = mv_xor_free_chan_resources;
dma_dev          1085 drivers/dma/mv_xor.c 	dma_dev->device_tx_status = mv_xor_status;
dma_dev          1086 drivers/dma/mv_xor.c 	dma_dev->device_issue_pending = mv_xor_issue_pending;
dma_dev          1089 drivers/dma/mv_xor.c 	if (dma_has_cap(DMA_INTERRUPT, dma_dev->cap_mask))
dma_dev          1090 drivers/dma/mv_xor.c 		dma_dev->device_prep_dma_interrupt = mv_xor_prep_dma_interrupt;
dma_dev          1091 drivers/dma/mv_xor.c 	if (dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask))
dma_dev          1092 drivers/dma/mv_xor.c 		dma_dev->device_prep_dma_memcpy = mv_xor_prep_dma_memcpy;
dma_dev          1093 drivers/dma/mv_xor.c 	if (dma_has_cap(DMA_XOR, dma_dev->cap_mask)) {
dma_dev          1094 drivers/dma/mv_xor.c 		dma_dev->max_xor = 8;
dma_dev          1095 drivers/dma/mv_xor.c 		dma_dev->device_prep_dma_xor = mv_xor_prep_dma_xor;
dma_dev          1123 drivers/dma/mv_xor.c 	mv_chan->dmachan.device = dma_dev;
dma_dev          1126 drivers/dma/mv_xor.c 	list_add_tail(&mv_chan->dmachan.device_node, &dma_dev->channels);
dma_dev          1128 drivers/dma/mv_xor.c 	if (dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask)) {
dma_dev          1135 drivers/dma/mv_xor.c 	if (dma_has_cap(DMA_XOR, dma_dev->cap_mask)) {
dma_dev          1144 drivers/dma/mv_xor.c 		 dma_has_cap(DMA_XOR, dma_dev->cap_mask) ? "xor " : "",
dma_dev          1145 drivers/dma/mv_xor.c 		 dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask) ? "cpy " : "",
dma_dev          1146 drivers/dma/mv_xor.c 		 dma_has_cap(DMA_INTERRUPT, dma_dev->cap_mask) ? "intr " : "");
dma_dev          1148 drivers/dma/mv_xor.c 	ret = dma_async_device_register(dma_dev);
dma_dev           714 drivers/dma/mv_xor_v2.c 	struct dma_device *dma_dev;
dma_dev           828 drivers/dma/mv_xor_v2.c 	dma_dev = &xor_dev->dmadev;
dma_dev           831 drivers/dma/mv_xor_v2.c 	dma_cap_zero(dma_dev->cap_mask);
dma_dev           832 drivers/dma/mv_xor_v2.c 	dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask);
dma_dev           833 drivers/dma/mv_xor_v2.c 	dma_cap_set(DMA_XOR, dma_dev->cap_mask);
dma_dev           834 drivers/dma/mv_xor_v2.c 	dma_cap_set(DMA_INTERRUPT, dma_dev->cap_mask);
dma_dev           837 drivers/dma/mv_xor_v2.c 	INIT_LIST_HEAD(&dma_dev->channels);
dma_dev           840 drivers/dma/mv_xor_v2.c 	dma_dev->device_tx_status = dma_cookie_status;
dma_dev           841 drivers/dma/mv_xor_v2.c 	dma_dev->device_issue_pending = mv_xor_v2_issue_pending;
dma_dev           842 drivers/dma/mv_xor_v2.c 	dma_dev->dev = &pdev->dev;
dma_dev           844 drivers/dma/mv_xor_v2.c 	dma_dev->device_prep_dma_memcpy = mv_xor_v2_prep_dma_memcpy;
dma_dev           845 drivers/dma/mv_xor_v2.c 	dma_dev->device_prep_dma_interrupt = mv_xor_v2_prep_dma_interrupt;
dma_dev           846 drivers/dma/mv_xor_v2.c 	dma_dev->max_xor = 8;
dma_dev           847 drivers/dma/mv_xor_v2.c 	dma_dev->device_prep_dma_xor = mv_xor_v2_prep_dma_xor;
dma_dev           849 drivers/dma/mv_xor_v2.c 	xor_dev->dmachan.device = dma_dev;
dma_dev           852 drivers/dma/mv_xor_v2.c 		      &dma_dev->channels);
dma_dev           858 drivers/dma/mv_xor_v2.c 	ret = dma_async_device_register(dma_dev);
dma_dev           223 drivers/dma/nbpfaxi.c 	struct dma_device dma_dev;
dma_dev           319 drivers/dma/nbpfaxi.c 	dev_dbg(nbpf->dma_dev.dev, "%s(0x%p + 0x%x) = 0x%x\n",
dma_dev           328 drivers/dma/nbpfaxi.c 	dev_dbg(nbpf->dma_dev.dev, "%s(0x%p + 0x%x) = 0x%x\n",
dma_dev           394 drivers/dma/nbpfaxi.c 	dev_dbg(chan->nbpf->dma_dev.dev, "%s(): next 0x%x, cur 0x%x\n", __func__,
dma_dev           850 drivers/dma/nbpfaxi.c 		dev_dbg(chan->nbpf->dma_dev.dev, "%s(): force-free desc %p cookie %d\n",
dma_dev          1091 drivers/dma/nbpfaxi.c 	dchan = dma_get_any_slave_channel(&nbpf->dma_dev);
dma_dev          1222 drivers/dma/nbpfaxi.c 	dev_warn(nbpf->dma_dev.dev, "DMA error IRQ %u\n", irq);
dma_dev          1240 drivers/dma/nbpfaxi.c 	struct dma_device *dma_dev = &nbpf->dma_dev;
dma_dev          1248 drivers/dma/nbpfaxi.c 	chan->dma_chan.device = dma_dev;
dma_dev          1252 drivers/dma/nbpfaxi.c 	dev_dbg(dma_dev->dev, "%s(): channel %d: -> %p\n", __func__, n, chan->base);
dma_dev          1257 drivers/dma/nbpfaxi.c 	ret = devm_request_irq(dma_dev->dev, chan->irq,
dma_dev          1265 drivers/dma/nbpfaxi.c 		      &dma_dev->channels);
dma_dev          1289 drivers/dma/nbpfaxi.c 	struct dma_device *dma_dev;
dma_dev          1311 drivers/dma/nbpfaxi.c 	dma_dev = &nbpf->dma_dev;
dma_dev          1312 drivers/dma/nbpfaxi.c 	dma_dev->dev = dev;
dma_dev          1390 drivers/dma/nbpfaxi.c 	INIT_LIST_HEAD(&dma_dev->channels);
dma_dev          1399 drivers/dma/nbpfaxi.c 	dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask);
dma_dev          1400 drivers/dma/nbpfaxi.c 	dma_cap_set(DMA_SLAVE, dma_dev->cap_mask);
dma_dev          1401 drivers/dma/nbpfaxi.c 	dma_cap_set(DMA_PRIVATE, dma_dev->cap_mask);
dma_dev          1404 drivers/dma/nbpfaxi.c 	dma_dev->device_alloc_chan_resources
dma_dev          1406 drivers/dma/nbpfaxi.c 	dma_dev->device_free_chan_resources = nbpf_free_chan_resources;
dma_dev          1407 drivers/dma/nbpfaxi.c 	dma_dev->device_prep_dma_memcpy = nbpf_prep_memcpy;
dma_dev          1408 drivers/dma/nbpfaxi.c 	dma_dev->device_tx_status = nbpf_tx_status;
dma_dev          1409 drivers/dma/nbpfaxi.c 	dma_dev->device_issue_pending = nbpf_issue_pending;
dma_dev          1420 drivers/dma/nbpfaxi.c 	dma_dev->device_prep_slave_sg = nbpf_prep_slave_sg;
dma_dev          1421 drivers/dma/nbpfaxi.c 	dma_dev->device_config = nbpf_config;
dma_dev          1422 drivers/dma/nbpfaxi.c 	dma_dev->device_pause = nbpf_pause;
dma_dev          1423 drivers/dma/nbpfaxi.c 	dma_dev->device_terminate_all = nbpf_terminate_all;
dma_dev          1425 drivers/dma/nbpfaxi.c 	dma_dev->src_addr_widths = NBPF_DMA_BUSWIDTHS;
dma_dev          1426 drivers/dma/nbpfaxi.c 	dma_dev->dst_addr_widths = NBPF_DMA_BUSWIDTHS;
dma_dev          1427 drivers/dma/nbpfaxi.c 	dma_dev->directions = BIT(DMA_DEV_TO_MEM) | BIT(DMA_MEM_TO_DEV);
dma_dev          1437 drivers/dma/nbpfaxi.c 	ret = dma_async_device_register(dma_dev);
dma_dev          1448 drivers/dma/nbpfaxi.c 	dma_async_device_unregister(dma_dev);
dma_dev          1471 drivers/dma/nbpfaxi.c 	dma_async_device_unregister(&nbpf->dma_dev);
dma_dev            37 drivers/dma/sh/shdma-base.c #define to_shdma_dev(d) container_of(d, struct shdma_dev, dma_dev)
dma_dev           306 drivers/dma/sh/shdma-base.c 		dev_warn(sdev->dma_dev.dev, "invalid slave ID passed to dma_request_slave\n");
dma_dev           961 drivers/dma/sh/shdma-base.c 	schan->dma_chan.device = &sdev->dma_dev;
dma_dev           964 drivers/dma/sh/shdma-base.c 	schan->dev = sdev->dma_dev.dev;
dma_dev           978 drivers/dma/sh/shdma-base.c 			&sdev->dma_dev.channels);
dma_dev           992 drivers/dma/sh/shdma-base.c 	struct dma_device *dma_dev = &sdev->dma_dev;
dma_dev          1015 drivers/dma/sh/shdma-base.c 	INIT_LIST_HEAD(&dma_dev->channels);
dma_dev          1018 drivers/dma/sh/shdma-base.c 	dma_dev->device_alloc_chan_resources
dma_dev          1020 drivers/dma/sh/shdma-base.c 	dma_dev->device_free_chan_resources = shdma_free_chan_resources;
dma_dev          1021 drivers/dma/sh/shdma-base.c 	dma_dev->device_prep_dma_memcpy = shdma_prep_memcpy;
dma_dev          1022 drivers/dma/sh/shdma-base.c 	dma_dev->device_tx_status = shdma_tx_status;
dma_dev          1023 drivers/dma/sh/shdma-base.c 	dma_dev->device_issue_pending = shdma_issue_pending;
dma_dev          1026 drivers/dma/sh/shdma-base.c 	dma_dev->device_prep_slave_sg = shdma_prep_slave_sg;
dma_dev          1027 drivers/dma/sh/shdma-base.c 	dma_dev->device_prep_dma_cyclic = shdma_prep_dma_cyclic;
dma_dev          1028 drivers/dma/sh/shdma-base.c 	dma_dev->device_config = shdma_config;
dma_dev          1029 drivers/dma/sh/shdma-base.c 	dma_dev->device_terminate_all = shdma_terminate_all;
dma_dev          1031 drivers/dma/sh/shdma-base.c 	dma_dev->dev = dev;
dma_dev            59 drivers/dma/sh/shdma.h 				     struct sh_dmae_device, shdma_dev.dma_dev)
dma_dev           165 drivers/dma/sh/shdmac.c 		dev_warn(shdev->shdma_dev.dma_dev.dev, "Can't initialize DMAOR.\n");
dma_dev           169 drivers/dma/sh/shdmac.c 		dev_warn(shdev->shdma_dev.dma_dev.dev,
dma_dev           522 drivers/dma/sh/shdmac.c 	struct platform_device *pdev = to_platform_device(sdev->dma_dev.dev);
dma_dev           527 drivers/dma/sh/shdmac.c 	sh_chan = devm_kzalloc(sdev->dma_dev.dev, sizeof(struct sh_dmae_chan),
dma_dev           549 drivers/dma/sh/shdmac.c 		dev_err(sdev->dma_dev.dev,
dma_dev           680 drivers/dma/sh/shdmac.c 	struct dma_device *dma_dev;
dma_dev           720 drivers/dma/sh/shdmac.c 	dma_dev = &shdev->shdma_dev.dma_dev;
dma_dev           731 drivers/dma/sh/shdmac.c 	dma_dev->src_addr_widths = widths;
dma_dev           732 drivers/dma/sh/shdmac.c 	dma_dev->dst_addr_widths = widths;
dma_dev           733 drivers/dma/sh/shdmac.c 	dma_dev->directions = BIT(DMA_MEM_TO_DEV) | BIT(DMA_DEV_TO_MEM);
dma_dev           734 drivers/dma/sh/shdmac.c 	dma_dev->residue_granularity = DMA_RESIDUE_GRANULARITY_DESCRIPTOR;
dma_dev           737 drivers/dma/sh/shdmac.c 		dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask);
dma_dev           739 drivers/dma/sh/shdmac.c 		dma_cap_set(DMA_SLAVE, dma_dev->cap_mask);
dma_dev           742 drivers/dma/sh/shdmac.c 	dma_dev->copy_align = LOG2_DEFAULT_XFER_SIZE;
dma_dev           859 drivers/dma/sh/shdmac.c 	err = dma_async_device_register(&shdev->shdma_dev.dma_dev);
dma_dev           890 drivers/dma/sh/shdmac.c 	struct dma_device *dma_dev = &shdev->shdma_dev.dma_dev;
dma_dev           892 drivers/dma/sh/shdmac.c 	dma_async_device_unregister(dma_dev);
dma_dev           206 drivers/dma/sprd-dma.c 	struct dma_device	dma_dev;
dma_dev           355 drivers/dma/sprd-dma.c 			dev_warn(sdev->dma_dev.dev,
dma_dev           419 drivers/dma/sprd-dma.c 		dev_warn(sdev->dma_dev.dev, "incorrect dma interrupt type\n");
dma_dev           478 drivers/dma/sprd-dma.c 		dev_err(sdev->dma_dev.dev, "invalid channel mode setting %d\n",
dma_dev           737 drivers/dma/sprd-dma.c 			dev_err(sdev->dma_dev.dev, "invalid source step\n");
dma_dev           753 drivers/dma/sprd-dma.c 			dev_err(sdev->dma_dev.dev, "invalid destination step\n");
dma_dev           761 drivers/dma/sprd-dma.c 		dev_err(sdev->dma_dev.dev, "invalid source datawidth\n");
dma_dev           767 drivers/dma/sprd-dma.c 		dev_err(sdev->dma_dev.dev, "invalid destination datawidth\n");
dma_dev          1134 drivers/dma/sprd-dma.c 	dma_cap_set(DMA_MEMCPY, sdev->dma_dev.cap_mask);
dma_dev          1136 drivers/dma/sprd-dma.c 	sdev->dma_dev.chancnt = chn_count;
dma_dev          1137 drivers/dma/sprd-dma.c 	INIT_LIST_HEAD(&sdev->dma_dev.channels);
dma_dev          1138 drivers/dma/sprd-dma.c 	INIT_LIST_HEAD(&sdev->dma_dev.global_node);
dma_dev          1139 drivers/dma/sprd-dma.c 	sdev->dma_dev.dev = &pdev->dev;
dma_dev          1140 drivers/dma/sprd-dma.c 	sdev->dma_dev.device_alloc_chan_resources = sprd_dma_alloc_chan_resources;
dma_dev          1141 drivers/dma/sprd-dma.c 	sdev->dma_dev.device_free_chan_resources = sprd_dma_free_chan_resources;
dma_dev          1142 drivers/dma/sprd-dma.c 	sdev->dma_dev.device_tx_status = sprd_dma_tx_status;
dma_dev          1143 drivers/dma/sprd-dma.c 	sdev->dma_dev.device_issue_pending = sprd_dma_issue_pending;
dma_dev          1144 drivers/dma/sprd-dma.c 	sdev->dma_dev.device_prep_dma_memcpy = sprd_dma_prep_dma_memcpy;
dma_dev          1145 drivers/dma/sprd-dma.c 	sdev->dma_dev.device_prep_slave_sg = sprd_dma_prep_slave_sg;
dma_dev          1146 drivers/dma/sprd-dma.c 	sdev->dma_dev.device_config = sprd_dma_slave_config;
dma_dev          1147 drivers/dma/sprd-dma.c 	sdev->dma_dev.device_pause = sprd_dma_pause;
dma_dev          1148 drivers/dma/sprd-dma.c 	sdev->dma_dev.device_resume = sprd_dma_resume;
dma_dev          1149 drivers/dma/sprd-dma.c 	sdev->dma_dev.device_terminate_all = sprd_dma_terminate_all;
dma_dev          1160 drivers/dma/sprd-dma.c 		vchan_init(&dma_chn->vc, &sdev->dma_dev);
dma_dev          1175 drivers/dma/sprd-dma.c 	ret = dma_async_device_register(&sdev->dma_dev);
dma_dev          1181 drivers/dma/sprd-dma.c 	sprd_dma_info.dma_cap = sdev->dma_dev.cap_mask;
dma_dev          1191 drivers/dma/sprd-dma.c 	dma_async_device_unregister(&sdev->dma_dev);
dma_dev          1214 drivers/dma/sprd-dma.c 	list_for_each_entry_safe(c, cn, &sdev->dma_dev.channels,
dma_dev          1221 drivers/dma/sprd-dma.c 	dma_async_device_unregister(&sdev->dma_dev);
dma_dev          1249 drivers/dma/sprd-dma.c 		dev_err(sdev->dma_dev.dev, "enable dma failed\n");
dma_dev           210 drivers/dma/tegra20-apb-dma.c 	struct dma_device		dma_dev;
dma_dev          1331 drivers/dma/tegra20-apb-dma.c 	chan = dma_get_any_slave_channel(&tdma->dma_dev);
dma_dev          1443 drivers/dma/tegra20-apb-dma.c 	INIT_LIST_HEAD(&tdma->dma_dev.channels);
dma_dev          1467 drivers/dma/tegra20-apb-dma.c 		tdc->dma_chan.device = &tdma->dma_dev;
dma_dev          1470 drivers/dma/tegra20-apb-dma.c 				&tdma->dma_dev.channels);
dma_dev          1485 drivers/dma/tegra20-apb-dma.c 	dma_cap_set(DMA_SLAVE, tdma->dma_dev.cap_mask);
dma_dev          1486 drivers/dma/tegra20-apb-dma.c 	dma_cap_set(DMA_PRIVATE, tdma->dma_dev.cap_mask);
dma_dev          1487 drivers/dma/tegra20-apb-dma.c 	dma_cap_set(DMA_CYCLIC, tdma->dma_dev.cap_mask);
dma_dev          1490 drivers/dma/tegra20-apb-dma.c 	tdma->dma_dev.dev = &pdev->dev;
dma_dev          1491 drivers/dma/tegra20-apb-dma.c 	tdma->dma_dev.device_alloc_chan_resources =
dma_dev          1493 drivers/dma/tegra20-apb-dma.c 	tdma->dma_dev.device_free_chan_resources =
dma_dev          1495 drivers/dma/tegra20-apb-dma.c 	tdma->dma_dev.device_prep_slave_sg = tegra_dma_prep_slave_sg;
dma_dev          1496 drivers/dma/tegra20-apb-dma.c 	tdma->dma_dev.device_prep_dma_cyclic = tegra_dma_prep_dma_cyclic;
dma_dev          1497 drivers/dma/tegra20-apb-dma.c 	tdma->dma_dev.src_addr_widths = BIT(DMA_SLAVE_BUSWIDTH_1_BYTE) |
dma_dev          1501 drivers/dma/tegra20-apb-dma.c 	tdma->dma_dev.dst_addr_widths = BIT(DMA_SLAVE_BUSWIDTH_1_BYTE) |
dma_dev          1505 drivers/dma/tegra20-apb-dma.c 	tdma->dma_dev.directions = BIT(DMA_DEV_TO_MEM) | BIT(DMA_MEM_TO_DEV);
dma_dev          1506 drivers/dma/tegra20-apb-dma.c 	tdma->dma_dev.residue_granularity = DMA_RESIDUE_GRANULARITY_BURST;
dma_dev          1507 drivers/dma/tegra20-apb-dma.c 	tdma->dma_dev.device_config = tegra_dma_slave_config;
dma_dev          1508 drivers/dma/tegra20-apb-dma.c 	tdma->dma_dev.device_terminate_all = tegra_dma_terminate_all;
dma_dev          1509 drivers/dma/tegra20-apb-dma.c 	tdma->dma_dev.device_tx_status = tegra_dma_tx_status;
dma_dev          1510 drivers/dma/tegra20-apb-dma.c 	tdma->dma_dev.device_issue_pending = tegra_dma_issue_pending;
dma_dev          1512 drivers/dma/tegra20-apb-dma.c 	ret = dma_async_device_register(&tdma->dma_dev);
dma_dev          1532 drivers/dma/tegra20-apb-dma.c 	dma_async_device_unregister(&tdma->dma_dev);
dma_dev          1553 drivers/dma/tegra20-apb-dma.c 	dma_async_device_unregister(&tdma->dma_dev);
dma_dev           153 drivers/dma/tegra210-adma.c 	struct dma_device		dma_dev;
dma_dev           702 drivers/dma/tegra210-adma.c 	chan = dma_get_any_slave_channel(&tdma->dma_dev);
dma_dev           851 drivers/dma/tegra210-adma.c 	INIT_LIST_HEAD(&tdma->dma_dev.channels);
dma_dev           864 drivers/dma/tegra210-adma.c 		vchan_init(&tdc->vc, &tdma->dma_dev);
dma_dev           879 drivers/dma/tegra210-adma.c 	dma_cap_set(DMA_SLAVE, tdma->dma_dev.cap_mask);
dma_dev           880 drivers/dma/tegra210-adma.c 	dma_cap_set(DMA_PRIVATE, tdma->dma_dev.cap_mask);
dma_dev           881 drivers/dma/tegra210-adma.c 	dma_cap_set(DMA_CYCLIC, tdma->dma_dev.cap_mask);
dma_dev           883 drivers/dma/tegra210-adma.c 	tdma->dma_dev.dev = &pdev->dev;
dma_dev           884 drivers/dma/tegra210-adma.c 	tdma->dma_dev.device_alloc_chan_resources =
dma_dev           886 drivers/dma/tegra210-adma.c 	tdma->dma_dev.device_free_chan_resources =
dma_dev           888 drivers/dma/tegra210-adma.c 	tdma->dma_dev.device_issue_pending = tegra_adma_issue_pending;
dma_dev           889 drivers/dma/tegra210-adma.c 	tdma->dma_dev.device_prep_dma_cyclic = tegra_adma_prep_dma_cyclic;
dma_dev           890 drivers/dma/tegra210-adma.c 	tdma->dma_dev.device_config = tegra_adma_slave_config;
dma_dev           891 drivers/dma/tegra210-adma.c 	tdma->dma_dev.device_tx_status = tegra_adma_tx_status;
dma_dev           892 drivers/dma/tegra210-adma.c 	tdma->dma_dev.device_terminate_all = tegra_adma_terminate_all;
dma_dev           893 drivers/dma/tegra210-adma.c 	tdma->dma_dev.src_addr_widths = BIT(DMA_SLAVE_BUSWIDTH_4_BYTES);
dma_dev           894 drivers/dma/tegra210-adma.c 	tdma->dma_dev.dst_addr_widths = BIT(DMA_SLAVE_BUSWIDTH_4_BYTES);
dma_dev           895 drivers/dma/tegra210-adma.c 	tdma->dma_dev.directions = BIT(DMA_DEV_TO_MEM) | BIT(DMA_MEM_TO_DEV);
dma_dev           896 drivers/dma/tegra210-adma.c 	tdma->dma_dev.residue_granularity = DMA_RESIDUE_GRANULARITY_SEGMENT;
dma_dev           897 drivers/dma/tegra210-adma.c 	tdma->dma_dev.device_pause = tegra_adma_pause;
dma_dev           898 drivers/dma/tegra210-adma.c 	tdma->dma_dev.device_resume = tegra_adma_resume;
dma_dev           900 drivers/dma/tegra210-adma.c 	ret = dma_async_device_register(&tdma->dma_dev);
dma_dev           921 drivers/dma/tegra210-adma.c 	dma_async_device_unregister(&tdma->dma_dev);
dma_dev           939 drivers/dma/tegra210-adma.c 	dma_async_device_unregister(&tdma->dma_dev);
dma_dev           308 drivers/dma/xgene-dma.c 	struct dma_device dma_dev[XGENE_DMA_MAX_CHANNEL];
dma_dev          1483 drivers/dma/xgene-dma.c 			       struct dma_device *dma_dev)
dma_dev          1486 drivers/dma/xgene-dma.c 	dma_cap_zero(dma_dev->cap_mask);
dma_dev          1501 drivers/dma/xgene-dma.c 		dma_cap_set(DMA_PQ, dma_dev->cap_mask);
dma_dev          1502 drivers/dma/xgene-dma.c 		dma_cap_set(DMA_XOR, dma_dev->cap_mask);
dma_dev          1505 drivers/dma/xgene-dma.c 		dma_cap_set(DMA_XOR, dma_dev->cap_mask);
dma_dev          1509 drivers/dma/xgene-dma.c 	dma_dev->dev = chan->dev;
dma_dev          1510 drivers/dma/xgene-dma.c 	dma_dev->device_alloc_chan_resources = xgene_dma_alloc_chan_resources;
dma_dev          1511 drivers/dma/xgene-dma.c 	dma_dev->device_free_chan_resources = xgene_dma_free_chan_resources;
dma_dev          1512 drivers/dma/xgene-dma.c 	dma_dev->device_issue_pending = xgene_dma_issue_pending;
dma_dev          1513 drivers/dma/xgene-dma.c 	dma_dev->device_tx_status = xgene_dma_tx_status;
dma_dev          1515 drivers/dma/xgene-dma.c 	if (dma_has_cap(DMA_XOR, dma_dev->cap_mask)) {
dma_dev          1516 drivers/dma/xgene-dma.c 		dma_dev->device_prep_dma_xor = xgene_dma_prep_xor;
dma_dev          1517 drivers/dma/xgene-dma.c 		dma_dev->max_xor = XGENE_DMA_MAX_XOR_SRC;
dma_dev          1518 drivers/dma/xgene-dma.c 		dma_dev->xor_align = DMAENGINE_ALIGN_64_BYTES;
dma_dev          1521 drivers/dma/xgene-dma.c 	if (dma_has_cap(DMA_PQ, dma_dev->cap_mask)) {
dma_dev          1522 drivers/dma/xgene-dma.c 		dma_dev->device_prep_dma_pq = xgene_dma_prep_pq;
dma_dev          1523 drivers/dma/xgene-dma.c 		dma_dev->max_pq = XGENE_DMA_MAX_XOR_SRC;
dma_dev          1524 drivers/dma/xgene-dma.c 		dma_dev->pq_align = DMAENGINE_ALIGN_64_BYTES;
dma_dev          1531 drivers/dma/xgene-dma.c 	struct dma_device *dma_dev = &pdma->dma_dev[id];
dma_dev          1534 drivers/dma/xgene-dma.c 	chan->dma_chan.device = dma_dev;
dma_dev          1548 drivers/dma/xgene-dma.c 	xgene_dma_set_caps(chan, dma_dev);
dma_dev          1551 drivers/dma/xgene-dma.c 	INIT_LIST_HEAD(&dma_dev->channels);
dma_dev          1552 drivers/dma/xgene-dma.c 	list_add_tail(&chan->dma_chan.device_node, &dma_dev->channels);
dma_dev          1555 drivers/dma/xgene-dma.c 	ret = dma_async_device_register(dma_dev);
dma_dev          1566 drivers/dma/xgene-dma.c 		 dma_has_cap(DMA_XOR, dma_dev->cap_mask) ? "XOR " : "",
dma_dev          1567 drivers/dma/xgene-dma.c 		 dma_has_cap(DMA_PQ, dma_dev->cap_mask) ? "PQ " : "");
dma_dev          1580 drivers/dma/xgene-dma.c 				dma_async_device_unregister(&pdma->dma_dev[j]);
dma_dev          1596 drivers/dma/xgene-dma.c 		dma_async_device_unregister(&pdma->dma_dev[i]);
dma_dev            66 drivers/gpu/drm/exynos/exynos_drm_dma.c 	if (get_dma_ops(priv->dma_dev) != get_dma_ops(subdrv_dev)) {
dma_dev           126 drivers/gpu/drm/exynos/exynos_drm_dma.c 	if (!priv->dma_dev) {
dma_dev           127 drivers/gpu/drm/exynos/exynos_drm_dma.c 		priv->dma_dev = dev;
dma_dev           142 drivers/gpu/drm/exynos/exynos_drm_dma.c 			mapping = iommu_get_domain_for_dev(priv->dma_dev);
dma_dev           168 drivers/gpu/drm/exynos/exynos_drm_dma.c 	priv->dma_dev = NULL;
dma_dev           203 drivers/gpu/drm/exynos/exynos_drm_drv.h 	struct device *dma_dev;
dma_dev           216 drivers/gpu/drm/exynos/exynos_drm_drv.h 	return priv->dma_dev;
dma_dev           211 drivers/gpu/drm/mediatek/mtk_drm_drv.c 	struct device *dma_dev;
dma_dev           274 drivers/gpu/drm/mediatek/mtk_drm_drv.c 	dma_dev = &pdev->dev;
dma_dev           275 drivers/gpu/drm/mediatek/mtk_drm_drv.c 	private->dma_dev = dma_dev;
dma_dev           281 drivers/gpu/drm/mediatek/mtk_drm_drv.c 	if (!dma_dev->dma_parms) {
dma_dev           283 drivers/gpu/drm/mediatek/mtk_drm_drv.c 		dma_dev->dma_parms =
dma_dev           284 drivers/gpu/drm/mediatek/mtk_drm_drv.c 			devm_kzalloc(drm->dev, sizeof(*dma_dev->dma_parms),
dma_dev           287 drivers/gpu/drm/mediatek/mtk_drm_drv.c 	if (!dma_dev->dma_parms) {
dma_dev           292 drivers/gpu/drm/mediatek/mtk_drm_drv.c 	ret = dma_set_max_seg_size(dma_dev, (unsigned int)DMA_BIT_MASK(32));
dma_dev           294 drivers/gpu/drm/mediatek/mtk_drm_drv.c 		dev_err(dma_dev, "Failed to set DMA segment size\n");
dma_dev           315 drivers/gpu/drm/mediatek/mtk_drm_drv.c 		dma_dev->dma_parms = NULL;
dma_dev           332 drivers/gpu/drm/mediatek/mtk_drm_drv.c 		private->dma_dev->dma_parms = NULL;
dma_dev           358 drivers/gpu/drm/mediatek/mtk_drm_drv.c 	return drm_gem_prime_import_dev(dev, dma_buf, private->dma_dev);
dma_dev            36 drivers/gpu/drm/mediatek/mtk_drm_drv.h 	struct device *dma_dev;
dma_dev            57 drivers/gpu/drm/mediatek/mtk_drm_gem.c 	mtk_gem->cookie = dma_alloc_attrs(priv->dma_dev, obj->size,
dma_dev            89 drivers/gpu/drm/mediatek/mtk_drm_gem.c 		dma_free_attrs(priv->dma_dev, obj->size, mtk_gem->cookie,
dma_dev           143 drivers/gpu/drm/mediatek/mtk_drm_gem.c 	ret = dma_mmap_attrs(priv->dma_dev, vma, mtk_gem->cookie,
dma_dev           199 drivers/gpu/drm/mediatek/mtk_drm_gem.c 	ret = dma_get_sgtable_attrs(priv->dma_dev, sgt, mtk_gem->cookie,
dma_dev            85 drivers/gpu/drm/omapdrm/omap_dmm_tiler.c 	struct dma_device *dma_dev = dmm->wa_dma_chan->device;
dma_dev            90 drivers/gpu/drm/omapdrm/omap_dmm_tiler.c 	tx = dma_dev->device_prep_dma_memcpy(dmm->wa_dma_chan, dst, src, 4, 0);
dma_dev           178 drivers/iommu/exynos-iommu.c static struct device *dma_dev;
dma_dev           659 drivers/iommu/exynos-iommu.c 	if (!dma_dev)
dma_dev           660 drivers/iommu/exynos-iommu.c 		dma_dev = &pdev->dev;
dma_dev           726 drivers/iommu/exynos-iommu.c 	dma_sync_single_for_cpu(dma_dev, virt_to_phys(ent), sizeof(*ent),
dma_dev           729 drivers/iommu/exynos-iommu.c 	dma_sync_single_for_device(dma_dev, virt_to_phys(ent), sizeof(*ent),
dma_dev           740 drivers/iommu/exynos-iommu.c 	BUG_ON(PG_ENT_SHIFT < 0 || !dma_dev);
dma_dev           765 drivers/iommu/exynos-iommu.c 	handle = dma_map_single(dma_dev, domain->pgtable, LV1TABLE_SIZE,
dma_dev           769 drivers/iommu/exynos-iommu.c 	if (dma_mapping_error(dma_dev, handle))
dma_dev           819 drivers/iommu/exynos-iommu.c 	dma_unmap_single(dma_dev, virt_to_phys(domain->pgtable), LV1TABLE_SIZE,
dma_dev           826 drivers/iommu/exynos-iommu.c 			dma_unmap_single(dma_dev, base, LV2TABLE_SIZE,
dma_dev           939 drivers/iommu/exynos-iommu.c 		handle = dma_map_single(dma_dev, pent, LV2TABLE_SIZE,
dma_dev           941 drivers/iommu/exynos-iommu.c 		if (dma_mapping_error(dma_dev, handle)) {
dma_dev          1027 drivers/iommu/exynos-iommu.c 		dma_sync_single_for_cpu(dma_dev, pent_base,
dma_dev          1039 drivers/iommu/exynos-iommu.c 		dma_sync_single_for_device(dma_dev, pent_base,
dma_dev          1186 drivers/iommu/exynos-iommu.c 	dma_sync_single_for_cpu(dma_dev, virt_to_phys(ent),
dma_dev          1190 drivers/iommu/exynos-iommu.c 	dma_sync_single_for_device(dma_dev, virt_to_phys(ent),
dma_dev           118 drivers/iommu/rockchip-iommu.c static struct device *dma_dev;
dma_dev           125 drivers/iommu/rockchip-iommu.c 	dma_sync_single_for_device(dma_dev, dma, size, DMA_TO_DEVICE);
dma_dev           675 drivers/iommu/rockchip-iommu.c 	pt_dma = dma_map_single(dma_dev, page_table, SPAGE_SIZE, DMA_TO_DEVICE);
dma_dev           676 drivers/iommu/rockchip-iommu.c 	if (dma_mapping_error(dma_dev, pt_dma)) {
dma_dev           677 drivers/iommu/rockchip-iommu.c 		dev_err(dma_dev, "DMA mapping error while allocating page table\n");
dma_dev           980 drivers/iommu/rockchip-iommu.c 	if (!dma_dev)
dma_dev          1000 drivers/iommu/rockchip-iommu.c 	rk_domain->dt_dma = dma_map_single(dma_dev, rk_domain->dt,
dma_dev          1002 drivers/iommu/rockchip-iommu.c 	if (dma_mapping_error(dma_dev, rk_domain->dt_dma)) {
dma_dev          1003 drivers/iommu/rockchip-iommu.c 		dev_err(dma_dev, "DMA map error for DT\n");
dma_dev          1042 drivers/iommu/rockchip-iommu.c 			dma_unmap_single(dma_dev, pt_phys,
dma_dev          1048 drivers/iommu/rockchip-iommu.c 	dma_unmap_single(dma_dev, rk_domain->dt_dma,
dma_dev          1108 drivers/iommu/rockchip-iommu.c 	data = devm_kzalloc(dma_dev, sizeof(*data), GFP_KERNEL);
dma_dev          1223 drivers/iommu/rockchip-iommu.c 	if (!dma_dev)
dma_dev          1224 drivers/iommu/rockchip-iommu.c 		dma_dev = &pdev->dev;
dma_dev           233 drivers/mcb/mcb-core.c 	dev->dma_dev = bus->carrier;
dma_dev           129 drivers/mmc/host/mmc_spi.c 	struct device		*dma_dev;
dma_dev           168 drivers/mmc/host/mmc_spi.c 	if (host->dma_dev)
dma_dev           169 drivers/mmc/host/mmc_spi.c 		dma_sync_single_for_device(host->dma_dev,
dma_dev           175 drivers/mmc/host/mmc_spi.c 	if (host->dma_dev)
dma_dev           176 drivers/mmc/host/mmc_spi.c 		dma_sync_single_for_cpu(host->dma_dev,
dma_dev           523 drivers/mmc/host/mmc_spi.c 	if (host->dma_dev) {
dma_dev           525 drivers/mmc/host/mmc_spi.c 		dma_sync_single_for_device(host->dma_dev,
dma_dev           531 drivers/mmc/host/mmc_spi.c 	if (host->dma_dev)
dma_dev           532 drivers/mmc/host/mmc_spi.c 		dma_sync_single_for_cpu(host->dma_dev,
dma_dev           665 drivers/mmc/host/mmc_spi.c 	if (host->dma_dev)
dma_dev           666 drivers/mmc/host/mmc_spi.c 		dma_sync_single_for_device(host->dma_dev,
dma_dev           677 drivers/mmc/host/mmc_spi.c 	if (host->dma_dev)
dma_dev           678 drivers/mmc/host/mmc_spi.c 		dma_sync_single_for_cpu(host->dma_dev,
dma_dev           730 drivers/mmc/host/mmc_spi.c 	if (host->dma_dev)
dma_dev           795 drivers/mmc/host/mmc_spi.c 	if (host->dma_dev) {
dma_dev           796 drivers/mmc/host/mmc_spi.c 		dma_sync_single_for_device(host->dma_dev,
dma_dev           799 drivers/mmc/host/mmc_spi.c 		dma_sync_single_for_device(host->dma_dev,
dma_dev           810 drivers/mmc/host/mmc_spi.c 	if (host->dma_dev) {
dma_dev           811 drivers/mmc/host/mmc_spi.c 		dma_sync_single_for_cpu(host->dma_dev,
dma_dev           814 drivers/mmc/host/mmc_spi.c 		dma_sync_single_for_cpu(host->dma_dev,
dma_dev           853 drivers/mmc/host/mmc_spi.c 	if (host->dma_dev)
dma_dev           869 drivers/mmc/host/mmc_spi.c 	struct device		*dma_dev = host->dma_dev;
dma_dev           904 drivers/mmc/host/mmc_spi.c 		if (dma_dev) {
dma_dev           910 drivers/mmc/host/mmc_spi.c 			dma_addr = dma_map_page(dma_dev, sg_page(sg), 0,
dma_dev           912 drivers/mmc/host/mmc_spi.c 			if (dma_mapping_error(dma_dev, dma_addr)) {
dma_dev           956 drivers/mmc/host/mmc_spi.c 		if (dma_dev)
dma_dev           957 drivers/mmc/host/mmc_spi.c 			dma_unmap_page(dma_dev, dma_addr, PAGE_SIZE, dir);
dma_dev           997 drivers/mmc/host/mmc_spi.c 		if (host->dma_dev)
dma_dev           998 drivers/mmc/host/mmc_spi.c 			dma_sync_single_for_device(host->dma_dev,
dma_dev          1004 drivers/mmc/host/mmc_spi.c 		if (host->dma_dev)
dma_dev          1005 drivers/mmc/host/mmc_spi.c 			dma_sync_single_for_cpu(host->dma_dev,
dma_dev          1382 drivers/mmc/host/mmc_spi.c 		host->dma_dev = dev;
dma_dev          1392 drivers/mmc/host/mmc_spi.c 		dma_sync_single_for_cpu(host->dma_dev,
dma_dev          1399 drivers/mmc/host/mmc_spi.c 	host->readback.is_dma_mapped = (host->dma_dev != NULL);
dma_dev          1452 drivers/mmc/host/mmc_spi.c 			host->dma_dev ? "" : ", no DMA",
dma_dev          1463 drivers/mmc/host/mmc_spi.c 	if (host->dma_dev)
dma_dev          1464 drivers/mmc/host/mmc_spi.c 		dma_unmap_single(host->dma_dev, host->data_dma,
dma_dev          1467 drivers/mmc/host/mmc_spi.c 	if (host->dma_dev)
dma_dev          1468 drivers/mmc/host/mmc_spi.c 		dma_unmap_single(host->dma_dev, host->ones_dma,
dma_dev          1494 drivers/mmc/host/mmc_spi.c 	if (host->dma_dev) {
dma_dev          1495 drivers/mmc/host/mmc_spi.c 		dma_unmap_single(host->dma_dev, host->ones_dma,
dma_dev          1497 drivers/mmc/host/mmc_spi.c 		dma_unmap_single(host->dma_dev, host->data_dma,
dma_dev           461 drivers/mtd/nand/raw/fsmc_nand.c 	struct dma_device *dma_dev;
dma_dev           476 drivers/mtd/nand/raw/fsmc_nand.c 	dma_dev = chan->device;
dma_dev           477 drivers/mtd/nand/raw/fsmc_nand.c 	dma_addr = dma_map_single(dma_dev->dev, buffer, len, direction);
dma_dev           487 drivers/mtd/nand/raw/fsmc_nand.c 	tx = dma_dev->device_prep_dma_memcpy(chan, dma_dst, dma_src,
dma_dev           520 drivers/mtd/nand/raw/fsmc_nand.c 	dma_unmap_single(dma_dev->dev, dma_addr, len, direction);
dma_dev           153 drivers/net/ethernet/broadcom/b44.c 	dma_sync_single_for_device(sdev->dma_dev, dma_base + offset,
dma_dev           162 drivers/net/ethernet/broadcom/b44.c 	dma_sync_single_for_cpu(sdev->dma_dev, dma_base + offset,
dma_dev           632 drivers/net/ethernet/broadcom/b44.c 		dma_unmap_single(bp->sdev->dma_dev,
dma_dev           677 drivers/net/ethernet/broadcom/b44.c 	mapping = dma_map_single(bp->sdev->dma_dev, skb->data,
dma_dev           683 drivers/net/ethernet/broadcom/b44.c 	if (dma_mapping_error(bp->sdev->dma_dev, mapping) ||
dma_dev           686 drivers/net/ethernet/broadcom/b44.c 		if (!dma_mapping_error(bp->sdev->dma_dev, mapping))
dma_dev           687 drivers/net/ethernet/broadcom/b44.c 			dma_unmap_single(bp->sdev->dma_dev, mapping,
dma_dev           693 drivers/net/ethernet/broadcom/b44.c 		mapping = dma_map_single(bp->sdev->dma_dev, skb->data,
dma_dev           696 drivers/net/ethernet/broadcom/b44.c 		if (dma_mapping_error(bp->sdev->dma_dev, mapping) ||
dma_dev           698 drivers/net/ethernet/broadcom/b44.c 			if (!dma_mapping_error(bp->sdev->dma_dev, mapping))
dma_dev           699 drivers/net/ethernet/broadcom/b44.c 				dma_unmap_single(bp->sdev->dma_dev, mapping, RX_PKT_BUF_SZ,DMA_FROM_DEVICE);
dma_dev           774 drivers/net/ethernet/broadcom/b44.c 	dma_sync_single_for_device(bp->sdev->dma_dev, dest_map->mapping,
dma_dev           796 drivers/net/ethernet/broadcom/b44.c 		dma_sync_single_for_cpu(bp->sdev->dma_dev, map,
dma_dev           830 drivers/net/ethernet/broadcom/b44.c 			dma_unmap_single(bp->sdev->dma_dev, map,
dma_dev           991 drivers/net/ethernet/broadcom/b44.c 	mapping = dma_map_single(bp->sdev->dma_dev, skb->data, len, DMA_TO_DEVICE);
dma_dev           992 drivers/net/ethernet/broadcom/b44.c 	if (dma_mapping_error(bp->sdev->dma_dev, mapping) || mapping + len > DMA_BIT_MASK(30)) {
dma_dev           996 drivers/net/ethernet/broadcom/b44.c 		if (!dma_mapping_error(bp->sdev->dma_dev, mapping))
dma_dev           997 drivers/net/ethernet/broadcom/b44.c 			dma_unmap_single(bp->sdev->dma_dev, mapping, len,
dma_dev          1004 drivers/net/ethernet/broadcom/b44.c 		mapping = dma_map_single(bp->sdev->dma_dev, bounce_skb->data,
dma_dev          1006 drivers/net/ethernet/broadcom/b44.c 		if (dma_mapping_error(bp->sdev->dma_dev, mapping) || mapping + len > DMA_BIT_MASK(30)) {
dma_dev          1007 drivers/net/ethernet/broadcom/b44.c 			if (!dma_mapping_error(bp->sdev->dma_dev, mapping))
dma_dev          1008 drivers/net/ethernet/broadcom/b44.c 				dma_unmap_single(bp->sdev->dma_dev, mapping,
dma_dev          1104 drivers/net/ethernet/broadcom/b44.c 		dma_unmap_single(bp->sdev->dma_dev, rp->mapping, RX_PKT_BUF_SZ,
dma_dev          1116 drivers/net/ethernet/broadcom/b44.c 		dma_unmap_single(bp->sdev->dma_dev, rp->mapping, rp->skb->len,
dma_dev          1139 drivers/net/ethernet/broadcom/b44.c 		dma_sync_single_for_device(bp->sdev->dma_dev, bp->rx_ring_dma,
dma_dev          1143 drivers/net/ethernet/broadcom/b44.c 		dma_sync_single_for_device(bp->sdev->dma_dev, bp->tx_ring_dma,
dma_dev          1164 drivers/net/ethernet/broadcom/b44.c 			dma_unmap_single(bp->sdev->dma_dev, bp->rx_ring_dma,
dma_dev          1168 drivers/net/ethernet/broadcom/b44.c 			dma_free_coherent(bp->sdev->dma_dev, DMA_TABLE_BYTES,
dma_dev          1175 drivers/net/ethernet/broadcom/b44.c 			dma_unmap_single(bp->sdev->dma_dev, bp->tx_ring_dma,
dma_dev          1179 drivers/net/ethernet/broadcom/b44.c 			dma_free_coherent(bp->sdev->dma_dev, DMA_TABLE_BYTES,
dma_dev          1205 drivers/net/ethernet/broadcom/b44.c 	bp->rx_ring = dma_alloc_coherent(bp->sdev->dma_dev, size,
dma_dev          1218 drivers/net/ethernet/broadcom/b44.c 		rx_ring_dma = dma_map_single(bp->sdev->dma_dev, rx_ring,
dma_dev          1222 drivers/net/ethernet/broadcom/b44.c 		if (dma_mapping_error(bp->sdev->dma_dev, rx_ring_dma) ||
dma_dev          1233 drivers/net/ethernet/broadcom/b44.c 	bp->tx_ring = dma_alloc_coherent(bp->sdev->dma_dev, size,
dma_dev          1246 drivers/net/ethernet/broadcom/b44.c 		tx_ring_dma = dma_map_single(bp->sdev->dma_dev, tx_ring,
dma_dev          1250 drivers/net/ethernet/broadcom/b44.c 		if (dma_mapping_error(bp->sdev->dma_dev, tx_ring_dma) ||
dma_dev          2394 drivers/net/ethernet/broadcom/b44.c 	if (dma_set_mask_and_coherent(sdev->dma_dev, DMA_BIT_MASK(30))) {
dma_dev           126 drivers/net/ethernet/broadcom/bgmac-bcma.c 	bgmac->dma_dev = core->dma_dev;
dma_dev           193 drivers/net/ethernet/broadcom/bgmac-platform.c 	bgmac->dma_dev = &pdev->dev;
dma_dev           135 drivers/net/ethernet/broadcom/bgmac.c 	struct device *dma_dev = bgmac->dma_dev;
dma_dev           162 drivers/net/ethernet/broadcom/bgmac.c 	slot->dma_addr = dma_map_single(dma_dev, skb->data, skb_headlen(skb),
dma_dev           164 drivers/net/ethernet/broadcom/bgmac.c 	if (unlikely(dma_mapping_error(dma_dev, slot->dma_addr)))
dma_dev           180 drivers/net/ethernet/broadcom/bgmac.c 		slot->dma_addr = skb_frag_dma_map(dma_dev, frag, 0,
dma_dev           182 drivers/net/ethernet/broadcom/bgmac.c 		if (unlikely(dma_mapping_error(dma_dev, slot->dma_addr)))
dma_dev           211 drivers/net/ethernet/broadcom/bgmac.c 	dma_unmap_single(dma_dev, slot->dma_addr, skb_headlen(skb),
dma_dev           220 drivers/net/ethernet/broadcom/bgmac.c 		dma_unmap_page(dma_dev, slot->dma_addr, len, DMA_TO_DEVICE);
dma_dev           237 drivers/net/ethernet/broadcom/bgmac.c 	struct device *dma_dev = bgmac->dma_dev;
dma_dev           262 drivers/net/ethernet/broadcom/bgmac.c 			dma_unmap_single(dma_dev, slot->dma_addr, len,
dma_dev           265 drivers/net/ethernet/broadcom/bgmac.c 			dma_unmap_page(dma_dev, slot->dma_addr, len,
dma_dev           336 drivers/net/ethernet/broadcom/bgmac.c 	struct device *dma_dev = bgmac->dma_dev;
dma_dev           352 drivers/net/ethernet/broadcom/bgmac.c 	dma_addr = dma_map_single(dma_dev, buf + BGMAC_RX_BUF_OFFSET,
dma_dev           354 drivers/net/ethernet/broadcom/bgmac.c 	if (dma_mapping_error(dma_dev, dma_addr)) {
dma_dev           399 drivers/net/ethernet/broadcom/bgmac.c static void bgmac_dma_rx_poison_buf(struct device *dma_dev,
dma_dev           404 drivers/net/ethernet/broadcom/bgmac.c 	dma_sync_single_for_cpu(dma_dev, slot->dma_addr, BGMAC_RX_BUF_SIZE,
dma_dev           408 drivers/net/ethernet/broadcom/bgmac.c 	dma_sync_single_for_device(dma_dev, slot->dma_addr, BGMAC_RX_BUF_SIZE,
dma_dev           425 drivers/net/ethernet/broadcom/bgmac.c 		struct device *dma_dev = bgmac->dma_dev;
dma_dev           436 drivers/net/ethernet/broadcom/bgmac.c 				bgmac_dma_rx_poison_buf(dma_dev, slot);
dma_dev           441 drivers/net/ethernet/broadcom/bgmac.c 			dma_unmap_single(dma_dev, dma_addr,
dma_dev           528 drivers/net/ethernet/broadcom/bgmac.c 	struct device *dma_dev = bgmac->dma_dev;
dma_dev           544 drivers/net/ethernet/broadcom/bgmac.c 			dma_unmap_single(dma_dev, slot->dma_addr,
dma_dev           547 drivers/net/ethernet/broadcom/bgmac.c 			dma_unmap_page(dma_dev, slot->dma_addr,
dma_dev           555 drivers/net/ethernet/broadcom/bgmac.c 	struct device *dma_dev = bgmac->dma_dev;
dma_dev           564 drivers/net/ethernet/broadcom/bgmac.c 		dma_unmap_single(dma_dev, slot->dma_addr,
dma_dev           576 drivers/net/ethernet/broadcom/bgmac.c 	struct device *dma_dev = bgmac->dma_dev;
dma_dev           584 drivers/net/ethernet/broadcom/bgmac.c 	dma_free_coherent(dma_dev, size, ring->cpu_base,
dma_dev           614 drivers/net/ethernet/broadcom/bgmac.c 	struct device *dma_dev = bgmac->dma_dev;
dma_dev           637 drivers/net/ethernet/broadcom/bgmac.c 		ring->cpu_base = dma_alloc_coherent(dma_dev, size,
dma_dev           662 drivers/net/ethernet/broadcom/bgmac.c 		ring->cpu_base = dma_alloc_coherent(dma_dev, size,
dma_dev           494 drivers/net/ethernet/broadcom/bgmac.h 	struct device *dma_dev;
dma_dev            57 drivers/net/ethernet/sfc/falcon/tx.c 		struct device *dma_dev = &tx_queue->efx->pci_dev->dev;
dma_dev            60 drivers/net/ethernet/sfc/falcon/tx.c 			dma_unmap_single(dma_dev, unmap_addr, buffer->unmap_len,
dma_dev            63 drivers/net/ethernet/sfc/falcon/tx.c 			dma_unmap_page(dma_dev, unmap_addr, buffer->unmap_len,
dma_dev           209 drivers/net/ethernet/sfc/falcon/tx.c 	struct device *dma_dev = &efx->pci_dev->dev;
dma_dev           220 drivers/net/ethernet/sfc/falcon/tx.c 	dma_addr = dma_map_single(dma_dev, skb->data, len, DMA_TO_DEVICE);
dma_dev           225 drivers/net/ethernet/sfc/falcon/tx.c 	if (unlikely(dma_mapping_error(dma_dev, dma_addr)))
dma_dev           254 drivers/net/ethernet/sfc/falcon/tx.c 		dma_addr = skb_frag_dma_map(dma_dev, fragment,
dma_dev           260 drivers/net/ethernet/sfc/falcon/tx.c 		if (unlikely(dma_mapping_error(dma_dev, dma_addr)))
dma_dev            65 drivers/net/ethernet/sfc/tx.c 		struct device *dma_dev = &tx_queue->efx->pci_dev->dev;
dma_dev            68 drivers/net/ethernet/sfc/tx.c 			dma_unmap_single(dma_dev, unmap_addr, buffer->unmap_len,
dma_dev            71 drivers/net/ethernet/sfc/tx.c 			dma_unmap_page(dma_dev, unmap_addr, buffer->unmap_len,
dma_dev           364 drivers/net/ethernet/sfc/tx.c 	struct device *dma_dev = &efx->pci_dev->dev;
dma_dev           375 drivers/net/ethernet/sfc/tx.c 	dma_addr = dma_map_single(dma_dev, skb->data, len, DMA_TO_DEVICE);
dma_dev           380 drivers/net/ethernet/sfc/tx.c 	if (unlikely(dma_mapping_error(dma_dev, dma_addr)))
dma_dev           424 drivers/net/ethernet/sfc/tx.c 		dma_addr = skb_frag_dma_map(dma_dev, fragment,
dma_dev           430 drivers/net/ethernet/sfc/tx.c 		if (unlikely(dma_mapping_error(dma_dev, dma_addr)))
dma_dev           173 drivers/net/ethernet/sfc/tx_tso.c 	struct device *dma_dev = &efx->pci_dev->dev;
dma_dev           198 drivers/net/ethernet/sfc/tx_tso.c 	dma_addr = dma_map_single(dma_dev, skb->data,
dma_dev           205 drivers/net/ethernet/sfc/tx_tso.c 	return unlikely(dma_mapping_error(dma_dev, dma_addr)) ? -ENOMEM : 0;
dma_dev            89 drivers/net/ethernet/sgi/ioc3-eth.c 	struct device *dma_dev;
dma_dev           150 drivers/net/ethernet/sgi/ioc3-eth.c 	d = dma_map_single(ip->dma_dev, new_skb->data,
dma_dev           153 drivers/net/ethernet/sgi/ioc3-eth.c 	if (dma_mapping_error(ip->dma_dev, d)) {
dma_dev           580 drivers/net/ethernet/sgi/ioc3-eth.c 			dma_unmap_single(ip->dma_dev, rxr[rx_entry],
dma_dev           812 drivers/net/ethernet/sgi/ioc3-eth.c 		dma_unmap_single(ip->dma_dev, be64_to_cpu(desc->p1),
dma_dev           817 drivers/net/ethernet/sgi/ioc3-eth.c 		dma_unmap_single(ip->dma_dev, be64_to_cpu(desc->p2),
dma_dev           851 drivers/net/ethernet/sgi/ioc3-eth.c 			dma_unmap_single(ip->dma_dev,
dma_dev          1218 drivers/net/ethernet/sgi/ioc3-eth.c 	ip->dma_dev = &pdev->dev;
dma_dev          1245 drivers/net/ethernet/sgi/ioc3-eth.c 	ip->rxr = dma_direct_alloc_pages(ip->dma_dev, RX_RING_SIZE,
dma_dev          1254 drivers/net/ethernet/sgi/ioc3-eth.c 	ip->txr = dma_direct_alloc_pages(ip->dma_dev, TX_RING_SIZE,
dma_dev          1316 drivers/net/ethernet/sgi/ioc3-eth.c 		dma_direct_free_pages(ip->dma_dev, RX_RING_SIZE, ip->rxr,
dma_dev          1319 drivers/net/ethernet/sgi/ioc3-eth.c 		dma_direct_free_pages(ip->dma_dev, TX_RING_SIZE, ip->txr,
dma_dev          1338 drivers/net/ethernet/sgi/ioc3-eth.c 	dma_direct_free_pages(ip->dma_dev, RX_RING_SIZE, ip->rxr,
dma_dev          1340 drivers/net/ethernet/sgi/ioc3-eth.c 	dma_direct_free_pages(ip->dma_dev, TX_RING_SIZE, ip->txr,
dma_dev          1449 drivers/net/ethernet/sgi/ioc3-eth.c 		d1 = dma_map_single(ip->dma_dev, skb->data, s1, DMA_TO_DEVICE);
dma_dev          1450 drivers/net/ethernet/sgi/ioc3-eth.c 		if (dma_mapping_error(ip->dma_dev, d1))
dma_dev          1452 drivers/net/ethernet/sgi/ioc3-eth.c 		d2 = dma_map_single(ip->dma_dev, (void *)b2, s1, DMA_TO_DEVICE);
dma_dev          1453 drivers/net/ethernet/sgi/ioc3-eth.c 		if (dma_mapping_error(ip->dma_dev, d2)) {
dma_dev          1454 drivers/net/ethernet/sgi/ioc3-eth.c 			dma_unmap_single(ip->dma_dev, d1, len, DMA_TO_DEVICE);
dma_dev          1465 drivers/net/ethernet/sgi/ioc3-eth.c 		d = dma_map_single(ip->dma_dev, skb->data, len, DMA_TO_DEVICE);
dma_dev          1466 drivers/net/ethernet/sgi/ioc3-eth.c 		if (dma_mapping_error(ip->dma_dev, d))
dma_dev           256 drivers/net/ethernet/sun/sunhme.c 	((__hp)->dma_map((__hp)->dma_dev, (__ptr), (__size), (__dir)))
dma_dev           258 drivers/net/ethernet/sun/sunhme.c 	((__hp)->dma_unmap((__hp)->dma_dev, (__addr), (__size), (__dir)))
dma_dev           260 drivers/net/ethernet/sun/sunhme.c 	((__hp)->dma_sync_for_cpu((__hp)->dma_dev, (__addr), (__size), (__dir)))
dma_dev           262 drivers/net/ethernet/sun/sunhme.c 	((__hp)->dma_sync_for_device((__hp)->dma_dev, (__addr), (__size), (__dir)))
dma_dev           282 drivers/net/ethernet/sun/sunhme.c 	dma_map_single((__hp)->dma_dev, (__ptr), (__size), (__dir))
dma_dev           284 drivers/net/ethernet/sun/sunhme.c 	dma_unmap_single((__hp)->dma_dev, (__addr), (__size), (__dir))
dma_dev           286 drivers/net/ethernet/sun/sunhme.c 	dma_dma_sync_single_for_cpu((__hp)->dma_dev, (__addr), (__size), (__dir))
dma_dev           288 drivers/net/ethernet/sun/sunhme.c 	dma_dma_sync_single_for_device((__hp)->dma_dev, (__addr), (__size), (__dir))
dma_dev           310 drivers/net/ethernet/sun/sunhme.c 	pci_map_single((__hp)->dma_dev, (__ptr), (__size), (__dir))
dma_dev           312 drivers/net/ethernet/sun/sunhme.c 	pci_unmap_single((__hp)->dma_dev, (__addr), (__size), (__dir))
dma_dev           314 drivers/net/ethernet/sun/sunhme.c 	pci_dma_sync_single_for_cpu((__hp)->dma_dev, (__addr), (__size), (__dir))
dma_dev           316 drivers/net/ethernet/sun/sunhme.c 	pci_dma_sync_single_for_device((__hp)->dma_dev, (__addr), (__size), (__dir))
dma_dev          1210 drivers/net/ethernet/sun/sunhme.c 			dma_unmap_single(hp->dma_dev, dma_addr,
dma_dev          1230 drivers/net/ethernet/sun/sunhme.c 					dma_unmap_single(hp->dma_dev, dma_addr,
dma_dev          1235 drivers/net/ethernet/sun/sunhme.c 					dma_unmap_page(hp->dma_dev, dma_addr,
dma_dev          1277 drivers/net/ethernet/sun/sunhme.c 		mapping = dma_map_single(hp->dma_dev, skb->data, RX_BUF_ALLOC_SIZE,
dma_dev          1279 drivers/net/ethernet/sun/sunhme.c 		if (dma_mapping_error(hp->dma_dev, mapping)) {
dma_dev          1957 drivers/net/ethernet/sun/sunhme.c 				dma_unmap_single(hp->dma_dev, dma_addr, dma_len, DMA_TO_DEVICE);
dma_dev          1959 drivers/net/ethernet/sun/sunhme.c 				dma_unmap_page(hp->dma_dev, dma_addr, dma_len, DMA_TO_DEVICE);
dma_dev          2039 drivers/net/ethernet/sun/sunhme.c 			mapping = dma_map_single(hp->dma_dev, new_skb->data,
dma_dev          2042 drivers/net/ethernet/sun/sunhme.c 			if (unlikely(dma_mapping_error(hp->dma_dev, mapping))) {
dma_dev          2048 drivers/net/ethernet/sun/sunhme.c 			dma_unmap_single(hp->dma_dev, dma_addr, RX_BUF_ALLOC_SIZE, DMA_FROM_DEVICE);
dma_dev          2067 drivers/net/ethernet/sun/sunhme.c 			dma_sync_single_for_cpu(hp->dma_dev, dma_addr, len, DMA_FROM_DEVICE);
dma_dev          2069 drivers/net/ethernet/sun/sunhme.c 			dma_sync_single_for_device(hp->dma_dev, dma_addr, len, DMA_FROM_DEVICE);
dma_dev          2272 drivers/net/ethernet/sun/sunhme.c 	dma_unmap_single(hp->dma_dev, first_mapping, first_len, DMA_TO_DEVICE);
dma_dev          2282 drivers/net/ethernet/sun/sunhme.c 		dma_unmap_page(hp->dma_dev, addr, len, DMA_TO_DEVICE);
dma_dev          2321 drivers/net/ethernet/sun/sunhme.c 		mapping = dma_map_single(hp->dma_dev, skb->data, len, DMA_TO_DEVICE);
dma_dev          2322 drivers/net/ethernet/sun/sunhme.c 		if (unlikely(dma_mapping_error(hp->dma_dev, mapping)))
dma_dev          2337 drivers/net/ethernet/sun/sunhme.c 		first_mapping = dma_map_single(hp->dma_dev, skb->data, first_len,
dma_dev          2339 drivers/net/ethernet/sun/sunhme.c 		if (unlikely(dma_mapping_error(hp->dma_dev, first_mapping)))
dma_dev          2348 drivers/net/ethernet/sun/sunhme.c 			mapping = skb_frag_dma_map(hp->dma_dev, this_frag,
dma_dev          2350 drivers/net/ethernet/sun/sunhme.c 			if (unlikely(dma_mapping_error(hp->dma_dev, mapping))) {
dma_dev          2743 drivers/net/ethernet/sun/sunhme.c 	hp->dma_dev = &op->dev;
dma_dev          2806 drivers/net/ethernet/sun/sunhme.c 	hp->happy_block = dma_alloc_coherent(hp->dma_dev,
dma_dev          2871 drivers/net/ethernet/sun/sunhme.c 	dma_free_coherent(hp->dma_dev,
dma_dev          3039 drivers/net/ethernet/sun/sunhme.c 	hp->dma_dev = &pdev->dev;
dma_dev          3223 drivers/net/ethernet/sun/sunhme.c 	dma_free_coherent(hp->dma_dev, PAGE_SIZE,
dma_dev          3300 drivers/net/ethernet/sun/sunhme.c 	dma_free_coherent(hp->dma_dev,
dma_dev           413 drivers/net/ethernet/sun/sunhme.h 	struct device		  *dma_dev;
dma_dev           326 drivers/net/thunderbolt.c 		struct device *dma_dev = tb_ring_dma_device(ring->ring);
dma_dev           346 drivers/net/thunderbolt.c 			dma_unmap_page(dma_dev, tf->frame.buffer_phy, size,
dma_dev           472 drivers/net/thunderbolt.c 		struct device *dma_dev = tb_ring_dma_device(ring->ring);
dma_dev           490 drivers/net/thunderbolt.c 		dma_addr = dma_map_page(dma_dev, tf->page, 0,
dma_dev           492 drivers/net/thunderbolt.c 		if (dma_mapping_error(dma_dev, dma_addr)) {
dma_dev           515 drivers/net/thunderbolt.c 	struct device *dma_dev = tb_ring_dma_device(ring->ring);
dma_dev           527 drivers/net/thunderbolt.c 	dma_sync_single_for_cpu(dma_dev, tf->frame.buffer_phy,
dma_dev           549 drivers/net/thunderbolt.c 	struct device *dma_dev = tb_ring_dma_device(ring->ring);
dma_dev           562 drivers/net/thunderbolt.c 		dma_addr = dma_map_page(dma_dev, tf->page, 0, TBNET_FRAME_SIZE,
dma_dev           564 drivers/net/thunderbolt.c 		if (dma_mapping_error(dma_dev, dma_addr)) {
dma_dev           744 drivers/net/thunderbolt.c 	struct device *dma_dev = tb_ring_dma_device(net->rx_ring.ring);
dma_dev           770 drivers/net/thunderbolt.c 		dma_unmap_page(dma_dev, frame->buffer_phy,
dma_dev           906 drivers/net/thunderbolt.c 	struct device *dma_dev = tb_ring_dma_device(net->tx_ring.ring);
dma_dev           921 drivers/net/thunderbolt.c 			dma_sync_single_for_device(dma_dev,
dma_dev           995 drivers/net/thunderbolt.c 		dma_sync_single_for_device(dma_dev, frames[i]->frame.buffer_phy,
dma_dev           104 drivers/net/wireless/broadcom/b43/bus.c 	dev->dma_dev = core->dma_dev;
dma_dev           203 drivers/net/wireless/broadcom/b43/bus.c 	dev->dma_dev = sdev->dma_dev;
dma_dev            40 drivers/net/wireless/broadcom/b43/bus.h 	struct device *dma_dev;
dma_dev           353 drivers/net/wireless/broadcom/b43/dma.c 		dmaaddr = dma_map_single(ring->dev->dev->dma_dev,
dma_dev           356 drivers/net/wireless/broadcom/b43/dma.c 		dmaaddr = dma_map_single(ring->dev->dev->dma_dev,
dma_dev           368 drivers/net/wireless/broadcom/b43/dma.c 		dma_unmap_single(ring->dev->dev->dma_dev,
dma_dev           371 drivers/net/wireless/broadcom/b43/dma.c 		dma_unmap_single(ring->dev->dev->dma_dev,
dma_dev           381 drivers/net/wireless/broadcom/b43/dma.c 	dma_sync_single_for_cpu(ring->dev->dev->dma_dev,
dma_dev           390 drivers/net/wireless/broadcom/b43/dma.c 	dma_sync_single_for_device(ring->dev->dev->dma_dev,
dma_dev           421 drivers/net/wireless/broadcom/b43/dma.c 	ring->descbase = dma_alloc_coherent(ring->dev->dev->dma_dev,
dma_dev           434 drivers/net/wireless/broadcom/b43/dma.c 	dma_free_coherent(ring->dev->dev->dma_dev, ring_mem_size,
dma_dev           542 drivers/net/wireless/broadcom/b43/dma.c 	if (unlikely(dma_mapping_error(ring->dev->dev->dma_dev, addr)))
dma_dev           898 drivers/net/wireless/broadcom/b43/dma.c 		dma_test = dma_map_single(dev->dev->dma_dev,
dma_dev           913 drivers/net/wireless/broadcom/b43/dma.c 			dma_test = dma_map_single(dev->dev->dma_dev,
dma_dev           927 drivers/net/wireless/broadcom/b43/dma.c 		dma_unmap_single(dev->dev->dma_dev,
dma_dev          1058 drivers/net/wireless/broadcom/b43/dma.c 	err = dma_set_mask_and_coherent(dev->dev->dma_dev, DMA_BIT_MASK(type));
dma_dev           255 drivers/net/wireless/broadcom/b43legacy/dma.c 		dmaaddr = dma_map_single(ring->dev->dev->dma_dev,
dma_dev           259 drivers/net/wireless/broadcom/b43legacy/dma.c 		dmaaddr = dma_map_single(ring->dev->dev->dma_dev,
dma_dev           273 drivers/net/wireless/broadcom/b43legacy/dma.c 		dma_unmap_single(ring->dev->dev->dma_dev,
dma_dev           277 drivers/net/wireless/broadcom/b43legacy/dma.c 		dma_unmap_single(ring->dev->dev->dma_dev,
dma_dev           289 drivers/net/wireless/broadcom/b43legacy/dma.c 	dma_sync_single_for_cpu(ring->dev->dev->dma_dev,
dma_dev           300 drivers/net/wireless/broadcom/b43legacy/dma.c 	dma_sync_single_for_device(ring->dev->dev->dma_dev,
dma_dev           321 drivers/net/wireless/broadcom/b43legacy/dma.c 	ring->descbase = dma_alloc_coherent(ring->dev->dev->dma_dev,
dma_dev           332 drivers/net/wireless/broadcom/b43legacy/dma.c 	dma_free_coherent(ring->dev->dev->dma_dev, B43legacy_DMA_RINGMEMSIZE,
dma_dev           416 drivers/net/wireless/broadcom/b43legacy/dma.c 	if (unlikely(dma_mapping_error(ring->dev->dev->dma_dev, addr)))
dma_dev           656 drivers/net/wireless/broadcom/b43legacy/dma.c 		dma_test = dma_map_single(dev->dev->dma_dev, ring->txhdr_cache,
dma_dev           670 drivers/net/wireless/broadcom/b43legacy/dma.c 			dma_test = dma_map_single(dev->dev->dma_dev,
dma_dev           680 drivers/net/wireless/broadcom/b43legacy/dma.c 		dma_unmap_single(dev->dev->dma_dev, dma_test,
dma_dev           783 drivers/net/wireless/broadcom/b43legacy/dma.c 	err = dma_set_mask_and_coherent(dev->dev->dma_dev, DMA_BIT_MASK(type));
dma_dev           590 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 	di->dmadev = core->dma_dev;
dma_dev           812 drivers/ntb/ntb_transport.c 			       struct device *dma_dev, size_t align)
dma_dev           818 drivers/ntb/ntb_transport.c 	alloc_addr = dma_alloc_coherent(dma_dev, mw->alloc_size,
dma_dev           821 drivers/ntb/ntb_transport.c 		dev_err(dma_dev, "Unable to alloc MW buff of size %zu\n",
dma_dev           850 drivers/ntb/ntb_transport.c 	dma_free_coherent(dma_dev, mw->alloc_size, alloc_addr, dma_addr);
dma_dev           783 drivers/ntb/test/ntb_perf.c 	struct device *dma_dev;
dma_dev           791 drivers/ntb/test/ntb_perf.c 	dma_dev = pthr->dma_chan->device->dev;
dma_dev           797 drivers/ntb/test/ntb_perf.c 	unmap = dmaengine_get_unmap_data(dma_dev, 2, GFP_NOWAIT);
dma_dev           802 drivers/ntb/test/ntb_perf.c 	unmap->addr[0] = dma_map_page(dma_dev, virt_to_page(src),
dma_dev           804 drivers/ntb/test/ntb_perf.c 	if (dma_mapping_error(dma_dev, unmap->addr[0])) {
dma_dev           810 drivers/ntb/test/ntb_perf.c 	unmap->addr[1] = dma_map_page(dma_dev, virt_to_page(dst),
dma_dev           812 drivers/ntb/test/ntb_perf.c 	if (dma_mapping_error(dma_dev, unmap->addr[1])) {
dma_dev          1099 drivers/s390/cio/css.c struct gen_pool *cio_gp_dma_create(struct device *dma_dev, int nr_pages)
dma_dev          1110 drivers/s390/cio/css.c 		cpu_addr = dma_alloc_coherent(dma_dev, PAGE_SIZE, &dma_addr,
dma_dev          1130 drivers/s390/cio/css.c void cio_gp_dma_destroy(struct gen_pool *gp_dma, struct device *dma_dev)
dma_dev          1135 drivers/s390/cio/css.c 	gen_pool_for_each_chunk(gp_dma, __gp_dma_free_dma, dma_dev);
dma_dev          1148 drivers/s390/cio/css.c void *cio_gp_dma_zalloc(struct gen_pool *gp_dma, struct device *dma_dev,
dma_dev          1160 drivers/s390/cio/css.c 		addr = (unsigned long) dma_alloc_coherent(dma_dev,
dma_dev           247 drivers/scsi/aha1542.c 	struct device *dev = cmd->device->host->dma_dev;
dma_dev           430 drivers/scsi/aha1542.c 		acmd->chain_handle = dma_map_single(sh->dma_dev, acmd->chain,
dma_dev           432 drivers/scsi/aha1542.c 		if (dma_mapping_error(sh->dma_dev, acmd->chain_handle))
dma_dev           830 drivers/scsi/aha1542.c 	struct device *dev = sh->dma_dev;
dma_dev           208 drivers/scsi/hosts.c 			   struct device *dma_dev)
dma_dev           232 drivers/scsi/hosts.c 	if (!dma_dev)
dma_dev           233 drivers/scsi/hosts.c 		dma_dev = shost->shost_gendev.parent;
dma_dev           235 drivers/scsi/hosts.c 	shost->dma_dev = dma_dev;
dma_dev           260 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c 		vio_disable_interrupts(vscsi->dma_dev);
dma_dev           817 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c 	rc = vio_enable_interrupts(vscsi->dma_dev);
dma_dev           889 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c 		vio_enable_interrupts(vscsi->dma_dev);
dma_dev          1230 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c 			vio_enable_interrupts(vscsi->dma_dev);
dma_dev          1516 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c 	info = dma_alloc_coherent(&vscsi->dma_dev->dev, sizeof(*info), &token,
dma_dev          1595 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c 	dma_free_coherent(&vscsi->dma_dev->dev, sizeof(*info), info, token);
dma_dev          1635 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c 	cap = dma_alloc_coherent(&vscsi->dma_dev->dev, olen, &token,
dma_dev          1647 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c 		strncpy(cap->name, dev_name(&vscsi->dma_dev->dev),
dma_dev          1711 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c 	dma_free_coherent(&vscsi->dma_dev->dev, olen, cap, token);
dma_dev          1944 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c 					rc = h_send_crq(vscsi->dma_dev->unit_address,
dma_dev          2595 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c 	struct vio_dev *vdev = vscsi->dma_dev;
dma_dev          2646 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c 		vdev = vscsi->dma_dev;
dma_dev          2965 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c 	vio_disable_interrupts(vscsi->dma_dev);
dma_dev          3015 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c 	struct vio_dev *vdev = vscsi->dma_dev;
dma_dev          3051 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c 	dma_unmap_single(&vscsi->dma_dev->dev, vscsi->cmd_q.crq_token,
dma_dev          3353 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c 		vio_enable_interrupts(vscsi->dma_dev);
dma_dev          3417 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c 			vio_enable_interrupts(vscsi->dma_dev);
dma_dev          3452 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c 	vscsi->dma_dev = vdev;
dma_dev          3606 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c 	dev_dbg(&vscsi->dev, "remove (%s)\n", dev_name(&vscsi->dma_dev->dev));
dma_dev          3650 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c 	return snprintf(buf, PAGE_SIZE, "%x\n", vscsi->dma_dev->unit_address);
dma_dev           295 drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.h 	struct vio_dev *dma_dev;
dma_dev          1784 drivers/scsi/scsi_lib.c 	struct device *dev = shost->dma_dev;
dma_dev            28 drivers/scsi/scsi_lib_dma.c 		struct device *dev = cmd->device->host->dma_dev;
dma_dev            46 drivers/scsi/scsi_lib_dma.c 		struct device *dev = cmd->device->host->dma_dev;
dma_dev            31 drivers/spi/spi-dw-mid.c 	if (s->dma_dev != chan->device->dev)
dma_dev            40 drivers/spi/spi-dw-mid.c 	struct pci_dev *dma_dev;
dma_dev            49 drivers/spi/spi-dw-mid.c 	dma_dev = pci_get_device(PCI_VENDOR_ID_INTEL, 0x0827, NULL);
dma_dev            50 drivers/spi/spi-dw-mid.c 	if (!dma_dev)
dma_dev            57 drivers/spi/spi-dw-mid.c 	rx->dma_dev = &dma_dev->dev;
dma_dev            64 drivers/spi/spi-dw-mid.c 	tx->dma_dev = &dma_dev->dev;
dma_dev            67 drivers/spi/spi-pxa2xx-pci.c 	if (dws->dma_dev != chan->device->dev)
dma_dev            76 drivers/spi/spi-pxa2xx-pci.c 	struct pci_dev *dma_dev;
dma_dev            81 drivers/spi/spi-pxa2xx-pci.c 	dma_dev = pci_get_slot(dev->bus, PCI_DEVFN(PCI_SLOT(dev->devfn), 0));
dma_dev            86 drivers/spi/spi-pxa2xx-pci.c 		slave->dma_dev = &dma_dev->dev;
dma_dev            94 drivers/spi/spi-pxa2xx-pci.c 		slave->dma_dev = &dma_dev->dev;
dma_dev           105 drivers/spi/spi-pxa2xx-pci.c 	struct pci_dev *dma_dev = pci_get_slot(dev->bus, PCI_DEVFN(21, 0));
dma_dev           132 drivers/spi/spi-pxa2xx-pci.c 	tx->dma_dev = &dma_dev->dev;
dma_dev           135 drivers/spi/spi-pxa2xx-pci.c 	rx->dma_dev = &dma_dev->dev;
dma_dev           833 drivers/spi/spi-topcliff-pch.c 	    (param->dma_dev == chan->device->dev)) {
dma_dev           845 drivers/spi/spi-topcliff-pch.c 	struct pci_dev *dma_dev;
dma_dev           860 drivers/spi/spi-topcliff-pch.c 	dma_dev = pci_get_slot(data->board_dat->pdev->bus,
dma_dev           865 drivers/spi/spi-topcliff-pch.c 	param->dma_dev = &dma_dev->dev;
dma_dev           880 drivers/spi/spi-topcliff-pch.c 	param->dma_dev = &dma_dev->dev;
dma_dev           499 drivers/ssb/main.c 			sdev->dma_dev = dev->parent;
dma_dev           515 drivers/ssb/main.c 			sdev->dma_dev = dev;
dma_dev           134 drivers/staging/ralink-gdma/ralink-gdma.c 	void (*init)(struct gdma_dma_dev *dma_dev);
dma_dev           155 drivers/staging/ralink-gdma/ralink-gdma.c static inline uint32_t gdma_dma_read(struct gdma_dma_dev *dma_dev,
dma_dev           158 drivers/staging/ralink-gdma/ralink-gdma.c 	return readl(dma_dev->base + reg);
dma_dev           161 drivers/staging/ralink-gdma/ralink-gdma.c static inline void gdma_dma_write(struct gdma_dma_dev *dma_dev,
dma_dev           164 drivers/staging/ralink-gdma/ralink-gdma.c 	writel(val, dma_dev->base + reg);
dma_dev           185 drivers/staging/ralink-gdma/ralink-gdma.c 	struct gdma_dma_dev *dma_dev = gdma_dma_chan_get_dev(chan);
dma_dev           188 drivers/staging/ralink-gdma/ralink-gdma.c 		dev_err(dma_dev->ddev.dev, "not support flow controller\n");
dma_dev           195 drivers/staging/ralink-gdma/ralink-gdma.c 			dev_err(dma_dev->ddev.dev, "only support 4 byte buswidth\n");
dma_dev           204 drivers/staging/ralink-gdma/ralink-gdma.c 			dev_err(dma_dev->ddev.dev, "only support 4 byte buswidth\n");
dma_dev           212 drivers/staging/ralink-gdma/ralink-gdma.c 		dev_err(dma_dev->ddev.dev, "direction type %d error\n",
dma_dev           223 drivers/staging/ralink-gdma/ralink-gdma.c 	struct gdma_dma_dev *dma_dev = gdma_dma_chan_get_dev(chan);
dma_dev           230 drivers/staging/ralink-gdma/ralink-gdma.c 	clear_bit(chan->id, &dma_dev->chan_issued);
dma_dev           238 drivers/staging/ralink-gdma/ralink-gdma.c 	while (gdma_dma_read(dma_dev, GDMA_REG_CTRL0(chan->id)) &
dma_dev           241 drivers/staging/ralink-gdma/ralink-gdma.c 			dev_err(dma_dev->ddev.dev, "chan %d wait timeout\n",
dma_dev           244 drivers/staging/ralink-gdma/ralink-gdma.c 			gdma_dma_write(dma_dev, GDMA_REG_CTRL0(chan->id), 0);
dma_dev           252 drivers/staging/ralink-gdma/ralink-gdma.c 		dev_dbg(dma_dev->ddev.dev, "terminate chan %d loops %d\n",
dma_dev           258 drivers/staging/ralink-gdma/ralink-gdma.c static void rt305x_dump_reg(struct gdma_dma_dev *dma_dev, int id)
dma_dev           260 drivers/staging/ralink-gdma/ralink-gdma.c 	dev_dbg(dma_dev->ddev.dev, "chan %d, src %08x, dst %08x, ctr0 %08x, ctr1 %08x, intr %08x, signal %08x\n",
dma_dev           262 drivers/staging/ralink-gdma/ralink-gdma.c 		gdma_dma_read(dma_dev, GDMA_REG_SRC_ADDR(id)),
dma_dev           263 drivers/staging/ralink-gdma/ralink-gdma.c 		gdma_dma_read(dma_dev, GDMA_REG_DST_ADDR(id)),
dma_dev           264 drivers/staging/ralink-gdma/ralink-gdma.c 		gdma_dma_read(dma_dev, GDMA_REG_CTRL0(id)),
dma_dev           265 drivers/staging/ralink-gdma/ralink-gdma.c 		gdma_dma_read(dma_dev, GDMA_REG_CTRL1(id)),
dma_dev           266 drivers/staging/ralink-gdma/ralink-gdma.c 		gdma_dma_read(dma_dev, GDMA_RT305X_STATUS_INT),
dma_dev           267 drivers/staging/ralink-gdma/ralink-gdma.c 		gdma_dma_read(dma_dev, GDMA_RT305X_STATUS_SIGNAL));
dma_dev           272 drivers/staging/ralink-gdma/ralink-gdma.c 	struct gdma_dma_dev *dma_dev = gdma_dma_chan_get_dev(chan);
dma_dev           278 drivers/staging/ralink-gdma/ralink-gdma.c 	ctrl0 = gdma_dma_read(dma_dev, GDMA_REG_CTRL0(chan->id));
dma_dev           280 drivers/staging/ralink-gdma/ralink-gdma.c 		dev_err(dma_dev->ddev.dev, "chan %d is start(%08x).\n",
dma_dev           282 drivers/staging/ralink-gdma/ralink-gdma.c 		rt305x_dump_reg(dma_dev, chan->id);
dma_dev           310 drivers/staging/ralink-gdma/ralink-gdma.c 		dev_err(dma_dev->ddev.dev, "direction type %d error\n",
dma_dev           321 drivers/staging/ralink-gdma/ralink-gdma.c 	gdma_dma_write(dma_dev, GDMA_REG_SRC_ADDR(chan->id), src_addr);
dma_dev           322 drivers/staging/ralink-gdma/ralink-gdma.c 	gdma_dma_write(dma_dev, GDMA_REG_DST_ADDR(chan->id), dst_addr);
dma_dev           323 drivers/staging/ralink-gdma/ralink-gdma.c 	gdma_dma_write(dma_dev, GDMA_REG_CTRL1(chan->id), ctrl1);
dma_dev           327 drivers/staging/ralink-gdma/ralink-gdma.c 	gdma_dma_write(dma_dev, GDMA_REG_CTRL0(chan->id), ctrl0);
dma_dev           332 drivers/staging/ralink-gdma/ralink-gdma.c static void rt3883_dump_reg(struct gdma_dma_dev *dma_dev, int id)
dma_dev           334 drivers/staging/ralink-gdma/ralink-gdma.c 	dev_dbg(dma_dev->ddev.dev, "chan %d, src %08x, dst %08x, ctr0 %08x, ctr1 %08x, unmask %08x, done %08x, req %08x, ack %08x, fin %08x\n",
dma_dev           336 drivers/staging/ralink-gdma/ralink-gdma.c 		gdma_dma_read(dma_dev, GDMA_REG_SRC_ADDR(id)),
dma_dev           337 drivers/staging/ralink-gdma/ralink-gdma.c 		gdma_dma_read(dma_dev, GDMA_REG_DST_ADDR(id)),
dma_dev           338 drivers/staging/ralink-gdma/ralink-gdma.c 		gdma_dma_read(dma_dev, GDMA_REG_CTRL0(id)),
dma_dev           339 drivers/staging/ralink-gdma/ralink-gdma.c 		gdma_dma_read(dma_dev, GDMA_REG_CTRL1(id)),
dma_dev           340 drivers/staging/ralink-gdma/ralink-gdma.c 		gdma_dma_read(dma_dev, GDMA_REG_UNMASK_INT),
dma_dev           341 drivers/staging/ralink-gdma/ralink-gdma.c 		gdma_dma_read(dma_dev, GDMA_REG_DONE_INT),
dma_dev           342 drivers/staging/ralink-gdma/ralink-gdma.c 		gdma_dma_read(dma_dev, GDMA_REG_REQSTS),
dma_dev           343 drivers/staging/ralink-gdma/ralink-gdma.c 		gdma_dma_read(dma_dev, GDMA_REG_ACKSTS),
dma_dev           344 drivers/staging/ralink-gdma/ralink-gdma.c 		gdma_dma_read(dma_dev, GDMA_REG_FINSTS));
dma_dev           349 drivers/staging/ralink-gdma/ralink-gdma.c 	struct gdma_dma_dev *dma_dev = gdma_dma_chan_get_dev(chan);
dma_dev           355 drivers/staging/ralink-gdma/ralink-gdma.c 	ctrl0 = gdma_dma_read(dma_dev, GDMA_REG_CTRL0(chan->id));
dma_dev           357 drivers/staging/ralink-gdma/ralink-gdma.c 		dev_err(dma_dev->ddev.dev, "chan %d is start(%08x).\n",
dma_dev           359 drivers/staging/ralink-gdma/ralink-gdma.c 		rt3883_dump_reg(dma_dev, chan->id);
dma_dev           385 drivers/staging/ralink-gdma/ralink-gdma.c 		dev_err(dma_dev->ddev.dev, "direction type %d error\n",
dma_dev           396 drivers/staging/ralink-gdma/ralink-gdma.c 	gdma_dma_write(dma_dev, GDMA_REG_SRC_ADDR(chan->id), src_addr);
dma_dev           397 drivers/staging/ralink-gdma/ralink-gdma.c 	gdma_dma_write(dma_dev, GDMA_REG_DST_ADDR(chan->id), dst_addr);
dma_dev           398 drivers/staging/ralink-gdma/ralink-gdma.c 	gdma_dma_write(dma_dev, GDMA_REG_CTRL1(chan->id), ctrl1);
dma_dev           402 drivers/staging/ralink-gdma/ralink-gdma.c 	gdma_dma_write(dma_dev, GDMA_REG_CTRL0(chan->id), ctrl0);
dma_dev           407 drivers/staging/ralink-gdma/ralink-gdma.c static inline int gdma_start_transfer(struct gdma_dma_dev *dma_dev,
dma_dev           410 drivers/staging/ralink-gdma/ralink-gdma.c 	return dma_dev->data->start_transfer(chan);
dma_dev           428 drivers/staging/ralink-gdma/ralink-gdma.c static void gdma_dma_chan_irq(struct gdma_dma_dev *dma_dev,
dma_dev           455 drivers/staging/ralink-gdma/ralink-gdma.c 		dev_dbg(dma_dev->ddev.dev, "chan %d no desc to complete\n",
dma_dev           459 drivers/staging/ralink-gdma/ralink-gdma.c 		set_bit(chan->id, &dma_dev->chan_issued);
dma_dev           465 drivers/staging/ralink-gdma/ralink-gdma.c 	struct gdma_dma_dev *dma_dev = devid;
dma_dev           469 drivers/staging/ralink-gdma/ralink-gdma.c 	done_reg = dma_dev->data->done_int_reg;
dma_dev           470 drivers/staging/ralink-gdma/ralink-gdma.c 	done = gdma_dma_read(dma_dev, done_reg);
dma_dev           475 drivers/staging/ralink-gdma/ralink-gdma.c 	gdma_dma_write(dma_dev, done_reg, done);
dma_dev           480 drivers/staging/ralink-gdma/ralink-gdma.c 			gdma_dma_chan_irq(dma_dev, &dma_dev->chan[i]);
dma_dev           481 drivers/staging/ralink-gdma/ralink-gdma.c 			atomic_dec(&dma_dev->cnt);
dma_dev           488 drivers/staging/ralink-gdma/ralink-gdma.c 	if (dma_dev->chan_issued)
dma_dev           489 drivers/staging/ralink-gdma/ralink-gdma.c 		tasklet_schedule(&dma_dev->task);
dma_dev           497 drivers/staging/ralink-gdma/ralink-gdma.c 	struct gdma_dma_dev *dma_dev = gdma_dma_chan_get_dev(chan);
dma_dev           503 drivers/staging/ralink-gdma/ralink-gdma.c 			set_bit(chan->id, &dma_dev->chan_issued);
dma_dev           504 drivers/staging/ralink-gdma/ralink-gdma.c 			tasklet_schedule(&dma_dev->task);
dma_dev           506 drivers/staging/ralink-gdma/ralink-gdma.c 			dev_dbg(dma_dev->ddev.dev, "chan %d no desc to issue\n",
dma_dev           706 drivers/staging/ralink-gdma/ralink-gdma.c 	struct gdma_dma_dev *dma_dev = (struct gdma_dma_dev *)arg;
dma_dev           713 drivers/staging/ralink-gdma/ralink-gdma.c 	chan_mask = dma_dev->data->chancnt - 1;
dma_dev           721 drivers/staging/ralink-gdma/ralink-gdma.c 		if (atomic_read(&dma_dev->cnt) >= 2) {
dma_dev           726 drivers/staging/ralink-gdma/ralink-gdma.c 		if (test_and_clear_bit(i, &dma_dev->chan_issued)) {
dma_dev           727 drivers/staging/ralink-gdma/ralink-gdma.c 			chan = &dma_dev->chan[i];
dma_dev           729 drivers/staging/ralink-gdma/ralink-gdma.c 				atomic_inc(&dma_dev->cnt);
dma_dev           730 drivers/staging/ralink-gdma/ralink-gdma.c 				gdma_start_transfer(dma_dev, chan);
dma_dev           732 drivers/staging/ralink-gdma/ralink-gdma.c 				dev_dbg(dma_dev->ddev.dev,
dma_dev           736 drivers/staging/ralink-gdma/ralink-gdma.c 			if (!dma_dev->chan_issued)
dma_dev           744 drivers/staging/ralink-gdma/ralink-gdma.c static void rt305x_gdma_init(struct gdma_dma_dev *dma_dev)
dma_dev           749 drivers/staging/ralink-gdma/ralink-gdma.c 	gdma_dma_write(dma_dev, GDMA_RT305X_GCT, GDMA_REG_GCT_ARBIT_RR);
dma_dev           751 drivers/staging/ralink-gdma/ralink-gdma.c 	gct = gdma_dma_read(dma_dev, GDMA_RT305X_GCT);
dma_dev           752 drivers/staging/ralink-gdma/ralink-gdma.c 	dev_info(dma_dev->ddev.dev, "revision: %d, channels: %d\n",
dma_dev           758 drivers/staging/ralink-gdma/ralink-gdma.c static void rt3883_gdma_init(struct gdma_dma_dev *dma_dev)
dma_dev           763 drivers/staging/ralink-gdma/ralink-gdma.c 	gdma_dma_write(dma_dev, GDMA_REG_GCT, GDMA_REG_GCT_ARBIT_RR);
dma_dev           765 drivers/staging/ralink-gdma/ralink-gdma.c 	gct = gdma_dma_read(dma_dev, GDMA_REG_GCT);
dma_dev           766 drivers/staging/ralink-gdma/ralink-gdma.c 	dev_info(dma_dev->ddev.dev, "revision: %d, channels: %d\n",
dma_dev           796 drivers/staging/ralink-gdma/ralink-gdma.c 	struct gdma_dma_dev *dma_dev;
dma_dev           814 drivers/staging/ralink-gdma/ralink-gdma.c 	dma_dev = devm_kzalloc(&pdev->dev,
dma_dev           815 drivers/staging/ralink-gdma/ralink-gdma.c 			       struct_size(dma_dev, chan, data->chancnt),
dma_dev           817 drivers/staging/ralink-gdma/ralink-gdma.c 	if (!dma_dev)
dma_dev           819 drivers/staging/ralink-gdma/ralink-gdma.c 	dma_dev->data = data;
dma_dev           825 drivers/staging/ralink-gdma/ralink-gdma.c 	dma_dev->base = base;
dma_dev           826 drivers/staging/ralink-gdma/ralink-gdma.c 	tasklet_init(&dma_dev->task, gdma_dma_tasklet, (unsigned long)dma_dev);
dma_dev           832 drivers/staging/ralink-gdma/ralink-gdma.c 			       0, dev_name(&pdev->dev), dma_dev);
dma_dev           840 drivers/staging/ralink-gdma/ralink-gdma.c 	dd = &dma_dev->ddev;
dma_dev           859 drivers/staging/ralink-gdma/ralink-gdma.c 	dd->dev->dma_parms = &dma_dev->dma_parms;
dma_dev           864 drivers/staging/ralink-gdma/ralink-gdma.c 		chan = &dma_dev->chan[i];
dma_dev           871 drivers/staging/ralink-gdma/ralink-gdma.c 	data->init(dma_dev);
dma_dev           880 drivers/staging/ralink-gdma/ralink-gdma.c 					 of_dma_xlate_by_chan_id, dma_dev);
dma_dev           886 drivers/staging/ralink-gdma/ralink-gdma.c 	platform_set_drvdata(pdev, dma_dev);
dma_dev           897 drivers/staging/ralink-gdma/ralink-gdma.c 	struct gdma_dma_dev *dma_dev = platform_get_drvdata(pdev);
dma_dev           899 drivers/staging/ralink-gdma/ralink-gdma.c 	tasklet_kill(&dma_dev->task);
dma_dev           901 drivers/staging/ralink-gdma/ralink-gdma.c 	dma_async_device_unregister(&dma_dev->ddev);
dma_dev           125 drivers/tty/serial/8250/8250_lpss.c 	struct pci_dev *dma_dev = pci_get_slot(pdev->bus, dma_devfn);
dma_dev           144 drivers/tty/serial/8250/8250_lpss.c 	param->dma_dev = &dma_dev->dev;
dma_dev           197 drivers/tty/serial/8250/8250_lpss.c 	param->dma_dev = &pdev->dev;
dma_dev           210 drivers/tty/serial/8250/8250_lpss.c 	if (!param->dma_dev)
dma_dev           248 drivers/tty/serial/8250/8250_lpss.c 	if (dws->dma_dev != chan->device->dev)
dma_dev           261 drivers/tty/serial/8250/8250_lpss.c 	if (!lpss->dma_param.dma_dev)
dma_dev           106 drivers/tty/serial/8250/8250_men_mcb.c 		data[i].uart.port.dev = mdev->dma_dev;
dma_dev            45 drivers/tty/serial/8250/8250_mid.c 	struct pci_dev *dma_dev;
dma_dev            71 drivers/tty/serial/8250/8250_mid.c 	mid->dma_dev = pci_get_slot(pdev->bus,
dma_dev            85 drivers/tty/serial/8250/8250_mid.c 	chip = pci_get_drvdata(mid->dma_dev);
dma_dev           121 drivers/tty/serial/8250/8250_mid.c 	mid->dma_dev = pci_get_slot(pdev->bus, PCI_DEVFN(5, 0));
dma_dev           184 drivers/tty/serial/8250/8250_mid.c 	mid->dma_dev = pdev;
dma_dev           192 drivers/tty/serial/8250/8250_mid.c 	if (!mid->dma_dev)
dma_dev           239 drivers/tty/serial/8250/8250_mid.c 	if (s->dma_dev != chan->device->dev || s->chan_id != chan->chan_id)
dma_dev           253 drivers/tty/serial/8250/8250_mid.c 	if (!mid->dma_dev)
dma_dev           270 drivers/tty/serial/8250/8250_mid.c 	rx_param->dma_dev = &mid->dma_dev->dev;
dma_dev           271 drivers/tty/serial/8250/8250_mid.c 	tx_param->dma_dev = &mid->dma_dev->dev;
dma_dev           602 drivers/tty/serial/amba-pl011.c 	struct dma_device *dma_dev = chan->device;
dma_dev           646 drivers/tty/serial/amba-pl011.c 	if (dma_map_sg(dma_dev->dev, &dmatx->sg, 1, DMA_TO_DEVICE) != 1) {
dma_dev           655 drivers/tty/serial/amba-pl011.c 		dma_unmap_sg(dma_dev->dev, &dmatx->sg, 1, DMA_TO_DEVICE);
dma_dev           673 drivers/tty/serial/amba-pl011.c 	dma_dev->device_issue_pending(chan);
dma_dev           703 drivers/tty/serial/pch_uart.c 	if ((chan->chan_id == param->chan_id) && (param->dma_dev ==
dma_dev           716 drivers/tty/serial/pch_uart.c 	struct pci_dev *dma_dev;
dma_dev           724 drivers/tty/serial/pch_uart.c 	dma_dev = pci_get_slot(priv->pdev->bus,
dma_dev           729 drivers/tty/serial/pch_uart.c 	param->dma_dev = &dma_dev->dev;
dma_dev           743 drivers/tty/serial/pch_uart.c 	param->dma_dev = &dma_dev->dev;
dma_dev           351 drivers/usb/host/bcma-hcd.c 	if (dma_set_mask_and_coherent(dev->dma_dev, DMA_BIT_MASK(32)))
dma_dev           165 drivers/usb/host/ssb-hcd.c 	if (dma_set_mask_and_coherent(dev->dma_dev, DMA_BIT_MASK(32)))
dma_dev           187 drivers/usb/renesas_usbhs/mod_gadget.c static int usbhsg_dma_map_ctrl(struct device *dma_dev, struct usbhs_pkt *pkt,
dma_dev           202 drivers/usb/renesas_usbhs/mod_gadget.c 		ret = usb_gadget_map_request_by_dev(dma_dev, req, dir);
dma_dev           208 drivers/usb/renesas_usbhs/mod_gadget.c 		usb_gadget_unmap_request_by_dev(dma_dev, req, dir);
dma_dev           917 drivers/usb/renesas_usbhs/mod_host.c static int usbhsh_dma_map_ctrl(struct device *dma_dev, struct usbhs_pkt *pkt,
dma_dev           672 drivers/usb/renesas_usbhs/pipe.c 		     int (*dma_map_ctrl)(struct device *dma_dev,
dma_dev            41 drivers/usb/renesas_usbhs/pipe.h 	int (*dma_map_ctrl)(struct device *dma_dev, struct usbhs_pkt *pkt,
dma_dev            79 drivers/usb/renesas_usbhs/pipe.h 		     int (*dma_map_ctrl)(struct device *dma_dev,
dma_dev          1539 drivers/video/fbdev/mx3fb.c 		mx3fb_pdata->dma_dev == chan->device->dev;
dma_dev            36 drivers/xen/gntdev-common.h 	struct device *dma_dev;
dma_dev            73 drivers/xen/gntdev-common.h 	struct device *dma_dev;
dma_dev           501 drivers/xen/gntdev-dmabuf.c 	args.dev = priv->dma_dev;
dma_dev           825 drivers/xen/gntdev-dmabuf.c 					   priv->dma_dev, op.fd,
dma_dev           103 drivers/xen/gntdev.c 		args.dev = map->dma_dev;
dma_dev           169 drivers/xen/gntdev.c 		add->dma_dev = priv->dma_dev;
dma_dev           171 drivers/xen/gntdev.c 		args.dev = priv->dma_dev;
dma_dev           625 drivers/xen/gntdev.c 	priv->dma_dev = gntdev_miscdev.this_device;
dma_dev           626 drivers/xen/gntdev.c 	dma_coerce_mask_and_coherent(priv->dma_dev, DMA_BIT_MASK(64));
dma_dev           272 include/linux/bcma/bcma.h 	struct device *dma_dev;
dma_dev           147 include/linux/iio/buffer-dma.h 	struct device *dma_dev, const struct iio_dma_buffer_ops *ops);
dma_dev            76 include/linux/mcb.h 	struct device *dma_dev;
dma_dev            18 include/linux/pch_dma.h 	struct device		*dma_dev;
dma_dev            28 include/linux/platform_data/dma-atmel.h 	struct device		*dma_dev;
dma_dev            27 include/linux/platform_data/dma-dw.h 	struct device		*dma_dev;
dma_dev            14 include/linux/platform_data/dma-hsu.h 	struct device	*dma_dev;
dma_dev            38 include/linux/platform_data/media/camera-mx3.h 	struct device *dma_dev;
dma_dev            43 include/linux/platform_data/video-mx3fb.h 	struct device			*dma_dev;
dma_dev           107 include/linux/shdma-base.h 	struct dma_device dma_dev;
dma_dev           114 include/linux/shdma-base.h 				i < (d)->dma_dev.chancnt; c = (d)->schan[++i])
dma_dev           273 include/linux/ssb/ssb.h 	struct device *dev, *dma_dev;
dma_dev           692 include/scsi/scsi_host.h 	struct device *dma_dev;
dma_dev           137 include/sound/dmaengine_pcm.h 	struct device *dma_dev;
dma_dev           375 sound/pci/trident/trident.h 	struct snd_dma_device dma_dev;
dma_dev           313 sound/soc/intel/baytrail/sst-baytrail-dsp.c 	ret = dma_coerce_mask_and_coherent(sst->dma_dev, DMA_BIT_MASK(32));
dma_dev           328 sound/soc/intel/baytrail/sst-baytrail-pcm.c 						      pdata->dma_dev,
dma_dev           102 sound/soc/intel/common/sst-acpi.c 	sst_pdata->dma_dev = dev;
dma_dev           265 sound/soc/intel/common/sst-dsp-priv.h 	struct device *dma_dev;
dma_dev           203 sound/soc/intel/common/sst-dsp.h 	struct device *dma_dev;
dma_dev           214 sound/soc/intel/common/sst-firmware.c 	return chan->device->dev == dsp->dma_dev;
dma_dev           299 sound/soc/intel/common/sst-firmware.c 	dma->chip = dw_probe(sst->dma_dev, &mem, sst_pdata->irq);
dma_dev           348 sound/soc/intel/common/sst-firmware.c 	sst_fw->dma_buf = dma_alloc_coherent(dsp->dma_dev, sst_fw->size,
dma_dev           385 sound/soc/intel/common/sst-firmware.c 	dma_free_coherent(dsp->dma_dev, sst_fw->size,
dma_dev           456 sound/soc/intel/common/sst-firmware.c 		dma_free_coherent(dsp->dma_dev, sst_fw->size, sst_fw->dma_buf,
dma_dev           936 sound/soc/intel/common/sst-firmware.c 	context->buffer = dma_alloc_coherent(dsp->dma_dev,
dma_dev          1007 sound/soc/intel/common/sst-firmware.c 	dma_free_coherent(dsp->dma_dev, module->persistent_size,
dma_dev          1218 sound/soc/intel/common/sst-firmware.c 	sst->dma_dev = pdata->dma_dev;
dma_dev           618 sound/soc/intel/haswell/sst-haswell-dsp.c 	dev = sst->dma_dev;
dma_dev          2142 sound/soc/intel/haswell/sst-haswell-ipc.c 	hsw->dx_context = dma_alloc_coherent(hsw->dsp->dma_dev,
dma_dev          2200 sound/soc/intel/haswell/sst-haswell-ipc.c 	dma_free_coherent(hsw->dsp->dma_dev, SST_HSW_DX_CONTEXT_SIZE,
dma_dev          2217 sound/soc/intel/haswell/sst-haswell-ipc.c 	dma_free_coherent(hsw->dsp->dma_dev, SST_HSW_DX_CONTEXT_SIZE,
dma_dev           939 sound/soc/intel/haswell/sst-haswell-pcm.c 	struct device *dev = pdata->dma_dev;
dma_dev          1044 sound/soc/intel/haswell/sst-haswell-pcm.c 	struct device *dma_dev, *dev;
dma_dev          1051 sound/soc/intel/haswell/sst-haswell-pcm.c 	dma_dev = pdata->dma_dev;
dma_dev          1064 sound/soc/intel/haswell/sst-haswell-pcm.c 			ret = snd_dma_alloc_pages(SNDRV_DMA_TYPE_DEV, dma_dev,
dma_dev          1073 sound/soc/intel/haswell/sst-haswell-pcm.c 			ret = snd_dma_alloc_pages(SNDRV_DMA_TYPE_DEV, dma_dev,
dma_dev            17 sound/soc/samsung/dma.h 				       struct device *dma_dev);
dma_dev            19 sound/soc/samsung/dmaengine.c 				       struct device *dma_dev)
dma_dev            29 sound/soc/samsung/dmaengine.c 	pcm_conf->dma_dev = dma_dev;
dma_dev           118 sound/soc/soc-generic-dmaengine-pcm.c 	struct device *dma_dev = dmaengine_dma_dev(pcm, substream);
dma_dev           141 sound/soc/soc-generic-dmaengine-pcm.c 	hw.period_bytes_max = dma_get_max_seg_size(dma_dev);
dma_dev           416 sound/soc/soc-generic-dmaengine-pcm.c 	    !(config && config->dma_dev && config->dma_dev->of_node)))
dma_dev           419 sound/soc/soc-generic-dmaengine-pcm.c 	if (config && config->dma_dev) {
dma_dev           427 sound/soc/soc-generic-dmaengine-pcm.c 			 dev_name(config->dma_dev));
dma_dev           428 sound/soc/soc-generic-dmaengine-pcm.c 		dev = config->dma_dev;
dma_dev            56 sound/soc/tegra/tegra_pcm.c 	config->dma_dev = dev->parent;