sdma              200 arch/mips/include/asm/octeon/cvmx-address.h 	} sdma;
sdma              265 arch/mips/include/asm/octeon/cvmx-address.h 	} sdma;
sdma              134 arch/powerpc/platforms/52xx/mpc52xx_pic.c static struct mpc52xx_sdma __iomem *sdma;
sdma              269 arch/powerpc/platforms/52xx/mpc52xx_pic.c 	io_be_setbit(&sdma->IntMask, l2irq);
sdma              275 arch/powerpc/platforms/52xx/mpc52xx_pic.c 	io_be_clrbit(&sdma->IntMask, l2irq);
sdma              281 arch/powerpc/platforms/52xx/mpc52xx_pic.c 	out_be32(&sdma->IntPend, 1 << l2irq);
sdma              416 arch/powerpc/platforms/52xx/mpc52xx_pic.c 	sdma = of_iomap(np, 0);
sdma              418 arch/powerpc/platforms/52xx/mpc52xx_pic.c 	if (!sdma)
sdma              425 arch/powerpc/platforms/52xx/mpc52xx_pic.c 	out_be32(&sdma->IntPend, 0xffffffff);	/* 1 means clear pending */
sdma              426 arch/powerpc/platforms/52xx/mpc52xx_pic.c 	out_be32(&sdma->IntMask, 0xffffffff);	/* 1 means disabled */
sdma              507 arch/powerpc/platforms/52xx/mpc52xx_pic.c 			status = in_be32(&sdma->IntPend);
sdma              108 arch/s390/include/asm/pci_clp.h 	u64 sdma;			/* start dma as */
sdma              152 arch/s390/pci/pci_clp.c 	zdev->start_dma = response->sdma;
sdma              363 drivers/dma/imx-sdma.c 	struct sdma_engine		*sdma;
sdma              617 drivers/dma/imx-sdma.c static inline u32 chnenbl_ofs(struct sdma_engine *sdma, unsigned int event)
sdma              619 drivers/dma/imx-sdma.c 	u32 chnenbl0 = sdma->drvdata->chnenbl0;
sdma              626 drivers/dma/imx-sdma.c 	struct sdma_engine *sdma = sdmac->sdma;
sdma              633 drivers/dma/imx-sdma.c 	evt = readl_relaxed(sdma->regs + SDMA_H_EVTOVR);
sdma              634 drivers/dma/imx-sdma.c 	mcu = readl_relaxed(sdma->regs + SDMA_H_HOSTOVR);
sdma              635 drivers/dma/imx-sdma.c 	dsp = readl_relaxed(sdma->regs + SDMA_H_DSPOVR);
sdma              652 drivers/dma/imx-sdma.c 	writel_relaxed(evt, sdma->regs + SDMA_H_EVTOVR);
sdma              653 drivers/dma/imx-sdma.c 	writel_relaxed(mcu, sdma->regs + SDMA_H_HOSTOVR);
sdma              654 drivers/dma/imx-sdma.c 	writel_relaxed(dsp, sdma->regs + SDMA_H_DSPOVR);
sdma              659 drivers/dma/imx-sdma.c static void sdma_enable_channel(struct sdma_engine *sdma, int channel)
sdma              661 drivers/dma/imx-sdma.c 	writel(BIT(channel), sdma->regs + SDMA_H_START);
sdma              667 drivers/dma/imx-sdma.c static int sdma_run_channel0(struct sdma_engine *sdma)
sdma              672 drivers/dma/imx-sdma.c 	sdma_enable_channel(sdma, 0);
sdma              674 drivers/dma/imx-sdma.c 	ret = readl_relaxed_poll_timeout_atomic(sdma->regs + SDMA_H_STATSTOP,
sdma              677 drivers/dma/imx-sdma.c 		dev_err(sdma->dev, "Timeout waiting for CH0 ready\n");
sdma              680 drivers/dma/imx-sdma.c 	reg = readl(sdma->regs + SDMA_H_CONFIG);
sdma              683 drivers/dma/imx-sdma.c 		writel_relaxed(reg, sdma->regs + SDMA_H_CONFIG);
sdma              689 drivers/dma/imx-sdma.c static int sdma_load_script(struct sdma_engine *sdma, void *buf, int size,
sdma              692 drivers/dma/imx-sdma.c 	struct sdma_buffer_descriptor *bd0 = sdma->bd0;
sdma              698 drivers/dma/imx-sdma.c 	buf_virt = dma_alloc_coherent(sdma->dev, size, &buf_phys, GFP_KERNEL);
sdma              703 drivers/dma/imx-sdma.c 	spin_lock_irqsave(&sdma->channel_0_lock, flags);
sdma              713 drivers/dma/imx-sdma.c 	ret = sdma_run_channel0(sdma);
sdma              715 drivers/dma/imx-sdma.c 	spin_unlock_irqrestore(&sdma->channel_0_lock, flags);
sdma              717 drivers/dma/imx-sdma.c 	dma_free_coherent(sdma->dev, size, buf_virt, buf_phys);
sdma              724 drivers/dma/imx-sdma.c 	struct sdma_engine *sdma = sdmac->sdma;
sdma              727 drivers/dma/imx-sdma.c 	u32 chnenbl = chnenbl_ofs(sdma, event);
sdma              729 drivers/dma/imx-sdma.c 	val = readl_relaxed(sdma->regs + chnenbl);
sdma              731 drivers/dma/imx-sdma.c 	writel_relaxed(val, sdma->regs + chnenbl);
sdma              736 drivers/dma/imx-sdma.c 	struct sdma_engine *sdma = sdmac->sdma;
sdma              738 drivers/dma/imx-sdma.c 	u32 chnenbl = chnenbl_ofs(sdma, event);
sdma              741 drivers/dma/imx-sdma.c 	val = readl_relaxed(sdma->regs + chnenbl);
sdma              743 drivers/dma/imx-sdma.c 	writel_relaxed(val, sdma->regs + chnenbl);
sdma              755 drivers/dma/imx-sdma.c 	struct sdma_engine *sdma = sdmac->sdma;
sdma              770 drivers/dma/imx-sdma.c 	sdma->channel_control[channel].base_bd_ptr = desc->bd_phys;
sdma              771 drivers/dma/imx-sdma.c 	sdma->channel_control[channel].current_bd_ptr = desc->bd_phys;
sdma              772 drivers/dma/imx-sdma.c 	sdma_enable_channel(sdma, sdmac->channel);
sdma              852 drivers/dma/imx-sdma.c 	struct sdma_engine *sdma = dev_id;
sdma              855 drivers/dma/imx-sdma.c 	stat = readl_relaxed(sdma->regs + SDMA_H_INTR);
sdma              856 drivers/dma/imx-sdma.c 	writel_relaxed(stat, sdma->regs + SDMA_H_INTR);
sdma              862 drivers/dma/imx-sdma.c 		struct sdma_channel *sdmac = &sdma->channel[channel];
sdma              890 drivers/dma/imx-sdma.c 	struct sdma_engine *sdma = sdmac->sdma;
sdma              905 drivers/dma/imx-sdma.c 		emi_2_emi = sdma->script_addrs->ap_2_ap_addr;
sdma              908 drivers/dma/imx-sdma.c 		emi_2_per = sdma->script_addrs->bp_2_ap_addr;
sdma              909 drivers/dma/imx-sdma.c 		per_2_emi = sdma->script_addrs->ap_2_bp_addr;
sdma              912 drivers/dma/imx-sdma.c 		per_2_emi = sdma->script_addrs->firi_2_mcu_addr;
sdma              913 drivers/dma/imx-sdma.c 		emi_2_per = sdma->script_addrs->mcu_2_firi_addr;
sdma              916 drivers/dma/imx-sdma.c 		per_2_emi = sdma->script_addrs->uart_2_mcu_addr;
sdma              917 drivers/dma/imx-sdma.c 		emi_2_per = sdma->script_addrs->mcu_2_app_addr;
sdma              920 drivers/dma/imx-sdma.c 		per_2_emi = sdma->script_addrs->uartsh_2_mcu_addr;
sdma              921 drivers/dma/imx-sdma.c 		emi_2_per = sdma->script_addrs->mcu_2_shp_addr;
sdma              924 drivers/dma/imx-sdma.c 		per_2_emi = sdma->script_addrs->ata_2_mcu_addr;
sdma              925 drivers/dma/imx-sdma.c 		emi_2_per = sdma->script_addrs->mcu_2_ata_addr;
sdma              931 drivers/dma/imx-sdma.c 		per_2_emi = sdma->script_addrs->app_2_mcu_addr;
sdma              932 drivers/dma/imx-sdma.c 		emi_2_per = sdma->script_addrs->mcu_2_app_addr;
sdma              935 drivers/dma/imx-sdma.c 		per_2_emi = sdma->script_addrs->ssish_2_mcu_addr;
sdma              936 drivers/dma/imx-sdma.c 		emi_2_per = sdma->script_addrs->mcu_2_ssish_addr;
sdma              944 drivers/dma/imx-sdma.c 		per_2_emi = sdma->script_addrs->shp_2_mcu_addr;
sdma              945 drivers/dma/imx-sdma.c 		emi_2_per = sdma->script_addrs->mcu_2_shp_addr;
sdma              948 drivers/dma/imx-sdma.c 		per_2_emi = sdma->script_addrs->asrc_2_mcu_addr;
sdma              949 drivers/dma/imx-sdma.c 		emi_2_per = sdma->script_addrs->asrc_2_mcu_addr;
sdma              950 drivers/dma/imx-sdma.c 		per_2_per = sdma->script_addrs->per_2_per_addr;
sdma              953 drivers/dma/imx-sdma.c 		per_2_emi = sdma->script_addrs->shp_2_mcu_addr;
sdma              954 drivers/dma/imx-sdma.c 		emi_2_per = sdma->script_addrs->mcu_2_shp_addr;
sdma              955 drivers/dma/imx-sdma.c 		per_2_per = sdma->script_addrs->per_2_per_addr;
sdma              958 drivers/dma/imx-sdma.c 		per_2_emi = sdma->script_addrs->mshc_2_mcu_addr;
sdma              959 drivers/dma/imx-sdma.c 		emi_2_per = sdma->script_addrs->mcu_2_mshc_addr;
sdma              962 drivers/dma/imx-sdma.c 		per_2_emi = sdma->script_addrs->dptc_dvfs_addr;
sdma              965 drivers/dma/imx-sdma.c 		per_2_emi = sdma->script_addrs->spdif_2_mcu_addr;
sdma              966 drivers/dma/imx-sdma.c 		emi_2_per = sdma->script_addrs->mcu_2_spdif_addr;
sdma              969 drivers/dma/imx-sdma.c 		emi_2_per = sdma->script_addrs->ext_mem_2_ipu_addr;
sdma              983 drivers/dma/imx-sdma.c 	struct sdma_engine *sdma = sdmac->sdma;
sdma              986 drivers/dma/imx-sdma.c 	struct sdma_context_data *context = sdma->context;
sdma              987 drivers/dma/imx-sdma.c 	struct sdma_buffer_descriptor *bd0 = sdma->bd0;
sdma             1006 drivers/dma/imx-sdma.c 	dev_dbg(sdma->dev, "load_address = %d\n", load_address);
sdma             1007 drivers/dma/imx-sdma.c 	dev_dbg(sdma->dev, "wml = 0x%08x\n", (u32)sdmac->watermark_level);
sdma             1008 drivers/dma/imx-sdma.c 	dev_dbg(sdma->dev, "shp_addr = 0x%08x\n", sdmac->shp_addr);
sdma             1009 drivers/dma/imx-sdma.c 	dev_dbg(sdma->dev, "per_addr = 0x%08x\n", sdmac->per_addr);
sdma             1010 drivers/dma/imx-sdma.c 	dev_dbg(sdma->dev, "event_mask0 = 0x%08x\n", (u32)sdmac->event_mask[0]);
sdma             1011 drivers/dma/imx-sdma.c 	dev_dbg(sdma->dev, "event_mask1 = 0x%08x\n", (u32)sdmac->event_mask[1]);
sdma             1013 drivers/dma/imx-sdma.c 	spin_lock_irqsave(&sdma->channel_0_lock, flags);
sdma             1030 drivers/dma/imx-sdma.c 	bd0->buffer_addr = sdma->context_phys;
sdma             1032 drivers/dma/imx-sdma.c 	ret = sdma_run_channel0(sdma);
sdma             1034 drivers/dma/imx-sdma.c 	spin_unlock_irqrestore(&sdma->channel_0_lock, flags);
sdma             1049 drivers/dma/imx-sdma.c 	struct sdma_engine *sdma = sdmac->sdma;
sdma             1052 drivers/dma/imx-sdma.c 	writel_relaxed(BIT(channel), sdma->regs + SDMA_H_STATSTOP);
sdma             1103 drivers/dma/imx-sdma.c 	struct sdma_engine *sdma = sdmac->sdma;
sdma             1130 drivers/dma/imx-sdma.c 	if (sdmac->per_address2 >= sdma->spba_start_addr &&
sdma             1131 drivers/dma/imx-sdma.c 			sdmac->per_address2 <= sdma->spba_end_addr)
sdma             1134 drivers/dma/imx-sdma.c 	if (sdmac->per_address >= sdma->spba_start_addr &&
sdma             1135 drivers/dma/imx-sdma.c 			sdmac->per_address <= sdma->spba_end_addr)
sdma             1192 drivers/dma/imx-sdma.c 	struct sdma_engine *sdma = sdmac->sdma;
sdma             1200 drivers/dma/imx-sdma.c 	writel_relaxed(priority, sdma->regs + SDMA_CHNPRI_0 + 4 * channel);
sdma             1205 drivers/dma/imx-sdma.c static int sdma_request_channel0(struct sdma_engine *sdma)
sdma             1209 drivers/dma/imx-sdma.c 	sdma->bd0 = dma_alloc_coherent(sdma->dev, PAGE_SIZE, &sdma->bd0_phys,
sdma             1211 drivers/dma/imx-sdma.c 	if (!sdma->bd0) {
sdma             1216 drivers/dma/imx-sdma.c 	sdma->channel_control[0].base_bd_ptr = sdma->bd0_phys;
sdma             1217 drivers/dma/imx-sdma.c 	sdma->channel_control[0].current_bd_ptr = sdma->bd0_phys;
sdma             1219 drivers/dma/imx-sdma.c 	sdma_set_channel_priority(&sdma->channel[0], MXC_SDMA_DEFAULT_PRIORITY);
sdma             1232 drivers/dma/imx-sdma.c 	desc->bd = dma_alloc_coherent(desc->sdmac->sdma->dev, bd_size,
sdma             1246 drivers/dma/imx-sdma.c 	dma_free_coherent(desc->sdmac->sdma->dev, bd_size, desc->bd,
sdma             1275 drivers/dma/imx-sdma.c 		dev_dbg(sdmac->sdma->dev, "MEMCPY in case?\n");
sdma             1302 drivers/dma/imx-sdma.c 	ret = clk_enable(sdmac->sdma->clk_ipg);
sdma             1305 drivers/dma/imx-sdma.c 	ret = clk_enable(sdmac->sdma->clk_ahb);
sdma             1316 drivers/dma/imx-sdma.c 	clk_disable(sdmac->sdma->clk_ahb);
sdma             1318 drivers/dma/imx-sdma.c 	clk_disable(sdmac->sdma->clk_ipg);
sdma             1325 drivers/dma/imx-sdma.c 	struct sdma_engine *sdma = sdmac->sdma;
sdma             1342 drivers/dma/imx-sdma.c 	clk_disable(sdma->clk_ipg);
sdma             1343 drivers/dma/imx-sdma.c 	clk_disable(sdma->clk_ahb);
sdma             1389 drivers/dma/imx-sdma.c 	struct sdma_engine *sdma = sdmac->sdma;
sdma             1399 drivers/dma/imx-sdma.c 	dev_dbg(sdma->dev, "memcpy: %pad->%pad, len=%zu, channel=%d.\n",
sdma             1429 drivers/dma/imx-sdma.c 		dev_dbg(sdma->dev, "entry %d: count: %zd dma: 0x%x %s%s\n",
sdma             1446 drivers/dma/imx-sdma.c 	struct sdma_engine *sdma = sdmac->sdma;
sdma             1458 drivers/dma/imx-sdma.c 	dev_dbg(sdma->dev, "setting up %d entries for channel %d.\n",
sdma             1470 drivers/dma/imx-sdma.c 			dev_err(sdma->dev, "SDMA channel %d: maximum bytes for sg entry exceeded: %d > %d\n",
sdma             1507 drivers/dma/imx-sdma.c 		dev_dbg(sdma->dev, "entry %d: count: %d dma: %#llx %s%s\n",
sdma             1530 drivers/dma/imx-sdma.c 	struct sdma_engine *sdma = sdmac->sdma;
sdma             1536 drivers/dma/imx-sdma.c 	dev_dbg(sdma->dev, "%s channel: %d\n", __func__, channel);
sdma             1549 drivers/dma/imx-sdma.c 		dev_err(sdma->dev, "SDMA channel %d: maximum period size exceeded: %zu > %d\n",
sdma             1573 drivers/dma/imx-sdma.c 		dev_dbg(sdma->dev, "entry %d: count: %zu dma: %#llx %s%s\n",
sdma             1633 drivers/dma/imx-sdma.c 		if (sdmac->event_id0 >= sdmac->sdma->drvdata->num_events)
sdma             1639 drivers/dma/imx-sdma.c 		if (sdmac->event_id1 >= sdmac->sdma->drvdata->num_events)
sdma             1700 drivers/dma/imx-sdma.c static void sdma_add_scripts(struct sdma_engine *sdma,
sdma             1704 drivers/dma/imx-sdma.c 	s32 *saddr_arr = (u32 *)sdma->script_addrs;
sdma             1708 drivers/dma/imx-sdma.c 	if (!sdma->script_number)
sdma             1709 drivers/dma/imx-sdma.c 		sdma->script_number = SDMA_SCRIPT_ADDRS_ARRAY_SIZE_V1;
sdma             1711 drivers/dma/imx-sdma.c 	if (sdma->script_number > sizeof(struct sdma_script_start_addrs)
sdma             1713 drivers/dma/imx-sdma.c 		dev_err(sdma->dev,
sdma             1715 drivers/dma/imx-sdma.c 			sdma->script_number);
sdma             1719 drivers/dma/imx-sdma.c 	for (i = 0; i < sdma->script_number; i++)
sdma             1726 drivers/dma/imx-sdma.c 	struct sdma_engine *sdma = context;
sdma             1732 drivers/dma/imx-sdma.c 		dev_info(sdma->dev, "external firmware not found, using ROM firmware\n");
sdma             1748 drivers/dma/imx-sdma.c 		sdma->script_number = SDMA_SCRIPT_ADDRS_ARRAY_SIZE_V1;
sdma             1751 drivers/dma/imx-sdma.c 		sdma->script_number = SDMA_SCRIPT_ADDRS_ARRAY_SIZE_V2;
sdma             1754 drivers/dma/imx-sdma.c 		sdma->script_number = SDMA_SCRIPT_ADDRS_ARRAY_SIZE_V3;
sdma             1757 drivers/dma/imx-sdma.c 		sdma->script_number = SDMA_SCRIPT_ADDRS_ARRAY_SIZE_V4;
sdma             1760 drivers/dma/imx-sdma.c 		dev_err(sdma->dev, "unknown firmware version\n");
sdma             1767 drivers/dma/imx-sdma.c 	clk_enable(sdma->clk_ipg);
sdma             1768 drivers/dma/imx-sdma.c 	clk_enable(sdma->clk_ahb);
sdma             1770 drivers/dma/imx-sdma.c 	sdma_load_script(sdma, ram_code,
sdma             1773 drivers/dma/imx-sdma.c 	clk_disable(sdma->clk_ipg);
sdma             1774 drivers/dma/imx-sdma.c 	clk_disable(sdma->clk_ahb);
sdma             1776 drivers/dma/imx-sdma.c 	sdma_add_scripts(sdma, addr);
sdma             1778 drivers/dma/imx-sdma.c 	dev_info(sdma->dev, "loaded firmware %d.%d\n",
sdma             1788 drivers/dma/imx-sdma.c static int sdma_event_remap(struct sdma_engine *sdma)
sdma             1790 drivers/dma/imx-sdma.c 	struct device_node *np = sdma->dev->of_node;
sdma             1804 drivers/dma/imx-sdma.c 		dev_dbg(sdma->dev, "no event needs to be remapped\n");
sdma             1807 drivers/dma/imx-sdma.c 		dev_err(sdma->dev, "the property %s must modulo %d\n",
sdma             1815 drivers/dma/imx-sdma.c 		dev_err(sdma->dev, "failed to get gpr regmap\n");
sdma             1823 drivers/dma/imx-sdma.c 			dev_err(sdma->dev, "failed to read property %s index %d\n",
sdma             1830 drivers/dma/imx-sdma.c 			dev_err(sdma->dev, "failed to read property %s index %d\n",
sdma             1837 drivers/dma/imx-sdma.c 			dev_err(sdma->dev, "failed to read property %s index %d\n",
sdma             1852 drivers/dma/imx-sdma.c static int sdma_get_firmware(struct sdma_engine *sdma,
sdma             1858 drivers/dma/imx-sdma.c 			FW_ACTION_HOTPLUG, fw_name, sdma->dev,
sdma             1859 drivers/dma/imx-sdma.c 			GFP_KERNEL, sdma, sdma_load_firmware);
sdma             1864 drivers/dma/imx-sdma.c static int sdma_init(struct sdma_engine *sdma)
sdma             1869 drivers/dma/imx-sdma.c 	ret = clk_enable(sdma->clk_ipg);
sdma             1872 drivers/dma/imx-sdma.c 	ret = clk_enable(sdma->clk_ahb);
sdma             1876 drivers/dma/imx-sdma.c 	if (sdma->drvdata->check_ratio &&
sdma             1877 drivers/dma/imx-sdma.c 	    (clk_get_rate(sdma->clk_ahb) == clk_get_rate(sdma->clk_ipg)))
sdma             1878 drivers/dma/imx-sdma.c 		sdma->clk_ratio = 1;
sdma             1881 drivers/dma/imx-sdma.c 	writel_relaxed(0, sdma->regs + SDMA_H_C0PTR);
sdma             1883 drivers/dma/imx-sdma.c 	sdma->channel_control = dma_alloc_coherent(sdma->dev,
sdma             1888 drivers/dma/imx-sdma.c 	if (!sdma->channel_control) {
sdma             1893 drivers/dma/imx-sdma.c 	sdma->context = (void *)sdma->channel_control +
sdma             1895 drivers/dma/imx-sdma.c 	sdma->context_phys = ccb_phys +
sdma             1899 drivers/dma/imx-sdma.c 	for (i = 0; i < sdma->drvdata->num_events; i++)
sdma             1900 drivers/dma/imx-sdma.c 		writel_relaxed(0, sdma->regs + chnenbl_ofs(sdma, i));
sdma             1904 drivers/dma/imx-sdma.c 		writel_relaxed(0, sdma->regs + SDMA_CHNPRI_0 + i * 4);
sdma             1906 drivers/dma/imx-sdma.c 	ret = sdma_request_channel0(sdma);
sdma             1910 drivers/dma/imx-sdma.c 	sdma_config_ownership(&sdma->channel[0], false, true, false);
sdma             1913 drivers/dma/imx-sdma.c 	writel_relaxed(0x4050, sdma->regs + SDMA_CHN0ADDR);
sdma             1916 drivers/dma/imx-sdma.c 	if (sdma->clk_ratio)
sdma             1917 drivers/dma/imx-sdma.c 		writel_relaxed(SDMA_H_CONFIG_ACR, sdma->regs + SDMA_H_CONFIG);
sdma             1919 drivers/dma/imx-sdma.c 		writel_relaxed(0, sdma->regs + SDMA_H_CONFIG);
sdma             1921 drivers/dma/imx-sdma.c 	writel_relaxed(ccb_phys, sdma->regs + SDMA_H_C0PTR);
sdma             1924 drivers/dma/imx-sdma.c 	sdma_set_channel_priority(&sdma->channel[0], 7);
sdma             1926 drivers/dma/imx-sdma.c 	clk_disable(sdma->clk_ipg);
sdma             1927 drivers/dma/imx-sdma.c 	clk_disable(sdma->clk_ahb);
sdma             1932 drivers/dma/imx-sdma.c 	clk_disable(sdma->clk_ahb);
sdma             1934 drivers/dma/imx-sdma.c 	clk_disable(sdma->clk_ipg);
sdma             1935 drivers/dma/imx-sdma.c 	dev_err(sdma->dev, "initialisation failed with %d\n", ret);
sdma             1956 drivers/dma/imx-sdma.c 	struct sdma_engine *sdma = ofdma->of_dma_data;
sdma             1957 drivers/dma/imx-sdma.c 	dma_cap_mask_t mask = sdma->dma_device.cap_mask;
sdma             1992 drivers/dma/imx-sdma.c 	struct sdma_engine *sdma;
sdma             2010 drivers/dma/imx-sdma.c 	sdma = devm_kzalloc(&pdev->dev, sizeof(*sdma), GFP_KERNEL);
sdma             2011 drivers/dma/imx-sdma.c 	if (!sdma)
sdma             2014 drivers/dma/imx-sdma.c 	spin_lock_init(&sdma->channel_0_lock);
sdma             2016 drivers/dma/imx-sdma.c 	sdma->dev = &pdev->dev;
sdma             2017 drivers/dma/imx-sdma.c 	sdma->drvdata = drvdata;
sdma             2024 drivers/dma/imx-sdma.c 	sdma->regs = devm_ioremap_resource(&pdev->dev, iores);
sdma             2025 drivers/dma/imx-sdma.c 	if (IS_ERR(sdma->regs))
sdma             2026 drivers/dma/imx-sdma.c 		return PTR_ERR(sdma->regs);
sdma             2028 drivers/dma/imx-sdma.c 	sdma->clk_ipg = devm_clk_get(&pdev->dev, "ipg");
sdma             2029 drivers/dma/imx-sdma.c 	if (IS_ERR(sdma->clk_ipg))
sdma             2030 drivers/dma/imx-sdma.c 		return PTR_ERR(sdma->clk_ipg);
sdma             2032 drivers/dma/imx-sdma.c 	sdma->clk_ahb = devm_clk_get(&pdev->dev, "ahb");
sdma             2033 drivers/dma/imx-sdma.c 	if (IS_ERR(sdma->clk_ahb))
sdma             2034 drivers/dma/imx-sdma.c 		return PTR_ERR(sdma->clk_ahb);
sdma             2036 drivers/dma/imx-sdma.c 	ret = clk_prepare(sdma->clk_ipg);
sdma             2040 drivers/dma/imx-sdma.c 	ret = clk_prepare(sdma->clk_ahb);
sdma             2045 drivers/dma/imx-sdma.c 			       sdma);
sdma             2049 drivers/dma/imx-sdma.c 	sdma->irq = irq;
sdma             2051 drivers/dma/imx-sdma.c 	sdma->script_addrs = kzalloc(sizeof(*sdma->script_addrs), GFP_KERNEL);
sdma             2052 drivers/dma/imx-sdma.c 	if (!sdma->script_addrs) {
sdma             2058 drivers/dma/imx-sdma.c 	saddr_arr = (s32 *)sdma->script_addrs;
sdma             2062 drivers/dma/imx-sdma.c 	dma_cap_set(DMA_SLAVE, sdma->dma_device.cap_mask);
sdma             2063 drivers/dma/imx-sdma.c 	dma_cap_set(DMA_CYCLIC, sdma->dma_device.cap_mask);
sdma             2064 drivers/dma/imx-sdma.c 	dma_cap_set(DMA_MEMCPY, sdma->dma_device.cap_mask);
sdma             2066 drivers/dma/imx-sdma.c 	INIT_LIST_HEAD(&sdma->dma_device.channels);
sdma             2069 drivers/dma/imx-sdma.c 		struct sdma_channel *sdmac = &sdma->channel[i];
sdma             2071 drivers/dma/imx-sdma.c 		sdmac->sdma = sdma;
sdma             2083 drivers/dma/imx-sdma.c 			vchan_init(&sdmac->vc, &sdma->dma_device);
sdma             2086 drivers/dma/imx-sdma.c 	ret = sdma_init(sdma);
sdma             2090 drivers/dma/imx-sdma.c 	ret = sdma_event_remap(sdma);
sdma             2094 drivers/dma/imx-sdma.c 	if (sdma->drvdata->script_addrs)
sdma             2095 drivers/dma/imx-sdma.c 		sdma_add_scripts(sdma, sdma->drvdata->script_addrs);
sdma             2097 drivers/dma/imx-sdma.c 		sdma_add_scripts(sdma, pdata->script_addrs);
sdma             2099 drivers/dma/imx-sdma.c 	sdma->dma_device.dev = &pdev->dev;
sdma             2101 drivers/dma/imx-sdma.c 	sdma->dma_device.device_alloc_chan_resources = sdma_alloc_chan_resources;
sdma             2102 drivers/dma/imx-sdma.c 	sdma->dma_device.device_free_chan_resources = sdma_free_chan_resources;
sdma             2103 drivers/dma/imx-sdma.c 	sdma->dma_device.device_tx_status = sdma_tx_status;
sdma             2104 drivers/dma/imx-sdma.c 	sdma->dma_device.device_prep_slave_sg = sdma_prep_slave_sg;
sdma             2105 drivers/dma/imx-sdma.c 	sdma->dma_device.device_prep_dma_cyclic = sdma_prep_dma_cyclic;
sdma             2106 drivers/dma/imx-sdma.c 	sdma->dma_device.device_config = sdma_config;
sdma             2107 drivers/dma/imx-sdma.c 	sdma->dma_device.device_terminate_all = sdma_disable_channel_async;
sdma             2108 drivers/dma/imx-sdma.c 	sdma->dma_device.device_synchronize = sdma_channel_synchronize;
sdma             2109 drivers/dma/imx-sdma.c 	sdma->dma_device.src_addr_widths = SDMA_DMA_BUSWIDTHS;
sdma             2110 drivers/dma/imx-sdma.c 	sdma->dma_device.dst_addr_widths = SDMA_DMA_BUSWIDTHS;
sdma             2111 drivers/dma/imx-sdma.c 	sdma->dma_device.directions = SDMA_DMA_DIRECTIONS;
sdma             2112 drivers/dma/imx-sdma.c 	sdma->dma_device.residue_granularity = DMA_RESIDUE_GRANULARITY_SEGMENT;
sdma             2113 drivers/dma/imx-sdma.c 	sdma->dma_device.device_prep_dma_memcpy = sdma_prep_memcpy;
sdma             2114 drivers/dma/imx-sdma.c 	sdma->dma_device.device_issue_pending = sdma_issue_pending;
sdma             2115 drivers/dma/imx-sdma.c 	sdma->dma_device.dev->dma_parms = &sdma->dma_parms;
sdma             2116 drivers/dma/imx-sdma.c 	sdma->dma_device.copy_align = 2;
sdma             2117 drivers/dma/imx-sdma.c 	dma_set_max_seg_size(sdma->dma_device.dev, SDMA_BD_MAX_CNT);
sdma             2119 drivers/dma/imx-sdma.c 	platform_set_drvdata(pdev, sdma);
sdma             2121 drivers/dma/imx-sdma.c 	ret = dma_async_device_register(&sdma->dma_device);
sdma             2128 drivers/dma/imx-sdma.c 		ret = of_dma_controller_register(np, sdma_xlate, sdma);
sdma             2137 drivers/dma/imx-sdma.c 			sdma->spba_start_addr = spba_res.start;
sdma             2138 drivers/dma/imx-sdma.c 			sdma->spba_end_addr = spba_res.end;
sdma             2150 drivers/dma/imx-sdma.c 		ret = sdma_get_firmware(sdma, pdata->fw_name);
sdma             2164 drivers/dma/imx-sdma.c 			ret = sdma_get_firmware(sdma, fw_name);
sdma             2173 drivers/dma/imx-sdma.c 	dma_async_device_unregister(&sdma->dma_device);
sdma             2175 drivers/dma/imx-sdma.c 	kfree(sdma->script_addrs);
sdma             2177 drivers/dma/imx-sdma.c 	clk_unprepare(sdma->clk_ahb);
sdma             2179 drivers/dma/imx-sdma.c 	clk_unprepare(sdma->clk_ipg);
sdma             2185 drivers/dma/imx-sdma.c 	struct sdma_engine *sdma = platform_get_drvdata(pdev);
sdma             2188 drivers/dma/imx-sdma.c 	devm_free_irq(&pdev->dev, sdma->irq, sdma);
sdma             2189 drivers/dma/imx-sdma.c 	dma_async_device_unregister(&sdma->dma_device);
sdma             2190 drivers/dma/imx-sdma.c 	kfree(sdma->script_addrs);
sdma             2191 drivers/dma/imx-sdma.c 	clk_unprepare(sdma->clk_ahb);
sdma             2192 drivers/dma/imx-sdma.c 	clk_unprepare(sdma->clk_ipg);
sdma             2195 drivers/dma/imx-sdma.c 		struct sdma_channel *sdmac = &sdma->channel[i];
sdma              238 drivers/dma/sirf-dma.c 	struct sirfsoc_dma *sdma = dma_chan_to_sirfsoc_dma(&schan->chan);
sdma              247 drivers/dma/sirf-dma.c 	base = sdma->base;
sdma              253 drivers/dma/sirf-dma.c 	if (sdma->type == SIRFSOC_DMA_VER_A7V2)
sdma              257 drivers/dma/sirf-dma.c 	sdma->exec_desc(sdesc, cid, schan->mode, base);
sdma              266 drivers/dma/sirf-dma.c 	struct sirfsoc_dma *sdma = data;
sdma              274 drivers/dma/sirf-dma.c 	switch (sdma->type) {
sdma              277 drivers/dma/sirf-dma.c 		is = readl(sdma->base + SIRFSOC_DMA_CH_INT);
sdma              278 drivers/dma/sirf-dma.c 		reg = sdma->base + SIRFSOC_DMA_CH_INT;
sdma              282 drivers/dma/sirf-dma.c 			schan = &sdma->channels[ch];
sdma              300 drivers/dma/sirf-dma.c 		is = readl(sdma->base + SIRFSOC_DMA_INT_ATLAS7);
sdma              302 drivers/dma/sirf-dma.c 		reg = sdma->base + SIRFSOC_DMA_INT_ATLAS7;
sdma              304 drivers/dma/sirf-dma.c 		schan = &sdma->channels[0];
sdma              332 drivers/dma/sirf-dma.c 	tasklet_schedule(&sdma->tasklet);
sdma              338 drivers/dma/sirf-dma.c static void sirfsoc_dma_process_completed(struct sirfsoc_dma *sdma)
sdma              349 drivers/dma/sirf-dma.c 	for (i = 0; i < sdma->dma.chancnt; i++) {
sdma              350 drivers/dma/sirf-dma.c 		schan = &sdma->channels[i];
sdma              398 drivers/dma/sirf-dma.c 	struct sirfsoc_dma *sdma = (void *)data;
sdma              400 drivers/dma/sirf-dma.c 	sirfsoc_dma_process_completed(sdma);
sdma              445 drivers/dma/sirf-dma.c 	struct sirfsoc_dma *sdma = dma_chan_to_sirfsoc_dma(&schan->chan);
sdma              451 drivers/dma/sirf-dma.c 	switch (sdma->type) {
sdma              453 drivers/dma/sirf-dma.c 		writel_relaxed(1 << cid, sdma->base + SIRFSOC_DMA_INT_EN_CLR);
sdma              454 drivers/dma/sirf-dma.c 		writel_relaxed(1 << cid, sdma->base + SIRFSOC_DMA_CH_INT);
sdma              456 drivers/dma/sirf-dma.c 			       sdma->base +
sdma              458 drivers/dma/sirf-dma.c 		writel_relaxed(1 << cid, sdma->base + SIRFSOC_DMA_CH_VALID);
sdma              461 drivers/dma/sirf-dma.c 		writel_relaxed(0, sdma->base + SIRFSOC_DMA_INT_EN_ATLAS7);
sdma              463 drivers/dma/sirf-dma.c 			       sdma->base + SIRFSOC_DMA_INT_ATLAS7);
sdma              464 drivers/dma/sirf-dma.c 		writel_relaxed(0, sdma->base + SIRFSOC_DMA_LOOP_CTRL_ATLAS7);
sdma              465 drivers/dma/sirf-dma.c 		writel_relaxed(0, sdma->base + SIRFSOC_DMA_VALID_ATLAS7);
sdma              468 drivers/dma/sirf-dma.c 		writel_relaxed(readl_relaxed(sdma->base + SIRFSOC_DMA_INT_EN) &
sdma              469 drivers/dma/sirf-dma.c 			       ~(1 << cid), sdma->base + SIRFSOC_DMA_INT_EN);
sdma              470 drivers/dma/sirf-dma.c 		writel_relaxed(readl_relaxed(sdma->base +
sdma              473 drivers/dma/sirf-dma.c 			       sdma->base + SIRFSOC_DMA_CH_LOOP_CTRL);
sdma              474 drivers/dma/sirf-dma.c 		writel_relaxed(1 << cid, sdma->base + SIRFSOC_DMA_CH_VALID);
sdma              491 drivers/dma/sirf-dma.c 	struct sirfsoc_dma *sdma = dma_chan_to_sirfsoc_dma(&schan->chan);
sdma              497 drivers/dma/sirf-dma.c 	switch (sdma->type) {
sdma              500 drivers/dma/sirf-dma.c 			       sdma->base +
sdma              504 drivers/dma/sirf-dma.c 		writel_relaxed(0, sdma->base + SIRFSOC_DMA_LOOP_CTRL_ATLAS7);
sdma              507 drivers/dma/sirf-dma.c 		writel_relaxed(readl_relaxed(sdma->base +
sdma              510 drivers/dma/sirf-dma.c 			       sdma->base + SIRFSOC_DMA_CH_LOOP_CTRL);
sdma              525 drivers/dma/sirf-dma.c 	struct sirfsoc_dma *sdma = dma_chan_to_sirfsoc_dma(&schan->chan);
sdma              530 drivers/dma/sirf-dma.c 	switch (sdma->type) {
sdma              533 drivers/dma/sirf-dma.c 			       sdma->base + SIRFSOC_DMA_CH_LOOP_CTRL_ATLAS7);
sdma              537 drivers/dma/sirf-dma.c 			       sdma->base + SIRFSOC_DMA_LOOP_CTRL_ATLAS7);
sdma              540 drivers/dma/sirf-dma.c 		writel_relaxed(readl_relaxed(sdma->base +
sdma              543 drivers/dma/sirf-dma.c 			       sdma->base + SIRFSOC_DMA_CH_LOOP_CTRL);
sdma              558 drivers/dma/sirf-dma.c 	struct sirfsoc_dma *sdma = dma_chan_to_sirfsoc_dma(chan);
sdma              565 drivers/dma/sirf-dma.c 	pm_runtime_get_sync(sdma->dma.dev);
sdma              571 drivers/dma/sirf-dma.c 			dev_notice(sdma->dma.dev, "Memory allocation error. "
sdma              599 drivers/dma/sirf-dma.c 	struct sirfsoc_dma *sdma = dma_chan_to_sirfsoc_dma(chan);
sdma              621 drivers/dma/sirf-dma.c 	pm_runtime_put(sdma->dma.dev);
sdma              643 drivers/dma/sirf-dma.c 	struct sirfsoc_dma *sdma = dma_chan_to_sirfsoc_dma(chan);
sdma              670 drivers/dma/sirf-dma.c 	if (sdma->type == SIRFSOC_DMA_VER_A7V2)
sdma              673 drivers/dma/sirf-dma.c 	if (sdma->type == SIRFSOC_DMA_VER_A7V2) {
sdma              674 drivers/dma/sirf-dma.c 		dma_pos = readl_relaxed(sdma->base + SIRFSOC_DMA_CUR_DATA_ADDR);
sdma              677 drivers/dma/sirf-dma.c 			sdma->base + cid * 0x10 + SIRFSOC_DMA_CH_ADDR) << 2;
sdma              692 drivers/dma/sirf-dma.c 	struct sirfsoc_dma *sdma = dma_chan_to_sirfsoc_dma(chan);
sdma              714 drivers/dma/sirf-dma.c 		sirfsoc_dma_process_completed(sdma);
sdma              830 drivers/dma/sirf-dma.c 	struct sirfsoc_dma *sdma = ofdma->of_dma_data;
sdma              836 drivers/dma/sirf-dma.c 	return dma_get_slave_channel(&sdma->channels[request].chan);
sdma              844 drivers/dma/sirf-dma.c 	struct sirfsoc_dma *sdma;
sdma              852 drivers/dma/sirf-dma.c 	sdma = devm_kzalloc(dev, sizeof(*sdma), GFP_KERNEL);
sdma              853 drivers/dma/sirf-dma.c 	if (!sdma)
sdma              859 drivers/dma/sirf-dma.c 	sdma->exec_desc = data->exec;
sdma              860 drivers/dma/sirf-dma.c 	sdma->type = data->type;
sdma              867 drivers/dma/sirf-dma.c 	sdma->irq = irq_of_parse_and_map(dn, 0);
sdma              868 drivers/dma/sirf-dma.c 	if (!sdma->irq) {
sdma              873 drivers/dma/sirf-dma.c 	sdma->clk = devm_clk_get(dev, NULL);
sdma              874 drivers/dma/sirf-dma.c 	if (IS_ERR(sdma->clk)) {
sdma              876 drivers/dma/sirf-dma.c 		return PTR_ERR(sdma->clk);
sdma              888 drivers/dma/sirf-dma.c 	sdma->base = devm_ioremap(dev, regs_start, regs_size);
sdma              889 drivers/dma/sirf-dma.c 	if (!sdma->base) {
sdma              895 drivers/dma/sirf-dma.c 	ret = request_irq(sdma->irq, &sirfsoc_dma_irq, 0, DRV_NAME, sdma);
sdma              902 drivers/dma/sirf-dma.c 	dma = &sdma->dma;
sdma              926 drivers/dma/sirf-dma.c 		schan = &sdma->channels[i];
sdma              941 drivers/dma/sirf-dma.c 	tasklet_init(&sdma->tasklet, sirfsoc_dma_tasklet, (unsigned long)sdma);
sdma              944 drivers/dma/sirf-dma.c 	dev_set_drvdata(dev, sdma);
sdma              951 drivers/dma/sirf-dma.c 	ret = of_dma_controller_register(dn, of_dma_sirfsoc_xlate, sdma);
sdma              965 drivers/dma/sirf-dma.c 	free_irq(sdma->irq, sdma);
sdma              967 drivers/dma/sirf-dma.c 	irq_dispose_mapping(sdma->irq);
sdma              974 drivers/dma/sirf-dma.c 	struct sirfsoc_dma *sdma = dev_get_drvdata(dev);
sdma              977 drivers/dma/sirf-dma.c 	dma_async_device_unregister(&sdma->dma);
sdma              978 drivers/dma/sirf-dma.c 	free_irq(sdma->irq, sdma);
sdma              979 drivers/dma/sirf-dma.c 	tasklet_kill(&sdma->tasklet);
sdma              980 drivers/dma/sirf-dma.c 	irq_dispose_mapping(sdma->irq);
sdma              990 drivers/dma/sirf-dma.c 	struct sirfsoc_dma *sdma = dev_get_drvdata(dev);
sdma              992 drivers/dma/sirf-dma.c 	clk_disable_unprepare(sdma->clk);
sdma              998 drivers/dma/sirf-dma.c 	struct sirfsoc_dma *sdma = dev_get_drvdata(dev);
sdma             1001 drivers/dma/sirf-dma.c 	ret = clk_prepare_enable(sdma->clk);
sdma             1011 drivers/dma/sirf-dma.c 	struct sirfsoc_dma *sdma = dev_get_drvdata(dev);
sdma             1012 drivers/dma/sirf-dma.c 	struct sirfsoc_dma_regs *save = &sdma->regs_save;
sdma             1029 drivers/dma/sirf-dma.c 	if (sdma->type == SIRFSOC_DMA_VER_A7V2) {
sdma             1042 drivers/dma/sirf-dma.c 		schan = &sdma->channels[ch];
sdma             1045 drivers/dma/sirf-dma.c 		save->ctrl[ch] = readl_relaxed(sdma->base +
sdma             1048 drivers/dma/sirf-dma.c 	save->interrupt_en = readl_relaxed(sdma->base + int_offset);
sdma             1058 drivers/dma/sirf-dma.c 	struct sirfsoc_dma *sdma = dev_get_drvdata(dev);
sdma             1059 drivers/dma/sirf-dma.c 	struct sirfsoc_dma_regs *save = &sdma->regs_save;
sdma             1073 drivers/dma/sirf-dma.c 	if (sdma->type == SIRFSOC_DMA_VER_A7V2) {
sdma             1083 drivers/dma/sirf-dma.c 	writel_relaxed(save->interrupt_en, sdma->base + int_offset);
sdma             1085 drivers/dma/sirf-dma.c 		schan = &sdma->channels[ch];
sdma             1092 drivers/dma/sirf-dma.c 			sdma->base + width_offset + ch * 4);
sdma             1094 drivers/dma/sirf-dma.c 			sdma->base + ch * 0x10 + SIRFSOC_DMA_CH_XLEN);
sdma             1096 drivers/dma/sirf-dma.c 			sdma->base + ch * 0x10 + SIRFSOC_DMA_CH_YLEN);
sdma             1098 drivers/dma/sirf-dma.c 			sdma->base + ch * 0x10 + SIRFSOC_DMA_CH_CTRL);
sdma             1099 drivers/dma/sirf-dma.c 		if (sdma->type == SIRFSOC_DMA_VER_A7V2) {
sdma             1101 drivers/dma/sirf-dma.c 				sdma->base + SIRFSOC_DMA_CH_ADDR);
sdma             1104 drivers/dma/sirf-dma.c 				sdma->base + ch * 0x10 + SIRFSOC_DMA_CH_ADDR);
sdma              929 drivers/gpu/drm/amd/amdgpu/amdgpu.h 	struct amdgpu_sdma		sdma;
sdma              410 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.c 		return adev->sdma.instance[0].fw_version;
sdma              413 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.c 		return adev->sdma.instance[1].fw_version;
sdma              616 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.c 		ring = &adev->sdma.instance[0].ring;
sdma              619 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.c 		ring = &adev->sdma.instance[1].ring;
sdma              167 drivers/gpu/drm/amd/amdgpu/amdgpu_cgs.c 			fw_version = adev->sdma.instance[0].fw_version;
sdma              170 drivers/gpu/drm/amd/amdgpu/amdgpu_cgs.c 			fw_version = adev->sdma.instance[1].fw_version;
sdma              140 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c 			for (j = 0; j < adev->sdma.num_instances; ++j)
sdma              141 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c 				rings[j] = &adev->sdma.instance[j].ring;
sdma              142 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c 			num_rings = adev->sdma.num_instances;
sdma              279 drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c 		if (query_fw->index >= adev->sdma.num_instances)
sdma              281 drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c 		fw_info->ver = adev->sdma.instance[query_fw->index].fw_version;
sdma              282 drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c 		fw_info->feature = adev->sdma.instance[query_fw->index].feature_version;
sdma              334 drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c 		for (i = 0; i < adev->sdma.num_instances; i++)
sdma              335 drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c 			if (adev->sdma.instance[i].ring.sched.ready)
sdma             1371 drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              965 drivers/gpu/drm/amd/amdgpu/amdgpu_psp.c 			adev->sdma.instance[ucode->ucode_id - AMDGPU_UCODE_ID_SDMA0].fw->data;
sdma              252 drivers/gpu/drm/amd/amdgpu/amdgpu_ring.c 	else if (ring == &adev->sdma.instance[0].page)
sdma               40 drivers/gpu/drm/amd/amdgpu/amdgpu_sdma.c 	for (i = 0; i < adev->sdma.num_instances; i++)
sdma               41 drivers/gpu/drm/amd/amdgpu/amdgpu_sdma.c 		if (ring == &adev->sdma.instance[i].ring ||
sdma               42 drivers/gpu/drm/amd/amdgpu/amdgpu_sdma.c 		    ring == &adev->sdma.instance[i].page)
sdma               43 drivers/gpu/drm/amd/amdgpu/amdgpu_sdma.c 			return &adev->sdma.instance[i];
sdma               53 drivers/gpu/drm/amd/amdgpu/amdgpu_sdma.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma               54 drivers/gpu/drm/amd/amdgpu/amdgpu_sdma.c 		if (ring == &adev->sdma.instance[i].ring ||
sdma               55 drivers/gpu/drm/amd/amdgpu/amdgpu_sdma.c 			ring == &adev->sdma.instance[i].page) {
sdma              410 drivers/gpu/drm/amd/amdgpu/amdgpu_ucode.c FW_VERSION_ATTR(sdma_fw_version, 0444, sdma.instance[0].fw_version);
sdma              411 drivers/gpu/drm/amd/amdgpu/amdgpu_ucode.c FW_VERSION_ATTR(sdma2_fw_version, 0444, sdma.instance[1].fw_version);
sdma              261 drivers/gpu/drm/amd/amdgpu/amdgpu_ucode.h 	struct sdma_firmware_header_v1_0 sdma;
sdma               76 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma               77 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 			release_firmware(adev->sdma.instance[i].fw);
sdma               78 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 			adev->sdma.instance[i].fw = NULL;
sdma              135 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              140 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 		err = request_firmware(&adev->sdma.instance[i].fw, fw_name, adev->dev);
sdma              143 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 		err = amdgpu_ucode_validate(adev->sdma.instance[i].fw);
sdma              148 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 		for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              149 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 			release_firmware(adev->sdma.instance[i].fw);
sdma              150 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 			adev->sdma.instance[i].fw = NULL;
sdma              203 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 	struct amdgpu_sdma_instance *sdma = amdgpu_sdma_get_instance_from_ring(ring);
sdma              207 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 		if (sdma && sdma->burst_nop && (i == 0))
sdma              308 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 	struct amdgpu_ring *sdma0 = &adev->sdma.instance[0].ring;
sdma              309 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 	struct amdgpu_ring *sdma1 = &adev->sdma.instance[1].ring;
sdma              317 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              376 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              414 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              440 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              441 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 		ring = &adev->sdma.instance[i].ring;
sdma              506 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              507 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 		ring = &adev->sdma.instance[i].ring;
sdma              551 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              552 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 		if (!adev->sdma.instance[i].fw)
sdma              554 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 		hdr = (const struct sdma_firmware_header_v1_0 *)adev->sdma.instance[i].fw->data;
sdma              557 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 		adev->sdma.instance[i].fw_version = le32_to_cpu(hdr->header.ucode_version);
sdma              558 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 		adev->sdma.instance[i].feature_version = le32_to_cpu(hdr->ucode_feature_version);
sdma              559 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 		if (adev->sdma.instance[i].feature_version >= 20)
sdma              560 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 			adev->sdma.instance[i].burst_nop = true;
sdma              562 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 			(adev->sdma.instance[i].fw->data + le32_to_cpu(hdr->header.ucode_array_offset_bytes));
sdma              566 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 		WREG32(mmSDMA0_UCODE_ADDR + sdma_offsets[i], adev->sdma.instance[i].fw_version);
sdma              810 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 	struct amdgpu_sdma_instance *sdma = amdgpu_sdma_get_instance_from_ring(ring);
sdma              816 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 		if (sdma && sdma->burst_nop && (i == 0))
sdma              935 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 	adev->sdma.num_instances = SDMA_MAX_INSTANCE;
sdma              959 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 			      &adev->sdma.trap_irq);
sdma              965 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 			      &adev->sdma.illegal_inst_irq);
sdma              971 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 			      &adev->sdma.illegal_inst_irq);
sdma              975 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              976 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 		ring = &adev->sdma.instance[i].ring;
sdma              980 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 				     &adev->sdma.trap_irq,
sdma              996 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 	for (i = 0; i < adev->sdma.num_instances; i++)
sdma              997 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 		amdgpu_ring_fini(&adev->sdma.instance[i].ring);
sdma             1170 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 			amdgpu_fence_process(&adev->sdma.instance[0].ring);
sdma             1183 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 			amdgpu_fence_process(&adev->sdma.instance[1].ring);
sdma             1206 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 	drm_sched_fault(&adev->sdma.instance[instance_id].ring.sched);
sdma             1279 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma             1280 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 		adev->sdma.instance[i].ring.funcs = &cik_sdma_ring_funcs;
sdma             1281 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 		adev->sdma.instance[i].ring.me = i;
sdma             1296 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 	adev->sdma.trap_irq.num_types = AMDGPU_SDMA_IRQ_LAST;
sdma             1297 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 	adev->sdma.trap_irq.funcs = &cik_sdma_trap_irq_funcs;
sdma             1298 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 	adev->sdma.illegal_inst_irq.funcs = &cik_sdma_illegal_inst_irq_funcs;
sdma             1362 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 	adev->mman.buffer_funcs_ring = &adev->sdma.instance[0].ring;
sdma             1379 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma             1380 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 		sched = &adev->sdma.instance[i].ring.sched;
sdma             1384 drivers/gpu/drm/amd/amdgpu/cik_sdma.c 	adev->vm_manager.vm_pte_num_rqs = adev->sdma.num_instances;
sdma             2183 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma             2185 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c 			adev->sdma.instance[i].fw->data;
sdma             2186 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c 		fw_data = (const __le32 *) (adev->sdma.instance[i].fw->data +
sdma              116 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              117 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 		release_firmware(adev->sdma.instance[i].fw);
sdma              118 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 		adev->sdma.instance[i].fw = NULL;
sdma              149 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              154 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 		err = request_firmware(&adev->sdma.instance[i].fw, fw_name, adev->dev);
sdma              157 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 		err = amdgpu_ucode_validate(adev->sdma.instance[i].fw);
sdma              160 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 		hdr = (const struct sdma_firmware_header_v1_0 *)adev->sdma.instance[i].fw->data;
sdma              161 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 		adev->sdma.instance[i].fw_version = le32_to_cpu(hdr->header.ucode_version);
sdma              162 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 		adev->sdma.instance[i].feature_version = le32_to_cpu(hdr->ucode_feature_version);
sdma              163 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 		if (adev->sdma.instance[i].feature_version >= 20)
sdma              164 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 			adev->sdma.instance[i].burst_nop = true;
sdma              169 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 			info->fw = adev->sdma.instance[i].fw;
sdma              179 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 		for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              180 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 			release_firmware(adev->sdma.instance[i].fw);
sdma              181 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 			adev->sdma.instance[i].fw = NULL;
sdma              231 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 	struct amdgpu_sdma_instance *sdma = amdgpu_sdma_get_instance_from_ring(ring);
sdma              235 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 		if (sdma && sdma->burst_nop && (i == 0))
sdma              341 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 	struct amdgpu_ring *sdma0 = &adev->sdma.instance[0].ring;
sdma              342 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 	struct amdgpu_ring *sdma1 = &adev->sdma.instance[1].ring;
sdma              350 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              392 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              418 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              419 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 		ring = &adev->sdma.instance[i].ring;
sdma              484 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              485 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 		ring = &adev->sdma.instance[i].ring;
sdma              749 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 	struct amdgpu_sdma_instance *sdma = amdgpu_sdma_get_instance_from_ring(ring);
sdma              755 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 		if (sdma && sdma->burst_nop && (i == 0))
sdma              828 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 	adev->sdma.num_instances = SDMA_MAX_INSTANCE;
sdma              846 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 			      &adev->sdma.trap_irq);
sdma              852 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 			      &adev->sdma.illegal_inst_irq);
sdma              858 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 			      &adev->sdma.illegal_inst_irq);
sdma              868 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              869 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 		ring = &adev->sdma.instance[i].ring;
sdma              874 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 				     &adev->sdma.trap_irq,
sdma              890 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 	for (i = 0; i < adev->sdma.num_instances; i++)
sdma              891 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 		amdgpu_ring_fini(&adev->sdma.instance[i].ring);
sdma             1063 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 			amdgpu_fence_process(&adev->sdma.instance[0].ring);
sdma             1076 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 			amdgpu_fence_process(&adev->sdma.instance[1].ring);
sdma             1101 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 		drm_sched_fault(&adev->sdma.instance[instance_id].ring.sched);
sdma             1166 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma             1167 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 		adev->sdma.instance[i].ring.funcs = &sdma_v2_4_ring_funcs;
sdma             1168 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 		adev->sdma.instance[i].ring.me = i;
sdma             1183 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 	adev->sdma.trap_irq.num_types = AMDGPU_SDMA_IRQ_LAST;
sdma             1184 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 	adev->sdma.trap_irq.funcs = &sdma_v2_4_trap_irq_funcs;
sdma             1185 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 	adev->sdma.illegal_inst_irq.funcs = &sdma_v2_4_illegal_inst_irq_funcs;
sdma             1250 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 	adev->mman.buffer_funcs_ring = &adev->sdma.instance[0].ring;
sdma             1267 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma             1268 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 		sched = &adev->sdma.instance[i].ring.sched;
sdma             1272 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c 	adev->vm_manager.vm_pte_num_rqs = adev->sdma.num_instances;
sdma              253 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              254 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 		release_firmware(adev->sdma.instance[i].fw);
sdma              255 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 		adev->sdma.instance[i].fw = NULL;
sdma              307 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              312 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 		err = request_firmware(&adev->sdma.instance[i].fw, fw_name, adev->dev);
sdma              315 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 		err = amdgpu_ucode_validate(adev->sdma.instance[i].fw);
sdma              318 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 		hdr = (const struct sdma_firmware_header_v1_0 *)adev->sdma.instance[i].fw->data;
sdma              319 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 		adev->sdma.instance[i].fw_version = le32_to_cpu(hdr->header.ucode_version);
sdma              320 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 		adev->sdma.instance[i].feature_version = le32_to_cpu(hdr->ucode_feature_version);
sdma              321 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 		if (adev->sdma.instance[i].feature_version >= 20)
sdma              322 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 			adev->sdma.instance[i].burst_nop = true;
sdma              326 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 		info->fw = adev->sdma.instance[i].fw;
sdma              335 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 		for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              336 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 			release_firmware(adev->sdma.instance[i].fw);
sdma              337 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 			adev->sdma.instance[i].fw = NULL;
sdma              405 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 	struct amdgpu_sdma_instance *sdma = amdgpu_sdma_get_instance_from_ring(ring);
sdma              409 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 		if (sdma && sdma->burst_nop && (i == 0))
sdma              515 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 	struct amdgpu_ring *sdma0 = &adev->sdma.instance[0].ring;
sdma              516 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 	struct amdgpu_ring *sdma1 = &adev->sdma.instance[1].ring;
sdma              524 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              585 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              627 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              655 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              656 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 		ring = &adev->sdma.instance[i].ring;
sdma              756 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              757 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 		ring = &adev->sdma.instance[i].ring;
sdma             1020 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 	struct amdgpu_sdma_instance *sdma = amdgpu_sdma_get_instance_from_ring(ring);
sdma             1026 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 		if (sdma && sdma->burst_nop && (i == 0))
sdma             1101 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 		adev->sdma.num_instances = 1;
sdma             1104 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 		adev->sdma.num_instances = SDMA_MAX_INSTANCE;
sdma             1124 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 			      &adev->sdma.trap_irq);
sdma             1130 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 			      &adev->sdma.illegal_inst_irq);
sdma             1136 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 			      &adev->sdma.illegal_inst_irq);
sdma             1146 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma             1147 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 		ring = &adev->sdma.instance[i].ring;
sdma             1158 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 				     &adev->sdma.trap_irq,
sdma             1174 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 	for (i = 0; i < adev->sdma.num_instances; i++)
sdma             1175 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 		amdgpu_ring_fini(&adev->sdma.instance[i].ring);
sdma             1261 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 		adev->sdma.srbm_soft_reset = srbm_soft_reset;
sdma             1264 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 		adev->sdma.srbm_soft_reset = 0;
sdma             1274 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 	if (!adev->sdma.srbm_soft_reset)
sdma             1277 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 	srbm_soft_reset = adev->sdma.srbm_soft_reset;
sdma             1293 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 	if (!adev->sdma.srbm_soft_reset)
sdma             1296 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 	srbm_soft_reset = adev->sdma.srbm_soft_reset;
sdma             1313 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 	if (!adev->sdma.srbm_soft_reset)
sdma             1316 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 	srbm_soft_reset = adev->sdma.srbm_soft_reset;
sdma             1397 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 			amdgpu_fence_process(&adev->sdma.instance[0].ring);
sdma             1410 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 			amdgpu_fence_process(&adev->sdma.instance[1].ring);
sdma             1435 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 		drm_sched_fault(&adev->sdma.instance[instance_id].ring.sched);
sdma             1447 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 		for (i = 0; i < adev->sdma.num_instances; i++) {
sdma             1461 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 		for (i = 0; i < adev->sdma.num_instances; i++) {
sdma             1486 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 		for (i = 0; i < adev->sdma.num_instances; i++) {
sdma             1494 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 		for (i = 0; i < adev->sdma.num_instances; i++) {
sdma             1604 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma             1605 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 		adev->sdma.instance[i].ring.funcs = &sdma_v3_0_ring_funcs;
sdma             1606 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 		adev->sdma.instance[i].ring.me = i;
sdma             1621 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 	adev->sdma.trap_irq.num_types = AMDGPU_SDMA_IRQ_LAST;
sdma             1622 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 	adev->sdma.trap_irq.funcs = &sdma_v3_0_trap_irq_funcs;
sdma             1623 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 	adev->sdma.illegal_inst_irq.funcs = &sdma_v3_0_illegal_inst_irq_funcs;
sdma             1688 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 	adev->mman.buffer_funcs_ring = &adev->sdma.instance[0].ring;
sdma             1705 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma             1706 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 		sched = &adev->sdma.instance[i].ring.sched;
sdma             1710 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c 	adev->vm_manager.vm_pte_num_rqs = adev->sdma.num_instances;
sdma              417 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              418 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		if (adev->sdma.instance[i].fw != NULL)
sdma              419 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 			release_firmware(adev->sdma.instance[i].fw);
sdma              427 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	memset((void*)adev->sdma.instance, 0,
sdma              483 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	err = request_firmware(&adev->sdma.instance[0].fw, fw_name, adev->dev);
sdma              487 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	err = sdma_v4_0_init_inst_ctx(&adev->sdma.instance[0]);
sdma              491 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	for (i = 1; i < adev->sdma.num_instances; i++) {
sdma              495 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 			memcpy((void*)&adev->sdma.instance[i],
sdma              496 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 			       (void*)&adev->sdma.instance[0],
sdma              502 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 			err = request_firmware(&adev->sdma.instance[i].fw, fw_name, adev->dev);
sdma              506 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 			err = sdma_v4_0_init_inst_ctx(&adev->sdma.instance[i]);
sdma              516 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              519 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 			info->fw = adev->sdma.instance[i].fw;
sdma              674 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	struct amdgpu_sdma_instance *sdma = amdgpu_sdma_get_instance_from_ring(ring);
sdma              678 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		if (sdma && sdma->burst_nop && (i == 0))
sdma              808 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	struct amdgpu_ring *sdma[AMDGPU_MAX_SDMA_INSTANCES];
sdma              812 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              813 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		sdma[i] = &adev->sdma.instance[i].ring;
sdma              815 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		if ((adev->mman.buffer_funcs_ring == sdma[i]) && unset != 1) {
sdma              827 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		sdma[i]->sched.ready = false;
sdma              852 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	struct amdgpu_ring *sdma[AMDGPU_MAX_SDMA_INSTANCES];
sdma              857 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              858 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		sdma[i] = &adev->sdma.instance[i].page;
sdma              860 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		if ((adev->mman.buffer_funcs_ring == sdma[i]) &&
sdma              875 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		sdma[i]->sched.ready = false;
sdma              916 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              946 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		if (adev->sdma.has_page_queue)
sdma              950 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              985 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	struct amdgpu_ring *ring = &adev->sdma.instance[i].ring;
sdma             1075 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	struct amdgpu_ring *ring = &adev->sdma.instance[i].page;
sdma             1252 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma             1253 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		if (!adev->sdma.instance[i].fw)
sdma             1256 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		hdr = (const struct sdma_firmware_header_v1_0 *)adev->sdma.instance[i].fw->data;
sdma             1261 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 			(adev->sdma.instance[i].fw->data +
sdma             1271 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 			    adev->sdma.instance[i].fw_version);
sdma             1308 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma             1313 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		if (adev->sdma.has_page_queue)
sdma             1338 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma             1339 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		ring = &adev->sdma.instance[i].ring;
sdma             1345 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		if (adev->sdma.has_page_queue) {
sdma             1346 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 			struct amdgpu_ring *page = &adev->sdma.instance[i].page;
sdma             1578 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	struct amdgpu_sdma_instance *sdma = amdgpu_sdma_get_instance_from_ring(ring);
sdma             1584 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		if (sdma && sdma->burst_nop && (i == 0))
sdma             1646 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	uint fw_version = adev->sdma.instance[0].fw_version;
sdma             1667 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		adev->sdma.num_instances = 1;
sdma             1669 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		adev->sdma.num_instances = 8;
sdma             1671 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		adev->sdma.num_instances = 2;
sdma             1681 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		adev->sdma.has_page_queue = false;
sdma             1683 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		adev->sdma.has_page_queue = true;
sdma             1700 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	struct ras_common_if **ras_if = &adev->sdma.ras_if;
sdma             1771 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma             1772 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		r = amdgpu_irq_get(adev, &adev->sdma.ecc_irq,
sdma             1799 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma             1802 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 				      &adev->sdma.trap_irq);
sdma             1808 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma             1811 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 				      &adev->sdma.ecc_irq);
sdma             1816 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma             1817 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		ring = &adev->sdma.instance[i].ring;
sdma             1828 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		r = amdgpu_ring_init(adev, ring, 1024, &adev->sdma.trap_irq,
sdma             1833 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		if (adev->sdma.has_page_queue) {
sdma             1834 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 			ring = &adev->sdma.instance[i].page;
sdma             1846 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 					     &adev->sdma.trap_irq,
sdma             1862 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 			adev->sdma.ras_if) {
sdma             1863 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		struct ras_common_if *ras_if = adev->sdma.ras_if;
sdma             1877 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma             1878 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		amdgpu_ring_fini(&adev->sdma.instance[i].ring);
sdma             1879 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		if (adev->sdma.has_page_queue)
sdma             1880 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 			amdgpu_ring_fini(&adev->sdma.instance[i].page);
sdma             1914 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma             1915 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		amdgpu_irq_put(adev, &adev->sdma.ecc_irq,
sdma             1949 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma             1962 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	u32 sdma[AMDGPU_MAX_SDMA_INSTANCES];
sdma             1966 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		for (j = 0; j < adev->sdma.num_instances; j++) {
sdma             1967 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 			sdma[j] = RREG32_SDMA(j, mmSDMA0_STATUS_REG);
sdma             1968 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 			if (!(sdma[j] & SDMA0_STATUS_REG__IDLE_MASK))
sdma             1971 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		if (j == adev->sdma.num_instances)
sdma             2010 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		amdgpu_fence_process(&adev->sdma.instance[instance].ring);
sdma             2014 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 			amdgpu_fence_process(&adev->sdma.instance[instance].page);
sdma             2021 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 			amdgpu_fence_process(&adev->sdma.instance[instance].page);
sdma             2060 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	struct ras_common_if *ras_if = adev->sdma.ras_if;
sdma             2088 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		drm_sched_fault(&adev->sdma.instance[instance].ring.sched);
sdma             2117 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		for (i = 0; i < adev->sdma.num_instances; i++) {
sdma             2131 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		for (i = 0; i < adev->sdma.num_instances; i++) {
sdma             2156 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		for (i = 0; i < adev->sdma.num_instances; i++) {
sdma             2164 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		for (i = 0; i < adev->sdma.num_instances; i++) {
sdma             2390 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma             2392 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 			adev->sdma.instance[i].ring.funcs =
sdma             2395 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 			adev->sdma.instance[i].ring.funcs =
sdma             2397 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		adev->sdma.instance[i].ring.me = i;
sdma             2398 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		if (adev->sdma.has_page_queue) {
sdma             2400 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 				adev->sdma.instance[i].page.funcs =
sdma             2403 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 				adev->sdma.instance[i].page.funcs =
sdma             2405 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 			adev->sdma.instance[i].page.me = i;
sdma             2428 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	switch (adev->sdma.num_instances) {
sdma             2430 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		adev->sdma.trap_irq.num_types = AMDGPU_SDMA_IRQ_INSTANCE1;
sdma             2431 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		adev->sdma.ecc_irq.num_types = AMDGPU_SDMA_IRQ_INSTANCE1;
sdma             2434 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		adev->sdma.trap_irq.num_types = AMDGPU_SDMA_IRQ_LAST;
sdma             2435 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		adev->sdma.ecc_irq.num_types = AMDGPU_SDMA_IRQ_LAST;
sdma             2439 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		adev->sdma.trap_irq.num_types = AMDGPU_SDMA_IRQ_INSTANCE2;
sdma             2440 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		adev->sdma.ecc_irq.num_types = AMDGPU_SDMA_IRQ_INSTANCE2;
sdma             2443 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	adev->sdma.trap_irq.funcs = &sdma_v4_0_trap_irq_funcs;
sdma             2444 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	adev->sdma.illegal_inst_irq.funcs = &sdma_v4_0_illegal_inst_irq_funcs;
sdma             2445 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	adev->sdma.ecc_irq.funcs = &sdma_v4_0_ecc_irq_funcs;
sdma             2510 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	if (adev->sdma.has_page_queue)
sdma             2511 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		adev->mman.buffer_funcs_ring = &adev->sdma.instance[0].page;
sdma             2513 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		adev->mman.buffer_funcs_ring = &adev->sdma.instance[0].ring;
sdma             2530 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma             2531 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 		if (adev->sdma.has_page_queue)
sdma             2532 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 			sched = &adev->sdma.instance[i].page.sched;
sdma             2534 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 			sched = &adev->sdma.instance[i].ring.sched;
sdma             2538 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c 	adev->vm_manager.vm_pte_num_rqs = adev->sdma.num_instances;
sdma              193 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              198 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 		err = request_firmware(&adev->sdma.instance[i].fw, fw_name, adev->dev);
sdma              201 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 		err = amdgpu_ucode_validate(adev->sdma.instance[i].fw);
sdma              204 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 		hdr = (const struct sdma_firmware_header_v1_0 *)adev->sdma.instance[i].fw->data;
sdma              205 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 		adev->sdma.instance[i].fw_version = le32_to_cpu(hdr->header.ucode_version);
sdma              206 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 		adev->sdma.instance[i].feature_version = le32_to_cpu(hdr->ucode_feature_version);
sdma              207 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 		if (adev->sdma.instance[i].feature_version >= 20)
sdma              208 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 			adev->sdma.instance[i].burst_nop = true;
sdma              215 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 			info->fw = adev->sdma.instance[i].fw;
sdma              224 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 		for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              225 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 			release_firmware(adev->sdma.instance[i].fw);
sdma              226 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 			adev->sdma.instance[i].fw = NULL;
sdma              358 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 	struct amdgpu_sdma_instance *sdma = amdgpu_sdma_get_instance_from_ring(ring);
sdma              362 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 		if (sdma && sdma->burst_nop && (i == 0))
sdma              500 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 	struct amdgpu_ring *sdma0 = &adev->sdma.instance[0].ring;
sdma              501 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 	struct amdgpu_ring *sdma1 = &adev->sdma.instance[1].ring;
sdma              509 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              571 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              606 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              634 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              635 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 		ring = &adev->sdma.instance[i].ring;
sdma              805 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              806 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 		if (!adev->sdma.instance[i].fw)
sdma              809 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 		hdr = (const struct sdma_firmware_header_v1_0 *)adev->sdma.instance[i].fw->data;
sdma              814 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 			(adev->sdma.instance[i].fw->data +
sdma              825 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 		WREG32(sdma_v5_0_get_reg_offset(adev, i, mmSDMA0_UCODE_ADDR), adev->sdma.instance[i].fw_version);
sdma             1115 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 	struct amdgpu_sdma_instance *sdma = amdgpu_sdma_get_instance_from_ring(ring);
sdma             1121 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 		if (sdma && sdma->burst_nop && (i == 0))
sdma             1209 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 	adev->sdma.num_instances = 2;
sdma             1229 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 			      &adev->sdma.trap_irq);
sdma             1236 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 			      &adev->sdma.trap_irq);
sdma             1246 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma             1247 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 		ring = &adev->sdma.instance[i].ring;
sdma             1260 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 				     &adev->sdma.trap_irq,
sdma             1276 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 	for (i = 0; i < adev->sdma.num_instances; i++)
sdma             1277 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 		amdgpu_ring_fini(&adev->sdma.instance[i].ring);
sdma             1326 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma             1435 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 			amdgpu_fence_process(&adev->sdma.instance[0].ring);
sdma             1451 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 			amdgpu_fence_process(&adev->sdma.instance[1].ring);
sdma             1481 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma             1518 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma             1644 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma             1645 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 		adev->sdma.instance[i].ring.funcs = &sdma_v5_0_ring_funcs;
sdma             1646 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 		adev->sdma.instance[i].ring.me = i;
sdma             1661 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 	adev->sdma.trap_irq.num_types = AMDGPU_SDMA_IRQ_INSTANCE0 +
sdma             1662 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 					adev->sdma.num_instances;
sdma             1663 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 	adev->sdma.trap_irq.funcs = &sdma_v5_0_trap_irq_funcs;
sdma             1664 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 	adev->sdma.illegal_inst_irq.funcs = &sdma_v5_0_illegal_inst_irq_funcs;
sdma             1730 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 		adev->mman.buffer_funcs_ring = &adev->sdma.instance[0].ring;
sdma             1748 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 		for (i = 0; i < adev->sdma.num_instances; i++) {
sdma             1749 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 			sched = &adev->sdma.instance[i].ring.sched;
sdma             1753 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 		adev->vm_manager.vm_pte_num_rqs = adev->sdma.num_instances;
sdma               49 drivers/gpu/drm/amd/amdgpu/si_dma.c 	u32 me = (ring == &adev->sdma.instance[0].ring) ? 0 : 1;
sdma               57 drivers/gpu/drm/amd/amdgpu/si_dma.c 	u32 me = (ring == &adev->sdma.instance[0].ring) ? 0 : 1;
sdma              118 drivers/gpu/drm/amd/amdgpu/si_dma.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              119 drivers/gpu/drm/amd/amdgpu/si_dma.c 		ring = &adev->sdma.instance[i].ring;
sdma              138 drivers/gpu/drm/amd/amdgpu/si_dma.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              139 drivers/gpu/drm/amd/amdgpu/si_dma.c 		ring = &adev->sdma.instance[i].ring;
sdma              470 drivers/gpu/drm/amd/amdgpu/si_dma.c 	adev->sdma.num_instances = 2;
sdma              488 drivers/gpu/drm/amd/amdgpu/si_dma.c 			      &adev->sdma.trap_irq);
sdma              494 drivers/gpu/drm/amd/amdgpu/si_dma.c 			      &adev->sdma.trap_irq);
sdma              498 drivers/gpu/drm/amd/amdgpu/si_dma.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              499 drivers/gpu/drm/amd/amdgpu/si_dma.c 		ring = &adev->sdma.instance[i].ring;
sdma              504 drivers/gpu/drm/amd/amdgpu/si_dma.c 				     &adev->sdma.trap_irq,
sdma              520 drivers/gpu/drm/amd/amdgpu/si_dma.c 	for (i = 0; i < adev->sdma.num_instances; i++)
sdma              521 drivers/gpu/drm/amd/amdgpu/si_dma.c 		amdgpu_ring_fini(&adev->sdma.instance[i].ring);
sdma              637 drivers/gpu/drm/amd/amdgpu/si_dma.c 		amdgpu_fence_process(&adev->sdma.instance[0].ring);
sdma              639 drivers/gpu/drm/amd/amdgpu/si_dma.c 		amdgpu_fence_process(&adev->sdma.instance[1].ring);
sdma              654 drivers/gpu/drm/amd/amdgpu/si_dma.c 		for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              666 drivers/gpu/drm/amd/amdgpu/si_dma.c 		for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              748 drivers/gpu/drm/amd/amdgpu/si_dma.c 	for (i = 0; i < adev->sdma.num_instances; i++)
sdma              749 drivers/gpu/drm/amd/amdgpu/si_dma.c 		adev->sdma.instance[i].ring.funcs = &si_dma_ring_funcs;
sdma              759 drivers/gpu/drm/amd/amdgpu/si_dma.c 	adev->sdma.trap_irq.num_types = AMDGPU_SDMA_IRQ_LAST;
sdma              760 drivers/gpu/drm/amd/amdgpu/si_dma.c 	adev->sdma.trap_irq.funcs = &si_dma_trap_irq_funcs;
sdma              824 drivers/gpu/drm/amd/amdgpu/si_dma.c 	adev->mman.buffer_funcs_ring = &adev->sdma.instance[0].ring;
sdma              841 drivers/gpu/drm/amd/amdgpu/si_dma.c 	for (i = 0; i < adev->sdma.num_instances; i++) {
sdma              842 drivers/gpu/drm/amd/amdgpu/si_dma.c 		sched = &adev->sdma.instance[i].ring.sched;
sdma              846 drivers/gpu/drm/amd/amdgpu/si_dma.c 	adev->vm_manager.vm_pte_num_rqs = adev->sdma.num_instances;
sdma             1263 drivers/gpu/drm/amd/amdgpu/soc15.c 		for (i = 0; i < adev->sdma.num_instances; i++) {
sdma             1264 drivers/gpu/drm/amd/amdgpu/soc15.c 			ring = &adev->sdma.instance[i].ring;
sdma              216 drivers/gpu/drm/radeon/radeon_ucode.h 	struct sdma_firmware_header_v1_0 sdma;
sdma              137 drivers/infiniband/hw/hfi1/vnic.h 	struct hfi1_vnic_sdma  sdma[HFI1_VNIC_MAX_TXQ];
sdma               72 drivers/infiniband/hw/hfi1/vnic_sdma.c 	struct hfi1_vnic_sdma   *sdma;
sdma               84 drivers/infiniband/hw/hfi1/vnic_sdma.c 	struct hfi1_vnic_sdma *vnic_sdma = tx->sdma;
sdma              171 drivers/infiniband/hw/hfi1/vnic_sdma.c 	struct hfi1_vnic_sdma *vnic_sdma = &vinfo->sdma[q_idx];
sdma              188 drivers/infiniband/hw/hfi1/vnic_sdma.c 	tx->sdma = vnic_sdma;
sdma              272 drivers/infiniband/hw/hfi1/vnic_sdma.c 	struct hfi1_vnic_sdma *vnic_sdma = &vinfo->sdma[q_idx];
sdma              282 drivers/infiniband/hw/hfi1/vnic_sdma.c 		struct hfi1_vnic_sdma *vnic_sdma = &vinfo->sdma[i];
sdma              170 drivers/media/pci/ddbridge/ddbridge-core.c 			     struct ddb_dma *sdma,
sdma              176 drivers/media/pci/ddbridge/ddbridge-core.c 	sdma->bufval = ddma->bufval;
sdma              177 drivers/media/pci/ddbridge/ddbridge-core.c 	base = sdma->bufregs;
sdma              809 drivers/scsi/isci/host.c 	       &ihost->scu_registers->sdma.unsolicited_frame_queue_control);
sdma              818 drivers/scsi/isci/host.c 	       &ihost->scu_registers->sdma.unsolicited_frame_get_pointer);
sdma              822 drivers/scsi/isci/host.c 	       &ihost->scu_registers->sdma.unsolicited_frame_put_pointer);
sdma             1557 drivers/scsi/isci/host.c 	writel(0, &ihost->scu_registers->sdma.unsolicited_frame_get_pointer);
sdma             2205 drivers/scsi/isci/host.c 	val = readl(&ihost->scu_registers->sdma.pdma_configuration);
sdma             2207 drivers/scsi/isci/host.c 	writel(val, &ihost->scu_registers->sdma.pdma_configuration);
sdma             2209 drivers/scsi/isci/host.c 	val = readl(&ihost->scu_registers->sdma.cdma_configuration);
sdma             2211 drivers/scsi/isci/host.c 	writel(val, &ihost->scu_registers->sdma.cdma_configuration);
sdma             2321 drivers/scsi/isci/host.c 		&ihost->scu_registers->sdma.uf_header_base_address_lower);
sdma             2323 drivers/scsi/isci/host.c 		&ihost->scu_registers->sdma.uf_header_base_address_upper);
sdma             2326 drivers/scsi/isci/host.c 		&ihost->scu_registers->sdma.uf_address_table_lower);
sdma             2328 drivers/scsi/isci/host.c 		&ihost->scu_registers->sdma.uf_address_table_upper);
sdma             2548 drivers/scsi/isci/host.c 			&ihost->scu_registers->sdma.unsolicited_frame_get_pointer);
sdma             1841 drivers/scsi/isci/registers.h 	struct scu_sdma_registers sdma;
sdma              366 drivers/soc/fsl/qe/qe.c 	struct sdma __iomem *sdma = &qe_immr->sdma;
sdma              369 drivers/soc/fsl/qe/qe.c 	if (!sdma)
sdma              380 drivers/soc/fsl/qe/qe.c 	out_be32(&sdma->sdebcr, (u32) sdma_buf_offset & QE_SDEBCR_BA_MASK);
sdma              381 drivers/soc/fsl/qe/qe.c  	out_be32(&sdma->sdmr, (QE_SDMR_GLB_1_MSK |
sdma              447 include/soc/fsl/qe/immap_qe.h 	struct sdma		sdma;		/* SDMA */
sdma              206 sound/parisc/harmony.c 			harmony_write(h, HARMONY_PNXTADD, h->sdma.addr);
sdma              312 sound/parisc/harmony.c 		harmony_write(h, HARMONY_PNXTADD, h->sdma.addr);
sdma              340 sound/parisc/harmony.c                 harmony_write(h, HARMONY_PNXTADD, h->sdma.addr);
sdma              386 sound/parisc/harmony.c 		snd_pcm_format_set_silence(fmt, h->sdma.area, SILENCE_BUFSZ / 
sdma              652 sound/parisc/harmony.c 				  &h->sdma);
sdma              871 sound/parisc/harmony.c         if (h->sdma.addr)
sdma              872 sound/parisc/harmony.c                 snd_dma_free_pages(&h->sdma);
sdma               39 sound/parisc/harmony.h         struct snd_dma_buffer sdma; /* silence */