sdma 200 arch/mips/include/asm/octeon/cvmx-address.h } sdma; sdma 265 arch/mips/include/asm/octeon/cvmx-address.h } sdma; sdma 134 arch/powerpc/platforms/52xx/mpc52xx_pic.c static struct mpc52xx_sdma __iomem *sdma; sdma 269 arch/powerpc/platforms/52xx/mpc52xx_pic.c io_be_setbit(&sdma->IntMask, l2irq); sdma 275 arch/powerpc/platforms/52xx/mpc52xx_pic.c io_be_clrbit(&sdma->IntMask, l2irq); sdma 281 arch/powerpc/platforms/52xx/mpc52xx_pic.c out_be32(&sdma->IntPend, 1 << l2irq); sdma 416 arch/powerpc/platforms/52xx/mpc52xx_pic.c sdma = of_iomap(np, 0); sdma 418 arch/powerpc/platforms/52xx/mpc52xx_pic.c if (!sdma) sdma 425 arch/powerpc/platforms/52xx/mpc52xx_pic.c out_be32(&sdma->IntPend, 0xffffffff); /* 1 means clear pending */ sdma 426 arch/powerpc/platforms/52xx/mpc52xx_pic.c out_be32(&sdma->IntMask, 0xffffffff); /* 1 means disabled */ sdma 507 arch/powerpc/platforms/52xx/mpc52xx_pic.c status = in_be32(&sdma->IntPend); sdma 108 arch/s390/include/asm/pci_clp.h u64 sdma; /* start dma as */ sdma 152 arch/s390/pci/pci_clp.c zdev->start_dma = response->sdma; sdma 363 drivers/dma/imx-sdma.c struct sdma_engine *sdma; sdma 617 drivers/dma/imx-sdma.c static inline u32 chnenbl_ofs(struct sdma_engine *sdma, unsigned int event) sdma 619 drivers/dma/imx-sdma.c u32 chnenbl0 = sdma->drvdata->chnenbl0; sdma 626 drivers/dma/imx-sdma.c struct sdma_engine *sdma = sdmac->sdma; sdma 633 drivers/dma/imx-sdma.c evt = readl_relaxed(sdma->regs + SDMA_H_EVTOVR); sdma 634 drivers/dma/imx-sdma.c mcu = readl_relaxed(sdma->regs + SDMA_H_HOSTOVR); sdma 635 drivers/dma/imx-sdma.c dsp = readl_relaxed(sdma->regs + SDMA_H_DSPOVR); sdma 652 drivers/dma/imx-sdma.c writel_relaxed(evt, sdma->regs + SDMA_H_EVTOVR); sdma 653 drivers/dma/imx-sdma.c writel_relaxed(mcu, sdma->regs + SDMA_H_HOSTOVR); sdma 654 drivers/dma/imx-sdma.c writel_relaxed(dsp, sdma->regs + SDMA_H_DSPOVR); sdma 659 drivers/dma/imx-sdma.c static void sdma_enable_channel(struct sdma_engine *sdma, int channel) sdma 661 drivers/dma/imx-sdma.c writel(BIT(channel), sdma->regs + SDMA_H_START); sdma 667 drivers/dma/imx-sdma.c static int sdma_run_channel0(struct sdma_engine *sdma) sdma 672 drivers/dma/imx-sdma.c sdma_enable_channel(sdma, 0); sdma 674 drivers/dma/imx-sdma.c ret = readl_relaxed_poll_timeout_atomic(sdma->regs + SDMA_H_STATSTOP, sdma 677 drivers/dma/imx-sdma.c dev_err(sdma->dev, "Timeout waiting for CH0 ready\n"); sdma 680 drivers/dma/imx-sdma.c reg = readl(sdma->regs + SDMA_H_CONFIG); sdma 683 drivers/dma/imx-sdma.c writel_relaxed(reg, sdma->regs + SDMA_H_CONFIG); sdma 689 drivers/dma/imx-sdma.c static int sdma_load_script(struct sdma_engine *sdma, void *buf, int size, sdma 692 drivers/dma/imx-sdma.c struct sdma_buffer_descriptor *bd0 = sdma->bd0; sdma 698 drivers/dma/imx-sdma.c buf_virt = dma_alloc_coherent(sdma->dev, size, &buf_phys, GFP_KERNEL); sdma 703 drivers/dma/imx-sdma.c spin_lock_irqsave(&sdma->channel_0_lock, flags); sdma 713 drivers/dma/imx-sdma.c ret = sdma_run_channel0(sdma); sdma 715 drivers/dma/imx-sdma.c spin_unlock_irqrestore(&sdma->channel_0_lock, flags); sdma 717 drivers/dma/imx-sdma.c dma_free_coherent(sdma->dev, size, buf_virt, buf_phys); sdma 724 drivers/dma/imx-sdma.c struct sdma_engine *sdma = sdmac->sdma; sdma 727 drivers/dma/imx-sdma.c u32 chnenbl = chnenbl_ofs(sdma, event); sdma 729 drivers/dma/imx-sdma.c val = readl_relaxed(sdma->regs + chnenbl); sdma 731 drivers/dma/imx-sdma.c writel_relaxed(val, sdma->regs + chnenbl); sdma 736 drivers/dma/imx-sdma.c struct sdma_engine *sdma = sdmac->sdma; sdma 738 drivers/dma/imx-sdma.c u32 chnenbl = chnenbl_ofs(sdma, event); sdma 741 drivers/dma/imx-sdma.c val = readl_relaxed(sdma->regs + chnenbl); sdma 743 drivers/dma/imx-sdma.c writel_relaxed(val, sdma->regs + chnenbl); sdma 755 drivers/dma/imx-sdma.c struct sdma_engine *sdma = sdmac->sdma; sdma 770 drivers/dma/imx-sdma.c sdma->channel_control[channel].base_bd_ptr = desc->bd_phys; sdma 771 drivers/dma/imx-sdma.c sdma->channel_control[channel].current_bd_ptr = desc->bd_phys; sdma 772 drivers/dma/imx-sdma.c sdma_enable_channel(sdma, sdmac->channel); sdma 852 drivers/dma/imx-sdma.c struct sdma_engine *sdma = dev_id; sdma 855 drivers/dma/imx-sdma.c stat = readl_relaxed(sdma->regs + SDMA_H_INTR); sdma 856 drivers/dma/imx-sdma.c writel_relaxed(stat, sdma->regs + SDMA_H_INTR); sdma 862 drivers/dma/imx-sdma.c struct sdma_channel *sdmac = &sdma->channel[channel]; sdma 890 drivers/dma/imx-sdma.c struct sdma_engine *sdma = sdmac->sdma; sdma 905 drivers/dma/imx-sdma.c emi_2_emi = sdma->script_addrs->ap_2_ap_addr; sdma 908 drivers/dma/imx-sdma.c emi_2_per = sdma->script_addrs->bp_2_ap_addr; sdma 909 drivers/dma/imx-sdma.c per_2_emi = sdma->script_addrs->ap_2_bp_addr; sdma 912 drivers/dma/imx-sdma.c per_2_emi = sdma->script_addrs->firi_2_mcu_addr; sdma 913 drivers/dma/imx-sdma.c emi_2_per = sdma->script_addrs->mcu_2_firi_addr; sdma 916 drivers/dma/imx-sdma.c per_2_emi = sdma->script_addrs->uart_2_mcu_addr; sdma 917 drivers/dma/imx-sdma.c emi_2_per = sdma->script_addrs->mcu_2_app_addr; sdma 920 drivers/dma/imx-sdma.c per_2_emi = sdma->script_addrs->uartsh_2_mcu_addr; sdma 921 drivers/dma/imx-sdma.c emi_2_per = sdma->script_addrs->mcu_2_shp_addr; sdma 924 drivers/dma/imx-sdma.c per_2_emi = sdma->script_addrs->ata_2_mcu_addr; sdma 925 drivers/dma/imx-sdma.c emi_2_per = sdma->script_addrs->mcu_2_ata_addr; sdma 931 drivers/dma/imx-sdma.c per_2_emi = sdma->script_addrs->app_2_mcu_addr; sdma 932 drivers/dma/imx-sdma.c emi_2_per = sdma->script_addrs->mcu_2_app_addr; sdma 935 drivers/dma/imx-sdma.c per_2_emi = sdma->script_addrs->ssish_2_mcu_addr; sdma 936 drivers/dma/imx-sdma.c emi_2_per = sdma->script_addrs->mcu_2_ssish_addr; sdma 944 drivers/dma/imx-sdma.c per_2_emi = sdma->script_addrs->shp_2_mcu_addr; sdma 945 drivers/dma/imx-sdma.c emi_2_per = sdma->script_addrs->mcu_2_shp_addr; sdma 948 drivers/dma/imx-sdma.c per_2_emi = sdma->script_addrs->asrc_2_mcu_addr; sdma 949 drivers/dma/imx-sdma.c emi_2_per = sdma->script_addrs->asrc_2_mcu_addr; sdma 950 drivers/dma/imx-sdma.c per_2_per = sdma->script_addrs->per_2_per_addr; sdma 953 drivers/dma/imx-sdma.c per_2_emi = sdma->script_addrs->shp_2_mcu_addr; sdma 954 drivers/dma/imx-sdma.c emi_2_per = sdma->script_addrs->mcu_2_shp_addr; sdma 955 drivers/dma/imx-sdma.c per_2_per = sdma->script_addrs->per_2_per_addr; sdma 958 drivers/dma/imx-sdma.c per_2_emi = sdma->script_addrs->mshc_2_mcu_addr; sdma 959 drivers/dma/imx-sdma.c emi_2_per = sdma->script_addrs->mcu_2_mshc_addr; sdma 962 drivers/dma/imx-sdma.c per_2_emi = sdma->script_addrs->dptc_dvfs_addr; sdma 965 drivers/dma/imx-sdma.c per_2_emi = sdma->script_addrs->spdif_2_mcu_addr; sdma 966 drivers/dma/imx-sdma.c emi_2_per = sdma->script_addrs->mcu_2_spdif_addr; sdma 969 drivers/dma/imx-sdma.c emi_2_per = sdma->script_addrs->ext_mem_2_ipu_addr; sdma 983 drivers/dma/imx-sdma.c struct sdma_engine *sdma = sdmac->sdma; sdma 986 drivers/dma/imx-sdma.c struct sdma_context_data *context = sdma->context; sdma 987 drivers/dma/imx-sdma.c struct sdma_buffer_descriptor *bd0 = sdma->bd0; sdma 1006 drivers/dma/imx-sdma.c dev_dbg(sdma->dev, "load_address = %d\n", load_address); sdma 1007 drivers/dma/imx-sdma.c dev_dbg(sdma->dev, "wml = 0x%08x\n", (u32)sdmac->watermark_level); sdma 1008 drivers/dma/imx-sdma.c dev_dbg(sdma->dev, "shp_addr = 0x%08x\n", sdmac->shp_addr); sdma 1009 drivers/dma/imx-sdma.c dev_dbg(sdma->dev, "per_addr = 0x%08x\n", sdmac->per_addr); sdma 1010 drivers/dma/imx-sdma.c dev_dbg(sdma->dev, "event_mask0 = 0x%08x\n", (u32)sdmac->event_mask[0]); sdma 1011 drivers/dma/imx-sdma.c dev_dbg(sdma->dev, "event_mask1 = 0x%08x\n", (u32)sdmac->event_mask[1]); sdma 1013 drivers/dma/imx-sdma.c spin_lock_irqsave(&sdma->channel_0_lock, flags); sdma 1030 drivers/dma/imx-sdma.c bd0->buffer_addr = sdma->context_phys; sdma 1032 drivers/dma/imx-sdma.c ret = sdma_run_channel0(sdma); sdma 1034 drivers/dma/imx-sdma.c spin_unlock_irqrestore(&sdma->channel_0_lock, flags); sdma 1049 drivers/dma/imx-sdma.c struct sdma_engine *sdma = sdmac->sdma; sdma 1052 drivers/dma/imx-sdma.c writel_relaxed(BIT(channel), sdma->regs + SDMA_H_STATSTOP); sdma 1103 drivers/dma/imx-sdma.c struct sdma_engine *sdma = sdmac->sdma; sdma 1130 drivers/dma/imx-sdma.c if (sdmac->per_address2 >= sdma->spba_start_addr && sdma 1131 drivers/dma/imx-sdma.c sdmac->per_address2 <= sdma->spba_end_addr) sdma 1134 drivers/dma/imx-sdma.c if (sdmac->per_address >= sdma->spba_start_addr && sdma 1135 drivers/dma/imx-sdma.c sdmac->per_address <= sdma->spba_end_addr) sdma 1192 drivers/dma/imx-sdma.c struct sdma_engine *sdma = sdmac->sdma; sdma 1200 drivers/dma/imx-sdma.c writel_relaxed(priority, sdma->regs + SDMA_CHNPRI_0 + 4 * channel); sdma 1205 drivers/dma/imx-sdma.c static int sdma_request_channel0(struct sdma_engine *sdma) sdma 1209 drivers/dma/imx-sdma.c sdma->bd0 = dma_alloc_coherent(sdma->dev, PAGE_SIZE, &sdma->bd0_phys, sdma 1211 drivers/dma/imx-sdma.c if (!sdma->bd0) { sdma 1216 drivers/dma/imx-sdma.c sdma->channel_control[0].base_bd_ptr = sdma->bd0_phys; sdma 1217 drivers/dma/imx-sdma.c sdma->channel_control[0].current_bd_ptr = sdma->bd0_phys; sdma 1219 drivers/dma/imx-sdma.c sdma_set_channel_priority(&sdma->channel[0], MXC_SDMA_DEFAULT_PRIORITY); sdma 1232 drivers/dma/imx-sdma.c desc->bd = dma_alloc_coherent(desc->sdmac->sdma->dev, bd_size, sdma 1246 drivers/dma/imx-sdma.c dma_free_coherent(desc->sdmac->sdma->dev, bd_size, desc->bd, sdma 1275 drivers/dma/imx-sdma.c dev_dbg(sdmac->sdma->dev, "MEMCPY in case?\n"); sdma 1302 drivers/dma/imx-sdma.c ret = clk_enable(sdmac->sdma->clk_ipg); sdma 1305 drivers/dma/imx-sdma.c ret = clk_enable(sdmac->sdma->clk_ahb); sdma 1316 drivers/dma/imx-sdma.c clk_disable(sdmac->sdma->clk_ahb); sdma 1318 drivers/dma/imx-sdma.c clk_disable(sdmac->sdma->clk_ipg); sdma 1325 drivers/dma/imx-sdma.c struct sdma_engine *sdma = sdmac->sdma; sdma 1342 drivers/dma/imx-sdma.c clk_disable(sdma->clk_ipg); sdma 1343 drivers/dma/imx-sdma.c clk_disable(sdma->clk_ahb); sdma 1389 drivers/dma/imx-sdma.c struct sdma_engine *sdma = sdmac->sdma; sdma 1399 drivers/dma/imx-sdma.c dev_dbg(sdma->dev, "memcpy: %pad->%pad, len=%zu, channel=%d.\n", sdma 1429 drivers/dma/imx-sdma.c dev_dbg(sdma->dev, "entry %d: count: %zd dma: 0x%x %s%s\n", sdma 1446 drivers/dma/imx-sdma.c struct sdma_engine *sdma = sdmac->sdma; sdma 1458 drivers/dma/imx-sdma.c dev_dbg(sdma->dev, "setting up %d entries for channel %d.\n", sdma 1470 drivers/dma/imx-sdma.c dev_err(sdma->dev, "SDMA channel %d: maximum bytes for sg entry exceeded: %d > %d\n", sdma 1507 drivers/dma/imx-sdma.c dev_dbg(sdma->dev, "entry %d: count: %d dma: %#llx %s%s\n", sdma 1530 drivers/dma/imx-sdma.c struct sdma_engine *sdma = sdmac->sdma; sdma 1536 drivers/dma/imx-sdma.c dev_dbg(sdma->dev, "%s channel: %d\n", __func__, channel); sdma 1549 drivers/dma/imx-sdma.c dev_err(sdma->dev, "SDMA channel %d: maximum period size exceeded: %zu > %d\n", sdma 1573 drivers/dma/imx-sdma.c dev_dbg(sdma->dev, "entry %d: count: %zu dma: %#llx %s%s\n", sdma 1633 drivers/dma/imx-sdma.c if (sdmac->event_id0 >= sdmac->sdma->drvdata->num_events) sdma 1639 drivers/dma/imx-sdma.c if (sdmac->event_id1 >= sdmac->sdma->drvdata->num_events) sdma 1700 drivers/dma/imx-sdma.c static void sdma_add_scripts(struct sdma_engine *sdma, sdma 1704 drivers/dma/imx-sdma.c s32 *saddr_arr = (u32 *)sdma->script_addrs; sdma 1708 drivers/dma/imx-sdma.c if (!sdma->script_number) sdma 1709 drivers/dma/imx-sdma.c sdma->script_number = SDMA_SCRIPT_ADDRS_ARRAY_SIZE_V1; sdma 1711 drivers/dma/imx-sdma.c if (sdma->script_number > sizeof(struct sdma_script_start_addrs) sdma 1713 drivers/dma/imx-sdma.c dev_err(sdma->dev, sdma 1715 drivers/dma/imx-sdma.c sdma->script_number); sdma 1719 drivers/dma/imx-sdma.c for (i = 0; i < sdma->script_number; i++) sdma 1726 drivers/dma/imx-sdma.c struct sdma_engine *sdma = context; sdma 1732 drivers/dma/imx-sdma.c dev_info(sdma->dev, "external firmware not found, using ROM firmware\n"); sdma 1748 drivers/dma/imx-sdma.c sdma->script_number = SDMA_SCRIPT_ADDRS_ARRAY_SIZE_V1; sdma 1751 drivers/dma/imx-sdma.c sdma->script_number = SDMA_SCRIPT_ADDRS_ARRAY_SIZE_V2; sdma 1754 drivers/dma/imx-sdma.c sdma->script_number = SDMA_SCRIPT_ADDRS_ARRAY_SIZE_V3; sdma 1757 drivers/dma/imx-sdma.c sdma->script_number = SDMA_SCRIPT_ADDRS_ARRAY_SIZE_V4; sdma 1760 drivers/dma/imx-sdma.c dev_err(sdma->dev, "unknown firmware version\n"); sdma 1767 drivers/dma/imx-sdma.c clk_enable(sdma->clk_ipg); sdma 1768 drivers/dma/imx-sdma.c clk_enable(sdma->clk_ahb); sdma 1770 drivers/dma/imx-sdma.c sdma_load_script(sdma, ram_code, sdma 1773 drivers/dma/imx-sdma.c clk_disable(sdma->clk_ipg); sdma 1774 drivers/dma/imx-sdma.c clk_disable(sdma->clk_ahb); sdma 1776 drivers/dma/imx-sdma.c sdma_add_scripts(sdma, addr); sdma 1778 drivers/dma/imx-sdma.c dev_info(sdma->dev, "loaded firmware %d.%d\n", sdma 1788 drivers/dma/imx-sdma.c static int sdma_event_remap(struct sdma_engine *sdma) sdma 1790 drivers/dma/imx-sdma.c struct device_node *np = sdma->dev->of_node; sdma 1804 drivers/dma/imx-sdma.c dev_dbg(sdma->dev, "no event needs to be remapped\n"); sdma 1807 drivers/dma/imx-sdma.c dev_err(sdma->dev, "the property %s must modulo %d\n", sdma 1815 drivers/dma/imx-sdma.c dev_err(sdma->dev, "failed to get gpr regmap\n"); sdma 1823 drivers/dma/imx-sdma.c dev_err(sdma->dev, "failed to read property %s index %d\n", sdma 1830 drivers/dma/imx-sdma.c dev_err(sdma->dev, "failed to read property %s index %d\n", sdma 1837 drivers/dma/imx-sdma.c dev_err(sdma->dev, "failed to read property %s index %d\n", sdma 1852 drivers/dma/imx-sdma.c static int sdma_get_firmware(struct sdma_engine *sdma, sdma 1858 drivers/dma/imx-sdma.c FW_ACTION_HOTPLUG, fw_name, sdma->dev, sdma 1859 drivers/dma/imx-sdma.c GFP_KERNEL, sdma, sdma_load_firmware); sdma 1864 drivers/dma/imx-sdma.c static int sdma_init(struct sdma_engine *sdma) sdma 1869 drivers/dma/imx-sdma.c ret = clk_enable(sdma->clk_ipg); sdma 1872 drivers/dma/imx-sdma.c ret = clk_enable(sdma->clk_ahb); sdma 1876 drivers/dma/imx-sdma.c if (sdma->drvdata->check_ratio && sdma 1877 drivers/dma/imx-sdma.c (clk_get_rate(sdma->clk_ahb) == clk_get_rate(sdma->clk_ipg))) sdma 1878 drivers/dma/imx-sdma.c sdma->clk_ratio = 1; sdma 1881 drivers/dma/imx-sdma.c writel_relaxed(0, sdma->regs + SDMA_H_C0PTR); sdma 1883 drivers/dma/imx-sdma.c sdma->channel_control = dma_alloc_coherent(sdma->dev, sdma 1888 drivers/dma/imx-sdma.c if (!sdma->channel_control) { sdma 1893 drivers/dma/imx-sdma.c sdma->context = (void *)sdma->channel_control + sdma 1895 drivers/dma/imx-sdma.c sdma->context_phys = ccb_phys + sdma 1899 drivers/dma/imx-sdma.c for (i = 0; i < sdma->drvdata->num_events; i++) sdma 1900 drivers/dma/imx-sdma.c writel_relaxed(0, sdma->regs + chnenbl_ofs(sdma, i)); sdma 1904 drivers/dma/imx-sdma.c writel_relaxed(0, sdma->regs + SDMA_CHNPRI_0 + i * 4); sdma 1906 drivers/dma/imx-sdma.c ret = sdma_request_channel0(sdma); sdma 1910 drivers/dma/imx-sdma.c sdma_config_ownership(&sdma->channel[0], false, true, false); sdma 1913 drivers/dma/imx-sdma.c writel_relaxed(0x4050, sdma->regs + SDMA_CHN0ADDR); sdma 1916 drivers/dma/imx-sdma.c if (sdma->clk_ratio) sdma 1917 drivers/dma/imx-sdma.c writel_relaxed(SDMA_H_CONFIG_ACR, sdma->regs + SDMA_H_CONFIG); sdma 1919 drivers/dma/imx-sdma.c writel_relaxed(0, sdma->regs + SDMA_H_CONFIG); sdma 1921 drivers/dma/imx-sdma.c writel_relaxed(ccb_phys, sdma->regs + SDMA_H_C0PTR); sdma 1924 drivers/dma/imx-sdma.c sdma_set_channel_priority(&sdma->channel[0], 7); sdma 1926 drivers/dma/imx-sdma.c clk_disable(sdma->clk_ipg); sdma 1927 drivers/dma/imx-sdma.c clk_disable(sdma->clk_ahb); sdma 1932 drivers/dma/imx-sdma.c clk_disable(sdma->clk_ahb); sdma 1934 drivers/dma/imx-sdma.c clk_disable(sdma->clk_ipg); sdma 1935 drivers/dma/imx-sdma.c dev_err(sdma->dev, "initialisation failed with %d\n", ret); sdma 1956 drivers/dma/imx-sdma.c struct sdma_engine *sdma = ofdma->of_dma_data; sdma 1957 drivers/dma/imx-sdma.c dma_cap_mask_t mask = sdma->dma_device.cap_mask; sdma 1992 drivers/dma/imx-sdma.c struct sdma_engine *sdma; sdma 2010 drivers/dma/imx-sdma.c sdma = devm_kzalloc(&pdev->dev, sizeof(*sdma), GFP_KERNEL); sdma 2011 drivers/dma/imx-sdma.c if (!sdma) sdma 2014 drivers/dma/imx-sdma.c spin_lock_init(&sdma->channel_0_lock); sdma 2016 drivers/dma/imx-sdma.c sdma->dev = &pdev->dev; sdma 2017 drivers/dma/imx-sdma.c sdma->drvdata = drvdata; sdma 2024 drivers/dma/imx-sdma.c sdma->regs = devm_ioremap_resource(&pdev->dev, iores); sdma 2025 drivers/dma/imx-sdma.c if (IS_ERR(sdma->regs)) sdma 2026 drivers/dma/imx-sdma.c return PTR_ERR(sdma->regs); sdma 2028 drivers/dma/imx-sdma.c sdma->clk_ipg = devm_clk_get(&pdev->dev, "ipg"); sdma 2029 drivers/dma/imx-sdma.c if (IS_ERR(sdma->clk_ipg)) sdma 2030 drivers/dma/imx-sdma.c return PTR_ERR(sdma->clk_ipg); sdma 2032 drivers/dma/imx-sdma.c sdma->clk_ahb = devm_clk_get(&pdev->dev, "ahb"); sdma 2033 drivers/dma/imx-sdma.c if (IS_ERR(sdma->clk_ahb)) sdma 2034 drivers/dma/imx-sdma.c return PTR_ERR(sdma->clk_ahb); sdma 2036 drivers/dma/imx-sdma.c ret = clk_prepare(sdma->clk_ipg); sdma 2040 drivers/dma/imx-sdma.c ret = clk_prepare(sdma->clk_ahb); sdma 2045 drivers/dma/imx-sdma.c sdma); sdma 2049 drivers/dma/imx-sdma.c sdma->irq = irq; sdma 2051 drivers/dma/imx-sdma.c sdma->script_addrs = kzalloc(sizeof(*sdma->script_addrs), GFP_KERNEL); sdma 2052 drivers/dma/imx-sdma.c if (!sdma->script_addrs) { sdma 2058 drivers/dma/imx-sdma.c saddr_arr = (s32 *)sdma->script_addrs; sdma 2062 drivers/dma/imx-sdma.c dma_cap_set(DMA_SLAVE, sdma->dma_device.cap_mask); sdma 2063 drivers/dma/imx-sdma.c dma_cap_set(DMA_CYCLIC, sdma->dma_device.cap_mask); sdma 2064 drivers/dma/imx-sdma.c dma_cap_set(DMA_MEMCPY, sdma->dma_device.cap_mask); sdma 2066 drivers/dma/imx-sdma.c INIT_LIST_HEAD(&sdma->dma_device.channels); sdma 2069 drivers/dma/imx-sdma.c struct sdma_channel *sdmac = &sdma->channel[i]; sdma 2071 drivers/dma/imx-sdma.c sdmac->sdma = sdma; sdma 2083 drivers/dma/imx-sdma.c vchan_init(&sdmac->vc, &sdma->dma_device); sdma 2086 drivers/dma/imx-sdma.c ret = sdma_init(sdma); sdma 2090 drivers/dma/imx-sdma.c ret = sdma_event_remap(sdma); sdma 2094 drivers/dma/imx-sdma.c if (sdma->drvdata->script_addrs) sdma 2095 drivers/dma/imx-sdma.c sdma_add_scripts(sdma, sdma->drvdata->script_addrs); sdma 2097 drivers/dma/imx-sdma.c sdma_add_scripts(sdma, pdata->script_addrs); sdma 2099 drivers/dma/imx-sdma.c sdma->dma_device.dev = &pdev->dev; sdma 2101 drivers/dma/imx-sdma.c sdma->dma_device.device_alloc_chan_resources = sdma_alloc_chan_resources; sdma 2102 drivers/dma/imx-sdma.c sdma->dma_device.device_free_chan_resources = sdma_free_chan_resources; sdma 2103 drivers/dma/imx-sdma.c sdma->dma_device.device_tx_status = sdma_tx_status; sdma 2104 drivers/dma/imx-sdma.c sdma->dma_device.device_prep_slave_sg = sdma_prep_slave_sg; sdma 2105 drivers/dma/imx-sdma.c sdma->dma_device.device_prep_dma_cyclic = sdma_prep_dma_cyclic; sdma 2106 drivers/dma/imx-sdma.c sdma->dma_device.device_config = sdma_config; sdma 2107 drivers/dma/imx-sdma.c sdma->dma_device.device_terminate_all = sdma_disable_channel_async; sdma 2108 drivers/dma/imx-sdma.c sdma->dma_device.device_synchronize = sdma_channel_synchronize; sdma 2109 drivers/dma/imx-sdma.c sdma->dma_device.src_addr_widths = SDMA_DMA_BUSWIDTHS; sdma 2110 drivers/dma/imx-sdma.c sdma->dma_device.dst_addr_widths = SDMA_DMA_BUSWIDTHS; sdma 2111 drivers/dma/imx-sdma.c sdma->dma_device.directions = SDMA_DMA_DIRECTIONS; sdma 2112 drivers/dma/imx-sdma.c sdma->dma_device.residue_granularity = DMA_RESIDUE_GRANULARITY_SEGMENT; sdma 2113 drivers/dma/imx-sdma.c sdma->dma_device.device_prep_dma_memcpy = sdma_prep_memcpy; sdma 2114 drivers/dma/imx-sdma.c sdma->dma_device.device_issue_pending = sdma_issue_pending; sdma 2115 drivers/dma/imx-sdma.c sdma->dma_device.dev->dma_parms = &sdma->dma_parms; sdma 2116 drivers/dma/imx-sdma.c sdma->dma_device.copy_align = 2; sdma 2117 drivers/dma/imx-sdma.c dma_set_max_seg_size(sdma->dma_device.dev, SDMA_BD_MAX_CNT); sdma 2119 drivers/dma/imx-sdma.c platform_set_drvdata(pdev, sdma); sdma 2121 drivers/dma/imx-sdma.c ret = dma_async_device_register(&sdma->dma_device); sdma 2128 drivers/dma/imx-sdma.c ret = of_dma_controller_register(np, sdma_xlate, sdma); sdma 2137 drivers/dma/imx-sdma.c sdma->spba_start_addr = spba_res.start; sdma 2138 drivers/dma/imx-sdma.c sdma->spba_end_addr = spba_res.end; sdma 2150 drivers/dma/imx-sdma.c ret = sdma_get_firmware(sdma, pdata->fw_name); sdma 2164 drivers/dma/imx-sdma.c ret = sdma_get_firmware(sdma, fw_name); sdma 2173 drivers/dma/imx-sdma.c dma_async_device_unregister(&sdma->dma_device); sdma 2175 drivers/dma/imx-sdma.c kfree(sdma->script_addrs); sdma 2177 drivers/dma/imx-sdma.c clk_unprepare(sdma->clk_ahb); sdma 2179 drivers/dma/imx-sdma.c clk_unprepare(sdma->clk_ipg); sdma 2185 drivers/dma/imx-sdma.c struct sdma_engine *sdma = platform_get_drvdata(pdev); sdma 2188 drivers/dma/imx-sdma.c devm_free_irq(&pdev->dev, sdma->irq, sdma); sdma 2189 drivers/dma/imx-sdma.c dma_async_device_unregister(&sdma->dma_device); sdma 2190 drivers/dma/imx-sdma.c kfree(sdma->script_addrs); sdma 2191 drivers/dma/imx-sdma.c clk_unprepare(sdma->clk_ahb); sdma 2192 drivers/dma/imx-sdma.c clk_unprepare(sdma->clk_ipg); sdma 2195 drivers/dma/imx-sdma.c struct sdma_channel *sdmac = &sdma->channel[i]; sdma 238 drivers/dma/sirf-dma.c struct sirfsoc_dma *sdma = dma_chan_to_sirfsoc_dma(&schan->chan); sdma 247 drivers/dma/sirf-dma.c base = sdma->base; sdma 253 drivers/dma/sirf-dma.c if (sdma->type == SIRFSOC_DMA_VER_A7V2) sdma 257 drivers/dma/sirf-dma.c sdma->exec_desc(sdesc, cid, schan->mode, base); sdma 266 drivers/dma/sirf-dma.c struct sirfsoc_dma *sdma = data; sdma 274 drivers/dma/sirf-dma.c switch (sdma->type) { sdma 277 drivers/dma/sirf-dma.c is = readl(sdma->base + SIRFSOC_DMA_CH_INT); sdma 278 drivers/dma/sirf-dma.c reg = sdma->base + SIRFSOC_DMA_CH_INT; sdma 282 drivers/dma/sirf-dma.c schan = &sdma->channels[ch]; sdma 300 drivers/dma/sirf-dma.c is = readl(sdma->base + SIRFSOC_DMA_INT_ATLAS7); sdma 302 drivers/dma/sirf-dma.c reg = sdma->base + SIRFSOC_DMA_INT_ATLAS7; sdma 304 drivers/dma/sirf-dma.c schan = &sdma->channels[0]; sdma 332 drivers/dma/sirf-dma.c tasklet_schedule(&sdma->tasklet); sdma 338 drivers/dma/sirf-dma.c static void sirfsoc_dma_process_completed(struct sirfsoc_dma *sdma) sdma 349 drivers/dma/sirf-dma.c for (i = 0; i < sdma->dma.chancnt; i++) { sdma 350 drivers/dma/sirf-dma.c schan = &sdma->channels[i]; sdma 398 drivers/dma/sirf-dma.c struct sirfsoc_dma *sdma = (void *)data; sdma 400 drivers/dma/sirf-dma.c sirfsoc_dma_process_completed(sdma); sdma 445 drivers/dma/sirf-dma.c struct sirfsoc_dma *sdma = dma_chan_to_sirfsoc_dma(&schan->chan); sdma 451 drivers/dma/sirf-dma.c switch (sdma->type) { sdma 453 drivers/dma/sirf-dma.c writel_relaxed(1 << cid, sdma->base + SIRFSOC_DMA_INT_EN_CLR); sdma 454 drivers/dma/sirf-dma.c writel_relaxed(1 << cid, sdma->base + SIRFSOC_DMA_CH_INT); sdma 456 drivers/dma/sirf-dma.c sdma->base + sdma 458 drivers/dma/sirf-dma.c writel_relaxed(1 << cid, sdma->base + SIRFSOC_DMA_CH_VALID); sdma 461 drivers/dma/sirf-dma.c writel_relaxed(0, sdma->base + SIRFSOC_DMA_INT_EN_ATLAS7); sdma 463 drivers/dma/sirf-dma.c sdma->base + SIRFSOC_DMA_INT_ATLAS7); sdma 464 drivers/dma/sirf-dma.c writel_relaxed(0, sdma->base + SIRFSOC_DMA_LOOP_CTRL_ATLAS7); sdma 465 drivers/dma/sirf-dma.c writel_relaxed(0, sdma->base + SIRFSOC_DMA_VALID_ATLAS7); sdma 468 drivers/dma/sirf-dma.c writel_relaxed(readl_relaxed(sdma->base + SIRFSOC_DMA_INT_EN) & sdma 469 drivers/dma/sirf-dma.c ~(1 << cid), sdma->base + SIRFSOC_DMA_INT_EN); sdma 470 drivers/dma/sirf-dma.c writel_relaxed(readl_relaxed(sdma->base + sdma 473 drivers/dma/sirf-dma.c sdma->base + SIRFSOC_DMA_CH_LOOP_CTRL); sdma 474 drivers/dma/sirf-dma.c writel_relaxed(1 << cid, sdma->base + SIRFSOC_DMA_CH_VALID); sdma 491 drivers/dma/sirf-dma.c struct sirfsoc_dma *sdma = dma_chan_to_sirfsoc_dma(&schan->chan); sdma 497 drivers/dma/sirf-dma.c switch (sdma->type) { sdma 500 drivers/dma/sirf-dma.c sdma->base + sdma 504 drivers/dma/sirf-dma.c writel_relaxed(0, sdma->base + SIRFSOC_DMA_LOOP_CTRL_ATLAS7); sdma 507 drivers/dma/sirf-dma.c writel_relaxed(readl_relaxed(sdma->base + sdma 510 drivers/dma/sirf-dma.c sdma->base + SIRFSOC_DMA_CH_LOOP_CTRL); sdma 525 drivers/dma/sirf-dma.c struct sirfsoc_dma *sdma = dma_chan_to_sirfsoc_dma(&schan->chan); sdma 530 drivers/dma/sirf-dma.c switch (sdma->type) { sdma 533 drivers/dma/sirf-dma.c sdma->base + SIRFSOC_DMA_CH_LOOP_CTRL_ATLAS7); sdma 537 drivers/dma/sirf-dma.c sdma->base + SIRFSOC_DMA_LOOP_CTRL_ATLAS7); sdma 540 drivers/dma/sirf-dma.c writel_relaxed(readl_relaxed(sdma->base + sdma 543 drivers/dma/sirf-dma.c sdma->base + SIRFSOC_DMA_CH_LOOP_CTRL); sdma 558 drivers/dma/sirf-dma.c struct sirfsoc_dma *sdma = dma_chan_to_sirfsoc_dma(chan); sdma 565 drivers/dma/sirf-dma.c pm_runtime_get_sync(sdma->dma.dev); sdma 571 drivers/dma/sirf-dma.c dev_notice(sdma->dma.dev, "Memory allocation error. " sdma 599 drivers/dma/sirf-dma.c struct sirfsoc_dma *sdma = dma_chan_to_sirfsoc_dma(chan); sdma 621 drivers/dma/sirf-dma.c pm_runtime_put(sdma->dma.dev); sdma 643 drivers/dma/sirf-dma.c struct sirfsoc_dma *sdma = dma_chan_to_sirfsoc_dma(chan); sdma 670 drivers/dma/sirf-dma.c if (sdma->type == SIRFSOC_DMA_VER_A7V2) sdma 673 drivers/dma/sirf-dma.c if (sdma->type == SIRFSOC_DMA_VER_A7V2) { sdma 674 drivers/dma/sirf-dma.c dma_pos = readl_relaxed(sdma->base + SIRFSOC_DMA_CUR_DATA_ADDR); sdma 677 drivers/dma/sirf-dma.c sdma->base + cid * 0x10 + SIRFSOC_DMA_CH_ADDR) << 2; sdma 692 drivers/dma/sirf-dma.c struct sirfsoc_dma *sdma = dma_chan_to_sirfsoc_dma(chan); sdma 714 drivers/dma/sirf-dma.c sirfsoc_dma_process_completed(sdma); sdma 830 drivers/dma/sirf-dma.c struct sirfsoc_dma *sdma = ofdma->of_dma_data; sdma 836 drivers/dma/sirf-dma.c return dma_get_slave_channel(&sdma->channels[request].chan); sdma 844 drivers/dma/sirf-dma.c struct sirfsoc_dma *sdma; sdma 852 drivers/dma/sirf-dma.c sdma = devm_kzalloc(dev, sizeof(*sdma), GFP_KERNEL); sdma 853 drivers/dma/sirf-dma.c if (!sdma) sdma 859 drivers/dma/sirf-dma.c sdma->exec_desc = data->exec; sdma 860 drivers/dma/sirf-dma.c sdma->type = data->type; sdma 867 drivers/dma/sirf-dma.c sdma->irq = irq_of_parse_and_map(dn, 0); sdma 868 drivers/dma/sirf-dma.c if (!sdma->irq) { sdma 873 drivers/dma/sirf-dma.c sdma->clk = devm_clk_get(dev, NULL); sdma 874 drivers/dma/sirf-dma.c if (IS_ERR(sdma->clk)) { sdma 876 drivers/dma/sirf-dma.c return PTR_ERR(sdma->clk); sdma 888 drivers/dma/sirf-dma.c sdma->base = devm_ioremap(dev, regs_start, regs_size); sdma 889 drivers/dma/sirf-dma.c if (!sdma->base) { sdma 895 drivers/dma/sirf-dma.c ret = request_irq(sdma->irq, &sirfsoc_dma_irq, 0, DRV_NAME, sdma); sdma 902 drivers/dma/sirf-dma.c dma = &sdma->dma; sdma 926 drivers/dma/sirf-dma.c schan = &sdma->channels[i]; sdma 941 drivers/dma/sirf-dma.c tasklet_init(&sdma->tasklet, sirfsoc_dma_tasklet, (unsigned long)sdma); sdma 944 drivers/dma/sirf-dma.c dev_set_drvdata(dev, sdma); sdma 951 drivers/dma/sirf-dma.c ret = of_dma_controller_register(dn, of_dma_sirfsoc_xlate, sdma); sdma 965 drivers/dma/sirf-dma.c free_irq(sdma->irq, sdma); sdma 967 drivers/dma/sirf-dma.c irq_dispose_mapping(sdma->irq); sdma 974 drivers/dma/sirf-dma.c struct sirfsoc_dma *sdma = dev_get_drvdata(dev); sdma 977 drivers/dma/sirf-dma.c dma_async_device_unregister(&sdma->dma); sdma 978 drivers/dma/sirf-dma.c free_irq(sdma->irq, sdma); sdma 979 drivers/dma/sirf-dma.c tasklet_kill(&sdma->tasklet); sdma 980 drivers/dma/sirf-dma.c irq_dispose_mapping(sdma->irq); sdma 990 drivers/dma/sirf-dma.c struct sirfsoc_dma *sdma = dev_get_drvdata(dev); sdma 992 drivers/dma/sirf-dma.c clk_disable_unprepare(sdma->clk); sdma 998 drivers/dma/sirf-dma.c struct sirfsoc_dma *sdma = dev_get_drvdata(dev); sdma 1001 drivers/dma/sirf-dma.c ret = clk_prepare_enable(sdma->clk); sdma 1011 drivers/dma/sirf-dma.c struct sirfsoc_dma *sdma = dev_get_drvdata(dev); sdma 1012 drivers/dma/sirf-dma.c struct sirfsoc_dma_regs *save = &sdma->regs_save; sdma 1029 drivers/dma/sirf-dma.c if (sdma->type == SIRFSOC_DMA_VER_A7V2) { sdma 1042 drivers/dma/sirf-dma.c schan = &sdma->channels[ch]; sdma 1045 drivers/dma/sirf-dma.c save->ctrl[ch] = readl_relaxed(sdma->base + sdma 1048 drivers/dma/sirf-dma.c save->interrupt_en = readl_relaxed(sdma->base + int_offset); sdma 1058 drivers/dma/sirf-dma.c struct sirfsoc_dma *sdma = dev_get_drvdata(dev); sdma 1059 drivers/dma/sirf-dma.c struct sirfsoc_dma_regs *save = &sdma->regs_save; sdma 1073 drivers/dma/sirf-dma.c if (sdma->type == SIRFSOC_DMA_VER_A7V2) { sdma 1083 drivers/dma/sirf-dma.c writel_relaxed(save->interrupt_en, sdma->base + int_offset); sdma 1085 drivers/dma/sirf-dma.c schan = &sdma->channels[ch]; sdma 1092 drivers/dma/sirf-dma.c sdma->base + width_offset + ch * 4); sdma 1094 drivers/dma/sirf-dma.c sdma->base + ch * 0x10 + SIRFSOC_DMA_CH_XLEN); sdma 1096 drivers/dma/sirf-dma.c sdma->base + ch * 0x10 + SIRFSOC_DMA_CH_YLEN); sdma 1098 drivers/dma/sirf-dma.c sdma->base + ch * 0x10 + SIRFSOC_DMA_CH_CTRL); sdma 1099 drivers/dma/sirf-dma.c if (sdma->type == SIRFSOC_DMA_VER_A7V2) { sdma 1101 drivers/dma/sirf-dma.c sdma->base + SIRFSOC_DMA_CH_ADDR); sdma 1104 drivers/dma/sirf-dma.c sdma->base + ch * 0x10 + SIRFSOC_DMA_CH_ADDR); sdma 929 drivers/gpu/drm/amd/amdgpu/amdgpu.h struct amdgpu_sdma sdma; sdma 410 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.c return adev->sdma.instance[0].fw_version; sdma 413 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.c return adev->sdma.instance[1].fw_version; sdma 616 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.c ring = &adev->sdma.instance[0].ring; sdma 619 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.c ring = &adev->sdma.instance[1].ring; sdma 167 drivers/gpu/drm/amd/amdgpu/amdgpu_cgs.c fw_version = adev->sdma.instance[0].fw_version; sdma 170 drivers/gpu/drm/amd/amdgpu/amdgpu_cgs.c fw_version = adev->sdma.instance[1].fw_version; sdma 140 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c for (j = 0; j < adev->sdma.num_instances; ++j) sdma 141 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c rings[j] = &adev->sdma.instance[j].ring; sdma 142 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c num_rings = adev->sdma.num_instances; sdma 279 drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c if (query_fw->index >= adev->sdma.num_instances) sdma 281 drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c fw_info->ver = adev->sdma.instance[query_fw->index].fw_version; sdma 282 drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c fw_info->feature = adev->sdma.instance[query_fw->index].feature_version; sdma 334 drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c for (i = 0; i < adev->sdma.num_instances; i++) sdma 335 drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c if (adev->sdma.instance[i].ring.sched.ready) sdma 1371 drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 965 drivers/gpu/drm/amd/amdgpu/amdgpu_psp.c adev->sdma.instance[ucode->ucode_id - AMDGPU_UCODE_ID_SDMA0].fw->data; sdma 252 drivers/gpu/drm/amd/amdgpu/amdgpu_ring.c else if (ring == &adev->sdma.instance[0].page) sdma 40 drivers/gpu/drm/amd/amdgpu/amdgpu_sdma.c for (i = 0; i < adev->sdma.num_instances; i++) sdma 41 drivers/gpu/drm/amd/amdgpu/amdgpu_sdma.c if (ring == &adev->sdma.instance[i].ring || sdma 42 drivers/gpu/drm/amd/amdgpu/amdgpu_sdma.c ring == &adev->sdma.instance[i].page) sdma 43 drivers/gpu/drm/amd/amdgpu/amdgpu_sdma.c return &adev->sdma.instance[i]; sdma 53 drivers/gpu/drm/amd/amdgpu/amdgpu_sdma.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 54 drivers/gpu/drm/amd/amdgpu/amdgpu_sdma.c if (ring == &adev->sdma.instance[i].ring || sdma 55 drivers/gpu/drm/amd/amdgpu/amdgpu_sdma.c ring == &adev->sdma.instance[i].page) { sdma 410 drivers/gpu/drm/amd/amdgpu/amdgpu_ucode.c FW_VERSION_ATTR(sdma_fw_version, 0444, sdma.instance[0].fw_version); sdma 411 drivers/gpu/drm/amd/amdgpu/amdgpu_ucode.c FW_VERSION_ATTR(sdma2_fw_version, 0444, sdma.instance[1].fw_version); sdma 261 drivers/gpu/drm/amd/amdgpu/amdgpu_ucode.h struct sdma_firmware_header_v1_0 sdma; sdma 76 drivers/gpu/drm/amd/amdgpu/cik_sdma.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 77 drivers/gpu/drm/amd/amdgpu/cik_sdma.c release_firmware(adev->sdma.instance[i].fw); sdma 78 drivers/gpu/drm/amd/amdgpu/cik_sdma.c adev->sdma.instance[i].fw = NULL; sdma 135 drivers/gpu/drm/amd/amdgpu/cik_sdma.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 140 drivers/gpu/drm/amd/amdgpu/cik_sdma.c err = request_firmware(&adev->sdma.instance[i].fw, fw_name, adev->dev); sdma 143 drivers/gpu/drm/amd/amdgpu/cik_sdma.c err = amdgpu_ucode_validate(adev->sdma.instance[i].fw); sdma 148 drivers/gpu/drm/amd/amdgpu/cik_sdma.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 149 drivers/gpu/drm/amd/amdgpu/cik_sdma.c release_firmware(adev->sdma.instance[i].fw); sdma 150 drivers/gpu/drm/amd/amdgpu/cik_sdma.c adev->sdma.instance[i].fw = NULL; sdma 203 drivers/gpu/drm/amd/amdgpu/cik_sdma.c struct amdgpu_sdma_instance *sdma = amdgpu_sdma_get_instance_from_ring(ring); sdma 207 drivers/gpu/drm/amd/amdgpu/cik_sdma.c if (sdma && sdma->burst_nop && (i == 0)) sdma 308 drivers/gpu/drm/amd/amdgpu/cik_sdma.c struct amdgpu_ring *sdma0 = &adev->sdma.instance[0].ring; sdma 309 drivers/gpu/drm/amd/amdgpu/cik_sdma.c struct amdgpu_ring *sdma1 = &adev->sdma.instance[1].ring; sdma 317 drivers/gpu/drm/amd/amdgpu/cik_sdma.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 376 drivers/gpu/drm/amd/amdgpu/cik_sdma.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 414 drivers/gpu/drm/amd/amdgpu/cik_sdma.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 440 drivers/gpu/drm/amd/amdgpu/cik_sdma.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 441 drivers/gpu/drm/amd/amdgpu/cik_sdma.c ring = &adev->sdma.instance[i].ring; sdma 506 drivers/gpu/drm/amd/amdgpu/cik_sdma.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 507 drivers/gpu/drm/amd/amdgpu/cik_sdma.c ring = &adev->sdma.instance[i].ring; sdma 551 drivers/gpu/drm/amd/amdgpu/cik_sdma.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 552 drivers/gpu/drm/amd/amdgpu/cik_sdma.c if (!adev->sdma.instance[i].fw) sdma 554 drivers/gpu/drm/amd/amdgpu/cik_sdma.c hdr = (const struct sdma_firmware_header_v1_0 *)adev->sdma.instance[i].fw->data; sdma 557 drivers/gpu/drm/amd/amdgpu/cik_sdma.c adev->sdma.instance[i].fw_version = le32_to_cpu(hdr->header.ucode_version); sdma 558 drivers/gpu/drm/amd/amdgpu/cik_sdma.c adev->sdma.instance[i].feature_version = le32_to_cpu(hdr->ucode_feature_version); sdma 559 drivers/gpu/drm/amd/amdgpu/cik_sdma.c if (adev->sdma.instance[i].feature_version >= 20) sdma 560 drivers/gpu/drm/amd/amdgpu/cik_sdma.c adev->sdma.instance[i].burst_nop = true; sdma 562 drivers/gpu/drm/amd/amdgpu/cik_sdma.c (adev->sdma.instance[i].fw->data + le32_to_cpu(hdr->header.ucode_array_offset_bytes)); sdma 566 drivers/gpu/drm/amd/amdgpu/cik_sdma.c WREG32(mmSDMA0_UCODE_ADDR + sdma_offsets[i], adev->sdma.instance[i].fw_version); sdma 810 drivers/gpu/drm/amd/amdgpu/cik_sdma.c struct amdgpu_sdma_instance *sdma = amdgpu_sdma_get_instance_from_ring(ring); sdma 816 drivers/gpu/drm/amd/amdgpu/cik_sdma.c if (sdma && sdma->burst_nop && (i == 0)) sdma 935 drivers/gpu/drm/amd/amdgpu/cik_sdma.c adev->sdma.num_instances = SDMA_MAX_INSTANCE; sdma 959 drivers/gpu/drm/amd/amdgpu/cik_sdma.c &adev->sdma.trap_irq); sdma 965 drivers/gpu/drm/amd/amdgpu/cik_sdma.c &adev->sdma.illegal_inst_irq); sdma 971 drivers/gpu/drm/amd/amdgpu/cik_sdma.c &adev->sdma.illegal_inst_irq); sdma 975 drivers/gpu/drm/amd/amdgpu/cik_sdma.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 976 drivers/gpu/drm/amd/amdgpu/cik_sdma.c ring = &adev->sdma.instance[i].ring; sdma 980 drivers/gpu/drm/amd/amdgpu/cik_sdma.c &adev->sdma.trap_irq, sdma 996 drivers/gpu/drm/amd/amdgpu/cik_sdma.c for (i = 0; i < adev->sdma.num_instances; i++) sdma 997 drivers/gpu/drm/amd/amdgpu/cik_sdma.c amdgpu_ring_fini(&adev->sdma.instance[i].ring); sdma 1170 drivers/gpu/drm/amd/amdgpu/cik_sdma.c amdgpu_fence_process(&adev->sdma.instance[0].ring); sdma 1183 drivers/gpu/drm/amd/amdgpu/cik_sdma.c amdgpu_fence_process(&adev->sdma.instance[1].ring); sdma 1206 drivers/gpu/drm/amd/amdgpu/cik_sdma.c drm_sched_fault(&adev->sdma.instance[instance_id].ring.sched); sdma 1279 drivers/gpu/drm/amd/amdgpu/cik_sdma.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 1280 drivers/gpu/drm/amd/amdgpu/cik_sdma.c adev->sdma.instance[i].ring.funcs = &cik_sdma_ring_funcs; sdma 1281 drivers/gpu/drm/amd/amdgpu/cik_sdma.c adev->sdma.instance[i].ring.me = i; sdma 1296 drivers/gpu/drm/amd/amdgpu/cik_sdma.c adev->sdma.trap_irq.num_types = AMDGPU_SDMA_IRQ_LAST; sdma 1297 drivers/gpu/drm/amd/amdgpu/cik_sdma.c adev->sdma.trap_irq.funcs = &cik_sdma_trap_irq_funcs; sdma 1298 drivers/gpu/drm/amd/amdgpu/cik_sdma.c adev->sdma.illegal_inst_irq.funcs = &cik_sdma_illegal_inst_irq_funcs; sdma 1362 drivers/gpu/drm/amd/amdgpu/cik_sdma.c adev->mman.buffer_funcs_ring = &adev->sdma.instance[0].ring; sdma 1379 drivers/gpu/drm/amd/amdgpu/cik_sdma.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 1380 drivers/gpu/drm/amd/amdgpu/cik_sdma.c sched = &adev->sdma.instance[i].ring.sched; sdma 1384 drivers/gpu/drm/amd/amdgpu/cik_sdma.c adev->vm_manager.vm_pte_num_rqs = adev->sdma.num_instances; sdma 2183 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 2185 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c adev->sdma.instance[i].fw->data; sdma 2186 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c fw_data = (const __le32 *) (adev->sdma.instance[i].fw->data + sdma 116 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 117 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c release_firmware(adev->sdma.instance[i].fw); sdma 118 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c adev->sdma.instance[i].fw = NULL; sdma 149 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 154 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c err = request_firmware(&adev->sdma.instance[i].fw, fw_name, adev->dev); sdma 157 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c err = amdgpu_ucode_validate(adev->sdma.instance[i].fw); sdma 160 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c hdr = (const struct sdma_firmware_header_v1_0 *)adev->sdma.instance[i].fw->data; sdma 161 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c adev->sdma.instance[i].fw_version = le32_to_cpu(hdr->header.ucode_version); sdma 162 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c adev->sdma.instance[i].feature_version = le32_to_cpu(hdr->ucode_feature_version); sdma 163 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c if (adev->sdma.instance[i].feature_version >= 20) sdma 164 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c adev->sdma.instance[i].burst_nop = true; sdma 169 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c info->fw = adev->sdma.instance[i].fw; sdma 179 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 180 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c release_firmware(adev->sdma.instance[i].fw); sdma 181 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c adev->sdma.instance[i].fw = NULL; sdma 231 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c struct amdgpu_sdma_instance *sdma = amdgpu_sdma_get_instance_from_ring(ring); sdma 235 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c if (sdma && sdma->burst_nop && (i == 0)) sdma 341 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c struct amdgpu_ring *sdma0 = &adev->sdma.instance[0].ring; sdma 342 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c struct amdgpu_ring *sdma1 = &adev->sdma.instance[1].ring; sdma 350 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 392 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 418 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 419 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c ring = &adev->sdma.instance[i].ring; sdma 484 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 485 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c ring = &adev->sdma.instance[i].ring; sdma 749 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c struct amdgpu_sdma_instance *sdma = amdgpu_sdma_get_instance_from_ring(ring); sdma 755 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c if (sdma && sdma->burst_nop && (i == 0)) sdma 828 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c adev->sdma.num_instances = SDMA_MAX_INSTANCE; sdma 846 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c &adev->sdma.trap_irq); sdma 852 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c &adev->sdma.illegal_inst_irq); sdma 858 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c &adev->sdma.illegal_inst_irq); sdma 868 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 869 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c ring = &adev->sdma.instance[i].ring; sdma 874 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c &adev->sdma.trap_irq, sdma 890 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c for (i = 0; i < adev->sdma.num_instances; i++) sdma 891 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c amdgpu_ring_fini(&adev->sdma.instance[i].ring); sdma 1063 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c amdgpu_fence_process(&adev->sdma.instance[0].ring); sdma 1076 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c amdgpu_fence_process(&adev->sdma.instance[1].ring); sdma 1101 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c drm_sched_fault(&adev->sdma.instance[instance_id].ring.sched); sdma 1166 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 1167 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c adev->sdma.instance[i].ring.funcs = &sdma_v2_4_ring_funcs; sdma 1168 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c adev->sdma.instance[i].ring.me = i; sdma 1183 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c adev->sdma.trap_irq.num_types = AMDGPU_SDMA_IRQ_LAST; sdma 1184 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c adev->sdma.trap_irq.funcs = &sdma_v2_4_trap_irq_funcs; sdma 1185 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c adev->sdma.illegal_inst_irq.funcs = &sdma_v2_4_illegal_inst_irq_funcs; sdma 1250 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c adev->mman.buffer_funcs_ring = &adev->sdma.instance[0].ring; sdma 1267 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 1268 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c sched = &adev->sdma.instance[i].ring.sched; sdma 1272 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c adev->vm_manager.vm_pte_num_rqs = adev->sdma.num_instances; sdma 253 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 254 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c release_firmware(adev->sdma.instance[i].fw); sdma 255 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c adev->sdma.instance[i].fw = NULL; sdma 307 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 312 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c err = request_firmware(&adev->sdma.instance[i].fw, fw_name, adev->dev); sdma 315 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c err = amdgpu_ucode_validate(adev->sdma.instance[i].fw); sdma 318 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c hdr = (const struct sdma_firmware_header_v1_0 *)adev->sdma.instance[i].fw->data; sdma 319 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c adev->sdma.instance[i].fw_version = le32_to_cpu(hdr->header.ucode_version); sdma 320 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c adev->sdma.instance[i].feature_version = le32_to_cpu(hdr->ucode_feature_version); sdma 321 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c if (adev->sdma.instance[i].feature_version >= 20) sdma 322 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c adev->sdma.instance[i].burst_nop = true; sdma 326 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c info->fw = adev->sdma.instance[i].fw; sdma 335 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 336 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c release_firmware(adev->sdma.instance[i].fw); sdma 337 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c adev->sdma.instance[i].fw = NULL; sdma 405 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c struct amdgpu_sdma_instance *sdma = amdgpu_sdma_get_instance_from_ring(ring); sdma 409 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c if (sdma && sdma->burst_nop && (i == 0)) sdma 515 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c struct amdgpu_ring *sdma0 = &adev->sdma.instance[0].ring; sdma 516 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c struct amdgpu_ring *sdma1 = &adev->sdma.instance[1].ring; sdma 524 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 585 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 627 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 655 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 656 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c ring = &adev->sdma.instance[i].ring; sdma 756 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 757 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c ring = &adev->sdma.instance[i].ring; sdma 1020 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c struct amdgpu_sdma_instance *sdma = amdgpu_sdma_get_instance_from_ring(ring); sdma 1026 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c if (sdma && sdma->burst_nop && (i == 0)) sdma 1101 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c adev->sdma.num_instances = 1; sdma 1104 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c adev->sdma.num_instances = SDMA_MAX_INSTANCE; sdma 1124 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c &adev->sdma.trap_irq); sdma 1130 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c &adev->sdma.illegal_inst_irq); sdma 1136 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c &adev->sdma.illegal_inst_irq); sdma 1146 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 1147 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c ring = &adev->sdma.instance[i].ring; sdma 1158 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c &adev->sdma.trap_irq, sdma 1174 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c for (i = 0; i < adev->sdma.num_instances; i++) sdma 1175 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c amdgpu_ring_fini(&adev->sdma.instance[i].ring); sdma 1261 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c adev->sdma.srbm_soft_reset = srbm_soft_reset; sdma 1264 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c adev->sdma.srbm_soft_reset = 0; sdma 1274 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c if (!adev->sdma.srbm_soft_reset) sdma 1277 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c srbm_soft_reset = adev->sdma.srbm_soft_reset; sdma 1293 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c if (!adev->sdma.srbm_soft_reset) sdma 1296 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c srbm_soft_reset = adev->sdma.srbm_soft_reset; sdma 1313 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c if (!adev->sdma.srbm_soft_reset) sdma 1316 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c srbm_soft_reset = adev->sdma.srbm_soft_reset; sdma 1397 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c amdgpu_fence_process(&adev->sdma.instance[0].ring); sdma 1410 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c amdgpu_fence_process(&adev->sdma.instance[1].ring); sdma 1435 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c drm_sched_fault(&adev->sdma.instance[instance_id].ring.sched); sdma 1447 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 1461 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 1486 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 1494 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 1604 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 1605 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c adev->sdma.instance[i].ring.funcs = &sdma_v3_0_ring_funcs; sdma 1606 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c adev->sdma.instance[i].ring.me = i; sdma 1621 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c adev->sdma.trap_irq.num_types = AMDGPU_SDMA_IRQ_LAST; sdma 1622 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c adev->sdma.trap_irq.funcs = &sdma_v3_0_trap_irq_funcs; sdma 1623 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c adev->sdma.illegal_inst_irq.funcs = &sdma_v3_0_illegal_inst_irq_funcs; sdma 1688 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c adev->mman.buffer_funcs_ring = &adev->sdma.instance[0].ring; sdma 1705 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 1706 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c sched = &adev->sdma.instance[i].ring.sched; sdma 1710 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c adev->vm_manager.vm_pte_num_rqs = adev->sdma.num_instances; sdma 417 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 418 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c if (adev->sdma.instance[i].fw != NULL) sdma 419 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c release_firmware(adev->sdma.instance[i].fw); sdma 427 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c memset((void*)adev->sdma.instance, 0, sdma 483 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c err = request_firmware(&adev->sdma.instance[0].fw, fw_name, adev->dev); sdma 487 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c err = sdma_v4_0_init_inst_ctx(&adev->sdma.instance[0]); sdma 491 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c for (i = 1; i < adev->sdma.num_instances; i++) { sdma 495 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c memcpy((void*)&adev->sdma.instance[i], sdma 496 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c (void*)&adev->sdma.instance[0], sdma 502 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c err = request_firmware(&adev->sdma.instance[i].fw, fw_name, adev->dev); sdma 506 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c err = sdma_v4_0_init_inst_ctx(&adev->sdma.instance[i]); sdma 516 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 519 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c info->fw = adev->sdma.instance[i].fw; sdma 674 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c struct amdgpu_sdma_instance *sdma = amdgpu_sdma_get_instance_from_ring(ring); sdma 678 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c if (sdma && sdma->burst_nop && (i == 0)) sdma 808 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c struct amdgpu_ring *sdma[AMDGPU_MAX_SDMA_INSTANCES]; sdma 812 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 813 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c sdma[i] = &adev->sdma.instance[i].ring; sdma 815 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c if ((adev->mman.buffer_funcs_ring == sdma[i]) && unset != 1) { sdma 827 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c sdma[i]->sched.ready = false; sdma 852 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c struct amdgpu_ring *sdma[AMDGPU_MAX_SDMA_INSTANCES]; sdma 857 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 858 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c sdma[i] = &adev->sdma.instance[i].page; sdma 860 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c if ((adev->mman.buffer_funcs_ring == sdma[i]) && sdma 875 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c sdma[i]->sched.ready = false; sdma 916 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 946 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c if (adev->sdma.has_page_queue) sdma 950 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 985 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c struct amdgpu_ring *ring = &adev->sdma.instance[i].ring; sdma 1075 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c struct amdgpu_ring *ring = &adev->sdma.instance[i].page; sdma 1252 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 1253 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c if (!adev->sdma.instance[i].fw) sdma 1256 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c hdr = (const struct sdma_firmware_header_v1_0 *)adev->sdma.instance[i].fw->data; sdma 1261 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c (adev->sdma.instance[i].fw->data + sdma 1271 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c adev->sdma.instance[i].fw_version); sdma 1308 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 1313 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c if (adev->sdma.has_page_queue) sdma 1338 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 1339 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c ring = &adev->sdma.instance[i].ring; sdma 1345 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c if (adev->sdma.has_page_queue) { sdma 1346 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c struct amdgpu_ring *page = &adev->sdma.instance[i].page; sdma 1578 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c struct amdgpu_sdma_instance *sdma = amdgpu_sdma_get_instance_from_ring(ring); sdma 1584 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c if (sdma && sdma->burst_nop && (i == 0)) sdma 1646 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c uint fw_version = adev->sdma.instance[0].fw_version; sdma 1667 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c adev->sdma.num_instances = 1; sdma 1669 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c adev->sdma.num_instances = 8; sdma 1671 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c adev->sdma.num_instances = 2; sdma 1681 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c adev->sdma.has_page_queue = false; sdma 1683 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c adev->sdma.has_page_queue = true; sdma 1700 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c struct ras_common_if **ras_if = &adev->sdma.ras_if; sdma 1771 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 1772 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c r = amdgpu_irq_get(adev, &adev->sdma.ecc_irq, sdma 1799 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 1802 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c &adev->sdma.trap_irq); sdma 1808 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 1811 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c &adev->sdma.ecc_irq); sdma 1816 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 1817 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c ring = &adev->sdma.instance[i].ring; sdma 1828 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c r = amdgpu_ring_init(adev, ring, 1024, &adev->sdma.trap_irq, sdma 1833 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c if (adev->sdma.has_page_queue) { sdma 1834 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c ring = &adev->sdma.instance[i].page; sdma 1846 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c &adev->sdma.trap_irq, sdma 1862 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c adev->sdma.ras_if) { sdma 1863 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c struct ras_common_if *ras_if = adev->sdma.ras_if; sdma 1877 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 1878 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c amdgpu_ring_fini(&adev->sdma.instance[i].ring); sdma 1879 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c if (adev->sdma.has_page_queue) sdma 1880 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c amdgpu_ring_fini(&adev->sdma.instance[i].page); sdma 1914 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 1915 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c amdgpu_irq_put(adev, &adev->sdma.ecc_irq, sdma 1949 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 1962 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c u32 sdma[AMDGPU_MAX_SDMA_INSTANCES]; sdma 1966 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c for (j = 0; j < adev->sdma.num_instances; j++) { sdma 1967 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c sdma[j] = RREG32_SDMA(j, mmSDMA0_STATUS_REG); sdma 1968 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c if (!(sdma[j] & SDMA0_STATUS_REG__IDLE_MASK)) sdma 1971 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c if (j == adev->sdma.num_instances) sdma 2010 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c amdgpu_fence_process(&adev->sdma.instance[instance].ring); sdma 2014 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c amdgpu_fence_process(&adev->sdma.instance[instance].page); sdma 2021 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c amdgpu_fence_process(&adev->sdma.instance[instance].page); sdma 2060 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c struct ras_common_if *ras_if = adev->sdma.ras_if; sdma 2088 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c drm_sched_fault(&adev->sdma.instance[instance].ring.sched); sdma 2117 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 2131 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 2156 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 2164 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 2390 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 2392 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c adev->sdma.instance[i].ring.funcs = sdma 2395 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c adev->sdma.instance[i].ring.funcs = sdma 2397 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c adev->sdma.instance[i].ring.me = i; sdma 2398 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c if (adev->sdma.has_page_queue) { sdma 2400 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c adev->sdma.instance[i].page.funcs = sdma 2403 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c adev->sdma.instance[i].page.funcs = sdma 2405 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c adev->sdma.instance[i].page.me = i; sdma 2428 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c switch (adev->sdma.num_instances) { sdma 2430 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c adev->sdma.trap_irq.num_types = AMDGPU_SDMA_IRQ_INSTANCE1; sdma 2431 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c adev->sdma.ecc_irq.num_types = AMDGPU_SDMA_IRQ_INSTANCE1; sdma 2434 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c adev->sdma.trap_irq.num_types = AMDGPU_SDMA_IRQ_LAST; sdma 2435 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c adev->sdma.ecc_irq.num_types = AMDGPU_SDMA_IRQ_LAST; sdma 2439 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c adev->sdma.trap_irq.num_types = AMDGPU_SDMA_IRQ_INSTANCE2; sdma 2440 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c adev->sdma.ecc_irq.num_types = AMDGPU_SDMA_IRQ_INSTANCE2; sdma 2443 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c adev->sdma.trap_irq.funcs = &sdma_v4_0_trap_irq_funcs; sdma 2444 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c adev->sdma.illegal_inst_irq.funcs = &sdma_v4_0_illegal_inst_irq_funcs; sdma 2445 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c adev->sdma.ecc_irq.funcs = &sdma_v4_0_ecc_irq_funcs; sdma 2510 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c if (adev->sdma.has_page_queue) sdma 2511 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c adev->mman.buffer_funcs_ring = &adev->sdma.instance[0].page; sdma 2513 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c adev->mman.buffer_funcs_ring = &adev->sdma.instance[0].ring; sdma 2530 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 2531 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c if (adev->sdma.has_page_queue) sdma 2532 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c sched = &adev->sdma.instance[i].page.sched; sdma 2534 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c sched = &adev->sdma.instance[i].ring.sched; sdma 2538 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c adev->vm_manager.vm_pte_num_rqs = adev->sdma.num_instances; sdma 193 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 198 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c err = request_firmware(&adev->sdma.instance[i].fw, fw_name, adev->dev); sdma 201 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c err = amdgpu_ucode_validate(adev->sdma.instance[i].fw); sdma 204 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c hdr = (const struct sdma_firmware_header_v1_0 *)adev->sdma.instance[i].fw->data; sdma 205 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c adev->sdma.instance[i].fw_version = le32_to_cpu(hdr->header.ucode_version); sdma 206 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c adev->sdma.instance[i].feature_version = le32_to_cpu(hdr->ucode_feature_version); sdma 207 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c if (adev->sdma.instance[i].feature_version >= 20) sdma 208 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c adev->sdma.instance[i].burst_nop = true; sdma 215 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c info->fw = adev->sdma.instance[i].fw; sdma 224 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 225 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c release_firmware(adev->sdma.instance[i].fw); sdma 226 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c adev->sdma.instance[i].fw = NULL; sdma 358 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c struct amdgpu_sdma_instance *sdma = amdgpu_sdma_get_instance_from_ring(ring); sdma 362 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c if (sdma && sdma->burst_nop && (i == 0)) sdma 500 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c struct amdgpu_ring *sdma0 = &adev->sdma.instance[0].ring; sdma 501 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c struct amdgpu_ring *sdma1 = &adev->sdma.instance[1].ring; sdma 509 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 571 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 606 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 634 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 635 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c ring = &adev->sdma.instance[i].ring; sdma 805 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 806 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c if (!adev->sdma.instance[i].fw) sdma 809 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c hdr = (const struct sdma_firmware_header_v1_0 *)adev->sdma.instance[i].fw->data; sdma 814 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c (adev->sdma.instance[i].fw->data + sdma 825 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c WREG32(sdma_v5_0_get_reg_offset(adev, i, mmSDMA0_UCODE_ADDR), adev->sdma.instance[i].fw_version); sdma 1115 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c struct amdgpu_sdma_instance *sdma = amdgpu_sdma_get_instance_from_ring(ring); sdma 1121 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c if (sdma && sdma->burst_nop && (i == 0)) sdma 1209 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c adev->sdma.num_instances = 2; sdma 1229 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c &adev->sdma.trap_irq); sdma 1236 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c &adev->sdma.trap_irq); sdma 1246 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 1247 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c ring = &adev->sdma.instance[i].ring; sdma 1260 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c &adev->sdma.trap_irq, sdma 1276 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c for (i = 0; i < adev->sdma.num_instances; i++) sdma 1277 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c amdgpu_ring_fini(&adev->sdma.instance[i].ring); sdma 1326 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 1435 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c amdgpu_fence_process(&adev->sdma.instance[0].ring); sdma 1451 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c amdgpu_fence_process(&adev->sdma.instance[1].ring); sdma 1481 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 1518 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 1644 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 1645 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c adev->sdma.instance[i].ring.funcs = &sdma_v5_0_ring_funcs; sdma 1646 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c adev->sdma.instance[i].ring.me = i; sdma 1661 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c adev->sdma.trap_irq.num_types = AMDGPU_SDMA_IRQ_INSTANCE0 + sdma 1662 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c adev->sdma.num_instances; sdma 1663 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c adev->sdma.trap_irq.funcs = &sdma_v5_0_trap_irq_funcs; sdma 1664 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c adev->sdma.illegal_inst_irq.funcs = &sdma_v5_0_illegal_inst_irq_funcs; sdma 1730 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c adev->mman.buffer_funcs_ring = &adev->sdma.instance[0].ring; sdma 1748 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 1749 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c sched = &adev->sdma.instance[i].ring.sched; sdma 1753 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c adev->vm_manager.vm_pte_num_rqs = adev->sdma.num_instances; sdma 49 drivers/gpu/drm/amd/amdgpu/si_dma.c u32 me = (ring == &adev->sdma.instance[0].ring) ? 0 : 1; sdma 57 drivers/gpu/drm/amd/amdgpu/si_dma.c u32 me = (ring == &adev->sdma.instance[0].ring) ? 0 : 1; sdma 118 drivers/gpu/drm/amd/amdgpu/si_dma.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 119 drivers/gpu/drm/amd/amdgpu/si_dma.c ring = &adev->sdma.instance[i].ring; sdma 138 drivers/gpu/drm/amd/amdgpu/si_dma.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 139 drivers/gpu/drm/amd/amdgpu/si_dma.c ring = &adev->sdma.instance[i].ring; sdma 470 drivers/gpu/drm/amd/amdgpu/si_dma.c adev->sdma.num_instances = 2; sdma 488 drivers/gpu/drm/amd/amdgpu/si_dma.c &adev->sdma.trap_irq); sdma 494 drivers/gpu/drm/amd/amdgpu/si_dma.c &adev->sdma.trap_irq); sdma 498 drivers/gpu/drm/amd/amdgpu/si_dma.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 499 drivers/gpu/drm/amd/amdgpu/si_dma.c ring = &adev->sdma.instance[i].ring; sdma 504 drivers/gpu/drm/amd/amdgpu/si_dma.c &adev->sdma.trap_irq, sdma 520 drivers/gpu/drm/amd/amdgpu/si_dma.c for (i = 0; i < adev->sdma.num_instances; i++) sdma 521 drivers/gpu/drm/amd/amdgpu/si_dma.c amdgpu_ring_fini(&adev->sdma.instance[i].ring); sdma 637 drivers/gpu/drm/amd/amdgpu/si_dma.c amdgpu_fence_process(&adev->sdma.instance[0].ring); sdma 639 drivers/gpu/drm/amd/amdgpu/si_dma.c amdgpu_fence_process(&adev->sdma.instance[1].ring); sdma 654 drivers/gpu/drm/amd/amdgpu/si_dma.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 666 drivers/gpu/drm/amd/amdgpu/si_dma.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 748 drivers/gpu/drm/amd/amdgpu/si_dma.c for (i = 0; i < adev->sdma.num_instances; i++) sdma 749 drivers/gpu/drm/amd/amdgpu/si_dma.c adev->sdma.instance[i].ring.funcs = &si_dma_ring_funcs; sdma 759 drivers/gpu/drm/amd/amdgpu/si_dma.c adev->sdma.trap_irq.num_types = AMDGPU_SDMA_IRQ_LAST; sdma 760 drivers/gpu/drm/amd/amdgpu/si_dma.c adev->sdma.trap_irq.funcs = &si_dma_trap_irq_funcs; sdma 824 drivers/gpu/drm/amd/amdgpu/si_dma.c adev->mman.buffer_funcs_ring = &adev->sdma.instance[0].ring; sdma 841 drivers/gpu/drm/amd/amdgpu/si_dma.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 842 drivers/gpu/drm/amd/amdgpu/si_dma.c sched = &adev->sdma.instance[i].ring.sched; sdma 846 drivers/gpu/drm/amd/amdgpu/si_dma.c adev->vm_manager.vm_pte_num_rqs = adev->sdma.num_instances; sdma 1263 drivers/gpu/drm/amd/amdgpu/soc15.c for (i = 0; i < adev->sdma.num_instances; i++) { sdma 1264 drivers/gpu/drm/amd/amdgpu/soc15.c ring = &adev->sdma.instance[i].ring; sdma 216 drivers/gpu/drm/radeon/radeon_ucode.h struct sdma_firmware_header_v1_0 sdma; sdma 137 drivers/infiniband/hw/hfi1/vnic.h struct hfi1_vnic_sdma sdma[HFI1_VNIC_MAX_TXQ]; sdma 72 drivers/infiniband/hw/hfi1/vnic_sdma.c struct hfi1_vnic_sdma *sdma; sdma 84 drivers/infiniband/hw/hfi1/vnic_sdma.c struct hfi1_vnic_sdma *vnic_sdma = tx->sdma; sdma 171 drivers/infiniband/hw/hfi1/vnic_sdma.c struct hfi1_vnic_sdma *vnic_sdma = &vinfo->sdma[q_idx]; sdma 188 drivers/infiniband/hw/hfi1/vnic_sdma.c tx->sdma = vnic_sdma; sdma 272 drivers/infiniband/hw/hfi1/vnic_sdma.c struct hfi1_vnic_sdma *vnic_sdma = &vinfo->sdma[q_idx]; sdma 282 drivers/infiniband/hw/hfi1/vnic_sdma.c struct hfi1_vnic_sdma *vnic_sdma = &vinfo->sdma[i]; sdma 170 drivers/media/pci/ddbridge/ddbridge-core.c struct ddb_dma *sdma, sdma 176 drivers/media/pci/ddbridge/ddbridge-core.c sdma->bufval = ddma->bufval; sdma 177 drivers/media/pci/ddbridge/ddbridge-core.c base = sdma->bufregs; sdma 809 drivers/scsi/isci/host.c &ihost->scu_registers->sdma.unsolicited_frame_queue_control); sdma 818 drivers/scsi/isci/host.c &ihost->scu_registers->sdma.unsolicited_frame_get_pointer); sdma 822 drivers/scsi/isci/host.c &ihost->scu_registers->sdma.unsolicited_frame_put_pointer); sdma 1557 drivers/scsi/isci/host.c writel(0, &ihost->scu_registers->sdma.unsolicited_frame_get_pointer); sdma 2205 drivers/scsi/isci/host.c val = readl(&ihost->scu_registers->sdma.pdma_configuration); sdma 2207 drivers/scsi/isci/host.c writel(val, &ihost->scu_registers->sdma.pdma_configuration); sdma 2209 drivers/scsi/isci/host.c val = readl(&ihost->scu_registers->sdma.cdma_configuration); sdma 2211 drivers/scsi/isci/host.c writel(val, &ihost->scu_registers->sdma.cdma_configuration); sdma 2321 drivers/scsi/isci/host.c &ihost->scu_registers->sdma.uf_header_base_address_lower); sdma 2323 drivers/scsi/isci/host.c &ihost->scu_registers->sdma.uf_header_base_address_upper); sdma 2326 drivers/scsi/isci/host.c &ihost->scu_registers->sdma.uf_address_table_lower); sdma 2328 drivers/scsi/isci/host.c &ihost->scu_registers->sdma.uf_address_table_upper); sdma 2548 drivers/scsi/isci/host.c &ihost->scu_registers->sdma.unsolicited_frame_get_pointer); sdma 1841 drivers/scsi/isci/registers.h struct scu_sdma_registers sdma; sdma 366 drivers/soc/fsl/qe/qe.c struct sdma __iomem *sdma = &qe_immr->sdma; sdma 369 drivers/soc/fsl/qe/qe.c if (!sdma) sdma 380 drivers/soc/fsl/qe/qe.c out_be32(&sdma->sdebcr, (u32) sdma_buf_offset & QE_SDEBCR_BA_MASK); sdma 381 drivers/soc/fsl/qe/qe.c out_be32(&sdma->sdmr, (QE_SDMR_GLB_1_MSK | sdma 447 include/soc/fsl/qe/immap_qe.h struct sdma sdma; /* SDMA */ sdma 206 sound/parisc/harmony.c harmony_write(h, HARMONY_PNXTADD, h->sdma.addr); sdma 312 sound/parisc/harmony.c harmony_write(h, HARMONY_PNXTADD, h->sdma.addr); sdma 340 sound/parisc/harmony.c harmony_write(h, HARMONY_PNXTADD, h->sdma.addr); sdma 386 sound/parisc/harmony.c snd_pcm_format_set_silence(fmt, h->sdma.area, SILENCE_BUFSZ / sdma 652 sound/parisc/harmony.c &h->sdma); sdma 871 sound/parisc/harmony.c if (h->sdma.addr) sdma 872 sound/parisc/harmony.c snd_dma_free_pages(&h->sdma); sdma 39 sound/parisc/harmony.h struct snd_dma_buffer sdma; /* silence */