Searched refs:dma_requests (Results 1 – 4 of 4) sorted by relevance
44 u32 dma_requests; /* number of DMA requests on eDMA */ member85 if (dma_spec->args[0] >= xbar->dma_requests) { in ti_am335x_xbar_route_allocate()153 &xbar->dma_requests)) { in ti_am335x_xbar_probe()157 xbar->dma_requests = TI_AM335X_XBAR_LINES; in ti_am335x_xbar_probe()181 for (i = 0; i < xbar->dma_requests; i++) in ti_am335x_xbar_probe()205 u32 dma_requests; /* number of DMA requests forwarded to DMA */ member258 map->xbar_out = idr_alloc(&xbar->map_idr, NULL, 0, xbar->dma_requests, in ti_dra7_xbar_route_allocate()317 &xbar->dma_requests)) { in ti_dra7_xbar_probe()321 xbar->dma_requests = TI_DRA7_XBAR_OUTPUTS; in ti_dra7_xbar_probe()349 for (i = 0; i < xbar->dma_requests; i++) in ti_dra7_xbar_probe()[all …]
130 u32 dma_requests; member742 if (request >= d->dma_requests) in zx_of_dma_simple_xlate()778 "dma-requests", &d->dma_requests); in zx_dma_probe()779 if (!d->dma_requests || !d->dma_channels) in zx_dma_probe()837 d->dma_requests * sizeof(struct zx_dma_chan), GFP_KERNEL); in zx_dma_probe()841 for (i = 0; i < d->dma_requests; i++) { in zx_dma_probe()
104 u32 dma_requests; member663 if (request > d->dma_requests) in k3_of_dma_simple_xlate()693 "dma-requests", &d->dma_requests); in k3_dma_probe()738 d->dma_requests * sizeof(struct k3_dma_chan), GFP_KERNEL); in k3_dma_probe()742 for (i = 0; i < d->dma_requests; i++) { in k3_dma_probe()
37 unsigned dma_requests; member1174 od->dma_requests = OMAP_SDMA_REQUESTS; in omap_dma_probe()1177 &od->dma_requests)) { in omap_dma_probe()1279 if (req <= od->dma_requests) { in omap_dma_filter_fn()