lcla_pool 590 drivers/dma/ste_dma40.c struct d40_lcla_pool lcla_pool; lcla_pool 708 drivers/dma/ste_dma40.c spin_lock_irqsave(&d40c->base->lcla_pool.lock, flags); lcla_pool 717 drivers/dma/ste_dma40.c if (!d40c->base->lcla_pool.alloc_map[idx]) { lcla_pool 718 drivers/dma/ste_dma40.c d40c->base->lcla_pool.alloc_map[idx] = d40d; lcla_pool 725 drivers/dma/ste_dma40.c spin_unlock_irqrestore(&d40c->base->lcla_pool.lock, flags); lcla_pool 740 drivers/dma/ste_dma40.c spin_lock_irqsave(&d40c->base->lcla_pool.lock, flags); lcla_pool 745 drivers/dma/ste_dma40.c if (d40c->base->lcla_pool.alloc_map[idx] == d40d) { lcla_pool 746 drivers/dma/ste_dma40.c d40c->base->lcla_pool.alloc_map[idx] = NULL; lcla_pool 755 drivers/dma/ste_dma40.c spin_unlock_irqrestore(&d40c->base->lcla_pool.lock, flags); lcla_pool 830 drivers/dma/ste_dma40.c struct d40_lcla_pool *pool = &chan->base->lcla_pool; lcla_pool 3271 drivers/dma/ste_dma40.c base->lcla_pool.alloc_map = kcalloc(num_phy_chans lcla_pool 3273 drivers/dma/ste_dma40.c sizeof(*base->lcla_pool.alloc_map), lcla_pool 3275 drivers/dma/ste_dma40.c if (!base->lcla_pool.alloc_map) lcla_pool 3295 drivers/dma/ste_dma40.c kfree(base->lcla_pool.alloc_map); lcla_pool 3375 drivers/dma/ste_dma40.c struct d40_lcla_pool *pool = &base->lcla_pool; lcla_pool 3392 drivers/dma/ste_dma40.c base->lcla_pool.pages = SZ_1K * base->num_phy_chans / PAGE_SIZE; lcla_pool 3396 drivers/dma/ste_dma40.c base->lcla_pool.pages); lcla_pool 3400 drivers/dma/ste_dma40.c base->lcla_pool.pages); lcla_pool 3404 drivers/dma/ste_dma40.c free_pages(page_list[j], base->lcla_pool.pages); lcla_pool 3414 drivers/dma/ste_dma40.c free_pages(page_list[j], base->lcla_pool.pages); lcla_pool 3417 drivers/dma/ste_dma40.c base->lcla_pool.base = (void *)page_list[i]; lcla_pool 3425 drivers/dma/ste_dma40.c __func__, base->lcla_pool.pages); lcla_pool 3426 drivers/dma/ste_dma40.c base->lcla_pool.base_unaligned = kmalloc(SZ_1K * lcla_pool 3430 drivers/dma/ste_dma40.c if (!base->lcla_pool.base_unaligned) { lcla_pool 3435 drivers/dma/ste_dma40.c base->lcla_pool.base = PTR_ALIGN(base->lcla_pool.base_unaligned, lcla_pool 3448 drivers/dma/ste_dma40.c writel(virt_to_phys(base->lcla_pool.base), lcla_pool 3582 drivers/dma/ste_dma40.c base->lcla_pool.base = ioremap(res->start, lcla_pool 3584 drivers/dma/ste_dma40.c if (!base->lcla_pool.base) { lcla_pool 3599 drivers/dma/ste_dma40.c spin_lock_init(&base->lcla_pool.lock); lcla_pool 3665 drivers/dma/ste_dma40.c if (base->lcla_pool.base && base->plat_data->use_esram_lcla) { lcla_pool 3666 drivers/dma/ste_dma40.c iounmap(base->lcla_pool.base); lcla_pool 3667 drivers/dma/ste_dma40.c base->lcla_pool.base = NULL; lcla_pool 3670 drivers/dma/ste_dma40.c if (base->lcla_pool.dma_addr) lcla_pool 3671 drivers/dma/ste_dma40.c dma_unmap_single(base->dev, base->lcla_pool.dma_addr, lcla_pool 3675 drivers/dma/ste_dma40.c if (!base->lcla_pool.base_unaligned && base->lcla_pool.base) lcla_pool 3676 drivers/dma/ste_dma40.c free_pages((unsigned long)base->lcla_pool.base, lcla_pool 3677 drivers/dma/ste_dma40.c base->lcla_pool.pages); lcla_pool 3679 drivers/dma/ste_dma40.c kfree(base->lcla_pool.base_unaligned); lcla_pool 3697 drivers/dma/ste_dma40.c kfree(base->lcla_pool.alloc_map);