lcla_pool         590 drivers/dma/ste_dma40.c 	struct d40_lcla_pool		  lcla_pool;
lcla_pool         708 drivers/dma/ste_dma40.c 	spin_lock_irqsave(&d40c->base->lcla_pool.lock, flags);
lcla_pool         717 drivers/dma/ste_dma40.c 		if (!d40c->base->lcla_pool.alloc_map[idx]) {
lcla_pool         718 drivers/dma/ste_dma40.c 			d40c->base->lcla_pool.alloc_map[idx] = d40d;
lcla_pool         725 drivers/dma/ste_dma40.c 	spin_unlock_irqrestore(&d40c->base->lcla_pool.lock, flags);
lcla_pool         740 drivers/dma/ste_dma40.c 	spin_lock_irqsave(&d40c->base->lcla_pool.lock, flags);
lcla_pool         745 drivers/dma/ste_dma40.c 		if (d40c->base->lcla_pool.alloc_map[idx] == d40d) {
lcla_pool         746 drivers/dma/ste_dma40.c 			d40c->base->lcla_pool.alloc_map[idx] = NULL;
lcla_pool         755 drivers/dma/ste_dma40.c 	spin_unlock_irqrestore(&d40c->base->lcla_pool.lock, flags);
lcla_pool         830 drivers/dma/ste_dma40.c 	struct d40_lcla_pool *pool = &chan->base->lcla_pool;
lcla_pool        3271 drivers/dma/ste_dma40.c 	base->lcla_pool.alloc_map = kcalloc(num_phy_chans
lcla_pool        3273 drivers/dma/ste_dma40.c 					    sizeof(*base->lcla_pool.alloc_map),
lcla_pool        3275 drivers/dma/ste_dma40.c 	if (!base->lcla_pool.alloc_map)
lcla_pool        3295 drivers/dma/ste_dma40.c 	kfree(base->lcla_pool.alloc_map);
lcla_pool        3375 drivers/dma/ste_dma40.c 	struct d40_lcla_pool *pool = &base->lcla_pool;
lcla_pool        3392 drivers/dma/ste_dma40.c 	base->lcla_pool.pages = SZ_1K * base->num_phy_chans / PAGE_SIZE;
lcla_pool        3396 drivers/dma/ste_dma40.c 						base->lcla_pool.pages);
lcla_pool        3400 drivers/dma/ste_dma40.c 				base->lcla_pool.pages);
lcla_pool        3404 drivers/dma/ste_dma40.c 				free_pages(page_list[j], base->lcla_pool.pages);
lcla_pool        3414 drivers/dma/ste_dma40.c 		free_pages(page_list[j], base->lcla_pool.pages);
lcla_pool        3417 drivers/dma/ste_dma40.c 		base->lcla_pool.base = (void *)page_list[i];
lcla_pool        3425 drivers/dma/ste_dma40.c 			 __func__, base->lcla_pool.pages);
lcla_pool        3426 drivers/dma/ste_dma40.c 		base->lcla_pool.base_unaligned = kmalloc(SZ_1K *
lcla_pool        3430 drivers/dma/ste_dma40.c 		if (!base->lcla_pool.base_unaligned) {
lcla_pool        3435 drivers/dma/ste_dma40.c 		base->lcla_pool.base = PTR_ALIGN(base->lcla_pool.base_unaligned,
lcla_pool        3448 drivers/dma/ste_dma40.c 	writel(virt_to_phys(base->lcla_pool.base),
lcla_pool        3582 drivers/dma/ste_dma40.c 		base->lcla_pool.base = ioremap(res->start,
lcla_pool        3584 drivers/dma/ste_dma40.c 		if (!base->lcla_pool.base) {
lcla_pool        3599 drivers/dma/ste_dma40.c 	spin_lock_init(&base->lcla_pool.lock);
lcla_pool        3665 drivers/dma/ste_dma40.c 	if (base->lcla_pool.base && base->plat_data->use_esram_lcla) {
lcla_pool        3666 drivers/dma/ste_dma40.c 		iounmap(base->lcla_pool.base);
lcla_pool        3667 drivers/dma/ste_dma40.c 		base->lcla_pool.base = NULL;
lcla_pool        3670 drivers/dma/ste_dma40.c 	if (base->lcla_pool.dma_addr)
lcla_pool        3671 drivers/dma/ste_dma40.c 		dma_unmap_single(base->dev, base->lcla_pool.dma_addr,
lcla_pool        3675 drivers/dma/ste_dma40.c 	if (!base->lcla_pool.base_unaligned && base->lcla_pool.base)
lcla_pool        3676 drivers/dma/ste_dma40.c 		free_pages((unsigned long)base->lcla_pool.base,
lcla_pool        3677 drivers/dma/ste_dma40.c 			   base->lcla_pool.pages);
lcla_pool        3679 drivers/dma/ste_dma40.c 	kfree(base->lcla_pool.base_unaligned);
lcla_pool        3697 drivers/dma/ste_dma40.c 	kfree(base->lcla_pool.alloc_map);