Lines Matching refs:chan

95 	struct dma_chan chan;  member
139 container_of(dchan, struct mmp_pdma_chan, chan)
241 struct mmp_pdma_device *pdev = to_mmp_pdma_dev(pchan->chan.device); in lookup_phy()
274 struct mmp_pdma_device *pdev = to_mmp_pdma_dev(pchan->chan.device); in mmp_pdma_free_phy()
295 static void start_pending_queue(struct mmp_pdma_chan *chan) in start_pending_queue() argument
300 if (!chan->idle) { in start_pending_queue()
301 dev_dbg(chan->dev, "DMA controller still busy\n"); in start_pending_queue()
305 if (list_empty(&chan->chain_pending)) { in start_pending_queue()
307 mmp_pdma_free_phy(chan); in start_pending_queue()
308 dev_dbg(chan->dev, "no pending list\n"); in start_pending_queue()
312 if (!chan->phy) { in start_pending_queue()
313 chan->phy = lookup_phy(chan); in start_pending_queue()
314 if (!chan->phy) { in start_pending_queue()
315 dev_dbg(chan->dev, "no free dma channel\n"); in start_pending_queue()
324 desc = list_first_entry(&chan->chain_pending, in start_pending_queue()
326 list_splice_tail_init(&chan->chain_pending, &chan->chain_running); in start_pending_queue()
332 set_desc(chan->phy, desc->async_tx.phys); in start_pending_queue()
333 enable_chan(chan->phy); in start_pending_queue()
334 chan->idle = false; in start_pending_queue()
341 struct mmp_pdma_chan *chan = to_mmp_pdma_chan(tx->chan); in mmp_pdma_tx_submit() local
347 spin_lock_irqsave(&chan->desc_lock, flags); in mmp_pdma_tx_submit()
354 list_splice_tail_init(&desc->tx_list, &chan->chain_pending); in mmp_pdma_tx_submit()
356 spin_unlock_irqrestore(&chan->desc_lock, flags); in mmp_pdma_tx_submit()
362 mmp_pdma_alloc_descriptor(struct mmp_pdma_chan *chan) in mmp_pdma_alloc_descriptor() argument
367 desc = dma_pool_alloc(chan->desc_pool, GFP_ATOMIC, &pdesc); in mmp_pdma_alloc_descriptor()
369 dev_err(chan->dev, "out of memory for link descriptor\n"); in mmp_pdma_alloc_descriptor()
375 dma_async_tx_descriptor_init(&desc->async_tx, &chan->chan); in mmp_pdma_alloc_descriptor()
393 struct mmp_pdma_chan *chan = to_mmp_pdma_chan(dchan); in mmp_pdma_alloc_chan_resources() local
395 if (chan->desc_pool) in mmp_pdma_alloc_chan_resources()
398 chan->desc_pool = dma_pool_create(dev_name(&dchan->dev->device), in mmp_pdma_alloc_chan_resources()
399 chan->dev, in mmp_pdma_alloc_chan_resources()
403 if (!chan->desc_pool) { in mmp_pdma_alloc_chan_resources()
404 dev_err(chan->dev, "unable to allocate descriptor pool\n"); in mmp_pdma_alloc_chan_resources()
408 mmp_pdma_free_phy(chan); in mmp_pdma_alloc_chan_resources()
409 chan->idle = true; in mmp_pdma_alloc_chan_resources()
410 chan->dev_addr = 0; in mmp_pdma_alloc_chan_resources()
414 static void mmp_pdma_free_desc_list(struct mmp_pdma_chan *chan, in mmp_pdma_free_desc_list() argument
421 dma_pool_free(chan->desc_pool, desc, desc->async_tx.phys); in mmp_pdma_free_desc_list()
427 struct mmp_pdma_chan *chan = to_mmp_pdma_chan(dchan); in mmp_pdma_free_chan_resources() local
430 spin_lock_irqsave(&chan->desc_lock, flags); in mmp_pdma_free_chan_resources()
431 mmp_pdma_free_desc_list(chan, &chan->chain_pending); in mmp_pdma_free_chan_resources()
432 mmp_pdma_free_desc_list(chan, &chan->chain_running); in mmp_pdma_free_chan_resources()
433 spin_unlock_irqrestore(&chan->desc_lock, flags); in mmp_pdma_free_chan_resources()
435 dma_pool_destroy(chan->desc_pool); in mmp_pdma_free_chan_resources()
436 chan->desc_pool = NULL; in mmp_pdma_free_chan_resources()
437 chan->idle = true; in mmp_pdma_free_chan_resources()
438 chan->dev_addr = 0; in mmp_pdma_free_chan_resources()
439 mmp_pdma_free_phy(chan); in mmp_pdma_free_chan_resources()
448 struct mmp_pdma_chan *chan; in mmp_pdma_prep_memcpy() local
458 chan = to_mmp_pdma_chan(dchan); in mmp_pdma_prep_memcpy()
459 chan->byte_align = false; in mmp_pdma_prep_memcpy()
461 if (!chan->dir) { in mmp_pdma_prep_memcpy()
462 chan->dir = DMA_MEM_TO_MEM; in mmp_pdma_prep_memcpy()
463 chan->dcmd = DCMD_INCTRGADDR | DCMD_INCSRCADDR; in mmp_pdma_prep_memcpy()
464 chan->dcmd |= DCMD_BURST32; in mmp_pdma_prep_memcpy()
469 new = mmp_pdma_alloc_descriptor(chan); in mmp_pdma_prep_memcpy()
471 dev_err(chan->dev, "no memory for desc\n"); in mmp_pdma_prep_memcpy()
477 chan->byte_align = true; in mmp_pdma_prep_memcpy()
479 new->desc.dcmd = chan->dcmd | (DCMD_LENGTH & copy); in mmp_pdma_prep_memcpy()
494 if (chan->dir == DMA_MEM_TO_DEV) { in mmp_pdma_prep_memcpy()
496 } else if (chan->dir == DMA_DEV_TO_MEM) { in mmp_pdma_prep_memcpy()
498 } else if (chan->dir == DMA_MEM_TO_MEM) { in mmp_pdma_prep_memcpy()
514 chan->cyclic_first = NULL; in mmp_pdma_prep_memcpy()
520 mmp_pdma_free_desc_list(chan, &first->tx_list); in mmp_pdma_prep_memcpy()
529 struct mmp_pdma_chan *chan = to_mmp_pdma_chan(dchan); in mmp_pdma_prep_slave_sg() local
539 chan->byte_align = false; in mmp_pdma_prep_slave_sg()
548 chan->byte_align = true; in mmp_pdma_prep_slave_sg()
551 new = mmp_pdma_alloc_descriptor(chan); in mmp_pdma_prep_slave_sg()
553 dev_err(chan->dev, "no memory for desc\n"); in mmp_pdma_prep_slave_sg()
557 new->desc.dcmd = chan->dcmd | (DCMD_LENGTH & len); in mmp_pdma_prep_slave_sg()
560 new->desc.dtadr = chan->dev_addr; in mmp_pdma_prep_slave_sg()
562 new->desc.dsadr = chan->dev_addr; in mmp_pdma_prep_slave_sg()
591 chan->dir = dir; in mmp_pdma_prep_slave_sg()
592 chan->cyclic_first = NULL; in mmp_pdma_prep_slave_sg()
598 mmp_pdma_free_desc_list(chan, &first->tx_list); in mmp_pdma_prep_slave_sg()
608 struct mmp_pdma_chan *chan; in mmp_pdma_prep_dma_cyclic() local
622 chan = to_mmp_pdma_chan(dchan); in mmp_pdma_prep_dma_cyclic()
627 dma_dst = chan->dev_addr; in mmp_pdma_prep_dma_cyclic()
631 dma_src = chan->dev_addr; in mmp_pdma_prep_dma_cyclic()
634 dev_err(chan->dev, "Unsupported direction for cyclic DMA\n"); in mmp_pdma_prep_dma_cyclic()
638 chan->dir = direction; in mmp_pdma_prep_dma_cyclic()
642 new = mmp_pdma_alloc_descriptor(chan); in mmp_pdma_prep_dma_cyclic()
644 dev_err(chan->dev, "no memory for desc\n"); in mmp_pdma_prep_dma_cyclic()
648 new->desc.dcmd = (chan->dcmd | DCMD_ENDIRQEN | in mmp_pdma_prep_dma_cyclic()
664 if (chan->dir == DMA_MEM_TO_DEV) in mmp_pdma_prep_dma_cyclic()
678 chan->cyclic_first = first; in mmp_pdma_prep_dma_cyclic()
684 mmp_pdma_free_desc_list(chan, &first->tx_list); in mmp_pdma_prep_dma_cyclic()
691 struct mmp_pdma_chan *chan = to_mmp_pdma_chan(dchan); in mmp_pdma_config() local
699 chan->dcmd = DCMD_INCTRGADDR | DCMD_FLOWSRC; in mmp_pdma_config()
704 chan->dcmd = DCMD_INCSRCADDR | DCMD_FLOWTRG; in mmp_pdma_config()
711 chan->dcmd |= DCMD_WIDTH1; in mmp_pdma_config()
713 chan->dcmd |= DCMD_WIDTH2; in mmp_pdma_config()
715 chan->dcmd |= DCMD_WIDTH4; in mmp_pdma_config()
718 chan->dcmd |= DCMD_BURST8; in mmp_pdma_config()
720 chan->dcmd |= DCMD_BURST16; in mmp_pdma_config()
722 chan->dcmd |= DCMD_BURST32; in mmp_pdma_config()
724 chan->dir = cfg->direction; in mmp_pdma_config()
725 chan->dev_addr = addr; in mmp_pdma_config()
731 chan->drcmr = cfg->slave_id; in mmp_pdma_config()
738 struct mmp_pdma_chan *chan = to_mmp_pdma_chan(dchan); in mmp_pdma_terminate_all() local
744 disable_chan(chan->phy); in mmp_pdma_terminate_all()
745 mmp_pdma_free_phy(chan); in mmp_pdma_terminate_all()
746 spin_lock_irqsave(&chan->desc_lock, flags); in mmp_pdma_terminate_all()
747 mmp_pdma_free_desc_list(chan, &chan->chain_pending); in mmp_pdma_terminate_all()
748 mmp_pdma_free_desc_list(chan, &chan->chain_running); in mmp_pdma_terminate_all()
749 spin_unlock_irqrestore(&chan->desc_lock, flags); in mmp_pdma_terminate_all()
750 chan->idle = true; in mmp_pdma_terminate_all()
755 static unsigned int mmp_pdma_residue(struct mmp_pdma_chan *chan, in mmp_pdma_residue() argument
761 bool cyclic = chan->cyclic_first != NULL; in mmp_pdma_residue()
767 if (!chan->phy) in mmp_pdma_residue()
770 if (chan->dir == DMA_DEV_TO_MEM) in mmp_pdma_residue()
771 curr = readl(chan->phy->base + DTADR(chan->phy->idx)); in mmp_pdma_residue()
773 curr = readl(chan->phy->base + DSADR(chan->phy->idx)); in mmp_pdma_residue()
775 list_for_each_entry(sw, &chan->chain_running, node) { in mmp_pdma_residue()
778 if (chan->dir == DMA_DEV_TO_MEM) in mmp_pdma_residue()
833 struct mmp_pdma_chan *chan = to_mmp_pdma_chan(dchan); in mmp_pdma_tx_status() local
838 dma_set_residue(txstate, mmp_pdma_residue(chan, cookie)); in mmp_pdma_tx_status()
849 struct mmp_pdma_chan *chan = to_mmp_pdma_chan(dchan); in mmp_pdma_issue_pending() local
852 spin_lock_irqsave(&chan->desc_lock, flags); in mmp_pdma_issue_pending()
853 start_pending_queue(chan); in mmp_pdma_issue_pending()
854 spin_unlock_irqrestore(&chan->desc_lock, flags); in mmp_pdma_issue_pending()
864 struct mmp_pdma_chan *chan = (struct mmp_pdma_chan *)data; in dma_do_tasklet() local
869 if (chan->cyclic_first) { in dma_do_tasklet()
873 spin_lock_irqsave(&chan->desc_lock, flags); in dma_do_tasklet()
874 desc = chan->cyclic_first; in dma_do_tasklet()
877 spin_unlock_irqrestore(&chan->desc_lock, flags); in dma_do_tasklet()
886 spin_lock_irqsave(&chan->desc_lock, flags); in dma_do_tasklet()
888 list_for_each_entry_safe(desc, _desc, &chan->chain_running, node) { in dma_do_tasklet()
903 dev_dbg(chan->dev, "completed_cookie=%d\n", cookie); in dma_do_tasklet()
912 chan->idle = list_empty(&chan->chain_running); in dma_do_tasklet()
915 start_pending_queue(chan); in dma_do_tasklet()
916 spin_unlock_irqrestore(&chan->desc_lock, flags); in dma_do_tasklet()
928 dma_pool_free(chan->desc_pool, desc, txd->phys); in dma_do_tasklet()
943 struct mmp_pdma_chan *chan; in mmp_pdma_chan_init() local
946 chan = devm_kzalloc(pdev->dev, sizeof(*chan), GFP_KERNEL); in mmp_pdma_chan_init()
947 if (chan == NULL) in mmp_pdma_chan_init()
962 spin_lock_init(&chan->desc_lock); in mmp_pdma_chan_init()
963 chan->dev = pdev->dev; in mmp_pdma_chan_init()
964 chan->chan.device = &pdev->device; in mmp_pdma_chan_init()
965 tasklet_init(&chan->tasklet, dma_do_tasklet, (unsigned long)chan); in mmp_pdma_chan_init()
966 INIT_LIST_HEAD(&chan->chain_pending); in mmp_pdma_chan_init()
967 INIT_LIST_HEAD(&chan->chain_running); in mmp_pdma_chan_init()
970 list_add_tail(&chan->chan.device_node, &pdev->device.channels); in mmp_pdma_chan_init()
985 struct dma_chan *chan; in mmp_pdma_dma_xlate() local
987 chan = dma_get_any_slave_channel(&d->device); in mmp_pdma_dma_xlate()
988 if (!chan) in mmp_pdma_dma_xlate()
991 to_mmp_pdma_chan(chan)->drcmr = dma_spec->args[0]; in mmp_pdma_dma_xlate()
993 return chan; in mmp_pdma_dma_xlate()
1120 bool mmp_pdma_filter_fn(struct dma_chan *chan, void *param) in mmp_pdma_filter_fn() argument
1122 struct mmp_pdma_chan *c = to_mmp_pdma_chan(chan); in mmp_pdma_filter_fn()
1124 if (chan->device->dev->driver != &mmp_pdma_driver.driver) in mmp_pdma_filter_fn()