Lines Matching refs:async_tx
398 set_desc_next(chan, &tail->hw, desc->async_tx.phys); in append_ld_queue()
430 cookie = dma_cookie_assign(&child->async_tx); in fsl_dma_tx_submit()
451 dma_pool_free(chan->desc_pool, desc, desc->async_tx.phys); in fsl_dma_free_descriptor()
473 dma_async_tx_descriptor_init(&desc->async_tx, &chan->common); in fsl_dma_alloc_descriptor()
474 desc->async_tx.tx_submit = fsl_dma_tx_submit; in fsl_dma_alloc_descriptor()
475 desc->async_tx.phys = pdesc; in fsl_dma_alloc_descriptor()
496 if (async_tx_test_ack(&desc->async_tx)) in fsldma_clean_completed_descriptor()
512 struct dma_async_tx_descriptor *txd = &desc->async_tx; in fsldma_run_tx_complete_actions()
552 if (!async_tx_test_ack(&desc->async_tx)) { in fsldma_clean_running_descriptor()
561 dma_pool_free(chan->desc_pool, desc, desc->async_tx.phys); in fsldma_clean_running_descriptor()
624 set_cdar(chan, desc->async_tx.phys); in fsl_chan_xfer_ld_queue()
663 if (desc->async_tx.phys == curr_phys) { in fsldma_cleanup_descriptors()
798 set_desc_next(chan, &prev->hw, new->async_tx.phys); in fsl_dma_prep_memcpy()
800 new->async_tx.cookie = 0; in fsl_dma_prep_memcpy()
801 async_tx_ack(&new->async_tx); in fsl_dma_prep_memcpy()
812 new->async_tx.flags = flags; /* client is in control of this ack */ in fsl_dma_prep_memcpy()
813 new->async_tx.cookie = -EBUSY; in fsl_dma_prep_memcpy()
818 return &first->async_tx; in fsl_dma_prep_memcpy()
881 set_desc_next(chan, &prev->hw, new->async_tx.phys); in fsl_dma_prep_sg()
883 new->async_tx.cookie = 0; in fsl_dma_prep_sg()
884 async_tx_ack(&new->async_tx); in fsl_dma_prep_sg()
928 new->async_tx.flags = flags; /* client is in control of this ack */ in fsl_dma_prep_sg()
929 new->async_tx.cookie = -EBUSY; in fsl_dma_prep_sg()
934 return &first->async_tx; in fsl_dma_prep_sg()