Lines Matching refs:async_tx

41 	container_of(tx, struct iop_adma_desc_slot, async_tx)
64 struct dma_async_tx_descriptor *tx = &desc->async_tx; in iop_adma_run_tx_complete_actions()
95 if (!async_tx_test_ack(&desc->async_tx)) in iop_adma_clean_slot()
130 iter->async_tx.cookie, iter->idx, busy, in __iop_adma_slot_cleanup()
131 iter->async_tx.phys, iop_desc_get_next_desc(iter), in __iop_adma_slot_cleanup()
132 async_tx_test_ack(&iter->async_tx)); in __iop_adma_slot_cleanup()
134 prefetch(&_iter->async_tx); in __iop_adma_slot_cleanup()
147 if (iter->async_tx.phys == current_desc) { in __iop_adma_slot_cleanup()
226 if (iter->xor_check_result && iter->async_tx.cookie) in __iop_adma_slot_cleanup()
289 prefetch(&_iter->async_tx); in iop_adma_alloc_slots()
321 iter->async_tx.phys, slots_per_op); in iop_adma_alloc_slots()
325 async_tx_ack(&iter->async_tx); in iop_adma_alloc_slots()
329 iter->async_tx.cookie = 0; in iop_adma_alloc_slots()
342 alloc_tail->async_tx.cookie = -EBUSY; in iop_adma_alloc_slots()
394 next_dma = grp_start->async_tx.phys; in iop_adma_tx_submit()
411 __func__, sw_desc->async_tx.cookie, sw_desc->idx); in iop_adma_tx_submit()
455 dma_async_tx_descriptor_init(&slot->async_tx, chan); in iop_adma_alloc_chan_resources()
456 slot->async_tx.tx_submit = iop_adma_tx_submit; in iop_adma_alloc_chan_resources()
461 slot->async_tx.phys = in iop_adma_alloc_chan_resources()
510 sw_desc->async_tx.flags = flags; in iop_adma_prep_dma_interrupt()
514 return sw_desc ? &sw_desc->async_tx : NULL; in iop_adma_prep_dma_interrupt()
541 sw_desc->async_tx.flags = flags; in iop_adma_prep_dma_memcpy()
545 return sw_desc ? &sw_desc->async_tx : NULL; in iop_adma_prep_dma_memcpy()
573 sw_desc->async_tx.flags = flags; in iop_adma_prep_dma_xor()
580 return sw_desc ? &sw_desc->async_tx : NULL; in iop_adma_prep_dma_xor()
608 sw_desc->async_tx.flags = flags; in iop_adma_prep_dma_xor_val()
615 return sw_desc ? &sw_desc->async_tx : NULL; in iop_adma_prep_dma_xor_val()
660 sw_desc->async_tx.flags = flags; in iop_adma_prep_dma_pq()
679 return sw_desc ? &sw_desc->async_tx : NULL; in iop_adma_prep_dma_pq()
714 sw_desc->async_tx.flags = flags; in iop_adma_prep_dma_pq_val()
723 return sw_desc ? &sw_desc->async_tx : NULL; in iop_adma_prep_dma_pq_val()
1461 async_tx_ack(&sw_desc->async_tx); in iop_chan_start_null_memcpy()
1467 cookie = dma_cookie_assign(&sw_desc->async_tx); in iop_chan_start_null_memcpy()
1484 iop_chan_set_next_descriptor(iop_chan, sw_desc->async_tx.phys); in iop_chan_start_null_memcpy()
1513 async_tx_ack(&sw_desc->async_tx); in iop_chan_start_null_xor()
1520 cookie = dma_cookie_assign(&sw_desc->async_tx); in iop_chan_start_null_xor()
1537 iop_chan_set_next_descriptor(iop_chan, sw_desc->async_tx.phys); in iop_chan_start_null_xor()