async_tx          159 drivers/dma/altera-msgdma.c 	struct dma_async_tx_descriptor async_tx;
async_tx          200 drivers/dma/altera-msgdma.c #define tx_to_desc(tx)	container_of(tx, struct msgdma_sw_desc, async_tx)
async_tx          368 drivers/dma/altera-msgdma.c 	async_tx_ack(&first->async_tx);
async_tx          369 drivers/dma/altera-msgdma.c 	first->async_tx.flags = flags;
async_tx          371 drivers/dma/altera-msgdma.c 	return &first->async_tx;
async_tx          453 drivers/dma/altera-msgdma.c 	first->async_tx.flags = flags;
async_tx          455 drivers/dma/altera-msgdma.c 	return &first->async_tx;
async_tx          590 drivers/dma/altera-msgdma.c 		callback = desc->async_tx.callback;
async_tx          591 drivers/dma/altera-msgdma.c 		callback_param = desc->async_tx.callback_param;
async_tx          616 drivers/dma/altera-msgdma.c 	dma_cookie_complete(&desc->async_tx);
async_tx          669 drivers/dma/altera-msgdma.c 		dma_async_tx_descriptor_init(&desc->async_tx, &mdev->dmachan);
async_tx          670 drivers/dma/altera-msgdma.c 		desc->async_tx.tx_submit = msgdma_tx_submit;
async_tx           83 drivers/dma/fsl_raid.c #define to_fsl_re_dma_desc(tx) container_of(tx, struct fsl_re_desc, async_tx)
async_tx          137 drivers/dma/fsl_raid.c 	dma_cookie_complete(&desc->async_tx);
async_tx          138 drivers/dma/fsl_raid.c 	dma_descriptor_unmap(&desc->async_tx);
async_tx          139 drivers/dma/fsl_raid.c 	dmaengine_desc_get_callback_invoke(&desc->async_tx, NULL);
async_tx          149 drivers/dma/fsl_raid.c 		if (async_tx_test_ack(&desc->async_tx))
async_tx          256 drivers/dma/fsl_raid.c 	desc->async_tx.tx_submit = fsl_re_tx_submit;
async_tx          257 drivers/dma/fsl_raid.c 	dma_async_tx_descriptor_init(&desc->async_tx, &re_chan->chan);
async_tx          289 drivers/dma/fsl_raid.c 		desc->async_tx.flags = flags;
async_tx          306 drivers/dma/fsl_raid.c 		desc->async_tx.flags = flags;
async_tx          384 drivers/dma/fsl_raid.c 	return &desc->async_tx;
async_tx          518 drivers/dma/fsl_raid.c 	return &desc->async_tx;
async_tx          570 drivers/dma/fsl_raid.c 	return &desc->async_tx;
async_tx          294 drivers/dma/fsl_raid.h 	struct dma_async_tx_descriptor async_tx;
async_tx          396 drivers/dma/fsldma.c 	set_desc_next(chan, &tail->hw, desc->async_tx.phys);
async_tx          428 drivers/dma/fsldma.c 		cookie = dma_cookie_assign(&child->async_tx);
async_tx          449 drivers/dma/fsldma.c 	dma_pool_free(chan->desc_pool, desc, desc->async_tx.phys);
async_tx          470 drivers/dma/fsldma.c 	dma_async_tx_descriptor_init(&desc->async_tx, &chan->common);
async_tx          471 drivers/dma/fsldma.c 	desc->async_tx.tx_submit = fsl_dma_tx_submit;
async_tx          472 drivers/dma/fsldma.c 	desc->async_tx.phys = pdesc;
async_tx          493 drivers/dma/fsldma.c 		if (async_tx_test_ack(&desc->async_tx))
async_tx          509 drivers/dma/fsldma.c 	struct dma_async_tx_descriptor *txd = &desc->async_tx;
async_tx          547 drivers/dma/fsldma.c 	if (!async_tx_test_ack(&desc->async_tx)) {
async_tx          556 drivers/dma/fsldma.c 	dma_pool_free(chan->desc_pool, desc, desc->async_tx.phys);
async_tx          619 drivers/dma/fsldma.c 	set_cdar(chan, desc->async_tx.phys);
async_tx          658 drivers/dma/fsldma.c 		if (desc->async_tx.phys == curr_phys) {
async_tx          793 drivers/dma/fsldma.c 			set_desc_next(chan, &prev->hw, new->async_tx.phys);
async_tx          795 drivers/dma/fsldma.c 		new->async_tx.cookie = 0;
async_tx          796 drivers/dma/fsldma.c 		async_tx_ack(&new->async_tx);
async_tx          807 drivers/dma/fsldma.c 	new->async_tx.flags = flags; /* client is in control of this ack */
async_tx          808 drivers/dma/fsldma.c 	new->async_tx.cookie = -EBUSY;
async_tx          813 drivers/dma/fsldma.c 	return &first->async_tx;
async_tx          104 drivers/dma/fsldma.h 	struct dma_async_tx_descriptor async_tx;
async_tx          192 drivers/dma/fsldma.h #define tx_to_fsl_desc(tx) container_of(tx, struct fsl_desc_sw, async_tx)
async_tx           32 drivers/dma/iop-adma.c 	container_of(tx, struct iop_adma_desc_slot, async_tx)
async_tx           55 drivers/dma/iop-adma.c 	struct dma_async_tx_descriptor *tx = &desc->async_tx;
async_tx           85 drivers/dma/iop-adma.c 	if (!async_tx_test_ack(&desc->async_tx))
async_tx          120 drivers/dma/iop-adma.c 			iter->async_tx.cookie, iter->idx, busy,
async_tx          121 drivers/dma/iop-adma.c 			&iter->async_tx.phys, (u64)iop_desc_get_next_desc(iter),
async_tx          122 drivers/dma/iop-adma.c 			async_tx_test_ack(&iter->async_tx));
async_tx          124 drivers/dma/iop-adma.c 		prefetch(&_iter->async_tx);
async_tx          137 drivers/dma/iop-adma.c 		if (iter->async_tx.phys == current_desc) {
async_tx          216 drivers/dma/iop-adma.c 		if (iter->xor_check_result && iter->async_tx.cookie)
async_tx          279 drivers/dma/iop-adma.c 		prefetch(&_iter->async_tx);
async_tx          311 drivers/dma/iop-adma.c 					(u64)iter->async_tx.phys, slots_per_op);
async_tx          315 drivers/dma/iop-adma.c 					async_tx_ack(&iter->async_tx);
async_tx          319 drivers/dma/iop-adma.c 				iter->async_tx.cookie = 0;
async_tx          332 drivers/dma/iop-adma.c 			alloc_tail->async_tx.cookie = -EBUSY;
async_tx          382 drivers/dma/iop-adma.c 	next_dma = grp_start->async_tx.phys;
async_tx          399 drivers/dma/iop-adma.c 		__func__, sw_desc->async_tx.cookie, sw_desc->idx);
async_tx          443 drivers/dma/iop-adma.c 		dma_async_tx_descriptor_init(&slot->async_tx, chan);
async_tx          444 drivers/dma/iop-adma.c 		slot->async_tx.tx_submit = iop_adma_tx_submit;
async_tx          449 drivers/dma/iop-adma.c 		slot->async_tx.phys =
async_tx          498 drivers/dma/iop-adma.c 		sw_desc->async_tx.flags = flags;
async_tx          502 drivers/dma/iop-adma.c 	return sw_desc ? &sw_desc->async_tx : NULL;
async_tx          529 drivers/dma/iop-adma.c 		sw_desc->async_tx.flags = flags;
async_tx          533 drivers/dma/iop-adma.c 	return sw_desc ? &sw_desc->async_tx : NULL;
async_tx          561 drivers/dma/iop-adma.c 		sw_desc->async_tx.flags = flags;
async_tx          568 drivers/dma/iop-adma.c 	return sw_desc ? &sw_desc->async_tx : NULL;
async_tx          596 drivers/dma/iop-adma.c 		sw_desc->async_tx.flags = flags;
async_tx          603 drivers/dma/iop-adma.c 	return sw_desc ? &sw_desc->async_tx : NULL;
async_tx          648 drivers/dma/iop-adma.c 		sw_desc->async_tx.flags = flags;
async_tx          667 drivers/dma/iop-adma.c 	return sw_desc ? &sw_desc->async_tx : NULL;
async_tx          702 drivers/dma/iop-adma.c 		sw_desc->async_tx.flags = flags;
async_tx          711 drivers/dma/iop-adma.c 	return sw_desc ? &sw_desc->async_tx : NULL;
async_tx         1450 drivers/dma/iop-adma.c 		async_tx_ack(&sw_desc->async_tx);
async_tx         1456 drivers/dma/iop-adma.c 		cookie = dma_cookie_assign(&sw_desc->async_tx);
async_tx         1473 drivers/dma/iop-adma.c 		iop_chan_set_next_descriptor(iop_chan, sw_desc->async_tx.phys);
async_tx         1502 drivers/dma/iop-adma.c 		async_tx_ack(&sw_desc->async_tx);
async_tx         1509 drivers/dma/iop-adma.c 		cookie = dma_cookie_assign(&sw_desc->async_tx);
async_tx         1526 drivers/dma/iop-adma.c 		iop_chan_set_next_descriptor(iop_chan, sw_desc->async_tx.phys);
async_tx          583 drivers/dma/iop-adma.h 				(u32) (desc->async_tx.phys + (i << 5));
async_tx           85 drivers/dma/mmp_pdma.c 	struct dma_async_tx_descriptor async_tx;
async_tx          133 drivers/dma/mmp_pdma.c 	container_of(tx, struct mmp_pdma_desc_sw, async_tx)
async_tx          334 drivers/dma/mmp_pdma.c 	set_desc(chan->phy, desc->async_tx.phys);
async_tx          352 drivers/dma/mmp_pdma.c 		cookie = dma_cookie_assign(&child->async_tx);
async_tx          376 drivers/dma/mmp_pdma.c 	dma_async_tx_descriptor_init(&desc->async_tx, &chan->chan);
async_tx          378 drivers/dma/mmp_pdma.c 	desc->async_tx.tx_submit = mmp_pdma_tx_submit;
async_tx          379 drivers/dma/mmp_pdma.c 	desc->async_tx.phys = pdesc;
async_tx          422 drivers/dma/mmp_pdma.c 		dma_pool_free(chan->desc_pool, desc, desc->async_tx.phys);
async_tx          487 drivers/dma/mmp_pdma.c 			prev->desc.ddadr = new->async_tx.phys;
async_tx          489 drivers/dma/mmp_pdma.c 		new->async_tx.cookie = 0;
async_tx          490 drivers/dma/mmp_pdma.c 		async_tx_ack(&new->async_tx);
async_tx          508 drivers/dma/mmp_pdma.c 	first->async_tx.flags = flags; /* client is in control of this ack */
async_tx          509 drivers/dma/mmp_pdma.c 	first->async_tx.cookie = -EBUSY;
async_tx          517 drivers/dma/mmp_pdma.c 	return &first->async_tx;
async_tx          572 drivers/dma/mmp_pdma.c 				prev->desc.ddadr = new->async_tx.phys;
async_tx          574 drivers/dma/mmp_pdma.c 			new->async_tx.cookie = 0;
async_tx          575 drivers/dma/mmp_pdma.c 			async_tx_ack(&new->async_tx);
async_tx          587 drivers/dma/mmp_pdma.c 	first->async_tx.cookie = -EBUSY;
async_tx          588 drivers/dma/mmp_pdma.c 	first->async_tx.flags = flags;
async_tx          597 drivers/dma/mmp_pdma.c 	return &first->async_tx;
async_tx          660 drivers/dma/mmp_pdma.c 			prev->desc.ddadr = new->async_tx.phys;
async_tx          662 drivers/dma/mmp_pdma.c 		new->async_tx.cookie = 0;
async_tx          663 drivers/dma/mmp_pdma.c 		async_tx_ack(&new->async_tx);
async_tx          677 drivers/dma/mmp_pdma.c 	first->async_tx.flags = flags; /* client is in control of this ack */
async_tx          678 drivers/dma/mmp_pdma.c 	first->async_tx.cookie = -EBUSY;
async_tx          681 drivers/dma/mmp_pdma.c 	new->desc.ddadr = first->async_tx.phys;
async_tx          684 drivers/dma/mmp_pdma.c 	return &first->async_tx;
async_tx          831 drivers/dma/mmp_pdma.c 		if (sw->async_tx.cookie == cookie) {
async_tx          887 drivers/dma/mmp_pdma.c 		dmaengine_desc_get_callback(&desc->async_tx, &cb);
async_tx          911 drivers/dma/mmp_pdma.c 			dma_cookie_t cookie = desc->async_tx.cookie;
async_tx          912 drivers/dma/mmp_pdma.c 			dma_cookie_complete(&desc->async_tx);
async_tx          930 drivers/dma/mmp_pdma.c 		struct dma_async_tx_descriptor *txd = &desc->async_tx;
async_tx           43 drivers/dma/mv_xor.c 	container_of(tx, struct mv_xor_desc_slot, async_tx)
async_tx          185 drivers/dma/mv_xor.c 	mv_chan_set_next_descriptor(mv_chan, sw_desc->async_tx.phys);
async_tx          196 drivers/dma/mv_xor.c 	BUG_ON(desc->async_tx.cookie < 0);
async_tx          198 drivers/dma/mv_xor.c 	if (desc->async_tx.cookie > 0) {
async_tx          199 drivers/dma/mv_xor.c 		cookie = desc->async_tx.cookie;
async_tx          201 drivers/dma/mv_xor.c 		dma_descriptor_unmap(&desc->async_tx);
async_tx          205 drivers/dma/mv_xor.c 		dmaengine_desc_get_callback_invoke(&desc->async_tx, NULL);
async_tx          209 drivers/dma/mv_xor.c 	dma_run_dependencies(&desc->async_tx);
async_tx          223 drivers/dma/mv_xor.c 		if (async_tx_test_ack(&iter->async_tx)) {
async_tx          239 drivers/dma/mv_xor.c 		__func__, __LINE__, desc, desc->async_tx.flags);
async_tx          244 drivers/dma/mv_xor.c 	if (!async_tx_test_ack(&desc->async_tx)) {
async_tx          293 drivers/dma/mv_xor.c 			if (iter->async_tx.phys == current_desc) {
async_tx          298 drivers/dma/mv_xor.c 			if (iter->async_tx.phys == current_desc) {
async_tx          365 drivers/dma/mv_xor.c 		async_tx_ack(&iter->async_tx);
async_tx          366 drivers/dma/mv_xor.c 		iter->async_tx.cookie = -EBUSY;
async_tx          392 drivers/dma/mv_xor.c 		__func__, sw_desc, &sw_desc->async_tx);
async_tx          408 drivers/dma/mv_xor.c 			&old_chain_tail->async_tx.phys);
async_tx          411 drivers/dma/mv_xor.c 		mv_desc_set_next_desc(old_chain_tail, sw_desc->async_tx.phys);
async_tx          420 drivers/dma/mv_xor.c 			if (current_desc == old_chain_tail->async_tx.phys)
async_tx          456 drivers/dma/mv_xor.c 		dma_async_tx_descriptor_init(&slot->async_tx, chan);
async_tx          457 drivers/dma/mv_xor.c 		slot->async_tx.tx_submit = mv_xor_tx_submit;
async_tx          461 drivers/dma/mv_xor.c 		slot->async_tx.phys = dma_desc + idx * MV_XOR_SLOT_SIZE;
async_tx          579 drivers/dma/mv_xor.c 		sw_desc->async_tx.flags = flags;
async_tx          594 drivers/dma/mv_xor.c 		__func__, sw_desc, &sw_desc->async_tx);
async_tx          595 drivers/dma/mv_xor.c 	return sw_desc ? &sw_desc->async_tx : NULL;
async_tx          147 drivers/dma/mv_xor.h 	struct dma_async_tx_descriptor	async_tx;
async_tx          179 drivers/dma/mv_xor_v2.c 	struct dma_async_tx_descriptor async_tx;
async_tx          304 drivers/dma/mv_xor_v2.c 		container_of(tx, struct mv_xor_v2_sw_desc, async_tx);
async_tx          310 drivers/dma/mv_xor_v2.c 		__func__, sw_desc, &sw_desc->async_tx);
async_tx          351 drivers/dma/mv_xor_v2.c 		if (async_tx_test_ack(&sw_desc->async_tx)) {
async_tx          391 drivers/dma/mv_xor_v2.c 	sw_desc->async_tx.flags = flags;
async_tx          420 drivers/dma/mv_xor_v2.c 	return &sw_desc->async_tx;
async_tx          447 drivers/dma/mv_xor_v2.c 	sw_desc->async_tx.flags = flags;
async_tx          479 drivers/dma/mv_xor_v2.c 	return &sw_desc->async_tx;
async_tx          509 drivers/dma/mv_xor_v2.c 	return &sw_desc->async_tx;
async_tx          573 drivers/dma/mv_xor_v2.c 		if (next_pending_sw_desc->async_tx.cookie > 0) {
async_tx          579 drivers/dma/mv_xor_v2.c 			dma_cookie_complete(&next_pending_sw_desc->async_tx);
async_tx          581 drivers/dma/mv_xor_v2.c 			dma_descriptor_unmap(&next_pending_sw_desc->async_tx);
async_tx          583 drivers/dma/mv_xor_v2.c 					&next_pending_sw_desc->async_tx, NULL);
async_tx          586 drivers/dma/mv_xor_v2.c 		dma_run_dependencies(&next_pending_sw_desc->async_tx);
async_tx          819 drivers/dma/mv_xor_v2.c 		dma_async_tx_descriptor_init(&sw_desc->async_tx,
async_tx          821 drivers/dma/mv_xor_v2.c 		sw_desc->async_tx.tx_submit = mv_xor_v2_tx_submit;
async_tx          822 drivers/dma/mv_xor_v2.c 		async_tx_ack(&sw_desc->async_tx);
async_tx          151 drivers/dma/nbpfaxi.c 	struct dma_async_tx_descriptor async_tx;
async_tx          630 drivers/dma/nbpfaxi.c 		running = chan->running ? chan->running->async_tx.cookie : -EINVAL;
async_tx          641 drivers/dma/nbpfaxi.c 				if (desc->async_tx.cookie == cookie) {
async_tx          648 drivers/dma/nbpfaxi.c 					if (desc->async_tx.cookie == cookie) {
async_tx          668 drivers/dma/nbpfaxi.c 	struct nbpf_desc *desc = container_of(tx, struct nbpf_desc, async_tx);
async_tx          716 drivers/dma/nbpfaxi.c 		dma_async_tx_descriptor_init(&desc->async_tx, dchan);
async_tx          717 drivers/dma/nbpfaxi.c 		desc->async_tx.tx_submit = nbpf_tx_submit;
async_tx          758 drivers/dma/nbpfaxi.c 		if (async_tx_test_ack(&desc->async_tx) && desc->user_wait) {
async_tx          851 drivers/dma/nbpfaxi.c 			__func__, desc, desc->async_tx.cookie);
async_tx          949 drivers/dma/nbpfaxi.c 	desc->async_tx.flags = flags;
async_tx          950 drivers/dma/nbpfaxi.c 	desc->async_tx.cookie = -EBUSY;
async_tx          979 drivers/dma/nbpfaxi.c 	return &desc->async_tx;
async_tx         1125 drivers/dma/nbpfaxi.c 			} else if (async_tx_test_ack(&desc->async_tx)) {
async_tx         1147 drivers/dma/nbpfaxi.c 		dma_cookie_complete(&desc->async_tx);
async_tx         1153 drivers/dma/nbpfaxi.c 		if (async_tx_test_ack(&desc->async_tx)) {
async_tx         1161 drivers/dma/nbpfaxi.c 		dmaengine_desc_get_callback(&desc->async_tx, &cb);
async_tx         1468 drivers/dma/ppc4xx/adma.c 	BUG_ON(desc->async_tx.cookie < 0);
async_tx         1469 drivers/dma/ppc4xx/adma.c 	if (desc->async_tx.cookie > 0) {
async_tx         1470 drivers/dma/ppc4xx/adma.c 		cookie = desc->async_tx.cookie;
async_tx         1471 drivers/dma/ppc4xx/adma.c 		desc->async_tx.cookie = 0;
async_tx         1473 drivers/dma/ppc4xx/adma.c 		dma_descriptor_unmap(&desc->async_tx);
async_tx         1477 drivers/dma/ppc4xx/adma.c 		dmaengine_desc_get_callback_invoke(&desc->async_tx, NULL);
async_tx         1481 drivers/dma/ppc4xx/adma.c 	dma_run_dependencies(&desc->async_tx);
async_tx         1495 drivers/dma/ppc4xx/adma.c 	if (!async_tx_test_ack(&desc->async_tx))
async_tx         1561 drivers/dma/ppc4xx/adma.c 		    iter->async_tx.cookie, iter->idx, busy, iter->phys,
async_tx         1563 drivers/dma/ppc4xx/adma.c 		    async_tx_test_ack(&iter->async_tx));
async_tx         1565 drivers/dma/ppc4xx/adma.c 		prefetch(&_iter->async_tx);
async_tx         1711 drivers/dma/ppc4xx/adma.c 		prefetch(&_iter->async_tx);
async_tx         1730 drivers/dma/ppc4xx/adma.c 					async_tx_ack(&iter->async_tx);
async_tx         1734 drivers/dma/ppc4xx/adma.c 				iter->async_tx.cookie = 0;
async_tx         1749 drivers/dma/ppc4xx/adma.c 			alloc_tail->async_tx.cookie = -EBUSY;
async_tx         1796 drivers/dma/ppc4xx/adma.c 		dma_async_tx_descriptor_init(&slot->async_tx, chan);
async_tx         1797 drivers/dma/ppc4xx/adma.c 		slot->async_tx.tx_submit = ppc440spe_adma_tx_submit;
async_tx         1938 drivers/dma/ppc4xx/adma.c 		sw_desc->async_tx.cookie, sw_desc->idx, sw_desc);
async_tx         1967 drivers/dma/ppc4xx/adma.c 		sw_desc->async_tx.flags = flags;
async_tx         1971 drivers/dma/ppc4xx/adma.c 	return sw_desc ? &sw_desc->async_tx : NULL;
async_tx         2008 drivers/dma/ppc4xx/adma.c 		sw_desc->async_tx.flags = flags;
async_tx         2012 drivers/dma/ppc4xx/adma.c 	return sw_desc ? &sw_desc->async_tx : NULL;
async_tx         2053 drivers/dma/ppc4xx/adma.c 		sw_desc->async_tx.flags = flags;
async_tx         2057 drivers/dma/ppc4xx/adma.c 	return sw_desc ? &sw_desc->async_tx : NULL;
async_tx         2106 drivers/dma/ppc4xx/adma.c 		chan = to_ppc440spe_adma_chan(sw_desc->async_tx.chan);
async_tx         2158 drivers/dma/ppc4xx/adma.c 		sw_desc->async_tx.flags = flags;
async_tx         2192 drivers/dma/ppc4xx/adma.c 		chan = to_ppc440spe_adma_chan(sw_desc->async_tx.chan);
async_tx         2267 drivers/dma/ppc4xx/adma.c 		sw_desc->async_tx.flags = flags;
async_tx         2407 drivers/dma/ppc4xx/adma.c 		sw_desc->async_tx.flags = flags;
async_tx         2447 drivers/dma/ppc4xx/adma.c 		sw_desc->async_tx.flags = flags;
async_tx         2535 drivers/dma/ppc4xx/adma.c 		return sw_desc ? &sw_desc->async_tx : NULL;
async_tx         2541 drivers/dma/ppc4xx/adma.c 		return sw_desc ? &sw_desc->async_tx : NULL;
async_tx         2578 drivers/dma/ppc4xx/adma.c 	return sw_desc ? &sw_desc->async_tx : NULL;
async_tx         2628 drivers/dma/ppc4xx/adma.c 		sw_desc->async_tx.flags = flags;
async_tx         2640 drivers/dma/ppc4xx/adma.c 			chan = to_ppc440spe_adma_chan(iter->async_tx.chan);
async_tx         2665 drivers/dma/ppc4xx/adma.c 			chan = to_ppc440spe_adma_chan(iter->async_tx.chan);
async_tx         2740 drivers/dma/ppc4xx/adma.c 			chan = to_ppc440spe_adma_chan(iter->async_tx.chan);
async_tx         2757 drivers/dma/ppc4xx/adma.c 	return sw_desc ? &sw_desc->async_tx : NULL;
async_tx         2792 drivers/dma/ppc4xx/adma.c 	chan = to_ppc440spe_adma_chan(sw_desc->async_tx.chan);
async_tx         2841 drivers/dma/ppc4xx/adma.c 	chan = to_ppc440spe_adma_chan(sw_desc->async_tx.chan);
async_tx         3030 drivers/dma/ppc4xx/adma.c 	chan = to_ppc440spe_adma_chan(sw_desc->async_tx.chan);
async_tx         3105 drivers/dma/ppc4xx/adma.c 	chan = to_ppc440spe_adma_chan(sw_desc->async_tx.chan);
async_tx         3213 drivers/dma/ppc4xx/adma.c 	chan = to_ppc440spe_adma_chan(sw_desc->async_tx.chan);
async_tx         3457 drivers/dma/ppc4xx/adma.c 	chan = to_ppc440spe_adma_chan(sw_desc->async_tx.chan);
async_tx         3671 drivers/dma/ppc4xx/adma.c 		async_tx_ack(&sw_desc->async_tx);
async_tx         3674 drivers/dma/ppc4xx/adma.c 		cookie = dma_cookie_assign(&sw_desc->async_tx);
async_tx         3745 drivers/dma/ppc4xx/adma.c 	async_tx_ack(&sw_desc->async_tx);
async_tx         3746 drivers/dma/ppc4xx/adma.c 	sw_desc->async_tx.callback = ppc440spe_test_callback;
async_tx         3747 drivers/dma/ppc4xx/adma.c 	sw_desc->async_tx.callback_param = NULL;
async_tx         3751 drivers/dma/ppc4xx/adma.c 	ppc440spe_adma_tx_submit(&sw_desc->async_tx);
async_tx           23 drivers/dma/ppc4xx/adma.h 		container_of(tx, struct ppc440spe_adma_desc_slot, async_tx)
async_tx          150 drivers/dma/ppc4xx/adma.h 	struct dma_async_tx_descriptor async_tx;
async_tx          137 drivers/dma/pxa_dma.c 	container_of(tx, struct pxad_desc_sw, async_tx)
async_tx           73 drivers/dma/sh/rcar-dmac.c 	struct dma_async_tx_descriptor async_tx;
async_tx           94 drivers/dma/sh/rcar-dmac.c #define to_rcar_dmac_desc(d)	container_of(d, struct rcar_dmac_desc, async_tx)
async_tx          402 drivers/dma/sh/rcar-dmac.c 		else if (desc->async_tx.callback)
async_tx          506 drivers/dma/sh/rcar-dmac.c 		dma_async_tx_descriptor_init(&desc->async_tx, &chan->chan);
async_tx          507 drivers/dma/sh/rcar-dmac.c 		desc->async_tx.tx_submit = rcar_dmac_tx_submit;
async_tx          561 drivers/dma/sh/rcar-dmac.c 		if (async_tx_test_ack(&desc->async_tx)) {
async_tx          916 drivers/dma/sh/rcar-dmac.c 	desc->async_tx.flags = dma_flags;
async_tx          917 drivers/dma/sh/rcar-dmac.c 	desc->async_tx.cookie = -EBUSY;
async_tx         1013 drivers/dma/sh/rcar-dmac.c 	return &desc->async_tx;
async_tx         1317 drivers/dma/sh/rcar-dmac.c 	if (cookie != desc->async_tx.cookie) {
async_tx         1319 drivers/dma/sh/rcar-dmac.c 			if (cookie == desc->async_tx.cookie)
async_tx         1323 drivers/dma/sh/rcar-dmac.c 			if (cookie == desc->async_tx.cookie)
async_tx         1327 drivers/dma/sh/rcar-dmac.c 			if (cookie == desc->async_tx.cookie)
async_tx         1604 drivers/dma/sh/rcar-dmac.c 		dmaengine_desc_get_callback(&desc->async_tx, &cb);
async_tx         1620 drivers/dma/sh/rcar-dmac.c 		dma_cookie_complete(&desc->async_tx);
async_tx         1623 drivers/dma/sh/rcar-dmac.c 		dmaengine_desc_get_callback(&desc->async_tx, &cb);
async_tx           73 drivers/dma/sh/shdma-base.c 		container_of(tx, struct shdma_desc, async_tx);
async_tx           92 drivers/dma/sh/shdma-base.c 				      chunk->async_tx.cookie > 0 ||
async_tx           93 drivers/dma/sh/shdma-base.c 				      chunk->async_tx.cookie == -EBUSY ||
async_tx           98 drivers/dma/sh/shdma-base.c 			chunk->async_tx.callback = callback;
async_tx           99 drivers/dma/sh/shdma-base.c 			chunk->async_tx.callback_param = tx->callback_param;
async_tx          102 drivers/dma/sh/shdma-base.c 			chunk->async_tx.callback = NULL;
async_tx          108 drivers/dma/sh/shdma-base.c 			tx->cookie, &chunk->async_tx, schan->id);
async_tx          237 drivers/dma/sh/shdma-base.c 		dma_async_tx_descriptor_init(&desc->async_tx,
async_tx          239 drivers/dma/sh/shdma-base.c 		desc->async_tx.tx_submit = shdma_tx_submit;
async_tx          337 drivers/dma/sh/shdma-base.c 		struct dma_async_tx_descriptor *tx = &desc->async_tx;
async_tx          389 drivers/dma/sh/shdma-base.c 					async_tx_ack(&desc->async_tx);
async_tx          398 drivers/dma/sh/shdma-base.c 		     async_tx_test_ack(&desc->async_tx)) || all) {
async_tx          524 drivers/dma/sh/shdma-base.c 		new->async_tx.cookie = -EBUSY;
async_tx          528 drivers/dma/sh/shdma-base.c 		new->async_tx.cookie = -EINVAL;
async_tx          533 drivers/dma/sh/shdma-base.c 		copy_size, *len, src, dst, &new->async_tx,
async_tx          534 drivers/dma/sh/shdma-base.c 		new->async_tx.cookie);
async_tx          537 drivers/dma/sh/shdma-base.c 	new->async_tx.flags = flags;
async_tx          620 drivers/dma/sh/shdma-base.c 		new->async_tx.cookie = -ENOSPC;
async_tx          627 drivers/dma/sh/shdma-base.c 	return &first->async_tx;
async_tx          884 drivers/dma/sh/shdma-base.c 			struct dma_async_tx_descriptor *tx = &sdesc->async_tx;
async_tx          929 drivers/dma/sh/shdma-base.c 				sdesc->async_tx.cookie, &sdesc->async_tx);
async_tx           57 drivers/dma/sh/shdma.h #define tx_to_sh_desc(tx) container_of(tx, struct sh_desc, async_tx)
async_tx          290 drivers/dma/sh/shdmac.c 		sdesc->async_tx.cookie, sh_chan->shdma_chan.id,
async_tx          308 drivers/dma/xilinx/xilinx_dma.c 	struct dma_async_tx_descriptor async_tx;
async_tx          455 drivers/dma/xilinx/xilinx_dma.c 	container_of(tx, struct xilinx_dma_tx_descriptor, async_tx)
async_tx          805 drivers/dma/xilinx/xilinx_dma.c 	callback = desc->async_tx.callback;
async_tx          806 drivers/dma/xilinx/xilinx_dma.c 	callback_param = desc->async_tx.callback_param;
async_tx          837 drivers/dma/xilinx/xilinx_dma.c 		dmaengine_desc_get_callback(&desc->async_tx, &cb);
async_tx          845 drivers/dma/xilinx/xilinx_dma.c 		dma_run_dependencies(&desc->async_tx);
async_tx         1238 drivers/dma/xilinx/xilinx_dma.c 			     head_desc->async_tx.phys);
async_tx         1306 drivers/dma/xilinx/xilinx_dma.c 			     head_desc->async_tx.phys);
async_tx         1311 drivers/dma/xilinx/xilinx_dma.c 				       head_desc->async_tx.phys);
async_tx         1315 drivers/dma/xilinx/xilinx_dma.c 				       head_desc->async_tx.phys);
async_tx         1319 drivers/dma/xilinx/xilinx_dma.c 				       head_desc->async_tx.phys);
async_tx         1404 drivers/dma/xilinx/xilinx_dma.c 			dma_cookie_complete(&desc->async_tx);
async_tx         1554 drivers/dma/xilinx/xilinx_dma.c 		tail_segment->hw.next_desc = (u32)desc->async_tx.phys;
async_tx         1559 drivers/dma/xilinx/xilinx_dma.c 		cdma_tail_segment->hw.next_desc = (u32)desc->async_tx.phys;
async_tx         1564 drivers/dma/xilinx/xilinx_dma.c 		axidma_tail_segment->hw.next_desc = (u32)desc->async_tx.phys;
async_tx         1659 drivers/dma/xilinx/xilinx_dma.c 	dma_async_tx_descriptor_init(&desc->async_tx, &chan->common);
async_tx         1660 drivers/dma/xilinx/xilinx_dma.c 	desc->async_tx.tx_submit = xilinx_dma_tx_submit;
async_tx         1661 drivers/dma/xilinx/xilinx_dma.c 	async_tx_ack(&desc->async_tx);
async_tx         1699 drivers/dma/xilinx/xilinx_dma.c 	desc->async_tx.phys = segment->phys;
async_tx         1701 drivers/dma/xilinx/xilinx_dma.c 	return &desc->async_tx;
async_tx         1734 drivers/dma/xilinx/xilinx_dma.c 	dma_async_tx_descriptor_init(&desc->async_tx, &chan->common);
async_tx         1735 drivers/dma/xilinx/xilinx_dma.c 	desc->async_tx.tx_submit = xilinx_dma_tx_submit;
async_tx         1754 drivers/dma/xilinx/xilinx_dma.c 	desc->async_tx.phys = segment->phys;
async_tx         1757 drivers/dma/xilinx/xilinx_dma.c 	return &desc->async_tx;
async_tx         1797 drivers/dma/xilinx/xilinx_dma.c 	dma_async_tx_descriptor_init(&desc->async_tx, &chan->common);
async_tx         1798 drivers/dma/xilinx/xilinx_dma.c 	desc->async_tx.tx_submit = xilinx_dma_tx_submit;
async_tx         1845 drivers/dma/xilinx/xilinx_dma.c 	desc->async_tx.phys = segment->phys;
async_tx         1856 drivers/dma/xilinx/xilinx_dma.c 	return &desc->async_tx;
async_tx         1904 drivers/dma/xilinx/xilinx_dma.c 	dma_async_tx_descriptor_init(&desc->async_tx, &chan->common);
async_tx         1905 drivers/dma/xilinx/xilinx_dma.c 	desc->async_tx.tx_submit = xilinx_dma_tx_submit;
async_tx         1945 drivers/dma/xilinx/xilinx_dma.c 	desc->async_tx.phys = head_segment->phys;
async_tx         1963 drivers/dma/xilinx/xilinx_dma.c 	return &desc->async_tx;
async_tx         2004 drivers/dma/xilinx/xilinx_dma.c 	dma_async_tx_descriptor_init(&desc->async_tx, &chan->common);
async_tx         2005 drivers/dma/xilinx/xilinx_dma.c 	desc->async_tx.tx_submit = xilinx_dma_tx_submit;
async_tx         2036 drivers/dma/xilinx/xilinx_dma.c 	desc->async_tx.phys = segment->phys;
async_tx         2047 drivers/dma/xilinx/xilinx_dma.c 	return &desc->async_tx;
async_tx          146 drivers/dma/xilinx/zynqmp_dma.c 					     async_tx)
async_tx          183 drivers/dma/xilinx/zynqmp_dma.c 	struct dma_async_tx_descriptor async_tx;
async_tx          484 drivers/dma/xilinx/zynqmp_dma.c 		dma_async_tx_descriptor_init(&desc->async_tx, &chan->common);
async_tx          485 drivers/dma/xilinx/zynqmp_dma.c 		desc->async_tx.tx_submit = zynqmp_dma_tx_submit;
async_tx          607 drivers/dma/xilinx/zynqmp_dma.c 		callback = desc->async_tx.callback;
async_tx          608 drivers/dma/xilinx/zynqmp_dma.c 		callback_param = desc->async_tx.callback_param;
async_tx          633 drivers/dma/xilinx/zynqmp_dma.c 	dma_cookie_complete(&desc->async_tx);
async_tx          843 drivers/dma/xilinx/zynqmp_dma.c 	async_tx_ack(&first->async_tx);
async_tx          844 drivers/dma/xilinx/zynqmp_dma.c 	first->async_tx.flags = flags;
async_tx          845 drivers/dma/xilinx/zynqmp_dma.c 	return &first->async_tx;
async_tx           92 include/linux/platform_data/dma-iop32x.h 	struct dma_async_tx_descriptor async_tx;
async_tx           48 include/linux/shdma-base.h 	struct dma_async_tx_descriptor async_tx;