async_tx 159 drivers/dma/altera-msgdma.c struct dma_async_tx_descriptor async_tx; async_tx 200 drivers/dma/altera-msgdma.c #define tx_to_desc(tx) container_of(tx, struct msgdma_sw_desc, async_tx) async_tx 368 drivers/dma/altera-msgdma.c async_tx_ack(&first->async_tx); async_tx 369 drivers/dma/altera-msgdma.c first->async_tx.flags = flags; async_tx 371 drivers/dma/altera-msgdma.c return &first->async_tx; async_tx 453 drivers/dma/altera-msgdma.c first->async_tx.flags = flags; async_tx 455 drivers/dma/altera-msgdma.c return &first->async_tx; async_tx 590 drivers/dma/altera-msgdma.c callback = desc->async_tx.callback; async_tx 591 drivers/dma/altera-msgdma.c callback_param = desc->async_tx.callback_param; async_tx 616 drivers/dma/altera-msgdma.c dma_cookie_complete(&desc->async_tx); async_tx 669 drivers/dma/altera-msgdma.c dma_async_tx_descriptor_init(&desc->async_tx, &mdev->dmachan); async_tx 670 drivers/dma/altera-msgdma.c desc->async_tx.tx_submit = msgdma_tx_submit; async_tx 83 drivers/dma/fsl_raid.c #define to_fsl_re_dma_desc(tx) container_of(tx, struct fsl_re_desc, async_tx) async_tx 137 drivers/dma/fsl_raid.c dma_cookie_complete(&desc->async_tx); async_tx 138 drivers/dma/fsl_raid.c dma_descriptor_unmap(&desc->async_tx); async_tx 139 drivers/dma/fsl_raid.c dmaengine_desc_get_callback_invoke(&desc->async_tx, NULL); async_tx 149 drivers/dma/fsl_raid.c if (async_tx_test_ack(&desc->async_tx)) async_tx 256 drivers/dma/fsl_raid.c desc->async_tx.tx_submit = fsl_re_tx_submit; async_tx 257 drivers/dma/fsl_raid.c dma_async_tx_descriptor_init(&desc->async_tx, &re_chan->chan); async_tx 289 drivers/dma/fsl_raid.c desc->async_tx.flags = flags; async_tx 306 drivers/dma/fsl_raid.c desc->async_tx.flags = flags; async_tx 384 drivers/dma/fsl_raid.c return &desc->async_tx; async_tx 518 drivers/dma/fsl_raid.c return &desc->async_tx; async_tx 570 drivers/dma/fsl_raid.c return &desc->async_tx; async_tx 294 drivers/dma/fsl_raid.h struct dma_async_tx_descriptor async_tx; async_tx 396 drivers/dma/fsldma.c set_desc_next(chan, &tail->hw, desc->async_tx.phys); async_tx 428 drivers/dma/fsldma.c cookie = dma_cookie_assign(&child->async_tx); async_tx 449 drivers/dma/fsldma.c dma_pool_free(chan->desc_pool, desc, desc->async_tx.phys); async_tx 470 drivers/dma/fsldma.c dma_async_tx_descriptor_init(&desc->async_tx, &chan->common); async_tx 471 drivers/dma/fsldma.c desc->async_tx.tx_submit = fsl_dma_tx_submit; async_tx 472 drivers/dma/fsldma.c desc->async_tx.phys = pdesc; async_tx 493 drivers/dma/fsldma.c if (async_tx_test_ack(&desc->async_tx)) async_tx 509 drivers/dma/fsldma.c struct dma_async_tx_descriptor *txd = &desc->async_tx; async_tx 547 drivers/dma/fsldma.c if (!async_tx_test_ack(&desc->async_tx)) { async_tx 556 drivers/dma/fsldma.c dma_pool_free(chan->desc_pool, desc, desc->async_tx.phys); async_tx 619 drivers/dma/fsldma.c set_cdar(chan, desc->async_tx.phys); async_tx 658 drivers/dma/fsldma.c if (desc->async_tx.phys == curr_phys) { async_tx 793 drivers/dma/fsldma.c set_desc_next(chan, &prev->hw, new->async_tx.phys); async_tx 795 drivers/dma/fsldma.c new->async_tx.cookie = 0; async_tx 796 drivers/dma/fsldma.c async_tx_ack(&new->async_tx); async_tx 807 drivers/dma/fsldma.c new->async_tx.flags = flags; /* client is in control of this ack */ async_tx 808 drivers/dma/fsldma.c new->async_tx.cookie = -EBUSY; async_tx 813 drivers/dma/fsldma.c return &first->async_tx; async_tx 104 drivers/dma/fsldma.h struct dma_async_tx_descriptor async_tx; async_tx 192 drivers/dma/fsldma.h #define tx_to_fsl_desc(tx) container_of(tx, struct fsl_desc_sw, async_tx) async_tx 32 drivers/dma/iop-adma.c container_of(tx, struct iop_adma_desc_slot, async_tx) async_tx 55 drivers/dma/iop-adma.c struct dma_async_tx_descriptor *tx = &desc->async_tx; async_tx 85 drivers/dma/iop-adma.c if (!async_tx_test_ack(&desc->async_tx)) async_tx 120 drivers/dma/iop-adma.c iter->async_tx.cookie, iter->idx, busy, async_tx 121 drivers/dma/iop-adma.c &iter->async_tx.phys, (u64)iop_desc_get_next_desc(iter), async_tx 122 drivers/dma/iop-adma.c async_tx_test_ack(&iter->async_tx)); async_tx 124 drivers/dma/iop-adma.c prefetch(&_iter->async_tx); async_tx 137 drivers/dma/iop-adma.c if (iter->async_tx.phys == current_desc) { async_tx 216 drivers/dma/iop-adma.c if (iter->xor_check_result && iter->async_tx.cookie) async_tx 279 drivers/dma/iop-adma.c prefetch(&_iter->async_tx); async_tx 311 drivers/dma/iop-adma.c (u64)iter->async_tx.phys, slots_per_op); async_tx 315 drivers/dma/iop-adma.c async_tx_ack(&iter->async_tx); async_tx 319 drivers/dma/iop-adma.c iter->async_tx.cookie = 0; async_tx 332 drivers/dma/iop-adma.c alloc_tail->async_tx.cookie = -EBUSY; async_tx 382 drivers/dma/iop-adma.c next_dma = grp_start->async_tx.phys; async_tx 399 drivers/dma/iop-adma.c __func__, sw_desc->async_tx.cookie, sw_desc->idx); async_tx 443 drivers/dma/iop-adma.c dma_async_tx_descriptor_init(&slot->async_tx, chan); async_tx 444 drivers/dma/iop-adma.c slot->async_tx.tx_submit = iop_adma_tx_submit; async_tx 449 drivers/dma/iop-adma.c slot->async_tx.phys = async_tx 498 drivers/dma/iop-adma.c sw_desc->async_tx.flags = flags; async_tx 502 drivers/dma/iop-adma.c return sw_desc ? &sw_desc->async_tx : NULL; async_tx 529 drivers/dma/iop-adma.c sw_desc->async_tx.flags = flags; async_tx 533 drivers/dma/iop-adma.c return sw_desc ? &sw_desc->async_tx : NULL; async_tx 561 drivers/dma/iop-adma.c sw_desc->async_tx.flags = flags; async_tx 568 drivers/dma/iop-adma.c return sw_desc ? &sw_desc->async_tx : NULL; async_tx 596 drivers/dma/iop-adma.c sw_desc->async_tx.flags = flags; async_tx 603 drivers/dma/iop-adma.c return sw_desc ? &sw_desc->async_tx : NULL; async_tx 648 drivers/dma/iop-adma.c sw_desc->async_tx.flags = flags; async_tx 667 drivers/dma/iop-adma.c return sw_desc ? &sw_desc->async_tx : NULL; async_tx 702 drivers/dma/iop-adma.c sw_desc->async_tx.flags = flags; async_tx 711 drivers/dma/iop-adma.c return sw_desc ? &sw_desc->async_tx : NULL; async_tx 1450 drivers/dma/iop-adma.c async_tx_ack(&sw_desc->async_tx); async_tx 1456 drivers/dma/iop-adma.c cookie = dma_cookie_assign(&sw_desc->async_tx); async_tx 1473 drivers/dma/iop-adma.c iop_chan_set_next_descriptor(iop_chan, sw_desc->async_tx.phys); async_tx 1502 drivers/dma/iop-adma.c async_tx_ack(&sw_desc->async_tx); async_tx 1509 drivers/dma/iop-adma.c cookie = dma_cookie_assign(&sw_desc->async_tx); async_tx 1526 drivers/dma/iop-adma.c iop_chan_set_next_descriptor(iop_chan, sw_desc->async_tx.phys); async_tx 583 drivers/dma/iop-adma.h (u32) (desc->async_tx.phys + (i << 5)); async_tx 85 drivers/dma/mmp_pdma.c struct dma_async_tx_descriptor async_tx; async_tx 133 drivers/dma/mmp_pdma.c container_of(tx, struct mmp_pdma_desc_sw, async_tx) async_tx 334 drivers/dma/mmp_pdma.c set_desc(chan->phy, desc->async_tx.phys); async_tx 352 drivers/dma/mmp_pdma.c cookie = dma_cookie_assign(&child->async_tx); async_tx 376 drivers/dma/mmp_pdma.c dma_async_tx_descriptor_init(&desc->async_tx, &chan->chan); async_tx 378 drivers/dma/mmp_pdma.c desc->async_tx.tx_submit = mmp_pdma_tx_submit; async_tx 379 drivers/dma/mmp_pdma.c desc->async_tx.phys = pdesc; async_tx 422 drivers/dma/mmp_pdma.c dma_pool_free(chan->desc_pool, desc, desc->async_tx.phys); async_tx 487 drivers/dma/mmp_pdma.c prev->desc.ddadr = new->async_tx.phys; async_tx 489 drivers/dma/mmp_pdma.c new->async_tx.cookie = 0; async_tx 490 drivers/dma/mmp_pdma.c async_tx_ack(&new->async_tx); async_tx 508 drivers/dma/mmp_pdma.c first->async_tx.flags = flags; /* client is in control of this ack */ async_tx 509 drivers/dma/mmp_pdma.c first->async_tx.cookie = -EBUSY; async_tx 517 drivers/dma/mmp_pdma.c return &first->async_tx; async_tx 572 drivers/dma/mmp_pdma.c prev->desc.ddadr = new->async_tx.phys; async_tx 574 drivers/dma/mmp_pdma.c new->async_tx.cookie = 0; async_tx 575 drivers/dma/mmp_pdma.c async_tx_ack(&new->async_tx); async_tx 587 drivers/dma/mmp_pdma.c first->async_tx.cookie = -EBUSY; async_tx 588 drivers/dma/mmp_pdma.c first->async_tx.flags = flags; async_tx 597 drivers/dma/mmp_pdma.c return &first->async_tx; async_tx 660 drivers/dma/mmp_pdma.c prev->desc.ddadr = new->async_tx.phys; async_tx 662 drivers/dma/mmp_pdma.c new->async_tx.cookie = 0; async_tx 663 drivers/dma/mmp_pdma.c async_tx_ack(&new->async_tx); async_tx 677 drivers/dma/mmp_pdma.c first->async_tx.flags = flags; /* client is in control of this ack */ async_tx 678 drivers/dma/mmp_pdma.c first->async_tx.cookie = -EBUSY; async_tx 681 drivers/dma/mmp_pdma.c new->desc.ddadr = first->async_tx.phys; async_tx 684 drivers/dma/mmp_pdma.c return &first->async_tx; async_tx 831 drivers/dma/mmp_pdma.c if (sw->async_tx.cookie == cookie) { async_tx 887 drivers/dma/mmp_pdma.c dmaengine_desc_get_callback(&desc->async_tx, &cb); async_tx 911 drivers/dma/mmp_pdma.c dma_cookie_t cookie = desc->async_tx.cookie; async_tx 912 drivers/dma/mmp_pdma.c dma_cookie_complete(&desc->async_tx); async_tx 930 drivers/dma/mmp_pdma.c struct dma_async_tx_descriptor *txd = &desc->async_tx; async_tx 43 drivers/dma/mv_xor.c container_of(tx, struct mv_xor_desc_slot, async_tx) async_tx 185 drivers/dma/mv_xor.c mv_chan_set_next_descriptor(mv_chan, sw_desc->async_tx.phys); async_tx 196 drivers/dma/mv_xor.c BUG_ON(desc->async_tx.cookie < 0); async_tx 198 drivers/dma/mv_xor.c if (desc->async_tx.cookie > 0) { async_tx 199 drivers/dma/mv_xor.c cookie = desc->async_tx.cookie; async_tx 201 drivers/dma/mv_xor.c dma_descriptor_unmap(&desc->async_tx); async_tx 205 drivers/dma/mv_xor.c dmaengine_desc_get_callback_invoke(&desc->async_tx, NULL); async_tx 209 drivers/dma/mv_xor.c dma_run_dependencies(&desc->async_tx); async_tx 223 drivers/dma/mv_xor.c if (async_tx_test_ack(&iter->async_tx)) { async_tx 239 drivers/dma/mv_xor.c __func__, __LINE__, desc, desc->async_tx.flags); async_tx 244 drivers/dma/mv_xor.c if (!async_tx_test_ack(&desc->async_tx)) { async_tx 293 drivers/dma/mv_xor.c if (iter->async_tx.phys == current_desc) { async_tx 298 drivers/dma/mv_xor.c if (iter->async_tx.phys == current_desc) { async_tx 365 drivers/dma/mv_xor.c async_tx_ack(&iter->async_tx); async_tx 366 drivers/dma/mv_xor.c iter->async_tx.cookie = -EBUSY; async_tx 392 drivers/dma/mv_xor.c __func__, sw_desc, &sw_desc->async_tx); async_tx 408 drivers/dma/mv_xor.c &old_chain_tail->async_tx.phys); async_tx 411 drivers/dma/mv_xor.c mv_desc_set_next_desc(old_chain_tail, sw_desc->async_tx.phys); async_tx 420 drivers/dma/mv_xor.c if (current_desc == old_chain_tail->async_tx.phys) async_tx 456 drivers/dma/mv_xor.c dma_async_tx_descriptor_init(&slot->async_tx, chan); async_tx 457 drivers/dma/mv_xor.c slot->async_tx.tx_submit = mv_xor_tx_submit; async_tx 461 drivers/dma/mv_xor.c slot->async_tx.phys = dma_desc + idx * MV_XOR_SLOT_SIZE; async_tx 579 drivers/dma/mv_xor.c sw_desc->async_tx.flags = flags; async_tx 594 drivers/dma/mv_xor.c __func__, sw_desc, &sw_desc->async_tx); async_tx 595 drivers/dma/mv_xor.c return sw_desc ? &sw_desc->async_tx : NULL; async_tx 147 drivers/dma/mv_xor.h struct dma_async_tx_descriptor async_tx; async_tx 179 drivers/dma/mv_xor_v2.c struct dma_async_tx_descriptor async_tx; async_tx 304 drivers/dma/mv_xor_v2.c container_of(tx, struct mv_xor_v2_sw_desc, async_tx); async_tx 310 drivers/dma/mv_xor_v2.c __func__, sw_desc, &sw_desc->async_tx); async_tx 351 drivers/dma/mv_xor_v2.c if (async_tx_test_ack(&sw_desc->async_tx)) { async_tx 391 drivers/dma/mv_xor_v2.c sw_desc->async_tx.flags = flags; async_tx 420 drivers/dma/mv_xor_v2.c return &sw_desc->async_tx; async_tx 447 drivers/dma/mv_xor_v2.c sw_desc->async_tx.flags = flags; async_tx 479 drivers/dma/mv_xor_v2.c return &sw_desc->async_tx; async_tx 509 drivers/dma/mv_xor_v2.c return &sw_desc->async_tx; async_tx 573 drivers/dma/mv_xor_v2.c if (next_pending_sw_desc->async_tx.cookie > 0) { async_tx 579 drivers/dma/mv_xor_v2.c dma_cookie_complete(&next_pending_sw_desc->async_tx); async_tx 581 drivers/dma/mv_xor_v2.c dma_descriptor_unmap(&next_pending_sw_desc->async_tx); async_tx 583 drivers/dma/mv_xor_v2.c &next_pending_sw_desc->async_tx, NULL); async_tx 586 drivers/dma/mv_xor_v2.c dma_run_dependencies(&next_pending_sw_desc->async_tx); async_tx 819 drivers/dma/mv_xor_v2.c dma_async_tx_descriptor_init(&sw_desc->async_tx, async_tx 821 drivers/dma/mv_xor_v2.c sw_desc->async_tx.tx_submit = mv_xor_v2_tx_submit; async_tx 822 drivers/dma/mv_xor_v2.c async_tx_ack(&sw_desc->async_tx); async_tx 151 drivers/dma/nbpfaxi.c struct dma_async_tx_descriptor async_tx; async_tx 630 drivers/dma/nbpfaxi.c running = chan->running ? chan->running->async_tx.cookie : -EINVAL; async_tx 641 drivers/dma/nbpfaxi.c if (desc->async_tx.cookie == cookie) { async_tx 648 drivers/dma/nbpfaxi.c if (desc->async_tx.cookie == cookie) { async_tx 668 drivers/dma/nbpfaxi.c struct nbpf_desc *desc = container_of(tx, struct nbpf_desc, async_tx); async_tx 716 drivers/dma/nbpfaxi.c dma_async_tx_descriptor_init(&desc->async_tx, dchan); async_tx 717 drivers/dma/nbpfaxi.c desc->async_tx.tx_submit = nbpf_tx_submit; async_tx 758 drivers/dma/nbpfaxi.c if (async_tx_test_ack(&desc->async_tx) && desc->user_wait) { async_tx 851 drivers/dma/nbpfaxi.c __func__, desc, desc->async_tx.cookie); async_tx 949 drivers/dma/nbpfaxi.c desc->async_tx.flags = flags; async_tx 950 drivers/dma/nbpfaxi.c desc->async_tx.cookie = -EBUSY; async_tx 979 drivers/dma/nbpfaxi.c return &desc->async_tx; async_tx 1125 drivers/dma/nbpfaxi.c } else if (async_tx_test_ack(&desc->async_tx)) { async_tx 1147 drivers/dma/nbpfaxi.c dma_cookie_complete(&desc->async_tx); async_tx 1153 drivers/dma/nbpfaxi.c if (async_tx_test_ack(&desc->async_tx)) { async_tx 1161 drivers/dma/nbpfaxi.c dmaengine_desc_get_callback(&desc->async_tx, &cb); async_tx 1468 drivers/dma/ppc4xx/adma.c BUG_ON(desc->async_tx.cookie < 0); async_tx 1469 drivers/dma/ppc4xx/adma.c if (desc->async_tx.cookie > 0) { async_tx 1470 drivers/dma/ppc4xx/adma.c cookie = desc->async_tx.cookie; async_tx 1471 drivers/dma/ppc4xx/adma.c desc->async_tx.cookie = 0; async_tx 1473 drivers/dma/ppc4xx/adma.c dma_descriptor_unmap(&desc->async_tx); async_tx 1477 drivers/dma/ppc4xx/adma.c dmaengine_desc_get_callback_invoke(&desc->async_tx, NULL); async_tx 1481 drivers/dma/ppc4xx/adma.c dma_run_dependencies(&desc->async_tx); async_tx 1495 drivers/dma/ppc4xx/adma.c if (!async_tx_test_ack(&desc->async_tx)) async_tx 1561 drivers/dma/ppc4xx/adma.c iter->async_tx.cookie, iter->idx, busy, iter->phys, async_tx 1563 drivers/dma/ppc4xx/adma.c async_tx_test_ack(&iter->async_tx)); async_tx 1565 drivers/dma/ppc4xx/adma.c prefetch(&_iter->async_tx); async_tx 1711 drivers/dma/ppc4xx/adma.c prefetch(&_iter->async_tx); async_tx 1730 drivers/dma/ppc4xx/adma.c async_tx_ack(&iter->async_tx); async_tx 1734 drivers/dma/ppc4xx/adma.c iter->async_tx.cookie = 0; async_tx 1749 drivers/dma/ppc4xx/adma.c alloc_tail->async_tx.cookie = -EBUSY; async_tx 1796 drivers/dma/ppc4xx/adma.c dma_async_tx_descriptor_init(&slot->async_tx, chan); async_tx 1797 drivers/dma/ppc4xx/adma.c slot->async_tx.tx_submit = ppc440spe_adma_tx_submit; async_tx 1938 drivers/dma/ppc4xx/adma.c sw_desc->async_tx.cookie, sw_desc->idx, sw_desc); async_tx 1967 drivers/dma/ppc4xx/adma.c sw_desc->async_tx.flags = flags; async_tx 1971 drivers/dma/ppc4xx/adma.c return sw_desc ? &sw_desc->async_tx : NULL; async_tx 2008 drivers/dma/ppc4xx/adma.c sw_desc->async_tx.flags = flags; async_tx 2012 drivers/dma/ppc4xx/adma.c return sw_desc ? &sw_desc->async_tx : NULL; async_tx 2053 drivers/dma/ppc4xx/adma.c sw_desc->async_tx.flags = flags; async_tx 2057 drivers/dma/ppc4xx/adma.c return sw_desc ? &sw_desc->async_tx : NULL; async_tx 2106 drivers/dma/ppc4xx/adma.c chan = to_ppc440spe_adma_chan(sw_desc->async_tx.chan); async_tx 2158 drivers/dma/ppc4xx/adma.c sw_desc->async_tx.flags = flags; async_tx 2192 drivers/dma/ppc4xx/adma.c chan = to_ppc440spe_adma_chan(sw_desc->async_tx.chan); async_tx 2267 drivers/dma/ppc4xx/adma.c sw_desc->async_tx.flags = flags; async_tx 2407 drivers/dma/ppc4xx/adma.c sw_desc->async_tx.flags = flags; async_tx 2447 drivers/dma/ppc4xx/adma.c sw_desc->async_tx.flags = flags; async_tx 2535 drivers/dma/ppc4xx/adma.c return sw_desc ? &sw_desc->async_tx : NULL; async_tx 2541 drivers/dma/ppc4xx/adma.c return sw_desc ? &sw_desc->async_tx : NULL; async_tx 2578 drivers/dma/ppc4xx/adma.c return sw_desc ? &sw_desc->async_tx : NULL; async_tx 2628 drivers/dma/ppc4xx/adma.c sw_desc->async_tx.flags = flags; async_tx 2640 drivers/dma/ppc4xx/adma.c chan = to_ppc440spe_adma_chan(iter->async_tx.chan); async_tx 2665 drivers/dma/ppc4xx/adma.c chan = to_ppc440spe_adma_chan(iter->async_tx.chan); async_tx 2740 drivers/dma/ppc4xx/adma.c chan = to_ppc440spe_adma_chan(iter->async_tx.chan); async_tx 2757 drivers/dma/ppc4xx/adma.c return sw_desc ? &sw_desc->async_tx : NULL; async_tx 2792 drivers/dma/ppc4xx/adma.c chan = to_ppc440spe_adma_chan(sw_desc->async_tx.chan); async_tx 2841 drivers/dma/ppc4xx/adma.c chan = to_ppc440spe_adma_chan(sw_desc->async_tx.chan); async_tx 3030 drivers/dma/ppc4xx/adma.c chan = to_ppc440spe_adma_chan(sw_desc->async_tx.chan); async_tx 3105 drivers/dma/ppc4xx/adma.c chan = to_ppc440spe_adma_chan(sw_desc->async_tx.chan); async_tx 3213 drivers/dma/ppc4xx/adma.c chan = to_ppc440spe_adma_chan(sw_desc->async_tx.chan); async_tx 3457 drivers/dma/ppc4xx/adma.c chan = to_ppc440spe_adma_chan(sw_desc->async_tx.chan); async_tx 3671 drivers/dma/ppc4xx/adma.c async_tx_ack(&sw_desc->async_tx); async_tx 3674 drivers/dma/ppc4xx/adma.c cookie = dma_cookie_assign(&sw_desc->async_tx); async_tx 3745 drivers/dma/ppc4xx/adma.c async_tx_ack(&sw_desc->async_tx); async_tx 3746 drivers/dma/ppc4xx/adma.c sw_desc->async_tx.callback = ppc440spe_test_callback; async_tx 3747 drivers/dma/ppc4xx/adma.c sw_desc->async_tx.callback_param = NULL; async_tx 3751 drivers/dma/ppc4xx/adma.c ppc440spe_adma_tx_submit(&sw_desc->async_tx); async_tx 23 drivers/dma/ppc4xx/adma.h container_of(tx, struct ppc440spe_adma_desc_slot, async_tx) async_tx 150 drivers/dma/ppc4xx/adma.h struct dma_async_tx_descriptor async_tx; async_tx 137 drivers/dma/pxa_dma.c container_of(tx, struct pxad_desc_sw, async_tx) async_tx 73 drivers/dma/sh/rcar-dmac.c struct dma_async_tx_descriptor async_tx; async_tx 94 drivers/dma/sh/rcar-dmac.c #define to_rcar_dmac_desc(d) container_of(d, struct rcar_dmac_desc, async_tx) async_tx 402 drivers/dma/sh/rcar-dmac.c else if (desc->async_tx.callback) async_tx 506 drivers/dma/sh/rcar-dmac.c dma_async_tx_descriptor_init(&desc->async_tx, &chan->chan); async_tx 507 drivers/dma/sh/rcar-dmac.c desc->async_tx.tx_submit = rcar_dmac_tx_submit; async_tx 561 drivers/dma/sh/rcar-dmac.c if (async_tx_test_ack(&desc->async_tx)) { async_tx 916 drivers/dma/sh/rcar-dmac.c desc->async_tx.flags = dma_flags; async_tx 917 drivers/dma/sh/rcar-dmac.c desc->async_tx.cookie = -EBUSY; async_tx 1013 drivers/dma/sh/rcar-dmac.c return &desc->async_tx; async_tx 1317 drivers/dma/sh/rcar-dmac.c if (cookie != desc->async_tx.cookie) { async_tx 1319 drivers/dma/sh/rcar-dmac.c if (cookie == desc->async_tx.cookie) async_tx 1323 drivers/dma/sh/rcar-dmac.c if (cookie == desc->async_tx.cookie) async_tx 1327 drivers/dma/sh/rcar-dmac.c if (cookie == desc->async_tx.cookie) async_tx 1604 drivers/dma/sh/rcar-dmac.c dmaengine_desc_get_callback(&desc->async_tx, &cb); async_tx 1620 drivers/dma/sh/rcar-dmac.c dma_cookie_complete(&desc->async_tx); async_tx 1623 drivers/dma/sh/rcar-dmac.c dmaengine_desc_get_callback(&desc->async_tx, &cb); async_tx 73 drivers/dma/sh/shdma-base.c container_of(tx, struct shdma_desc, async_tx); async_tx 92 drivers/dma/sh/shdma-base.c chunk->async_tx.cookie > 0 || async_tx 93 drivers/dma/sh/shdma-base.c chunk->async_tx.cookie == -EBUSY || async_tx 98 drivers/dma/sh/shdma-base.c chunk->async_tx.callback = callback; async_tx 99 drivers/dma/sh/shdma-base.c chunk->async_tx.callback_param = tx->callback_param; async_tx 102 drivers/dma/sh/shdma-base.c chunk->async_tx.callback = NULL; async_tx 108 drivers/dma/sh/shdma-base.c tx->cookie, &chunk->async_tx, schan->id); async_tx 237 drivers/dma/sh/shdma-base.c dma_async_tx_descriptor_init(&desc->async_tx, async_tx 239 drivers/dma/sh/shdma-base.c desc->async_tx.tx_submit = shdma_tx_submit; async_tx 337 drivers/dma/sh/shdma-base.c struct dma_async_tx_descriptor *tx = &desc->async_tx; async_tx 389 drivers/dma/sh/shdma-base.c async_tx_ack(&desc->async_tx); async_tx 398 drivers/dma/sh/shdma-base.c async_tx_test_ack(&desc->async_tx)) || all) { async_tx 524 drivers/dma/sh/shdma-base.c new->async_tx.cookie = -EBUSY; async_tx 528 drivers/dma/sh/shdma-base.c new->async_tx.cookie = -EINVAL; async_tx 533 drivers/dma/sh/shdma-base.c copy_size, *len, src, dst, &new->async_tx, async_tx 534 drivers/dma/sh/shdma-base.c new->async_tx.cookie); async_tx 537 drivers/dma/sh/shdma-base.c new->async_tx.flags = flags; async_tx 620 drivers/dma/sh/shdma-base.c new->async_tx.cookie = -ENOSPC; async_tx 627 drivers/dma/sh/shdma-base.c return &first->async_tx; async_tx 884 drivers/dma/sh/shdma-base.c struct dma_async_tx_descriptor *tx = &sdesc->async_tx; async_tx 929 drivers/dma/sh/shdma-base.c sdesc->async_tx.cookie, &sdesc->async_tx); async_tx 57 drivers/dma/sh/shdma.h #define tx_to_sh_desc(tx) container_of(tx, struct sh_desc, async_tx) async_tx 290 drivers/dma/sh/shdmac.c sdesc->async_tx.cookie, sh_chan->shdma_chan.id, async_tx 308 drivers/dma/xilinx/xilinx_dma.c struct dma_async_tx_descriptor async_tx; async_tx 455 drivers/dma/xilinx/xilinx_dma.c container_of(tx, struct xilinx_dma_tx_descriptor, async_tx) async_tx 805 drivers/dma/xilinx/xilinx_dma.c callback = desc->async_tx.callback; async_tx 806 drivers/dma/xilinx/xilinx_dma.c callback_param = desc->async_tx.callback_param; async_tx 837 drivers/dma/xilinx/xilinx_dma.c dmaengine_desc_get_callback(&desc->async_tx, &cb); async_tx 845 drivers/dma/xilinx/xilinx_dma.c dma_run_dependencies(&desc->async_tx); async_tx 1238 drivers/dma/xilinx/xilinx_dma.c head_desc->async_tx.phys); async_tx 1306 drivers/dma/xilinx/xilinx_dma.c head_desc->async_tx.phys); async_tx 1311 drivers/dma/xilinx/xilinx_dma.c head_desc->async_tx.phys); async_tx 1315 drivers/dma/xilinx/xilinx_dma.c head_desc->async_tx.phys); async_tx 1319 drivers/dma/xilinx/xilinx_dma.c head_desc->async_tx.phys); async_tx 1404 drivers/dma/xilinx/xilinx_dma.c dma_cookie_complete(&desc->async_tx); async_tx 1554 drivers/dma/xilinx/xilinx_dma.c tail_segment->hw.next_desc = (u32)desc->async_tx.phys; async_tx 1559 drivers/dma/xilinx/xilinx_dma.c cdma_tail_segment->hw.next_desc = (u32)desc->async_tx.phys; async_tx 1564 drivers/dma/xilinx/xilinx_dma.c axidma_tail_segment->hw.next_desc = (u32)desc->async_tx.phys; async_tx 1659 drivers/dma/xilinx/xilinx_dma.c dma_async_tx_descriptor_init(&desc->async_tx, &chan->common); async_tx 1660 drivers/dma/xilinx/xilinx_dma.c desc->async_tx.tx_submit = xilinx_dma_tx_submit; async_tx 1661 drivers/dma/xilinx/xilinx_dma.c async_tx_ack(&desc->async_tx); async_tx 1699 drivers/dma/xilinx/xilinx_dma.c desc->async_tx.phys = segment->phys; async_tx 1701 drivers/dma/xilinx/xilinx_dma.c return &desc->async_tx; async_tx 1734 drivers/dma/xilinx/xilinx_dma.c dma_async_tx_descriptor_init(&desc->async_tx, &chan->common); async_tx 1735 drivers/dma/xilinx/xilinx_dma.c desc->async_tx.tx_submit = xilinx_dma_tx_submit; async_tx 1754 drivers/dma/xilinx/xilinx_dma.c desc->async_tx.phys = segment->phys; async_tx 1757 drivers/dma/xilinx/xilinx_dma.c return &desc->async_tx; async_tx 1797 drivers/dma/xilinx/xilinx_dma.c dma_async_tx_descriptor_init(&desc->async_tx, &chan->common); async_tx 1798 drivers/dma/xilinx/xilinx_dma.c desc->async_tx.tx_submit = xilinx_dma_tx_submit; async_tx 1845 drivers/dma/xilinx/xilinx_dma.c desc->async_tx.phys = segment->phys; async_tx 1856 drivers/dma/xilinx/xilinx_dma.c return &desc->async_tx; async_tx 1904 drivers/dma/xilinx/xilinx_dma.c dma_async_tx_descriptor_init(&desc->async_tx, &chan->common); async_tx 1905 drivers/dma/xilinx/xilinx_dma.c desc->async_tx.tx_submit = xilinx_dma_tx_submit; async_tx 1945 drivers/dma/xilinx/xilinx_dma.c desc->async_tx.phys = head_segment->phys; async_tx 1963 drivers/dma/xilinx/xilinx_dma.c return &desc->async_tx; async_tx 2004 drivers/dma/xilinx/xilinx_dma.c dma_async_tx_descriptor_init(&desc->async_tx, &chan->common); async_tx 2005 drivers/dma/xilinx/xilinx_dma.c desc->async_tx.tx_submit = xilinx_dma_tx_submit; async_tx 2036 drivers/dma/xilinx/xilinx_dma.c desc->async_tx.phys = segment->phys; async_tx 2047 drivers/dma/xilinx/xilinx_dma.c return &desc->async_tx; async_tx 146 drivers/dma/xilinx/zynqmp_dma.c async_tx) async_tx 183 drivers/dma/xilinx/zynqmp_dma.c struct dma_async_tx_descriptor async_tx; async_tx 484 drivers/dma/xilinx/zynqmp_dma.c dma_async_tx_descriptor_init(&desc->async_tx, &chan->common); async_tx 485 drivers/dma/xilinx/zynqmp_dma.c desc->async_tx.tx_submit = zynqmp_dma_tx_submit; async_tx 607 drivers/dma/xilinx/zynqmp_dma.c callback = desc->async_tx.callback; async_tx 608 drivers/dma/xilinx/zynqmp_dma.c callback_param = desc->async_tx.callback_param; async_tx 633 drivers/dma/xilinx/zynqmp_dma.c dma_cookie_complete(&desc->async_tx); async_tx 843 drivers/dma/xilinx/zynqmp_dma.c async_tx_ack(&first->async_tx); async_tx 844 drivers/dma/xilinx/zynqmp_dma.c first->async_tx.flags = flags; async_tx 845 drivers/dma/xilinx/zynqmp_dma.c return &first->async_tx; async_tx 92 include/linux/platform_data/dma-iop32x.h struct dma_async_tx_descriptor async_tx; async_tx 48 include/linux/shdma-base.h struct dma_async_tx_descriptor async_tx;