txd 238 arch/arm/mach-imx/mach-pcm043.c iomux_v3_cfg_t txd = MX35_PAD_STXD4__AUDMUX_AUD4_TXD; txd 266 arch/arm/mach-imx/mach-pcm043.c mxc_iomux_v3_setup_pad(txd); txd 704 drivers/ata/pata_ep93xx.c struct dma_async_tx_descriptor *txd; txd 712 drivers/ata/pata_ep93xx.c txd = dmaengine_prep_slave_sg(channel, qc->sg, qc->n_elem, qc->dma_dir, txd 714 drivers/ata/pata_ep93xx.c if (!txd) { txd 718 drivers/ata/pata_ep93xx.c txd->callback = NULL; txd 719 drivers/ata/pata_ep93xx.c txd->callback_param = NULL; txd 721 drivers/ata/pata_ep93xx.c if (dmaengine_submit(txd) < 0) { txd 532 drivers/dma/amba-pl08x.c struct pl08x_txd *txd = to_pl08x_txd(&vd->tx); txd 535 drivers/dma/amba-pl08x.c list_del(&txd->vd.node); txd 537 drivers/dma/amba-pl08x.c plchan->at = txd; txd 543 drivers/dma/amba-pl08x.c pl08x_write_lli(pl08x, phychan, &txd->llis_va[0], txd->ccfg); txd 759 drivers/dma/amba-pl08x.c struct pl08x_txd *txd; txd 765 drivers/dma/amba-pl08x.c txd = plchan->at; txd 767 drivers/dma/amba-pl08x.c if (!ch || !txd) txd 782 drivers/dma/amba-pl08x.c llis_va = txd->llis_va; txd 783 drivers/dma/amba-pl08x.c llis_bus = txd->llis_bus; txd 1107 drivers/dma/amba-pl08x.c struct pl08x_txd *txd; txd 1173 drivers/dma/amba-pl08x.c u32 *llis_va = bd->txd->llis_va + offset; txd 1174 drivers/dma/amba-pl08x.c dma_addr_t llis_bus = bd->txd->llis_bus; txd 1258 drivers/dma/amba-pl08x.c struct pl08x_txd *txd) txd 1268 drivers/dma/amba-pl08x.c txd->llis_va = dma_pool_alloc(pl08x->pool, GFP_NOWAIT, &txd->llis_bus); txd 1269 drivers/dma/amba-pl08x.c if (!txd->llis_va) { txd 1274 drivers/dma/amba-pl08x.c bd.txd = txd; txd 1276 drivers/dma/amba-pl08x.c cctl = txd->cctl; txd 1284 drivers/dma/amba-pl08x.c list_for_each_entry(dsg, &txd->dsg_list, node) { txd 1286 drivers/dma/amba-pl08x.c cctl = txd->cctl; txd 1337 drivers/dma/amba-pl08x.c fc = (txd->ccfg & PL080_CONFIG_FLOW_CONTROL_MASK) >> txd 1471 drivers/dma/amba-pl08x.c llis_va = txd->llis_va; txd 1474 drivers/dma/amba-pl08x.c if (txd->cyclic) { txd 1476 drivers/dma/amba-pl08x.c last_lli[PL080_LLI_LLI] = txd->llis_bus | bd.lli_bus; txd 1493 drivers/dma/amba-pl08x.c struct pl08x_txd *txd) txd 1497 drivers/dma/amba-pl08x.c if (txd->llis_va) txd 1498 drivers/dma/amba-pl08x.c dma_pool_free(pl08x->pool, txd->llis_va, txd->llis_bus); txd 1500 drivers/dma/amba-pl08x.c list_for_each_entry_safe(dsg, _dsg, &txd->dsg_list, node) { txd 1505 drivers/dma/amba-pl08x.c kfree(txd); txd 1510 drivers/dma/amba-pl08x.c struct pl08x_txd *txd = to_pl08x_txd(&vd->tx); txd 1514 drivers/dma/amba-pl08x.c if (!txd->done) txd 1517 drivers/dma/amba-pl08x.c pl08x_free_txd(plchan->host, txd); txd 1580 drivers/dma/amba-pl08x.c struct pl08x_txd *txd = to_pl08x_txd(&vd->tx); txd 1583 drivers/dma/amba-pl08x.c list_for_each_entry(dsg, &txd->dsg_list, node) txd 1754 drivers/dma/amba-pl08x.c struct pl08x_txd *txd = kzalloc(sizeof(*txd), GFP_NOWAIT); txd 1756 drivers/dma/amba-pl08x.c if (txd) txd 1757 drivers/dma/amba-pl08x.c INIT_LIST_HEAD(&txd->dsg_list); txd 1758 drivers/dma/amba-pl08x.c return txd; txd 1895 drivers/dma/amba-pl08x.c struct pl08x_txd *txd; txd 1899 drivers/dma/amba-pl08x.c txd = pl08x_get_txd(plchan); txd 1900 drivers/dma/amba-pl08x.c if (!txd) { txd 1908 drivers/dma/amba-pl08x.c pl08x_free_txd(pl08x, txd); txd 1911 drivers/dma/amba-pl08x.c list_add_tail(&dsg->node, &txd->dsg_list); txd 1918 drivers/dma/amba-pl08x.c txd->ccfg = 0; txd 1919 drivers/dma/amba-pl08x.c txd->cctl = pl08x_ftdmac020_memcpy_cctl(pl08x); txd 1921 drivers/dma/amba-pl08x.c txd->ccfg = PL080_CONFIG_ERR_IRQ_MASK | txd 1924 drivers/dma/amba-pl08x.c txd->cctl = pl08x_memcpy_cctl(pl08x); txd 1927 drivers/dma/amba-pl08x.c ret = pl08x_fill_llis_for_desc(plchan->host, txd); txd 1929 drivers/dma/amba-pl08x.c pl08x_free_txd(pl08x, txd); txd 1933 drivers/dma/amba-pl08x.c return vchan_tx_prep(&plchan->vc, &txd->vd, flags); txd 1943 drivers/dma/amba-pl08x.c struct pl08x_txd *txd; txd 1949 drivers/dma/amba-pl08x.c txd = pl08x_get_txd(plchan); txd 1950 drivers/dma/amba-pl08x.c if (!txd) { txd 1975 drivers/dma/amba-pl08x.c pl08x_free_txd(pl08x, txd); txd 1983 drivers/dma/amba-pl08x.c pl08x_free_txd(pl08x, txd); txd 1989 drivers/dma/amba-pl08x.c txd->cctl = cctl | pl08x_select_bus(false, src_buses, dst_buses); txd 1998 drivers/dma/amba-pl08x.c txd->ccfg = PL080_CONFIG_ERR_IRQ_MASK | txd 2004 drivers/dma/amba-pl08x.c pl08x_free_txd(pl08x, txd); txd 2016 drivers/dma/amba-pl08x.c txd->ccfg |= plchan->signal << PL080_CONFIG_DST_SEL_SHIFT; txd 2018 drivers/dma/amba-pl08x.c txd->ccfg |= plchan->signal << PL080_CONFIG_SRC_SEL_SHIFT; txd 2020 drivers/dma/amba-pl08x.c return txd; txd 2023 drivers/dma/amba-pl08x.c static int pl08x_tx_add_sg(struct pl08x_txd *txd, txd 2035 drivers/dma/amba-pl08x.c list_add_tail(&dsg->node, &txd->dsg_list); txd 2056 drivers/dma/amba-pl08x.c struct pl08x_txd *txd; txd 2064 drivers/dma/amba-pl08x.c txd = pl08x_init_txd(chan, direction, &slave_addr); txd 2065 drivers/dma/amba-pl08x.c if (!txd) txd 2069 drivers/dma/amba-pl08x.c ret = pl08x_tx_add_sg(txd, direction, slave_addr, txd 2074 drivers/dma/amba-pl08x.c pl08x_free_txd(pl08x, txd); txd 2081 drivers/dma/amba-pl08x.c ret = pl08x_fill_llis_for_desc(plchan->host, txd); txd 2084 drivers/dma/amba-pl08x.c pl08x_free_txd(pl08x, txd); txd 2088 drivers/dma/amba-pl08x.c return vchan_tx_prep(&plchan->vc, &txd->vd, flags); txd 2098 drivers/dma/amba-pl08x.c struct pl08x_txd *txd; txd 2108 drivers/dma/amba-pl08x.c txd = pl08x_init_txd(chan, direction, &slave_addr); txd 2109 drivers/dma/amba-pl08x.c if (!txd) txd 2112 drivers/dma/amba-pl08x.c txd->cyclic = true; txd 2113 drivers/dma/amba-pl08x.c txd->cctl |= PL080_CONTROL_TC_IRQ_EN; txd 2115 drivers/dma/amba-pl08x.c ret = pl08x_tx_add_sg(txd, direction, slave_addr, txd 2119 drivers/dma/amba-pl08x.c pl08x_free_txd(pl08x, txd); txd 2124 drivers/dma/amba-pl08x.c ret = pl08x_fill_llis_for_desc(plchan->host, txd); txd 2127 drivers/dma/amba-pl08x.c pl08x_free_txd(pl08x, txd); txd 2131 drivers/dma/amba-pl08x.c return vchan_tx_prep(&plchan->vc, &txd->vd, flags); txd 112 drivers/dma/at_hdmac.c dma_async_tx_descriptor_init(&desc->txd, chan); txd 114 drivers/dma/at_hdmac.c desc->txd.flags = DMA_CTRL_ACK; txd 115 drivers/dma/at_hdmac.c desc->txd.tx_submit = atc_tx_submit; txd 116 drivers/dma/at_hdmac.c desc->txd.phys = phys; txd 136 drivers/dma/at_hdmac.c if (async_tx_test_ack(&desc->txd)) { txd 203 drivers/dma/at_hdmac.c (*prev)->lli.dscr = desc->txd.phys; txd 244 drivers/dma/at_hdmac.c channel_writel(atchan, DSCR, first->txd.phys); txd 265 drivers/dma/at_hdmac.c if (desc->txd.cookie == cookie) txd 270 drivers/dma/at_hdmac.c if (desc->txd.cookie == cookie) txd 443 drivers/dma/at_hdmac.c struct dma_async_tx_descriptor *txd = &desc->txd; txd 447 drivers/dma/at_hdmac.c "descriptor %u complete\n", txd->cookie); txd 451 drivers/dma/at_hdmac.c dma_cookie_complete(txd); txd 465 drivers/dma/at_hdmac.c dma_descriptor_unmap(txd); txd 473 drivers/dma/at_hdmac.c dmaengine_desc_get_callback_invoke(txd, NULL); txd 476 drivers/dma/at_hdmac.c dma_run_dependencies(txd); txd 571 drivers/dma/at_hdmac.c " cookie: %d\n", bad_desc->txd.cookie); txd 589 drivers/dma/at_hdmac.c struct dma_async_tx_descriptor *txd = &first->txd; txd 595 drivers/dma/at_hdmac.c dmaengine_desc_get_callback_invoke(txd, NULL); txd 679 drivers/dma/at_hdmac.c desc->txd.cookie); txd 684 drivers/dma/at_hdmac.c desc->txd.cookie); txd 781 drivers/dma/at_hdmac.c desc->txd.cookie = -EBUSY; txd 787 drivers/dma/at_hdmac.c desc->txd.flags = flags; /* client is in control of this ack */ txd 789 drivers/dma/at_hdmac.c return &desc->txd; txd 850 drivers/dma/at_hdmac.c desc->txd.cookie = 0; txd 857 drivers/dma/at_hdmac.c first->txd.cookie = -EBUSY; txd 863 drivers/dma/at_hdmac.c first->txd.flags = flags; /* client is in control of this ack */ txd 865 drivers/dma/at_hdmac.c return &first->txd; txd 906 drivers/dma/at_hdmac.c desc->txd.cookie = 0; txd 962 drivers/dma/at_hdmac.c desc->txd.cookie = -EBUSY; txd 968 drivers/dma/at_hdmac.c desc->txd.flags = flags; txd 970 drivers/dma/at_hdmac.c return &desc->txd; txd 1039 drivers/dma/at_hdmac.c first->txd.cookie = -EBUSY; txd 1045 drivers/dma/at_hdmac.c first->txd.flags = flags; txd 1047 drivers/dma/at_hdmac.c return &first->txd; txd 1186 drivers/dma/at_hdmac.c first->txd.cookie = -EBUSY; txd 1190 drivers/dma/at_hdmac.c first->txd.flags = flags; /* client is in control of this ack */ txd 1192 drivers/dma/at_hdmac.c return &first->txd; txd 1342 drivers/dma/at_hdmac.c prev->lli.dscr = first->txd.phys; txd 1345 drivers/dma/at_hdmac.c first->txd.cookie = -EBUSY; txd 1348 drivers/dma/at_hdmac.c return &first->txd; txd 1628 drivers/dma/at_hdmac.c dma_pool_free(atdma->dma_desc_pool, desc, desc->txd.phys); txd 190 drivers/dma/at_hdmac_regs.h struct dma_async_tx_descriptor txd; txd 207 drivers/dma/at_hdmac_regs.h txd_to_at_desc(struct dma_async_tx_descriptor *txd) txd 209 drivers/dma/at_hdmac_regs.h return container_of(txd, struct at_desc, txd); txd 274 drivers/dma/at_xdmac.c static inline struct at_xdmac_desc *txd_to_at_desc(struct dma_async_tx_descriptor *txd) txd 276 drivers/dma/at_xdmac.c return container_of(txd, struct at_xdmac_desc, tx_dma_desc); txd 1569 drivers/dma/at_xdmac.c struct dma_async_tx_descriptor *txd; txd 1574 drivers/dma/at_xdmac.c txd = &desc->tx_dma_desc; txd 1576 drivers/dma/at_xdmac.c if (txd->flags & DMA_PREP_INTERRUPT) txd 1577 drivers/dma/at_xdmac.c dmaengine_desc_get_callback_invoke(txd, NULL); txd 1638 drivers/dma/at_xdmac.c struct dma_async_tx_descriptor *txd; txd 1654 drivers/dma/at_xdmac.c txd = &desc->tx_dma_desc; txd 1660 drivers/dma/at_xdmac.c dma_cookie_complete(txd); txd 1661 drivers/dma/at_xdmac.c if (txd->flags & DMA_PREP_INTERRUPT) txd 1662 drivers/dma/at_xdmac.c dmaengine_desc_get_callback_invoke(txd, NULL); txd 1665 drivers/dma/at_xdmac.c dma_run_dependencies(txd); txd 75 drivers/dma/dw/core.c __func__, desc->txd.cookie); txd 92 drivers/dma/dw/core.c dma_async_tx_descriptor_init(&desc->txd, &dwc->chan); txd 93 drivers/dma/dw/core.c desc->txd.tx_submit = dwc_tx_submit; txd 94 drivers/dma/dw/core.c desc->txd.flags = DMA_CTRL_ACK; txd 95 drivers/dma/dw/core.c desc->txd.phys = phys; txd 109 drivers/dma/dw/core.c dma_pool_free(dw->desc_pool, child, child->txd.phys); txd 113 drivers/dma/dw/core.c dma_pool_free(dw->desc_pool, desc, desc->txd.phys); txd 218 drivers/dma/dw/core.c channel_writel(dwc, LLP, first->txd.phys | lms); txd 233 drivers/dma/dw/core.c dev_vdbg(chan2dev(&dwc->chan), "%s: started %u\n", __func__, desc->txd.cookie); txd 243 drivers/dma/dw/core.c struct dma_async_tx_descriptor *txd = &desc->txd; txd 248 drivers/dma/dw/core.c dev_vdbg(chan2dev(&dwc->chan), "descriptor %u complete\n", txd->cookie); txd 251 drivers/dma/dw/core.c dma_cookie_complete(txd); txd 253 drivers/dma/dw/core.c dmaengine_desc_get_callback(txd, &cb); txd 259 drivers/dma/dw/core.c async_tx_ack(&child->txd); txd 260 drivers/dma/dw/core.c async_tx_ack(&desc->txd); txd 375 drivers/dma/dw/core.c if (desc->txd.phys == DWC_LLP_LOC(llp)) { txd 460 drivers/dma/dw/core.c " cookie: %d\n", bad_desc->txd.cookie); txd 599 drivers/dma/dw/core.c lli_write(prev, llp, desc->txd.phys | lms); txd 611 drivers/dma/dw/core.c first->txd.flags = flags; txd 614 drivers/dma/dw/core.c return &first->txd; txd 689 drivers/dma/dw/core.c lli_write(prev, llp, desc->txd.phys | lms); txd 738 drivers/dma/dw/core.c lli_write(prev, llp, desc->txd.phys | lms); txd 763 drivers/dma/dw/core.c return &first->txd; txd 885 drivers/dma/dw/core.c if (desc->txd.cookie == c) txd 394 drivers/dma/dw/regs.h struct dma_async_tx_descriptor txd; txd 403 drivers/dma/dw/regs.h txd_to_dw_desc(struct dma_async_tx_descriptor *txd) txd 405 drivers/dma/dw/regs.h return container_of(txd, struct dw_desc, txd); txd 127 drivers/dma/ep93xx_dma.c struct dma_async_tx_descriptor txd; txd 256 drivers/dma/ep93xx_dma.c d->txd.callback = desc->txd.callback; txd 257 drivers/dma/ep93xx_dma.c d->txd.callback_param = desc->txd.callback_param; txd 300 drivers/dma/ep93xx_dma.c return !desc->txd.cookie; txd 428 drivers/dma/ep93xx_dma.c desc->txd.cookie, desc->src_addr, desc->dst_addr, txd 619 drivers/dma/ep93xx_dma.c last_done = !desc || desc->txd.cookie; txd 683 drivers/dma/ep93xx_dma.c if (async_tx_test_ack(&desc->txd)) { txd 691 drivers/dma/ep93xx_dma.c desc->txd.cookie = 0; txd 692 drivers/dma/ep93xx_dma.c desc->txd.callback = NULL; txd 693 drivers/dma/ep93xx_dma.c desc->txd.callback_param = NULL; txd 765 drivers/dma/ep93xx_dma.c dma_cookie_complete(&desc->txd); txd 768 drivers/dma/ep93xx_dma.c dmaengine_desc_get_callback(&desc->txd, &cb); txd 777 drivers/dma/ep93xx_dma.c dma_descriptor_unmap(&desc->txd); txd 839 drivers/dma/ep93xx_dma.c desc = container_of(tx, struct ep93xx_dma_desc, txd); txd 925 drivers/dma/ep93xx_dma.c dma_async_tx_descriptor_init(&desc->txd, chan); txd 926 drivers/dma/ep93xx_dma.c desc->txd.flags = DMA_CTRL_ACK; txd 927 drivers/dma/ep93xx_dma.c desc->txd.tx_submit = ep93xx_dma_tx_submit; txd 1012 drivers/dma/ep93xx_dma.c first->txd.cookie = -EBUSY; txd 1013 drivers/dma/ep93xx_dma.c first->txd.flags = flags; txd 1015 drivers/dma/ep93xx_dma.c return &first->txd; txd 1087 drivers/dma/ep93xx_dma.c first->txd.cookie = -EBUSY; txd 1088 drivers/dma/ep93xx_dma.c first->txd.flags = flags; txd 1090 drivers/dma/ep93xx_dma.c return &first->txd; txd 1168 drivers/dma/ep93xx_dma.c first->txd.cookie = -EBUSY; txd 1170 drivers/dma/ep93xx_dma.c return &first->txd; txd 509 drivers/dma/fsldma.c struct dma_async_tx_descriptor *txd = &desc->async_tx; txd 512 drivers/dma/fsldma.c BUG_ON(txd->cookie < 0); txd 514 drivers/dma/fsldma.c if (txd->cookie > 0) { txd 515 drivers/dma/fsldma.c ret = txd->cookie; txd 517 drivers/dma/fsldma.c dma_descriptor_unmap(txd); txd 519 drivers/dma/fsldma.c dmaengine_desc_get_callback_invoke(txd, NULL); txd 523 drivers/dma/fsldma.c dma_run_dependencies(txd); txd 220 drivers/dma/ioat/dma.c async_tx_ack(&desc->txd); txd 221 drivers/dma/ioat/dma.c ioat_set_chainaddr(ioat_chan, desc->txd.phys); txd 253 drivers/dma/ioat/dma.c ioat_set_chainaddr(ioat_chan, desc->txd.phys); txd 347 drivers/dma/ioat/dma.c dma_async_tx_descriptor_init(&desc->txd, chan); txd 348 drivers/dma/ioat/dma.c desc->txd.tx_submit = ioat_tx_submit_unlock; txd 350 drivers/dma/ioat/dma.c desc->txd.phys = phys; txd 426 drivers/dma/ioat/dma.c hw->next = next->txd.phys; txd 428 drivers/dma/ioat/dma.c ring[i]->hw->next = ring[0]->txd.phys; txd 609 drivers/dma/ioat/dma.c tx = &desc->txd; txd 727 drivers/dma/ioat/dma.c tx = &desc->txd; txd 756 drivers/dma/ioat/dma.c ioat_chan->last_completion = *ioat_chan->completion = desc->txd.phys; txd 832 drivers/dma/ioat/dma.c tx = &desc->txd; txd 842 drivers/dma/ioat/dma.c *ioat_chan->completion = desc->txd.phys; txd 184 drivers/dma/ioat/dma.h struct dma_async_tx_descriptor txd; txd 230 drivers/dma/ioat/dma.h ({ if (d) __dump_desc_dbg(c, d->hw, &d->txd, desc_id(d)); 0; }) txd 139 drivers/dma/ioat/prep.c desc->txd.flags = flags; txd 147 drivers/dma/ioat/prep.c return &desc->txd; txd 223 drivers/dma/ioat/prep.c desc->txd.flags = flags; txd 231 drivers/dma/ioat/prep.c compl_desc->txd.flags = flags & DMA_PREP_INTERRUPT; txd 241 drivers/dma/ioat/prep.c return &compl_desc->txd; txd 289 drivers/dma/ioat/prep.c desc_id(desc), (unsigned long long) desc->txd.phys, txd 291 drivers/dma/ioat/prep.c desc->txd.flags, pq->size, pq->ctl, pq->ctl_f.op, txd 322 drivers/dma/ioat/prep.c desc_id(desc), (unsigned long long) desc->txd.phys, txd 324 drivers/dma/ioat/prep.c desc->txd.flags, pq->size, pq->ctl, txd 433 drivers/dma/ioat/prep.c desc->txd.flags = flags; txd 447 drivers/dma/ioat/prep.c compl_desc->txd.flags = flags & DMA_PREP_INTERRUPT; txd 459 drivers/dma/ioat/prep.c return &compl_desc->txd; txd 547 drivers/dma/ioat/prep.c desc->txd.flags = flags; txd 560 drivers/dma/ioat/prep.c return &desc->txd; txd 729 drivers/dma/ioat/prep.c desc->txd.flags = flags; txd 735 drivers/dma/ioat/prep.c return &desc->txd; txd 779 drivers/dma/ipu/ipu_idmac.c if (async_tx_test_ack(&desc->txd)) txd 926 drivers/dma/ipu/ipu_idmac.c struct dma_async_tx_descriptor *txd = &desc->txd; txd 928 drivers/dma/ipu/ipu_idmac.c memset(txd, 0, sizeof(*txd)); txd 929 drivers/dma/ipu/ipu_idmac.c dma_async_tx_descriptor_init(txd, &ichan->dma_chan); txd 930 drivers/dma/ipu/ipu_idmac.c txd->tx_submit = idmac_tx_submit; txd 1278 drivers/dma/ipu/ipu_idmac.c dmaengine_desc_get_callback(&descnew->txd, &cb); txd 1290 drivers/dma/ipu/ipu_idmac.c dma_cookie_complete(&desc->txd); txd 1292 drivers/dma/ipu/ipu_idmac.c dmaengine_desc_get_callback(&desc->txd, &cb); txd 1296 drivers/dma/ipu/ipu_idmac.c if (done && (desc->txd.flags & DMA_PREP_INTERRUPT)) txd 1317 drivers/dma/ipu/ipu_idmac.c if (async_tx_test_ack(&desc->txd)) { txd 1325 drivers/dma/ipu/ipu_idmac.c async_tx_clear_ack(&desc->txd); txd 1340 drivers/dma/ipu/ipu_idmac.c struct dma_async_tx_descriptor *txd = NULL; txd 1364 drivers/dma/ipu/ipu_idmac.c txd = &desc->txd; txd 1365 drivers/dma/ipu/ipu_idmac.c txd->flags = tx_flags; txd 1373 drivers/dma/ipu/ipu_idmac.c return txd; txd 1450 drivers/dma/ipu/ipu_idmac.c async_tx_clear_ack(&desc->txd); txd 166 drivers/dma/mediatek/mtk-hsdma.c struct mtk_hsdma_pdesc *txd; txd 327 drivers/dma/mediatek/mtk-hsdma.c pc->sz_ring = 2 * MTK_DMA_SIZE * sizeof(*ring->txd); txd 328 drivers/dma/mediatek/mtk-hsdma.c ring->txd = dma_alloc_coherent(hsdma2dev(hsdma), pc->sz_ring, txd 330 drivers/dma/mediatek/mtk-hsdma.c if (!ring->txd) txd 333 drivers/dma/mediatek/mtk-hsdma.c ring->rxd = &ring->txd[MTK_DMA_SIZE]; txd 334 drivers/dma/mediatek/mtk-hsdma.c ring->rphys = ring->tphys + MTK_DMA_SIZE * sizeof(*ring->txd); txd 384 drivers/dma/mediatek/mtk-hsdma.c pc->sz_ring, ring->txd, ring->tphys); txd 409 drivers/dma/mediatek/mtk-hsdma.c pc->sz_ring, ring->txd, ring->tphys); txd 417 drivers/dma/mediatek/mtk-hsdma.c struct mtk_hsdma_pdesc *txd, *rxd; txd 450 drivers/dma/mediatek/mtk-hsdma.c txd = &ring->txd[ring->cur_tptr]; txd 451 drivers/dma/mediatek/mtk-hsdma.c WRITE_ONCE(txd->desc1, hvd->src); txd 452 drivers/dma/mediatek/mtk-hsdma.c WRITE_ONCE(txd->desc2, txd 930 drivers/dma/mmp_pdma.c struct dma_async_tx_descriptor *txd = &desc->async_tx; txd 935 drivers/dma/mmp_pdma.c dmaengine_desc_get_callback(txd, &cb); txd 938 drivers/dma/mmp_pdma.c dma_pool_free(chan->desc_pool, desc, txd->phys); txd 460 drivers/dma/mpc512x_dma.c static dma_cookie_t mpc_dma_tx_submit(struct dma_async_tx_descriptor *txd) txd 462 drivers/dma/mpc512x_dma.c struct mpc_dma_chan *mchan = dma_chan_to_mpc_dma_chan(txd->chan); txd 467 drivers/dma/mpc512x_dma.c mdesc = container_of(txd, struct mpc_dma_desc, desc); txd 479 drivers/dma/mpc512x_dma.c cookie = dma_cookie_assign(txd); txd 196 drivers/dma/owl-dma.c struct owl_dma_txd *txd; txd 343 drivers/dma/owl-dma.c static struct owl_dma_lli *owl_dma_add_lli(struct owl_dma_txd *txd, txd 349 drivers/dma/owl-dma.c list_add_tail(&next->node, &txd->lli_list); txd 500 drivers/dma/owl-dma.c struct owl_dma_txd *txd = to_owl_txd(&vd->tx); txd 507 drivers/dma/owl-dma.c vchan->txd = txd; txd 513 drivers/dma/owl-dma.c lli = list_first_entry(&txd->lli_list, txd 516 drivers/dma/owl-dma.c if (txd->cyclic) txd 601 drivers/dma/owl-dma.c struct owl_dma_txd *txd; txd 614 drivers/dma/owl-dma.c txd = vchan->txd; txd 615 drivers/dma/owl-dma.c if (txd) { txd 616 drivers/dma/owl-dma.c vchan->txd = NULL; txd 618 drivers/dma/owl-dma.c vchan_cookie_complete(&txd->vd); txd 636 drivers/dma/owl-dma.c static void owl_dma_free_txd(struct owl_dma *od, struct owl_dma_txd *txd) txd 640 drivers/dma/owl-dma.c if (unlikely(!txd)) txd 643 drivers/dma/owl-dma.c list_for_each_entry_safe(lli, _lli, &txd->lli_list, node) txd 646 drivers/dma/owl-dma.c kfree(txd); txd 652 drivers/dma/owl-dma.c struct owl_dma_txd *txd = to_owl_txd(&vd->tx); txd 654 drivers/dma/owl-dma.c owl_dma_free_txd(od, txd); txd 669 drivers/dma/owl-dma.c if (vchan->txd) { txd 670 drivers/dma/owl-dma.c owl_dma_desc_free(&vchan->txd->vd); txd 671 drivers/dma/owl-dma.c vchan->txd = NULL; txd 716 drivers/dma/owl-dma.c if (!vchan->pchan && !vchan->txd) txd 733 drivers/dma/owl-dma.c struct owl_dma_txd *txd; txd 739 drivers/dma/owl-dma.c txd = vchan->txd; txd 741 drivers/dma/owl-dma.c if (!pchan || !txd) txd 750 drivers/dma/owl-dma.c list_for_each_entry(lli, &txd->lli_list, node) { txd 753 drivers/dma/owl-dma.c list_for_each_entry(lli, &txd->lli_list, node) txd 770 drivers/dma/owl-dma.c struct owl_dma_txd *txd; txd 783 drivers/dma/owl-dma.c txd = to_owl_txd(&vd->tx); txd 784 drivers/dma/owl-dma.c list_for_each_entry(lli, &txd->lli_list, node) txd 832 drivers/dma/owl-dma.c struct owl_dma_txd *txd; txd 840 drivers/dma/owl-dma.c txd = kzalloc(sizeof(*txd), GFP_NOWAIT); txd 841 drivers/dma/owl-dma.c if (!txd) txd 844 drivers/dma/owl-dma.c INIT_LIST_HEAD(&txd->lli_list); txd 858 drivers/dma/owl-dma.c &vchan->cfg, txd->cyclic); txd 864 drivers/dma/owl-dma.c prev = owl_dma_add_lli(txd, prev, lli, false); txd 867 drivers/dma/owl-dma.c return vchan_tx_prep(&vchan->vc, &txd->vd, flags); txd 870 drivers/dma/owl-dma.c owl_dma_free_txd(od, txd); txd 884 drivers/dma/owl-dma.c struct owl_dma_txd *txd; txd 891 drivers/dma/owl-dma.c txd = kzalloc(sizeof(*txd), GFP_NOWAIT); txd 892 drivers/dma/owl-dma.c if (!txd) txd 895 drivers/dma/owl-dma.c INIT_LIST_HEAD(&txd->lli_list); txd 922 drivers/dma/owl-dma.c txd->cyclic); txd 928 drivers/dma/owl-dma.c prev = owl_dma_add_lli(txd, prev, lli, false); txd 931 drivers/dma/owl-dma.c return vchan_tx_prep(&vchan->vc, &txd->vd, flags); txd 934 drivers/dma/owl-dma.c owl_dma_free_txd(od, txd); txd 949 drivers/dma/owl-dma.c struct owl_dma_txd *txd; txd 955 drivers/dma/owl-dma.c txd = kzalloc(sizeof(*txd), GFP_NOWAIT); txd 956 drivers/dma/owl-dma.c if (!txd) txd 959 drivers/dma/owl-dma.c INIT_LIST_HEAD(&txd->lli_list); txd 960 drivers/dma/owl-dma.c txd->cyclic = true; txd 978 drivers/dma/owl-dma.c dir, sconfig, txd->cyclic); txd 987 drivers/dma/owl-dma.c prev = owl_dma_add_lli(txd, prev, lli, false); txd 991 drivers/dma/owl-dma.c owl_dma_add_lli(txd, prev, first, true); txd 993 drivers/dma/owl-dma.c return vchan_tx_prep(&vchan->vc, &txd->vd, flags); txd 996 drivers/dma/owl-dma.c owl_dma_free_txd(od, txd); txd 85 drivers/dma/pch_dma.c struct dma_async_tx_descriptor txd; txd 138 drivers/dma/pch_dma.c struct pch_dma_desc *to_pd_desc(struct dma_async_tx_descriptor *txd) txd 140 drivers/dma/pch_dma.c return container_of(txd, struct pch_dma_desc, txd); txd 343 drivers/dma/pch_dma.c channel_writel(pd_chan, NEXT, desc->txd.phys); txd 351 drivers/dma/pch_dma.c struct dma_async_tx_descriptor *txd = &desc->txd; txd 354 drivers/dma/pch_dma.c dmaengine_desc_get_callback(txd, &cb); txd 392 drivers/dma/pch_dma.c bad_desc->txd.cookie); txd 408 drivers/dma/pch_dma.c static dma_cookie_t pd_tx_submit(struct dma_async_tx_descriptor *txd) txd 410 drivers/dma/pch_dma.c struct pch_dma_desc *desc = to_pd_desc(txd); txd 411 drivers/dma/pch_dma.c struct pch_dma_chan *pd_chan = to_pd_chan(txd->chan); txd 435 drivers/dma/pch_dma.c dma_async_tx_descriptor_init(&desc->txd, chan); txd 436 drivers/dma/pch_dma.c desc->txd.tx_submit = pd_tx_submit; txd 437 drivers/dma/pch_dma.c desc->txd.flags = DMA_CTRL_ACK; txd 438 drivers/dma/pch_dma.c desc->txd.phys = addr; txd 453 drivers/dma/pch_dma.c if (async_tx_test_ack(&desc->txd)) { txd 544 drivers/dma/pch_dma.c dma_pool_free(pd->pool, desc, desc->txd.phys); txd 629 drivers/dma/pch_dma.c prev->regs.next |= desc->txd.phys; txd 641 drivers/dma/pch_dma.c first->txd.cookie = -EBUSY; txd 642 drivers/dma/pch_dma.c desc->txd.flags = flags; txd 644 drivers/dma/pch_dma.c return &first->txd; txd 520 drivers/dma/pl330.c struct dma_async_tx_descriptor txd; txd 2027 drivers/dma/pl330.c return container_of(tx, struct dma_pl330_desc, txd); txd 2051 drivers/dma/pl330.c __func__, __LINE__, desc->txd.cookie); txd 2070 drivers/dma/pl330.c dma_cookie_complete(&desc->txd); txd 2096 drivers/dma/pl330.c dmaengine_desc_get_callback(&desc->txd, &cb); txd 2113 drivers/dma/pl330.c dma_descriptor_unmap(&desc->txd); txd 2294 drivers/dma/pl330.c dma_cookie_complete(&desc->txd); txd 2299 drivers/dma/pl330.c dma_cookie_complete(&desc->txd); txd 2436 drivers/dma/pl330.c if (desc->txd.cookie == cookie) { txd 2502 drivers/dma/pl330.c desc->txd.callback = last->txd.callback; txd 2503 drivers/dma/pl330.c desc->txd.callback_param = last->txd.callback_param; txd 2507 drivers/dma/pl330.c dma_cookie_assign(&desc->txd); txd 2513 drivers/dma/pl330.c cookie = dma_cookie_assign(&last->txd); txd 2525 drivers/dma/pl330.c desc->txd.tx_submit = pl330_tx_submit; txd 2569 drivers/dma/pl330.c desc->txd.callback = NULL; txd 2600 drivers/dma/pl330.c desc->txd.cookie = 0; txd 2601 drivers/dma/pl330.c async_tx_ack(&desc->txd); txd 2606 drivers/dma/pl330.c dma_async_tx_descriptor_init(&desc->txd, &pch->chan); txd 2749 drivers/dma/pl330.c desc->txd.flags = flags; txd 2751 drivers/dma/pl330.c return &desc->txd; txd 2801 drivers/dma/pl330.c desc->txd.flags = flags; txd 2803 drivers/dma/pl330.c return &desc->txd; txd 2886 drivers/dma/pl330.c desc->txd.flags = flg; txd 2887 drivers/dma/pl330.c return &desc->txd; txd 317 drivers/dma/qcom/hidma.c static dma_cookie_t hidma_tx_submit(struct dma_async_tx_descriptor *txd) txd 319 drivers/dma/qcom/hidma.c struct hidma_chan *mchan = to_hidma_chan(txd->chan); txd 334 drivers/dma/qcom/hidma.c mdesc = container_of(txd, struct hidma_desc, desc); txd 341 drivers/dma/qcom/hidma.c cookie = dma_cookie_assign(txd); txd 495 drivers/dma/qcom/hidma.c struct dma_async_tx_descriptor *txd = &mdesc->desc; txd 497 drivers/dma/qcom/hidma.c dma_descriptor_unmap(txd); txd 498 drivers/dma/qcom/hidma.c dmaengine_desc_get_callback_invoke(txd, NULL); txd 499 drivers/dma/qcom/hidma.c dma_run_dependencies(txd); txd 373 drivers/dma/s3c24xx-dma.c struct s3c24xx_txd *txd = s3cchan->at; txd 376 drivers/dma/s3c24xx-dma.c return tc * txd->width; txd 417 drivers/dma/s3c24xx-dma.c struct s3c24xx_txd *txd = kzalloc(sizeof(*txd), GFP_NOWAIT); txd 419 drivers/dma/s3c24xx-dma.c if (txd) { txd 420 drivers/dma/s3c24xx-dma.c INIT_LIST_HEAD(&txd->dsg_list); txd 421 drivers/dma/s3c24xx-dma.c txd->dcon = S3C24XX_DCON_INT | S3C24XX_DCON_NORELOAD; txd 424 drivers/dma/s3c24xx-dma.c return txd; txd 427 drivers/dma/s3c24xx-dma.c static void s3c24xx_dma_free_txd(struct s3c24xx_txd *txd) txd 431 drivers/dma/s3c24xx-dma.c list_for_each_entry_safe(dsg, _dsg, &txd->dsg_list, node) { txd 436 drivers/dma/s3c24xx-dma.c kfree(txd); txd 440 drivers/dma/s3c24xx-dma.c struct s3c24xx_txd *txd) txd 445 drivers/dma/s3c24xx-dma.c struct s3c24xx_sg *dsg = list_entry(txd->at, struct s3c24xx_sg, node); txd 446 drivers/dma/s3c24xx-dma.c u32 dcon = txd->dcon; txd 450 drivers/dma/s3c24xx-dma.c switch (txd->width) { txd 484 drivers/dma/s3c24xx-dma.c writel_relaxed(txd->disrcc, phy->base + S3C24XX_DISRCC); txd 486 drivers/dma/s3c24xx-dma.c writel_relaxed(txd->didstc, phy->base + S3C24XX_DIDSTC); txd 507 drivers/dma/s3c24xx-dma.c struct s3c24xx_txd *txd = to_s3c24xx_txd(&vd->tx); txd 509 drivers/dma/s3c24xx-dma.c list_del(&txd->vd.node); txd 511 drivers/dma/s3c24xx-dma.c s3cchan->at = txd; txd 518 drivers/dma/s3c24xx-dma.c txd->at = txd->dsg_list.next; txd 519 drivers/dma/s3c24xx-dma.c s3c24xx_dma_start_next_sg(s3cchan, txd); txd 637 drivers/dma/s3c24xx-dma.c struct s3c24xx_txd *txd = to_s3c24xx_txd(&vd->tx); txd 643 drivers/dma/s3c24xx-dma.c s3c24xx_dma_free_txd(txd); txd 650 drivers/dma/s3c24xx-dma.c struct s3c24xx_txd *txd; txd 671 drivers/dma/s3c24xx-dma.c txd = s3cchan->at; txd 672 drivers/dma/s3c24xx-dma.c if (txd) { txd 674 drivers/dma/s3c24xx-dma.c if (!list_is_last(txd->at, &txd->dsg_list)) { txd 675 drivers/dma/s3c24xx-dma.c txd->at = txd->at->next; txd 676 drivers/dma/s3c24xx-dma.c if (txd->cyclic) txd 677 drivers/dma/s3c24xx-dma.c vchan_cyclic_callback(&txd->vd); txd 678 drivers/dma/s3c24xx-dma.c s3c24xx_dma_start_next_sg(s3cchan, txd); txd 679 drivers/dma/s3c24xx-dma.c } else if (!txd->cyclic) { txd 681 drivers/dma/s3c24xx-dma.c vchan_cookie_complete(&txd->vd); txd 692 drivers/dma/s3c24xx-dma.c vchan_cyclic_callback(&txd->vd); txd 695 drivers/dma/s3c24xx-dma.c txd->at = txd->dsg_list.next; txd 696 drivers/dma/s3c24xx-dma.c s3c24xx_dma_start_next_sg(s3cchan, txd); txd 761 drivers/dma/s3c24xx-dma.c struct s3c24xx_txd *txd; txd 783 drivers/dma/s3c24xx-dma.c txd = to_s3c24xx_txd(&vd->tx); txd 785 drivers/dma/s3c24xx-dma.c list_for_each_entry(dsg, &txd->dsg_list, node) txd 792 drivers/dma/s3c24xx-dma.c txd = s3cchan->at; txd 794 drivers/dma/s3c24xx-dma.c dsg = list_entry(txd->at, struct s3c24xx_sg, node); txd 795 drivers/dma/s3c24xx-dma.c list_for_each_entry_from(dsg, &txd->dsg_list, node) txd 821 drivers/dma/s3c24xx-dma.c struct s3c24xx_txd *txd; txd 833 drivers/dma/s3c24xx-dma.c txd = s3c24xx_dma_get_txd(); txd 834 drivers/dma/s3c24xx-dma.c if (!txd) txd 839 drivers/dma/s3c24xx-dma.c s3c24xx_dma_free_txd(txd); txd 842 drivers/dma/s3c24xx-dma.c list_add_tail(&dsg->node, &txd->dsg_list); txd 858 drivers/dma/s3c24xx-dma.c txd->width = (src_mod == 0 && dest_mod == 0) ? 4 : 1; txd 861 drivers/dma/s3c24xx-dma.c txd->width = ((src_mod == 2 || src_mod == 0) && txd 865 drivers/dma/s3c24xx-dma.c txd->width = 1; txd 869 drivers/dma/s3c24xx-dma.c txd->disrcc = S3C24XX_DISRCC_LOC_AHB | S3C24XX_DISRCC_INC_INCREMENT; txd 870 drivers/dma/s3c24xx-dma.c txd->didstc = S3C24XX_DIDSTC_LOC_AHB | S3C24XX_DIDSTC_INC_INCREMENT; txd 871 drivers/dma/s3c24xx-dma.c txd->dcon |= S3C24XX_DCON_DEMAND | S3C24XX_DCON_SYNC_HCLK | txd 874 drivers/dma/s3c24xx-dma.c return vchan_tx_prep(&s3cchan->vc, &txd->vd, flags); txd 885 drivers/dma/s3c24xx-dma.c struct s3c24xx_txd *txd; txd 902 drivers/dma/s3c24xx-dma.c txd = s3c24xx_dma_get_txd(); txd 903 drivers/dma/s3c24xx-dma.c if (!txd) txd 906 drivers/dma/s3c24xx-dma.c txd->cyclic = 1; txd 909 drivers/dma/s3c24xx-dma.c txd->dcon |= S3C24XX_DCON_HANDSHAKE; txd 913 drivers/dma/s3c24xx-dma.c txd->dcon |= S3C24XX_DCON_SYNC_PCLK; txd 917 drivers/dma/s3c24xx-dma.c txd->dcon |= S3C24XX_DCON_SYNC_HCLK; txd 932 drivers/dma/s3c24xx-dma.c txd->dcon |= S3C24XX_DCON_SERV_SINGLE; txd 935 drivers/dma/s3c24xx-dma.c txd->disrcc = S3C24XX_DISRCC_LOC_AHB | txd 937 drivers/dma/s3c24xx-dma.c txd->didstc = hwcfg; txd 939 drivers/dma/s3c24xx-dma.c txd->width = s3cchan->cfg.dst_addr_width; txd 941 drivers/dma/s3c24xx-dma.c txd->disrcc = hwcfg; txd 942 drivers/dma/s3c24xx-dma.c txd->didstc = S3C24XX_DIDSTC_LOC_AHB | txd 945 drivers/dma/s3c24xx-dma.c txd->width = s3cchan->cfg.src_addr_width; txd 953 drivers/dma/s3c24xx-dma.c s3c24xx_dma_free_txd(txd); txd 956 drivers/dma/s3c24xx-dma.c list_add_tail(&dsg->node, &txd->dsg_list); txd 971 drivers/dma/s3c24xx-dma.c return vchan_tx_prep(&s3cchan->vc, &txd->vd, flags); txd 983 drivers/dma/s3c24xx-dma.c struct s3c24xx_txd *txd; txd 993 drivers/dma/s3c24xx-dma.c txd = s3c24xx_dma_get_txd(); txd 994 drivers/dma/s3c24xx-dma.c if (!txd) txd 998 drivers/dma/s3c24xx-dma.c txd->dcon |= S3C24XX_DCON_HANDSHAKE; txd 1002 drivers/dma/s3c24xx-dma.c txd->dcon |= S3C24XX_DCON_SYNC_PCLK; txd 1006 drivers/dma/s3c24xx-dma.c txd->dcon |= S3C24XX_DCON_SYNC_HCLK; txd 1021 drivers/dma/s3c24xx-dma.c txd->dcon |= S3C24XX_DCON_SERV_SINGLE; txd 1024 drivers/dma/s3c24xx-dma.c txd->disrcc = S3C24XX_DISRCC_LOC_AHB | txd 1026 drivers/dma/s3c24xx-dma.c txd->didstc = hwcfg; txd 1028 drivers/dma/s3c24xx-dma.c txd->width = s3cchan->cfg.dst_addr_width; txd 1030 drivers/dma/s3c24xx-dma.c txd->disrcc = hwcfg; txd 1031 drivers/dma/s3c24xx-dma.c txd->didstc = S3C24XX_DIDSTC_LOC_AHB | txd 1034 drivers/dma/s3c24xx-dma.c txd->width = s3cchan->cfg.src_addr_width; txd 1036 drivers/dma/s3c24xx-dma.c s3c24xx_dma_free_txd(txd); txd 1045 drivers/dma/s3c24xx-dma.c s3c24xx_dma_free_txd(txd); txd 1048 drivers/dma/s3c24xx-dma.c list_add_tail(&dsg->node, &txd->dsg_list); txd 1060 drivers/dma/s3c24xx-dma.c return vchan_tx_prep(&s3cchan->vc, &txd->vd, flags); txd 148 drivers/dma/sa11x0-dma.c static void sa11x0_dma_start_desc(struct sa11x0_dma_phy *p, struct sa11x0_dma_desc *txd) txd 150 drivers/dma/sa11x0-dma.c list_del(&txd->vd.node); txd 151 drivers/dma/sa11x0-dma.c p->txd_load = txd; txd 155 drivers/dma/sa11x0-dma.c p->num, &txd->vd, txd->vd.tx.cookie, txd->ddar); txd 161 drivers/dma/sa11x0-dma.c struct sa11x0_dma_desc *txd = p->txd_load; txd 167 drivers/dma/sa11x0-dma.c if (!txd) txd 176 drivers/dma/sa11x0-dma.c if (p->sg_load == txd->sglen) { txd 177 drivers/dma/sa11x0-dma.c if (!txd->cyclic) { txd 185 drivers/dma/sa11x0-dma.c if (txn && txn->ddar == txd->ddar) { txd 186 drivers/dma/sa11x0-dma.c txd = txn; txd 198 drivers/dma/sa11x0-dma.c sg = &txd->sg[p->sg_load++]; txd 225 drivers/dma/sa11x0-dma.c struct sa11x0_dma_desc *txd = p->txd_done; txd 227 drivers/dma/sa11x0-dma.c if (++p->sg_done == txd->sglen) { txd 228 drivers/dma/sa11x0-dma.c if (!txd->cyclic) { txd 229 drivers/dma/sa11x0-dma.c vchan_cookie_complete(&txd->vd); txd 237 drivers/dma/sa11x0-dma.c if ((p->sg_done % txd->period) == 0) txd 238 drivers/dma/sa11x0-dma.c vchan_cyclic_callback(&txd->vd); txd 301 drivers/dma/sa11x0-dma.c struct sa11x0_dma_desc *txd = sa11x0_dma_next_desc(c); txd 304 drivers/dma/sa11x0-dma.c if (txd) { txd 307 drivers/dma/sa11x0-dma.c sa11x0_dma_start_desc(p, txd); txd 308 drivers/dma/sa11x0-dma.c p->txd_done = txd; txd 318 drivers/dma/sa11x0-dma.c writel_relaxed(txd->ddar, p->base + DMA_DDAR); txd 447 drivers/dma/sa11x0-dma.c struct sa11x0_dma_desc *txd; txd 451 drivers/dma/sa11x0-dma.c txd = p->txd_done; txd 453 drivers/dma/sa11x0-dma.c txd = p->txd_load; txd 455 drivers/dma/sa11x0-dma.c txd = NULL; txd 458 drivers/dma/sa11x0-dma.c if (txd) { txd 464 drivers/dma/sa11x0-dma.c for (i = 0; i < txd->sglen; i++) { txd 466 drivers/dma/sa11x0-dma.c i, txd->sg[i].addr, txd->sg[i].len); txd 467 drivers/dma/sa11x0-dma.c if (addr >= txd->sg[i].addr && txd 468 drivers/dma/sa11x0-dma.c addr < txd->sg[i].addr + txd->sg[i].len) { txd 471 drivers/dma/sa11x0-dma.c len = txd->sg[i].len - txd 472 drivers/dma/sa11x0-dma.c (addr - txd->sg[i].addr); txd 480 drivers/dma/sa11x0-dma.c for (; i < txd->sglen; i++) { txd 482 drivers/dma/sa11x0-dma.c i, txd->sg[i].addr, txd->sg[i].len); txd 483 drivers/dma/sa11x0-dma.c bytes += txd->sg[i].len; txd 527 drivers/dma/sa11x0-dma.c struct sa11x0_dma_desc *txd; txd 556 drivers/dma/sa11x0-dma.c txd = kzalloc(struct_size(txd, sg, j), GFP_ATOMIC); txd 557 drivers/dma/sa11x0-dma.c if (!txd) { txd 585 drivers/dma/sa11x0-dma.c txd->sg[j].addr = addr; txd 586 drivers/dma/sa11x0-dma.c txd->sg[j].len = tlen; txd 594 drivers/dma/sa11x0-dma.c txd->ddar = c->ddar; txd 595 drivers/dma/sa11x0-dma.c txd->size = size; txd 596 drivers/dma/sa11x0-dma.c txd->sglen = j; txd 599 drivers/dma/sa11x0-dma.c &c->vc, &txd->vd, txd->size, txd->sglen); txd 601 drivers/dma/sa11x0-dma.c return vchan_tx_prep(&c->vc, &txd->vd, flags); txd 609 drivers/dma/sa11x0-dma.c struct sa11x0_dma_desc *txd; txd 626 drivers/dma/sa11x0-dma.c txd = kzalloc(struct_size(txd, sg, sglen), GFP_ATOMIC); txd 627 drivers/dma/sa11x0-dma.c if (!txd) { txd 643 drivers/dma/sa11x0-dma.c txd->sg[k].addr = addr; txd 644 drivers/dma/sa11x0-dma.c txd->sg[k].len = tlen; txd 654 drivers/dma/sa11x0-dma.c txd->ddar = c->ddar; txd 655 drivers/dma/sa11x0-dma.c txd->size = size; txd 656 drivers/dma/sa11x0-dma.c txd->sglen = sglen; txd 657 drivers/dma/sa11x0-dma.c txd->cyclic = 1; txd 658 drivers/dma/sa11x0-dma.c txd->period = sgperiod; txd 660 drivers/dma/sa11x0-dma.c return vchan_tx_prep(&c->vc, &txd->vd, DMA_PREP_INTERRUPT | DMA_CTRL_ACK); txd 1049 drivers/dma/sa11x0-dma.c struct sa11x0_dma_desc *txd = NULL; txd 1055 drivers/dma/sa11x0-dma.c txd = p->txd_done; txd 1057 drivers/dma/sa11x0-dma.c txd = p->txd_load; txd 1059 drivers/dma/sa11x0-dma.c if (!txd) txd 1062 drivers/dma/sa11x0-dma.c writel_relaxed(txd->ddar, p->base + DMA_DDAR); txd 404 drivers/dma/sirf-dma.c static dma_cookie_t sirfsoc_dma_tx_submit(struct dma_async_tx_descriptor *txd) txd 406 drivers/dma/sirf-dma.c struct sirfsoc_dma_chan *schan = dma_chan_to_sirfsoc_dma_chan(txd->chan); txd 411 drivers/dma/sirf-dma.c sdesc = container_of(txd, struct sirfsoc_dma_desc, desc); txd 418 drivers/dma/sirf-dma.c cookie = dma_cookie_assign(txd); txd 373 drivers/dma/ste_dma40.c struct dma_async_tx_descriptor txd; txd 775 drivers/dma/ste_dma40.c if (async_tx_test_ack(&d->txd)) { txd 1470 drivers/dma/ste_dma40.c struct d40_desc *d40d = container_of(tx, struct d40_desc, txd); txd 1594 drivers/dma/ste_dma40.c dma_cookie_complete(&d40d->txd); txd 1606 drivers/dma/ste_dma40.c callback_active = !!(d40d->txd.flags & DMA_PREP_INTERRUPT); txd 1607 drivers/dma/ste_dma40.c dmaengine_desc_get_callback(&d40d->txd, &cb); txd 1610 drivers/dma/ste_dma40.c if (async_tx_test_ack(&d40d->txd)) { txd 2197 drivers/dma/ste_dma40.c desc->txd.flags = dma_flags; txd 2198 drivers/dma/ste_dma40.c desc->txd.tx_submit = d40_tx_submit; txd 2200 drivers/dma/ste_dma40.c dma_async_tx_descriptor_init(&desc->txd, &chan->chan); txd 2264 drivers/dma/ste_dma40.c return &desc->txd; txd 2514 drivers/dma/ste_dma40.c struct dma_async_tx_descriptor *txd; txd 2530 drivers/dma/ste_dma40.c txd = d40_prep_sg(chan, sg, sg, periods, direction, txd 2535 drivers/dma/ste_dma40.c return txd; txd 343 drivers/dma/sun6i-dma.c struct sun6i_desc *txd = pchan->desc; txd 354 drivers/dma/sun6i-dma.c for (lli = txd->v_lli; lli; lli = lli->v_lli_next) { txd 368 drivers/dma/sun6i-dma.c struct sun6i_desc *txd) txd 370 drivers/dma/sun6i-dma.c if ((!prev && !txd) || !next) txd 374 drivers/dma/sun6i-dma.c txd->p_lli = next_phy; txd 375 drivers/dma/sun6i-dma.c txd->v_lli = next; txd 403 drivers/dma/sun6i-dma.c struct sun6i_desc *txd = to_sun6i_desc(&vd->tx); txd 408 drivers/dma/sun6i-dma.c if (unlikely(!txd)) txd 411 drivers/dma/sun6i-dma.c p_lli = txd->p_lli; txd 412 drivers/dma/sun6i-dma.c v_lli = txd->v_lli; txd 424 drivers/dma/sun6i-dma.c kfree(txd); txd 636 drivers/dma/sun6i-dma.c struct sun6i_desc *txd; txd 647 drivers/dma/sun6i-dma.c txd = kzalloc(sizeof(*txd), GFP_NOWAIT); txd 648 drivers/dma/sun6i-dma.c if (!txd) txd 671 drivers/dma/sun6i-dma.c sun6i_dma_lli_add(NULL, v_lli, p_lli, txd); txd 675 drivers/dma/sun6i-dma.c return vchan_tx_prep(&vchan->vc, &txd->vd, flags); txd 678 drivers/dma/sun6i-dma.c kfree(txd); txd 691 drivers/dma/sun6i-dma.c struct sun6i_desc *txd; txd 706 drivers/dma/sun6i-dma.c txd = kzalloc(sizeof(*txd), GFP_NOWAIT); txd 707 drivers/dma/sun6i-dma.c if (!txd) txd 745 drivers/dma/sun6i-dma.c prev = sun6i_dma_lli_add(prev, v_lli, p_lli, txd); txd 748 drivers/dma/sun6i-dma.c dev_dbg(chan2dev(chan), "First: %pad\n", &txd->p_lli); txd 749 drivers/dma/sun6i-dma.c for (prev = txd->v_lli; prev; prev = prev->v_lli_next) txd 752 drivers/dma/sun6i-dma.c return vchan_tx_prep(&vchan->vc, &txd->vd, flags); txd 755 drivers/dma/sun6i-dma.c for (prev = txd->v_lli; prev; prev = prev->v_lli_next) txd 757 drivers/dma/sun6i-dma.c kfree(txd); txd 773 drivers/dma/sun6i-dma.c struct sun6i_desc *txd; txd 785 drivers/dma/sun6i-dma.c txd = kzalloc(sizeof(*txd), GFP_NOWAIT); txd 786 drivers/dma/sun6i-dma.c if (!txd) txd 813 drivers/dma/sun6i-dma.c prev = sun6i_dma_lli_add(prev, v_lli, p_lli, txd); txd 816 drivers/dma/sun6i-dma.c prev->p_lli_next = txd->p_lli; /* cyclic list */ txd 820 drivers/dma/sun6i-dma.c return vchan_tx_prep(&vchan->vc, &txd->vd, flags); txd 823 drivers/dma/sun6i-dma.c for (prev = txd->v_lli; prev; prev = prev->v_lli_next) txd 825 drivers/dma/sun6i-dma.c kfree(txd); txd 935 drivers/dma/sun6i-dma.c struct sun6i_desc *txd; txd 947 drivers/dma/sun6i-dma.c txd = to_sun6i_desc(&vd->tx); txd 950 drivers/dma/sun6i-dma.c for (lli = txd->v_lli; lli != NULL; lli = lli->v_lli_next) txd 164 drivers/dma/tegra20-apb-dma.c struct dma_async_tx_descriptor txd; txd 261 drivers/dma/tegra20-apb-dma.c return container_of(td, struct tegra_dma_desc, txd); txd 284 drivers/dma/tegra20-apb-dma.c if (async_tx_test_ack(&dma_desc->txd) && !dma_desc->cb_count) { txd 287 drivers/dma/tegra20-apb-dma.c dma_desc->txd.flags = 0; txd 299 drivers/dma/tegra20-apb-dma.c dma_async_tx_descriptor_init(&dma_desc->txd, &tdc->dma_chan); txd 300 drivers/dma/tegra20-apb-dma.c dma_desc->txd.tx_submit = tegra_dma_tx_submit; txd 301 drivers/dma/tegra20-apb-dma.c dma_desc->txd.flags = 0; txd 610 drivers/dma/tegra20-apb-dma.c dma_cookie_complete(&dma_desc->txd); txd 669 drivers/dma/tegra20-apb-dma.c dmaengine_desc_get_callback(&dma_desc->txd, &cb); txd 706 drivers/dma/tegra20-apb-dma.c static dma_cookie_t tegra_dma_tx_submit(struct dma_async_tx_descriptor *txd) txd 708 drivers/dma/tegra20-apb-dma.c struct tegra_dma_desc *dma_desc = txd_to_tegra_dma_desc(txd); txd 709 drivers/dma/tegra20-apb-dma.c struct tegra_dma_channel *tdc = to_tegra_dma_chan(txd->chan); txd 715 drivers/dma/tegra20-apb-dma.c cookie = dma_cookie_assign(&dma_desc->txd); txd 879 drivers/dma/tegra20-apb-dma.c if (dma_desc->txd.cookie == cookie) { txd 888 drivers/dma/tegra20-apb-dma.c if (dma_desc->txd.cookie == cookie) { txd 1103 drivers/dma/tegra20-apb-dma.c dma_desc->txd.flags = DMA_CTRL_ACK; txd 1120 drivers/dma/tegra20-apb-dma.c return &dma_desc->txd; txd 1239 drivers/dma/tegra20-apb-dma.c dma_desc->txd.flags = DMA_CTRL_ACK; txd 1256 drivers/dma/tegra20-apb-dma.c return &dma_desc->txd; txd 87 drivers/dma/ti/cppi41.c struct dma_async_tx_descriptor txd; txd 355 drivers/dma/ti/cppi41.c dma_cookie_complete(&c->txd); txd 356 drivers/dma/ti/cppi41.c dmaengine_desc_get_callback_invoke(&c->txd, NULL); txd 387 drivers/dma/ti/cppi41.c dma_async_tx_descriptor_init(&c->txd, chan); txd 388 drivers/dma/ti/cppi41.c c->txd.tx_submit = cppi41_tx_submit; txd 589 drivers/dma/ti/cppi41.c struct dma_async_tx_descriptor *txd = NULL; txd 627 drivers/dma/ti/cppi41.c txd = &c->txd; txd 633 drivers/dma/ti/cppi41.c return txd; txd 730 drivers/dma/ti/cppi41.c dma_cookie_complete(&c->txd); txd 731 drivers/dma/ti/cppi41.c dmaengine_desc_get_callback_invoke(&c->txd, &abort_result); txd 64 drivers/dma/timb_dma.c struct dma_async_tx_descriptor txd; txd 199 drivers/dma/timb_dma.c iowrite32(td_desc->txd.phys, td_chan->membase + txd 209 drivers/dma/timb_dma.c iowrite32(td_desc->txd.phys, td_chan->membase + txd 222 drivers/dma/timb_dma.c struct dma_async_tx_descriptor *txd; txd 231 drivers/dma/timb_dma.c txd = &td_desc->txd; txd 234 drivers/dma/timb_dma.c txd->cookie); txd 243 drivers/dma/timb_dma.c dma_cookie_complete(txd); txd 246 drivers/dma/timb_dma.c dmaengine_desc_get_callback(txd, &cb); txd 250 drivers/dma/timb_dma.c dma_descriptor_unmap(txd); txd 288 drivers/dma/timb_dma.c __func__, td_desc->txd.cookie); txd 294 drivers/dma/timb_dma.c static dma_cookie_t td_tx_submit(struct dma_async_tx_descriptor *txd) txd 296 drivers/dma/timb_dma.c struct timb_dma_desc *td_desc = container_of(txd, struct timb_dma_desc, txd 297 drivers/dma/timb_dma.c txd); txd 298 drivers/dma/timb_dma.c struct timb_dma_chan *td_chan = container_of(txd->chan, txd 303 drivers/dma/timb_dma.c cookie = dma_cookie_assign(txd); txd 306 drivers/dma/timb_dma.c dev_dbg(chan2dev(txd->chan), "%s: started %u\n", __func__, txd 307 drivers/dma/timb_dma.c txd->cookie); txd 311 drivers/dma/timb_dma.c dev_dbg(chan2dev(txd->chan), "tx_submit: queued %u\n", txd 312 drivers/dma/timb_dma.c txd->cookie); txd 338 drivers/dma/timb_dma.c dma_async_tx_descriptor_init(&td_desc->txd, chan); txd 339 drivers/dma/timb_dma.c td_desc->txd.tx_submit = td_tx_submit; txd 340 drivers/dma/timb_dma.c td_desc->txd.flags = DMA_CTRL_ACK; txd 342 drivers/dma/timb_dma.c td_desc->txd.phys = dma_map_single(chan2dmadev(chan), txd 345 drivers/dma/timb_dma.c err = dma_mapping_error(chan2dmadev(chan), td_desc->txd.phys); txd 362 drivers/dma/timb_dma.c dev_dbg(chan2dev(td_desc->txd.chan), "Freeing desc: %p\n", td_desc); txd 363 drivers/dma/timb_dma.c dma_unmap_single(chan2dmadev(td_desc->txd.chan), td_desc->txd.phys, txd 388 drivers/dma/timb_dma.c if (async_tx_test_ack(&td_desc->txd)) { txd 539 drivers/dma/timb_dma.c dma_sync_single_for_device(chan2dmadev(chan), td_desc->txd.phys, txd 542 drivers/dma/timb_dma.c return &td_desc->txd; txd 139 drivers/dma/txx9dmac.c txd_to_txx9dmac_desc(struct dma_async_tx_descriptor *txd) txd 141 drivers/dma/txx9dmac.c return container_of(txd, struct txx9dmac_desc, txd); txd 199 drivers/dma/txx9dmac.c dma_async_tx_descriptor_init(&desc->txd, &dc->chan); txd 200 drivers/dma/txx9dmac.c desc->txd.tx_submit = txx9dmac_tx_submit; txd 202 drivers/dma/txx9dmac.c desc->txd.flags = DMA_CTRL_ACK; txd 203 drivers/dma/txx9dmac.c desc->txd.phys = dma_map_single(chan2parent(&dc->chan), &desc->hwdesc, txd 216 drivers/dma/txx9dmac.c if (async_tx_test_ack(&desc->txd)) { txd 249 drivers/dma/txx9dmac.c child->txd.phys, ddev->descsize, txd 252 drivers/dma/txx9dmac.c desc->txd.phys, ddev->descsize, txd 337 drivers/dma/txx9dmac.c first->txd.cookie, first); txd 367 drivers/dma/txx9dmac.c channel64_write_CHAR(dc, first->txd.phys); txd 388 drivers/dma/txx9dmac.c channel32_writel(dc, CHAR, first->txd.phys); txd 390 drivers/dma/txx9dmac.c channel32_writel(dc, CHAR, first->txd.phys); txd 403 drivers/dma/txx9dmac.c struct dma_async_tx_descriptor *txd = &desc->txd; txd 406 drivers/dma/txx9dmac.c txd->cookie, desc); txd 408 drivers/dma/txx9dmac.c dma_cookie_complete(txd); txd 409 drivers/dma/txx9dmac.c dmaengine_desc_get_callback(txd, &cb); txd 415 drivers/dma/txx9dmac.c dma_descriptor_unmap(txd); txd 421 drivers/dma/txx9dmac.c dma_run_dependencies(txd); txd 434 drivers/dma/txx9dmac.c desc_write_CHAR(dc, prev, desc->txd.phys); txd 436 drivers/dma/txx9dmac.c prev->txd.phys, ddev->descsize, txd 442 drivers/dma/txx9dmac.c if ((desc->txd.flags & DMA_PREP_INTERRUPT) && txd 528 drivers/dma/txx9dmac.c bad_desc->txd.cookie); txd 700 drivers/dma/txx9dmac.c desc->txd.cookie, desc); txd 778 drivers/dma/txx9dmac.c desc_write_CHAR(dc, prev, desc->txd.phys); txd 780 drivers/dma/txx9dmac.c prev->txd.phys, ddev->descsize, txd 793 drivers/dma/txx9dmac.c prev->txd.phys, ddev->descsize, txd 796 drivers/dma/txx9dmac.c first->txd.flags = flags; txd 799 drivers/dma/txx9dmac.c return &first->txd; txd 872 drivers/dma/txx9dmac.c desc_write_CHAR(dc, prev, desc->txd.phys); txd 874 drivers/dma/txx9dmac.c prev->txd.phys, txd 888 drivers/dma/txx9dmac.c prev->txd.phys, ddev->descsize, txd 891 drivers/dma/txx9dmac.c first->txd.flags = flags; txd 894 drivers/dma/txx9dmac.c return &first->txd; txd 949 drivers/dma/txx9dmac.c desc_write_CHAR(dc, prev, desc->txd.phys); txd 951 drivers/dma/txx9dmac.c prev->txd.phys, ddev->descsize, txd 954 drivers/dma/txx9dmac.c channel_read_CHAR(dc) == prev->txd.phys) txd 956 drivers/dma/txx9dmac.c channel_write_CHAR(dc, desc->txd.phys); txd 975 drivers/dma/txx9dmac.c if (!(prev->txd.flags & DMA_PREP_INTERRUPT) || txd 1065 drivers/dma/txx9dmac.c dma_unmap_single(chan2parent(chan), desc->txd.phys, txd 231 drivers/dma/txx9dmac.h struct dma_async_tx_descriptor txd; txd 729 drivers/i2c/busses/i2c-qup.c struct dma_async_tx_descriptor *txd, *rxd = NULL; txd 755 drivers/i2c/busses/i2c-qup.c txd = dmaengine_prep_slave_sg(qup->btx.dma, qup->btx.sg, tx_cnt, txd 758 drivers/i2c/busses/i2c-qup.c if (!txd) { txd 765 drivers/i2c/busses/i2c-qup.c txd->callback = qup_i2c_bam_cb; txd 766 drivers/i2c/busses/i2c-qup.c txd->callback_param = qup; txd 769 drivers/i2c/busses/i2c-qup.c cookie_tx = dmaengine_submit(txd); txd 552 drivers/mailbox/bcm-pdc-mailbox.c struct dma64dd *txd = &pdcs->txd_64[pdcs->txout]; txd 558 drivers/mailbox/bcm-pdc-mailbox.c txd->addrlow = cpu_to_le32(lower_32_bits(dma_addr)); txd 559 drivers/mailbox/bcm-pdc-mailbox.c txd->addrhigh = cpu_to_le32(upper_32_bits(dma_addr)); txd 560 drivers/mailbox/bcm-pdc-mailbox.c txd->ctrl1 = cpu_to_le32(flags); txd 561 drivers/mailbox/bcm-pdc-mailbox.c txd->ctrl2 = cpu_to_le32(buf_len); txd 75 drivers/media/pci/mantis/mantis_i2c.c u32 txd = 0, stat, trials; txd 82 drivers/media/pci/mantis/mantis_i2c.c txd = (msg->addr << 25) | (msg->buf[i] << 8) txd 88 drivers/media/pci/mantis/mantis_i2c.c txd &= ~MANTIS_I2C_STOP; txd 91 drivers/media/pci/mantis/mantis_i2c.c mmwrite(txd, MANTIS_I2CDATA_CTL); txd 119 drivers/media/pci/mantis/mantis_i2c.c u32 stat, data, txd; txd 142 drivers/media/pci/mantis/mantis_i2c.c txd = msgs[i].addr << 25 | (0x1 << 24) txd 146 drivers/media/pci/mantis/mantis_i2c.c mmwrite(txd, MANTIS_I2CDATA_CTL); txd 716 drivers/net/ethernet/3com/typhoon.c struct tx_desc *txd, *first_txd; txd 786 drivers/net/ethernet/3com/typhoon.c txd = (struct tx_desc *) (txRing->ringBase + txRing->lastWrite); txd 795 drivers/net/ethernet/3com/typhoon.c txd->flags = TYPHOON_FRAG_DESC | TYPHOON_DESC_VALID; txd 796 drivers/net/ethernet/3com/typhoon.c txd->len = cpu_to_le16(skb->len); txd 797 drivers/net/ethernet/3com/typhoon.c txd->frag.addr = cpu_to_le32(skb_dma); txd 798 drivers/net/ethernet/3com/typhoon.c txd->frag.addrHi = 0; txd 806 drivers/net/ethernet/3com/typhoon.c txd->flags = TYPHOON_FRAG_DESC | TYPHOON_DESC_VALID; txd 807 drivers/net/ethernet/3com/typhoon.c txd->len = cpu_to_le16(len); txd 808 drivers/net/ethernet/3com/typhoon.c txd->frag.addr = cpu_to_le32(skb_dma); txd 809 drivers/net/ethernet/3com/typhoon.c txd->frag.addrHi = 0; txd 816 drivers/net/ethernet/3com/typhoon.c txd = (struct tx_desc *) (txRing->ringBase + txd 824 drivers/net/ethernet/3com/typhoon.c txd->flags = TYPHOON_FRAG_DESC | TYPHOON_DESC_VALID; txd 825 drivers/net/ethernet/3com/typhoon.c txd->len = cpu_to_le16(len); txd 826 drivers/net/ethernet/3com/typhoon.c txd->frag.addr = cpu_to_le32(skb_dma); txd 827 drivers/net/ethernet/3com/typhoon.c txd->frag.addrHi = 0; txd 436 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c struct hw_atl_txd_s *txd = NULL; txd 446 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c txd = (struct hw_atl_txd_s *)&ring->dx_ring[ring->sw_tail * txd 448 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c txd->ctl = 0; txd 449 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c txd->ctl2 = 0; txd 450 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c txd->buf_addr = 0; txd 455 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c txd->ctl |= (buff->len_l3 << 31) | txd 459 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c txd->ctl2 |= (buff->mss << 16) | txd 469 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c txd->ctl |= HW_ATL_A0_TXD_CTL_CMD_IPV6; txd 473 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c txd->buf_addr = buff->pa; txd 474 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c txd->ctl |= (HW_ATL_A0_TXD_CTL_BLEN & txd 476 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c txd->ctl |= HW_ATL_A0_TXD_CTL_DESC_TYPE_TXD; txd 478 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c txd->ctl2 |= HW_ATL_A0_TXD_CTL2_LEN & (pkt_len << 14); txd 481 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c txd->ctl |= HW_ATL_A0_TXD_CTL_CMD_LSO; txd 482 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c txd->ctl2 |= HW_ATL_A0_TXD_CTL2_CTX_EN; txd 487 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c txd->ctl |= HW_ATL_A0_TXD_CTL_CMD_IPCSO; txd 490 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c txd->ctl |= HW_ATL_A0_TXD_CTL_CMD_TUCSO; txd 493 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c txd->ctl |= HW_ATL_A0_TXD_CTL_EOP; txd 494 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c txd->ctl |= HW_ATL_A0_TXD_CTL_CMD_WB; txd 491 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c struct hw_atl_txd_s *txd = NULL; txd 502 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c txd = (struct hw_atl_txd_s *)&ring->dx_ring[ring->sw_tail * txd 504 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c txd->ctl = 0; txd 505 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c txd->ctl2 = 0; txd 506 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c txd->buf_addr = 0; txd 511 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c txd->ctl |= HW_ATL_B0_TXD_CTL_CMD_TCP; txd 512 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c txd->ctl |= HW_ATL_B0_TXD_CTL_DESC_TYPE_TXC; txd 513 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c txd->ctl |= (buff->len_l3 << 31) | txd 515 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c txd->ctl2 |= (buff->mss << 16); txd 522 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c txd->ctl |= HW_ATL_B0_TXD_CTL_CMD_IPV6; txd 523 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c txd->ctl2 |= (buff->len_l4 << 8) | txd 527 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c txd->ctl |= HW_ATL_B0_TXD_CTL_DESC_TYPE_TXC; txd 528 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c txd->ctl |= buff->vlan_tx_tag << 4; txd 534 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c txd->buf_addr = buff->pa; txd 535 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c txd->ctl |= (HW_ATL_B0_TXD_CTL_BLEN & txd 537 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c txd->ctl |= HW_ATL_B0_TXD_CTL_DESC_TYPE_TXD; txd 540 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c txd->ctl2 |= HW_ATL_B0_TXD_CTL2_LEN & (pkt_len << 14); txd 544 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c txd->ctl2 |= HW_ATL_B0_TXD_CTL2_CTX_EN; txd 547 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c txd->ctl |= HW_ATL_B0_TXD_CTL_CMD_LSO; txd 551 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c txd->ctl |= HW_ATL_B0_TXD_CTL_CMD_IPCSO; txd 554 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c txd->ctl |= HW_ATL_B0_TXD_CTL_CMD_TUCSO; txd 557 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c txd->ctl |= HW_ATL_B0_TXD_CTL_CMD_VLAN; txd 560 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c txd->ctl |= HW_ATL_B0_TXD_CTL_EOP; txd 561 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c txd->ctl |= HW_ATL_B0_TXD_CTL_CMD_WB; txd 390 drivers/net/ethernet/aurora/nb8800.c struct nb8800_tx_desc *txd; txd 425 drivers/net/ethernet/aurora/nb8800.c txd = &priv->tx_descs[next]; txd 426 drivers/net/ethernet/aurora/nb8800.c desc = &txd->desc[0]; txd 431 drivers/net/ethernet/aurora/nb8800.c memcpy(txd->buf, skb->data, align); txd 435 drivers/net/ethernet/aurora/nb8800.c desc->n_addr = txb->dma_desc + sizeof(txd->desc[0]); txd 496 drivers/net/ethernet/aurora/nb8800.c struct nb8800_tx_desc *txd = &priv->tx_descs[done]; txd 500 drivers/net/ethernet/aurora/nb8800.c if (!txd->report) txd 509 drivers/net/ethernet/aurora/nb8800.c if (IS_TX_ERROR(txd->report)) { txd 510 drivers/net/ethernet/aurora/nb8800.c nb8800_tx_error(dev, txd->report); txd 517 drivers/net/ethernet/aurora/nb8800.c dev->stats.tx_bytes += TX_BYTES_TRANSFERRED(txd->report); txd 518 drivers/net/ethernet/aurora/nb8800.c dev->stats.collisions += TX_EARLY_COLLISIONS(txd->report); txd 522 drivers/net/ethernet/aurora/nb8800.c txd->report = 0; txd 785 drivers/net/ethernet/aurora/nb8800.c struct nb8800_tx_desc *txd; txd 809 drivers/net/ethernet/aurora/nb8800.c txd = &priv->tx_descs[i]; txd 810 drivers/net/ethernet/aurora/nb8800.c txd->desc[0].r_addr = r_dma; txd 811 drivers/net/ethernet/aurora/nb8800.c txd->desc[1].r_addr = r_dma; txd 812 drivers/net/ethernet/aurora/nb8800.c txd->report = 0; txd 875 drivers/net/ethernet/aurora/nb8800.c struct nb8800_tx_desc *txd = &priv->tx_descs[0]; txd 901 drivers/net/ethernet/aurora/nb8800.c txd->desc[0].s_addr = txd 903 drivers/net/ethernet/aurora/nb8800.c txd->desc[0].config = DESC_BTS(2) | DESC_DS | DESC_EOF | DESC_EOC | 8; txd 904 drivers/net/ethernet/aurora/nb8800.c memset(txd->buf, 0, sizeof(txd->buf)); txd 334 drivers/net/ethernet/cirrus/ep93xx_eth.c struct ep93xx_tdesc *txd; txd 346 drivers/net/ethernet/cirrus/ep93xx_eth.c txd = &ep->descs->tdesc[entry]; txd 348 drivers/net/ethernet/cirrus/ep93xx_eth.c txd->tdesc1 = TDESC1_EOF | (entry << 16) | (skb->len & 0xfff); txd 349 drivers/net/ethernet/cirrus/ep93xx_eth.c dma_sync_single_for_cpu(dev->dev.parent, txd->buf_addr, skb->len, txd 352 drivers/net/ethernet/cirrus/ep93xx_eth.c dma_sync_single_for_device(dev->dev.parent, txd->buf_addr, skb->len, txd 618 drivers/net/ethernet/cortina/gemini.c struct gmac_txdesc *txd; txd 629 drivers/net/ethernet/cortina/gemini.c txd = txq->ring + c; txd 630 drivers/net/ethernet/cortina/gemini.c word0 = txd->word0; txd 631 drivers/net/ethernet/cortina/gemini.c word1 = txd->word1; txd 632 drivers/net/ethernet/cortina/gemini.c mapping = txd->word2.buf_adr; txd 633 drivers/net/ethernet/cortina/gemini.c word3 = txd->word3.bits32; txd 653 drivers/net/ethernet/cortina/gemini.c bytes += txd->word1.bits.byte_count; txd 1147 drivers/net/ethernet/cortina/gemini.c struct gmac_txdesc *txd; txd 1201 drivers/net/ethernet/cortina/gemini.c txd = txq->ring + w; txd 1202 drivers/net/ethernet/cortina/gemini.c txd->word0.bits32 = buflen; txd 1203 drivers/net/ethernet/cortina/gemini.c txd->word1.bits32 = word1; txd 1204 drivers/net/ethernet/cortina/gemini.c txd->word2.buf_adr = mapping; txd 1205 drivers/net/ethernet/cortina/gemini.c txd->word3.bits32 = word3; txd 609 drivers/net/ethernet/dec/tulip/de2104x.c struct de_desc *txd; txd 623 drivers/net/ethernet/dec/tulip/de2104x.c txd = &de->tx_ring[entry]; txd 632 drivers/net/ethernet/dec/tulip/de2104x.c txd->opts2 = cpu_to_le32(flags); txd 633 drivers/net/ethernet/dec/tulip/de2104x.c txd->addr1 = cpu_to_le32(mapping); txd 639 drivers/net/ethernet/dec/tulip/de2104x.c txd->opts1 = cpu_to_le32(DescOwn); txd 724 drivers/net/ethernet/dec/tulip/de2104x.c struct de_desc *txd; txd 773 drivers/net/ethernet/dec/tulip/de2104x.c txd = &de->tx_ring[entry]; txd 775 drivers/net/ethernet/dec/tulip/de2104x.c txd->opts2 = cpu_to_le32(SetupFrame | RingEnd | sizeof (de->setup_frame)); txd 777 drivers/net/ethernet/dec/tulip/de2104x.c txd->opts2 = cpu_to_le32(SetupFrame | sizeof (de->setup_frame)); txd 778 drivers/net/ethernet/dec/tulip/de2104x.c txd->addr1 = cpu_to_le32(mapping); txd 781 drivers/net/ethernet/dec/tulip/de2104x.c txd->opts1 = cpu_to_le32(DescOwn); txd 60 drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c struct fm10k_tx_desc *txd = FM10K_TX_DESC(ring, i); txd 63 drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c i, txd->buffer_addr, txd->buflen, txd->vlan, txd 64 drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c txd->mss, txd->hdrlen, txd->flags); txd 497 drivers/net/ethernet/intel/i40e/i40e_debugfs.c struct i40e_tx_desc *txd; txd 530 drivers/net/ethernet/intel/i40e/i40e_debugfs.c txd = I40E_TX_DESC(ring, i); txd 533 drivers/net/ethernet/intel/i40e/i40e_debugfs.c i, txd->buffer_addr, txd 534 drivers/net/ethernet/intel/i40e/i40e_debugfs.c txd->cmd_type_offset_bsz); txd 551 drivers/net/ethernet/intel/i40e/i40e_debugfs.c txd = I40E_TX_DESC(ring, desc_n); txd 555 drivers/net/ethernet/intel/i40e/i40e_debugfs.c txd->buffer_addr, txd->cmd_type_offset_bsz); txd 2074 drivers/net/ethernet/marvell/mv643xx_eth.c struct tx_desc *txd = tx_desc + i; txd 2081 drivers/net/ethernet/marvell/mv643xx_eth.c txd->cmd_sts = 0; txd 2082 drivers/net/ethernet/marvell/mv643xx_eth.c txd->next_desc_ptr = txq->tx_desc_dma + txd 829 drivers/net/ethernet/mediatek/mtk_eth_soc.c struct mtk_tx_dma *txd) txd 831 drivers/net/ethernet/mediatek/mtk_eth_soc.c int idx = txd - ring->dma; txd 885 drivers/net/ethernet/mediatek/mtk_eth_soc.c struct mtk_tx_dma *txd, dma_addr_t mapped_addr, txd 893 drivers/net/ethernet/mediatek/mtk_eth_soc.c txd->txd3 = mapped_addr; txd 894 drivers/net/ethernet/mediatek/mtk_eth_soc.c txd->txd2 |= TX_DMA_PLEN1(size); txd 899 drivers/net/ethernet/mediatek/mtk_eth_soc.c txd->txd1 = mapped_addr; txd 900 drivers/net/ethernet/mediatek/mtk_eth_soc.c txd->txd2 = TX_DMA_PLEN0(size); txd 912 drivers/net/ethernet/mediatek/mtk_eth_soc.c struct mtk_tx_dma *itxd, *txd; txd 957 drivers/net/ethernet/mediatek/mtk_eth_soc.c txd = itxd; txd 958 drivers/net/ethernet/mediatek/mtk_eth_soc.c txd_pdma = qdma_to_pdma(ring, txd); txd 973 drivers/net/ethernet/mediatek/mtk_eth_soc.c txd = mtk_qdma_phys_to_virt(ring, txd->txd2); txd 974 drivers/net/ethernet/mediatek/mtk_eth_soc.c txd_pdma = qdma_to_pdma(ring, txd); txd 975 drivers/net/ethernet/mediatek/mtk_eth_soc.c if (txd == ring->last_free) txd 995 drivers/net/ethernet/mediatek/mtk_eth_soc.c WRITE_ONCE(txd->txd1, mapped_addr); txd 996 drivers/net/ethernet/mediatek/mtk_eth_soc.c WRITE_ONCE(txd->txd3, (TX_DMA_SWC | txd 999 drivers/net/ethernet/mediatek/mtk_eth_soc.c WRITE_ONCE(txd->txd4, fport); txd 1001 drivers/net/ethernet/mediatek/mtk_eth_soc.c tx_buf = mtk_desc_to_tx_buf(ring, txd); txd 1033 drivers/net/ethernet/mediatek/mtk_eth_soc.c ring->next_free = mtk_qdma_phys_to_virt(ring, txd->txd2); txd 1044 drivers/net/ethernet/mediatek/mtk_eth_soc.c mtk_w32(eth, txd->txd2, MTK_QTX_CTX_PTR); txd 1046 drivers/net/ethernet/mediatek/mtk_eth_soc.c int next_idx = NEXT_DESP_IDX(txd_to_idx(ring, txd), txd 1066 drivers/net/ethernet/mediatek/mtk_eth_soc.c } while (itxd != txd); txd 111 drivers/net/ethernet/micrel/ks8851.c u8 txd[8]; txd 233 drivers/net/ethernet/micrel/ks8851.c __le16 *txb = (__le16 *)ks->txd; txd 727 drivers/net/ethernet/netronome/nfp/nfp_net_common.c struct nfp_net_tx_desc *txd, struct sk_buff *skb, txd 751 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->l3_offset = l3_offset - md_bytes; txd 752 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->l4_offset = l4_offset - md_bytes; txd 753 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->lso_hdrlen = hdrlen - md_bytes; txd 754 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->mss = cpu_to_le16(mss); txd 755 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->flags |= PCIE_DESC_TX_LSO; txd 776 drivers/net/ethernet/netronome/nfp/nfp_net_common.c struct nfp_net_tx_desc *txd, struct sk_buff *skb) txd 788 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->flags |= PCIE_DESC_TX_CSUM; txd 790 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->flags |= PCIE_DESC_TX_ENCAP; txd 796 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->flags |= PCIE_DESC_TX_IP4_CSUM; txd 807 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->flags |= PCIE_DESC_TX_TCP_CSUM; txd 810 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->flags |= PCIE_DESC_TX_UDP_CSUM; txd 983 drivers/net/ethernet/netronome/nfp/nfp_net_common.c struct nfp_net_tx_desc *txd; txd 1037 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd = &tx_ring->txds[wr_idx]; txd 1038 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->offset_eop = (nr_frags ? 0 : PCIE_DESC_TX_EOP) | md_bytes; txd 1039 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->dma_len = cpu_to_le16(skb_headlen(skb)); txd 1040 drivers/net/ethernet/netronome/nfp/nfp_net_common.c nfp_desc_set_dma_addr(txd, dma_addr); txd 1041 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->data_len = cpu_to_le16(skb->len); txd 1043 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->flags = 0; txd 1044 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->mss = 0; txd 1045 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->lso_hdrlen = 0; txd 1048 drivers/net/ethernet/netronome/nfp/nfp_net_common.c nfp_net_tx_tso(r_vec, txbuf, txd, skb, md_bytes); txd 1049 drivers/net/ethernet/netronome/nfp/nfp_net_common.c nfp_net_tx_csum(dp, r_vec, txbuf, txd, skb); txd 1051 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->flags |= PCIE_DESC_TX_VLAN; txd 1052 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->vlan = cpu_to_le16(skb_vlan_tag_get(skb)); txd 1060 drivers/net/ethernet/netronome/nfp/nfp_net_common.c second_half = txd->vals8[1]; txd 1076 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd = &tx_ring->txds[wr_idx]; txd 1077 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->dma_len = cpu_to_le16(fsize); txd 1078 drivers/net/ethernet/netronome/nfp/nfp_net_common.c nfp_desc_set_dma_addr(txd, dma_addr); txd 1079 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->offset_eop = md_bytes | txd 1081 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->vals8[1] = second_half; txd 1743 drivers/net/ethernet/netronome/nfp/nfp_net_common.c struct nfp_net_tx_desc *txd; txd 1776 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd = &tx_ring->txds[wr_idx]; txd 1777 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->offset_eop = PCIE_DESC_TX_EOP; txd 1778 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->dma_len = cpu_to_le16(pkt_len); txd 1779 drivers/net/ethernet/netronome/nfp/nfp_net_common.c nfp_desc_set_dma_addr(txd, rxbuf->dma_addr + dma_off); txd 1780 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->data_len = cpu_to_le16(pkt_len); txd 1782 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->flags = 0; txd 1783 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->mss = 0; txd 1784 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->lso_hdrlen = 0; txd 2073 drivers/net/ethernet/netronome/nfp/nfp_net_common.c struct nfp_net_tx_desc *txd; txd 2124 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd = &tx_ring->txds[wr_idx]; txd 2125 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->offset_eop = meta_len | PCIE_DESC_TX_EOP; txd 2126 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->dma_len = cpu_to_le16(skb_headlen(skb)); txd 2127 drivers/net/ethernet/netronome/nfp/nfp_net_common.c nfp_desc_set_dma_addr(txd, dma_addr); txd 2128 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->data_len = cpu_to_le16(skb->len); txd 2130 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->flags = 0; txd 2131 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->mss = 0; txd 2132 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->lso_hdrlen = 0; txd 77 drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c struct nfp_net_tx_desc *txd; txd 106 drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c txd = &tx_ring->txds[i]; txd 108 drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c txd->vals[0], txd->vals[1], txd 109 drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c txd->vals[2], txd->vals[3]); txd 655 drivers/net/ethernet/realtek/8139cp.c struct cp_desc *txd = cp->tx_ring + tx_tail; txd 660 drivers/net/ethernet/realtek/8139cp.c status = le32_to_cpu(txd->opts1); txd 667 drivers/net/ethernet/realtek/8139cp.c dma_unmap_single(&cp->pdev->dev, le64_to_cpu(txd->addr), txd 719 drivers/net/ethernet/realtek/8139cp.c struct cp_desc *txd; txd 724 drivers/net/ethernet/realtek/8139cp.c txd = &cp->tx_ring[index]; txd 726 drivers/net/ethernet/realtek/8139cp.c dma_unmap_single(&cp->pdev->dev, le64_to_cpu(txd->addr), txd 779 drivers/net/ethernet/realtek/8139cp.c struct cp_desc *txd = &cp->tx_ring[entry]; txd 788 drivers/net/ethernet/realtek/8139cp.c txd->opts2 = opts2; txd 789 drivers/net/ethernet/realtek/8139cp.c txd->addr = cpu_to_le64(mapping); txd 794 drivers/net/ethernet/realtek/8139cp.c txd->opts1 = cpu_to_le32(opts1); txd 802 drivers/net/ethernet/realtek/8139cp.c struct cp_desc *txd; txd 842 drivers/net/ethernet/realtek/8139cp.c txd = &cp->tx_ring[entry]; txd 843 drivers/net/ethernet/realtek/8139cp.c txd->opts2 = opts2; txd 844 drivers/net/ethernet/realtek/8139cp.c txd->addr = cpu_to_le64(mapping); txd 847 drivers/net/ethernet/realtek/8139cp.c txd->opts1 = cpu_to_le32(ctrl); txd 854 drivers/net/ethernet/realtek/8139cp.c txd = &cp->tx_ring[first_entry]; txd 855 drivers/net/ethernet/realtek/8139cp.c txd->opts2 = opts2; txd 856 drivers/net/ethernet/realtek/8139cp.c txd->addr = cpu_to_le64(first_mapping); txd 860 drivers/net/ethernet/realtek/8139cp.c txd->opts1 = cpu_to_le32(ctrl); txd 5701 drivers/net/ethernet/realtek/r8169_main.c struct TxDesc *uninitialized_var(txd); txd 5713 drivers/net/ethernet/realtek/r8169_main.c txd = tp->TxDescArray + entry; txd 5724 drivers/net/ethernet/realtek/r8169_main.c txd->opts1 = rtl8169_get_txd_opts1(opts[0], len, entry); txd 5725 drivers/net/ethernet/realtek/r8169_main.c txd->opts2 = cpu_to_le32(opts[1]); txd 5726 drivers/net/ethernet/realtek/r8169_main.c txd->addr = cpu_to_le64(mapping); txd 5733 drivers/net/ethernet/realtek/r8169_main.c txd->opts1 |= cpu_to_le32(LastFrag); txd 5885 drivers/net/ethernet/realtek/r8169_main.c struct TxDesc *txd = tp->TxDescArray + entry; txd 5898 drivers/net/ethernet/realtek/r8169_main.c if (unlikely(le32_to_cpu(txd->opts1) & DescOwn)) txd 5920 drivers/net/ethernet/realtek/r8169_main.c txd->addr = cpu_to_le64(mapping); txd 5932 drivers/net/ethernet/realtek/r8169_main.c txd->opts2 = cpu_to_le32(opts[1]); txd 5941 drivers/net/ethernet/realtek/r8169_main.c txd->opts1 = rtl8169_get_txd_opts1(opts[0], len, entry); txd 5977 drivers/net/ethernet/realtek/r8169_main.c rtl8169_unmap_tx_skb(d, tp->tx_skb + entry, txd); txd 2300 drivers/net/ethernet/sfc/ef10.c return efx_nic_alloc_buffer(tx_queue->efx, &tx_queue->txd.buf, txd 2308 drivers/net/ethernet/sfc/ef10.c const efx_qword_t *txd) txd 2315 drivers/net/ethernet/sfc/ef10.c reg.qword[0] = *txd; txd 2411 drivers/net/ethernet/sfc/ef10.c size_t entries = tx_queue->txd.buf.len / EFX_BUF_SIZE; txd 2418 drivers/net/ethernet/sfc/ef10.c efx_qword_t *txd; txd 2454 drivers/net/ethernet/sfc/ef10.c dma_addr = tx_queue->txd.buf.dma_addr; txd 2501 drivers/net/ethernet/sfc/ef10.c txd = efx_tx_desc(tx_queue, 0); txd 2502 drivers/net/ethernet/sfc/ef10.c EFX_POPULATE_QWORD_5(*txd, txd 2520 drivers/net/ethernet/sfc/ef10.c efx_ef10_push_tx_desc(tx_queue, txd); txd 2555 drivers/net/ethernet/sfc/ef10.c efx_nic_free_buffer(tx_queue->efx, &tx_queue->txd.buf); txd 2594 drivers/net/ethernet/sfc/ef10.c efx_qword_t *txd; txd 2603 drivers/net/ethernet/sfc/ef10.c txd = efx_tx_desc(tx_queue, write_ptr); txd 2608 drivers/net/ethernet/sfc/ef10.c *txd = buffer->option; txd 2609 drivers/net/ethernet/sfc/ef10.c if (EFX_QWORD_FIELD(*txd, ESF_DZ_TX_OPTION_TYPE) == 1) txd 2616 drivers/net/ethernet/sfc/ef10.c *txd, txd 2627 drivers/net/ethernet/sfc/ef10.c txd = efx_tx_desc(tx_queue, txd 2629 drivers/net/ethernet/sfc/ef10.c efx_ef10_push_tx_desc(tx_queue, txd); txd 523 drivers/net/ethernet/sfc/efx.c memset(&tx_queue->txd, 0, sizeof(tx_queue->txd)); txd 834 drivers/net/ethernet/sfc/efx.c tx_queue->txd.index + txd 835 drivers/net/ethernet/sfc/efx.c tx_queue->txd.entries); txd 482 drivers/net/ethernet/sfc/falcon/efx.c memset(&tx_queue->txd, 0, sizeof(tx_queue->txd)); txd 793 drivers/net/ethernet/sfc/falcon/efx.c tx_queue->txd.index + txd 794 drivers/net/ethernet/sfc/falcon/efx.c tx_queue->txd.entries); txd 285 drivers/net/ethernet/sfc/falcon/farch.c const ef4_qword_t *txd) txd 296 drivers/net/ethernet/sfc/falcon/farch.c reg.qword[0] = *txd; txd 309 drivers/net/ethernet/sfc/falcon/farch.c ef4_qword_t *txd; txd 320 drivers/net/ethernet/sfc/falcon/farch.c txd = ef4_tx_desc(tx_queue, write_ptr); txd 327 drivers/net/ethernet/sfc/falcon/farch.c EF4_POPULATE_QWORD_4(*txd, txd 338 drivers/net/ethernet/sfc/falcon/farch.c txd = ef4_tx_desc(tx_queue, txd 340 drivers/net/ethernet/sfc/falcon/farch.c ef4_farch_push_tx_desc(tx_queue, txd); txd 369 drivers/net/ethernet/sfc/falcon/farch.c return ef4_alloc_special_buffer(efx, &tx_queue->txd, txd 379 drivers/net/ethernet/sfc/falcon/farch.c ef4_init_special_buffer(efx, &tx_queue->txd); txd 386 drivers/net/ethernet/sfc/falcon/farch.c FRF_AZ_TX_DESCQ_BUF_BASE_ID, tx_queue->txd.index, txd 392 drivers/net/ethernet/sfc/falcon/farch.c __ffs(tx_queue->txd.entries), txd 454 drivers/net/ethernet/sfc/falcon/farch.c ef4_fini_special_buffer(efx, &tx_queue->txd); txd 460 drivers/net/ethernet/sfc/falcon/farch.c ef4_free_special_buffer(tx_queue->efx, &tx_queue->txd); txd 220 drivers/net/ethernet/sfc/falcon/net_driver.h struct ef4_special_buffer txd; txd 65 drivers/net/ethernet/sfc/falcon/nic.h return ((ef4_qword_t *) (tx_queue->txd.buf.addr)) + index; txd 294 drivers/net/ethernet/sfc/farch.c const efx_qword_t *txd) txd 305 drivers/net/ethernet/sfc/farch.c reg.qword[0] = *txd; txd 318 drivers/net/ethernet/sfc/farch.c efx_qword_t *txd; txd 329 drivers/net/ethernet/sfc/farch.c txd = efx_tx_desc(tx_queue, write_ptr); txd 336 drivers/net/ethernet/sfc/farch.c EFX_POPULATE_QWORD_4(*txd, txd 347 drivers/net/ethernet/sfc/farch.c txd = efx_tx_desc(tx_queue, txd 349 drivers/net/ethernet/sfc/farch.c efx_farch_push_tx_desc(tx_queue, txd); txd 375 drivers/net/ethernet/sfc/farch.c return efx_alloc_special_buffer(efx, &tx_queue->txd, txd 386 drivers/net/ethernet/sfc/farch.c efx_init_special_buffer(efx, &tx_queue->txd); txd 393 drivers/net/ethernet/sfc/farch.c FRF_AZ_TX_DESCQ_BUF_BASE_ID, tx_queue->txd.index, txd 399 drivers/net/ethernet/sfc/farch.c __ffs(tx_queue->txd.entries), txd 442 drivers/net/ethernet/sfc/farch.c efx_fini_special_buffer(efx, &tx_queue->txd); txd 448 drivers/net/ethernet/sfc/farch.c efx_free_special_buffer(tx_queue->efx, &tx_queue->txd); txd 247 drivers/net/ethernet/sfc/net_driver.h struct efx_special_buffer txd; txd 63 drivers/net/ethernet/sfc/nic.h return ((efx_qword_t *) (tx_queue->txd.buf.addr)) + index; txd 88 drivers/net/ethernet/sfc/tx_tso.c ptr = (char *) (((efx_qword_t *)tx_queue->txd.buf.addr) + insert_ptr); txd 700 drivers/net/ethernet/sis/sis190.c struct TxDesc *txd = tp->TxDescRing + entry; txd 701 drivers/net/ethernet/sis/sis190.c u32 status = le32_to_cpu(txd->status); txd 715 drivers/net/ethernet/sis/sis190.c sis190_unmap_tx_skb(tp->pci_dev, skb, txd); txd 1881 drivers/net/ethernet/sun/cassini.c struct cas_tx_desc *txd = txds + entry; txd 1883 drivers/net/ethernet/sun/cassini.c daddr = le64_to_cpu(txd->buffer); txd 1885 drivers/net/ethernet/sun/cassini.c le64_to_cpu(txd->control)); txd 2728 drivers/net/ethernet/sun/cassini.c struct cas_tx_desc *txd = cp->init_txds[ring] + entry; txd 2735 drivers/net/ethernet/sun/cassini.c txd->control = cpu_to_le64(ctrl); txd 2736 drivers/net/ethernet/sun/cassini.c txd->buffer = cpu_to_le64(mapping); txd 3881 drivers/net/ethernet/sun/cassini.c struct cas_tx_desc *txd = cp->init_txds[ring]; txd 3902 drivers/net/ethernet/sun/cassini.c daddr = le64_to_cpu(txd[ent].buffer); txd 3904 drivers/net/ethernet/sun/cassini.c le64_to_cpu(txd[ent].control)); txd 651 drivers/net/ethernet/sun/sungem.c struct gem_txd *txd; txd 680 drivers/net/ethernet/sun/sungem.c txd = &gp->init_block->txd[entry]; txd 682 drivers/net/ethernet/sun/sungem.c dma_addr = le64_to_cpu(txd->buffer); txd 683 drivers/net/ethernet/sun/sungem.c dma_len = le64_to_cpu(txd->control_word) & TXDCTRL_BUFSZ; txd 1030 drivers/net/ethernet/sun/sungem.c struct gem_txd *txd = &gp->init_block->txd[entry]; txd 1042 drivers/net/ethernet/sun/sungem.c txd->buffer = cpu_to_le64(mapping); txd 1044 drivers/net/ethernet/sun/sungem.c txd->control_word = cpu_to_le64(ctrl); txd 1047 drivers/net/ethernet/sun/sungem.c struct gem_txd *txd; txd 1079 drivers/net/ethernet/sun/sungem.c txd = &gp->init_block->txd[entry]; txd 1080 drivers/net/ethernet/sun/sungem.c txd->buffer = cpu_to_le64(mapping); txd 1082 drivers/net/ethernet/sun/sungem.c txd->control_word = cpu_to_le64(this_ctrl | len); txd 1089 drivers/net/ethernet/sun/sungem.c txd = &gp->init_block->txd[first_entry]; txd 1090 drivers/net/ethernet/sun/sungem.c txd->buffer = cpu_to_le64(first_mapping); txd 1092 drivers/net/ethernet/sun/sungem.c txd->control_word = txd 1602 drivers/net/ethernet/sun/sungem.c struct gem_txd *txd; txd 1611 drivers/net/ethernet/sun/sungem.c txd = &gb->txd[ent]; txd 1612 drivers/net/ethernet/sun/sungem.c dma_addr = le64_to_cpu(txd->buffer); txd 1614 drivers/net/ethernet/sun/sungem.c le64_to_cpu(txd->control_word) & txd 1664 drivers/net/ethernet/sun/sungem.c struct gem_txd *txd = &gb->txd[i]; txd 1666 drivers/net/ethernet/sun/sungem.c txd->control_word = 0; txd 1668 drivers/net/ethernet/sun/sungem.c txd->buffer = 0; txd 956 drivers/net/ethernet/sun/sungem.h struct gem_txd txd[INIT_BLOCK_TX_RING_SIZE]; txd 204 drivers/net/ethernet/sun/sunhme.c static void sbus_hme_write_txd(struct happy_meal_txd *txd, u32 flags, u32 addr) txd 206 drivers/net/ethernet/sun/sunhme.c txd->tx_addr = (__force hme32)addr; txd 208 drivers/net/ethernet/sun/sunhme.c txd->tx_flags = (__force hme32)flags; txd 233 drivers/net/ethernet/sun/sunhme.c static void pci_hme_write_txd(struct happy_meal_txd *txd, u32 flags, u32 addr) txd 235 drivers/net/ethernet/sun/sunhme.c txd->tx_addr = (__force hme32)cpu_to_le32(addr); txd 237 drivers/net/ethernet/sun/sunhme.c txd->tx_flags = (__force hme32)cpu_to_le32(flags); txd 1220 drivers/net/ethernet/sun/sunhme.c struct happy_meal_txd *txd; txd 1227 drivers/net/ethernet/sun/sunhme.c txd = &hp->happy_block->happy_meal_txd[i]; txd 1228 drivers/net/ethernet/sun/sunhme.c dma_addr = hme_read_desc32(hp, &txd->tx_addr); txd 1231 drivers/net/ethernet/sun/sunhme.c (hme_read_desc32(hp, &txd->tx_flags) txd 1236 drivers/net/ethernet/sun/sunhme.c (hme_read_desc32(hp, &txd->tx_flags) txd 209 drivers/net/fddi/skfp/h/hwmtm.h #define HWM_GET_TX_PHYS(txd) (u_long)AIX_REVERSE((txd)->txd_tbadr) txd 225 drivers/net/fddi/skfp/h/hwmtm.h #define HWM_GET_TX_LEN(txd) ((int)AIX_REVERSE((txd)->txd_tbctrl)& RD_LENGTH) txd 98 drivers/net/fddi/skfp/hwmtm.c volatile struct s_smt_fp_txd *txd); txd 121 drivers/net/fddi/skfp/skfddi.c void mac_drv_tx_complete(struct s_smc *smc, volatile struct s_smt_fp_txd *txd); txd 1106 drivers/net/fddi/skfp/skfddi.c struct s_smt_fp_txd *txd; // Current TxD. txd 1170 drivers/net/fddi/skfp/skfddi.c txd = (struct s_smt_fp_txd *) HWM_GET_CURR_TXD(smc, queue); txd 1175 drivers/net/fddi/skfp/skfddi.c txd->txd_os.skb = skb; // save skb txd 1176 drivers/net/fddi/skfp/skfddi.c txd->txd_os.dma_addr = dma_address; // save dma mapping txd 1487 drivers/net/fddi/skfp/skfddi.c void mac_drv_tx_complete(struct s_smc *smc, volatile struct s_smt_fp_txd *txd) txd 1494 drivers/net/fddi/skfp/skfddi.c if (!(skb = txd->txd_os.skb)) { txd 1498 drivers/net/fddi/skfp/skfddi.c txd->txd_os.skb = NULL; txd 1501 drivers/net/fddi/skfp/skfddi.c pci_unmap_single(&smc->os.pdev, txd->txd_os.dma_addr, txd 1503 drivers/net/fddi/skfp/skfddi.c txd->txd_os.dma_addr = 0; txd 111 drivers/net/hamradio/yam.c int txd; /* tx delay */ txd 595 drivers/net/hamradio/yam.c if ((yp->tx_state == TX_TAIL) || (yp->txd == 0)) txd 598 drivers/net/hamradio/yam.c yp->tx_count = (yp->bitrate * yp->txd) / 8000; txd 810 drivers/net/hamradio/yam.c seq_printf(seq, " TxDelay %u\n", yp->txd); txd 1015 drivers/net/hamradio/yam.c yp->txd = yi.cfg.txdelay; txd 1043 drivers/net/hamradio/yam.c yi.cfg.txdelay = yp->txd; txd 1092 drivers/net/hamradio/yam.c yp->txd = DEFAULT_TXD; txd 353 drivers/net/vmxnet3/vmxnet3_defs.h struct Vmxnet3_TxDesc txd; txd 337 drivers/net/vmxnet3/vmxnet3_drv.c BUG_ON(VMXNET3_TXDESC_GET_EOP(&(tq->tx_ring.base[eop_idx].txd)) != 1); txd 694 drivers/net/vmxnet3/vmxnet3_drv.c ctx->sop_txd->txd.addr = cpu_to_le64(tq->data_ring.basePA + txd 706 drivers/net/vmxnet3/vmxnet3_drv.c le64_to_cpu(ctx->sop_txd->txd.addr), txd 739 drivers/net/vmxnet3/vmxnet3_drv.c BUG_ON(gdesc->txd.gen == tq->tx_ring.gen); txd 741 drivers/net/vmxnet3/vmxnet3_drv.c gdesc->txd.addr = cpu_to_le64(tbi->dma_addr); txd 747 drivers/net/vmxnet3/vmxnet3_drv.c tq->tx_ring.next2fill, le64_to_cpu(gdesc->txd.addr), txd 781 drivers/net/vmxnet3/vmxnet3_drv.c BUG_ON(gdesc->txd.gen == tq->tx_ring.gen); txd 783 drivers/net/vmxnet3/vmxnet3_drv.c gdesc->txd.addr = cpu_to_le64(tbi->dma_addr); txd 789 drivers/net/vmxnet3/vmxnet3_drv.c tq->tx_ring.next2fill, le64_to_cpu(gdesc->txd.addr), txd 1086 drivers/net/vmxnet3/vmxnet3_drv.c gdesc->txd.hlen = ctx.eth_ip_hdr_size + ctx.l4_hdr_size; txd 1087 drivers/net/vmxnet3/vmxnet3_drv.c gdesc->txd.om = VMXNET3_OM_TSO; txd 1088 drivers/net/vmxnet3/vmxnet3_drv.c gdesc->txd.msscof = ctx.mss; txd 1089 drivers/net/vmxnet3/vmxnet3_drv.c num_pkts = (skb->len - gdesc->txd.hlen + ctx.mss - 1) / ctx.mss; txd 1092 drivers/net/vmxnet3/vmxnet3_drv.c gdesc->txd.hlen = ctx.eth_ip_hdr_size; txd 1093 drivers/net/vmxnet3/vmxnet3_drv.c gdesc->txd.om = VMXNET3_OM_CSUM; txd 1094 drivers/net/vmxnet3/vmxnet3_drv.c gdesc->txd.msscof = ctx.eth_ip_hdr_size + txd 1097 drivers/net/vmxnet3/vmxnet3_drv.c gdesc->txd.om = 0; txd 1098 drivers/net/vmxnet3/vmxnet3_drv.c gdesc->txd.msscof = 0; txd 1106 drivers/net/vmxnet3/vmxnet3_drv.c gdesc->txd.ti = 1; txd 1107 drivers/net/vmxnet3/vmxnet3_drv.c gdesc->txd.tci = skb_vlan_tag_get(skb); txd 1129 drivers/net/vmxnet3/vmxnet3_drv.c tq->tx_ring.base), le64_to_cpu(gdesc->txd.addr), txd 325 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c return txd(di, i + 1); txd 330 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c return txd(di, i - 1); txd 340 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c return txd(di, t-h); txd 162 drivers/net/wireless/marvell/mwl8k.c struct mwl8k_tx_desc *txd; txd 1463 drivers/net/wireless/marvell/mwl8k.c txq->txd = pci_zalloc_consistent(priv->pdev, size, &txq->txd_dma); txd 1464 drivers/net/wireless/marvell/mwl8k.c if (txq->txd == NULL) { txd 1471 drivers/net/wireless/marvell/mwl8k.c pci_free_consistent(priv->pdev, size, txq->txd, txq->txd_dma); txd 1479 drivers/net/wireless/marvell/mwl8k.c tx_desc = txq->txd + i; txd 1512 drivers/net/wireless/marvell/mwl8k.c struct mwl8k_tx_desc *tx_desc = txq->txd + desc; txd 1688 drivers/net/wireless/marvell/mwl8k.c tx_desc = txq->txd + tx; txd 1769 drivers/net/wireless/marvell/mwl8k.c if (txq->txd == NULL) txd 1779 drivers/net/wireless/marvell/mwl8k.c txq->txd, txq->txd_dma); txd 1780 drivers/net/wireless/marvell/mwl8k.c txq->txd = NULL; txd 2090 drivers/net/wireless/marvell/mwl8k.c tx = txq->txd + txq->tail; txd 33 drivers/net/wireless/mediatek/mt76/mt7603/dma.c __le32 *txd = (__le32 *)skb->data; txd 46 drivers/net/wireless/mediatek/mt76/mt7603/dma.c val = le32_to_cpu(txd[1]); txd 58 drivers/net/wireless/mediatek/mt76/mt7603/dma.c val = le32_to_cpu(txd[0]); txd 63 drivers/net/wireless/mediatek/mt76/mt7603/dma.c txd[0] = cpu_to_le32(val); txd 22 drivers/net/wireless/mediatek/mt76/mt7603/mcu.c struct mt7603_mcu_txd *txd; txd 29 drivers/net/wireless/mediatek/mt76/mt7603/mcu.c txd = (struct mt7603_mcu_txd *)skb_push(skb, hdrlen); txd 30 drivers/net/wireless/mediatek/mt76/mt7603/mcu.c memset(txd, 0, hdrlen); txd 32 drivers/net/wireless/mediatek/mt76/mt7603/mcu.c txd->len = cpu_to_le16(skb->len); txd 34 drivers/net/wireless/mediatek/mt76/mt7603/mcu.c txd->pq_id = cpu_to_le16(MCU_PORT_QUEUE_FW); txd 36 drivers/net/wireless/mediatek/mt76/mt7603/mcu.c txd->pq_id = cpu_to_le16(MCU_PORT_QUEUE); txd 37 drivers/net/wireless/mediatek/mt76/mt7603/mcu.c txd->pkt_type = MCU_PKT_ID; txd 38 drivers/net/wireless/mediatek/mt76/mt7603/mcu.c txd->seq = seq; txd 41 drivers/net/wireless/mediatek/mt76/mt7603/mcu.c txd->cid = -cmd; txd 42 drivers/net/wireless/mediatek/mt76/mt7603/mcu.c txd->set_query = MCU_Q_NA; txd 44 drivers/net/wireless/mediatek/mt76/mt7603/mcu.c txd->cid = MCU_CMD_EXT_CID; txd 45 drivers/net/wireless/mediatek/mt76/mt7603/mcu.c txd->ext_cid = cmd; txd 46 drivers/net/wireless/mediatek/mt76/mt7603/mcu.c txd->set_query = MCU_Q_SET; txd 47 drivers/net/wireless/mediatek/mt76/mt7603/mcu.c txd->ext_cid_ack = 1; txd 58 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c __le32 *txd; txd 76 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c txd = mcu_txd->txd; txd 81 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c txd[0] = cpu_to_le32(val); txd 86 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c txd[1] = cpu_to_le32(val); txd 88 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c mcu_txd->len = cpu_to_le16(skb->len - sizeof(mcu_txd->txd)); txd 8 drivers/net/wireless/mediatek/mt76/mt7615/mcu.h __le32 txd[8]; txd 1099 drivers/net/wireless/ralink/rt2x00/rt2400pci.c __le32 *txd = entry_priv->desc; txd 1105 drivers/net/wireless/ralink/rt2x00/rt2400pci.c word = rt2x00_desc_read(txd, 1); txd 1107 drivers/net/wireless/ralink/rt2x00/rt2400pci.c rt2x00_desc_write(txd, 1, word); txd 1109 drivers/net/wireless/ralink/rt2x00/rt2400pci.c word = rt2x00_desc_read(txd, 2); txd 1112 drivers/net/wireless/ralink/rt2x00/rt2400pci.c rt2x00_desc_write(txd, 2, word); txd 1114 drivers/net/wireless/ralink/rt2x00/rt2400pci.c word = rt2x00_desc_read(txd, 3); txd 1121 drivers/net/wireless/ralink/rt2x00/rt2400pci.c rt2x00_desc_write(txd, 3, word); txd 1123 drivers/net/wireless/ralink/rt2x00/rt2400pci.c word = rt2x00_desc_read(txd, 4); txd 1132 drivers/net/wireless/ralink/rt2x00/rt2400pci.c rt2x00_desc_write(txd, 4, word); txd 1139 drivers/net/wireless/ralink/rt2x00/rt2400pci.c word = rt2x00_desc_read(txd, 0); txd 1153 drivers/net/wireless/ralink/rt2x00/rt2400pci.c rt2x00_desc_write(txd, 0, word); txd 1158 drivers/net/wireless/ralink/rt2x00/rt2400pci.c skbdesc->desc = txd; txd 1252 drivers/net/wireless/ralink/rt2x00/rt2500pci.c __le32 *txd = entry_priv->desc; txd 1258 drivers/net/wireless/ralink/rt2x00/rt2500pci.c word = rt2x00_desc_read(txd, 1); txd 1260 drivers/net/wireless/ralink/rt2x00/rt2500pci.c rt2x00_desc_write(txd, 1, word); txd 1262 drivers/net/wireless/ralink/rt2x00/rt2500pci.c word = rt2x00_desc_read(txd, 2); txd 1267 drivers/net/wireless/ralink/rt2x00/rt2500pci.c rt2x00_desc_write(txd, 2, word); txd 1269 drivers/net/wireless/ralink/rt2x00/rt2500pci.c word = rt2x00_desc_read(txd, 3); txd 1276 drivers/net/wireless/ralink/rt2x00/rt2500pci.c rt2x00_desc_write(txd, 3, word); txd 1278 drivers/net/wireless/ralink/rt2x00/rt2500pci.c word = rt2x00_desc_read(txd, 10); txd 1281 drivers/net/wireless/ralink/rt2x00/rt2500pci.c rt2x00_desc_write(txd, 10, word); txd 1288 drivers/net/wireless/ralink/rt2x00/rt2500pci.c word = rt2x00_desc_read(txd, 0); txd 1305 drivers/net/wireless/ralink/rt2x00/rt2500pci.c rt2x00_desc_write(txd, 0, word); txd 1310 drivers/net/wireless/ralink/rt2x00/rt2500pci.c skbdesc->desc = txd; txd 1060 drivers/net/wireless/ralink/rt2x00/rt2500usb.c __le32 *txd = (__le32 *) entry->skb->data; txd 1066 drivers/net/wireless/ralink/rt2x00/rt2500usb.c word = rt2x00_desc_read(txd, 0); txd 1082 drivers/net/wireless/ralink/rt2x00/rt2500usb.c rt2x00_desc_write(txd, 0, word); txd 1084 drivers/net/wireless/ralink/rt2x00/rt2500usb.c word = rt2x00_desc_read(txd, 1); txd 1089 drivers/net/wireless/ralink/rt2x00/rt2500usb.c rt2x00_desc_write(txd, 1, word); txd 1091 drivers/net/wireless/ralink/rt2x00/rt2500usb.c word = rt2x00_desc_read(txd, 2); txd 1098 drivers/net/wireless/ralink/rt2x00/rt2500usb.c rt2x00_desc_write(txd, 2, word); txd 1101 drivers/net/wireless/ralink/rt2x00/rt2500usb.c _rt2x00_desc_write(txd, 3, skbdesc->iv[0]); txd 1102 drivers/net/wireless/ralink/rt2x00/rt2500usb.c _rt2x00_desc_write(txd, 4, skbdesc->iv[1]); txd 1109 drivers/net/wireless/ralink/rt2x00/rt2500usb.c skbdesc->desc = txd; txd 72 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c __le32 *txd = entry_priv->desc; txd 89 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c rt2x00_desc_write(txd, 0, word); txd 100 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c rt2x00_desc_write(txd, 1, word); txd 105 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c rt2x00_desc_write(txd, 2, word); txd 111 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c rt2x00_desc_write(txd, 3, word); txd 116 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c skbdesc->desc = txd; txd 1780 drivers/net/wireless/ralink/rt2x00/rt61pci.c __le32 *txd = entry_priv->desc; txd 1786 drivers/net/wireless/ralink/rt2x00/rt61pci.c word = rt2x00_desc_read(txd, 1); txd 1795 drivers/net/wireless/ralink/rt2x00/rt61pci.c rt2x00_desc_write(txd, 1, word); txd 1797 drivers/net/wireless/ralink/rt2x00/rt61pci.c word = rt2x00_desc_read(txd, 2); txd 1804 drivers/net/wireless/ralink/rt2x00/rt61pci.c rt2x00_desc_write(txd, 2, word); txd 1807 drivers/net/wireless/ralink/rt2x00/rt61pci.c _rt2x00_desc_write(txd, 3, skbdesc->iv[0]); txd 1808 drivers/net/wireless/ralink/rt2x00/rt61pci.c _rt2x00_desc_write(txd, 4, skbdesc->iv[1]); txd 1811 drivers/net/wireless/ralink/rt2x00/rt61pci.c word = rt2x00_desc_read(txd, 5); txd 1817 drivers/net/wireless/ralink/rt2x00/rt61pci.c rt2x00_desc_write(txd, 5, word); txd 1820 drivers/net/wireless/ralink/rt2x00/rt61pci.c word = rt2x00_desc_read(txd, 6); txd 1823 drivers/net/wireless/ralink/rt2x00/rt61pci.c rt2x00_desc_write(txd, 6, word); txd 1825 drivers/net/wireless/ralink/rt2x00/rt61pci.c word = rt2x00_desc_read(txd, 11); txd 1828 drivers/net/wireless/ralink/rt2x00/rt61pci.c rt2x00_desc_write(txd, 11, word); txd 1836 drivers/net/wireless/ralink/rt2x00/rt61pci.c word = rt2x00_desc_read(txd, 0); txd 1859 drivers/net/wireless/ralink/rt2x00/rt61pci.c rt2x00_desc_write(txd, 0, word); txd 1864 drivers/net/wireless/ralink/rt2x00/rt61pci.c skbdesc->desc = txd; txd 1448 drivers/net/wireless/ralink/rt2x00/rt73usb.c __le32 *txd = (__le32 *) entry->skb->data; txd 1454 drivers/net/wireless/ralink/rt2x00/rt73usb.c word = rt2x00_desc_read(txd, 0); txd 1478 drivers/net/wireless/ralink/rt2x00/rt73usb.c rt2x00_desc_write(txd, 0, word); txd 1480 drivers/net/wireless/ralink/rt2x00/rt73usb.c word = rt2x00_desc_read(txd, 1); txd 1488 drivers/net/wireless/ralink/rt2x00/rt73usb.c rt2x00_desc_write(txd, 1, word); txd 1490 drivers/net/wireless/ralink/rt2x00/rt73usb.c word = rt2x00_desc_read(txd, 2); txd 1497 drivers/net/wireless/ralink/rt2x00/rt73usb.c rt2x00_desc_write(txd, 2, word); txd 1500 drivers/net/wireless/ralink/rt2x00/rt73usb.c _rt2x00_desc_write(txd, 3, skbdesc->iv[0]); txd 1501 drivers/net/wireless/ralink/rt2x00/rt73usb.c _rt2x00_desc_write(txd, 4, skbdesc->iv[1]); txd 1504 drivers/net/wireless/ralink/rt2x00/rt73usb.c word = rt2x00_desc_read(txd, 5); txd 1508 drivers/net/wireless/ralink/rt2x00/rt73usb.c rt2x00_desc_write(txd, 5, word); txd 1514 drivers/net/wireless/ralink/rt2x00/rt73usb.c skbdesc->desc = txd; txd 1524 drivers/ntb/ntb_transport.c struct dma_async_tx_descriptor *txd; txd 1560 drivers/ntb/ntb_transport.c txd = device->device_prep_dma_memcpy(chan, unmap->addr[1], txd 1563 drivers/ntb/ntb_transport.c if (!txd) txd 1566 drivers/ntb/ntb_transport.c txd->callback_result = ntb_rx_copy_callback; txd 1567 drivers/ntb/ntb_transport.c txd->callback_param = entry; txd 1568 drivers/ntb/ntb_transport.c dma_set_unmap(txd, unmap); txd 1570 drivers/ntb/ntb_transport.c cookie = dmaengine_submit(txd); txd 1804 drivers/ntb/ntb_transport.c struct dma_async_tx_descriptor *txd; txd 1834 drivers/ntb/ntb_transport.c txd = device->device_prep_dma_memcpy(chan, dest, unmap->addr[0], len, txd 1836 drivers/ntb/ntb_transport.c if (!txd) txd 1839 drivers/ntb/ntb_transport.c txd->callback_result = ntb_tx_copy_callback; txd 1840 drivers/ntb/ntb_transport.c txd->callback_param = entry; txd 1841 drivers/ntb/ntb_transport.c dma_set_unmap(txd, unmap); txd 1843 drivers/ntb/ntb_transport.c cookie = dmaengine_submit(txd); txd 675 drivers/rapidio/devices/tsi721.h struct dma_async_tx_descriptor txd; txd 58 drivers/rapidio/devices/tsi721_dma.c struct tsi721_tx_desc *to_tsi721_desc(struct dma_async_tx_descriptor *txd) txd 60 drivers/rapidio/devices/tsi721_dma.c return container_of(txd, struct tsi721_tx_desc, txd); txd 375 drivers/rapidio/devices/tsi721_dma.c struct dma_async_tx_descriptor *txd = &desc->txd; txd 376 drivers/rapidio/devices/tsi721_dma.c dma_async_tx_callback callback = txd->callback; txd 377 drivers/rapidio/devices/tsi721_dma.c void *param = txd->callback_param; txd 411 drivers/rapidio/devices/tsi721_dma.c struct dma_chan *dchan = desc->txd.chan; txd 641 drivers/rapidio/devices/tsi721_dma.c dma_cookie_complete(&desc->txd); txd 667 drivers/rapidio/devices/tsi721_dma.c dma_cookie_complete(&desc->txd); txd 668 drivers/rapidio/devices/tsi721_dma.c if (desc->txd.flags & DMA_PREP_INTERRUPT) { txd 669 drivers/rapidio/devices/tsi721_dma.c callback = desc->txd.callback; txd 670 drivers/rapidio/devices/tsi721_dma.c param = desc->txd.callback_param; txd 691 drivers/rapidio/devices/tsi721_dma.c static dma_cookie_t tsi721_tx_submit(struct dma_async_tx_descriptor *txd) txd 693 drivers/rapidio/devices/tsi721_dma.c struct tsi721_tx_desc *desc = to_tsi721_desc(txd); txd 694 drivers/rapidio/devices/tsi721_dma.c struct tsi721_bdma_chan *bdma_chan = to_tsi721_chan(txd->chan); txd 701 drivers/rapidio/devices/tsi721_dma.c bdma_chan->id, txd); txd 712 drivers/rapidio/devices/tsi721_dma.c cookie = dma_cookie_assign(txd); txd 750 drivers/rapidio/devices/tsi721_dma.c dma_async_tx_descriptor_init(&desc[i].txd, dchan); txd 751 drivers/rapidio/devices/tsi721_dma.c desc[i].txd.tx_submit = tsi721_tx_submit; txd 752 drivers/rapidio/devices/tsi721_dma.c desc[i].txd.flags = DMA_CTRL_ACK; txd 833 drivers/rapidio/devices/tsi721_dma.c struct dma_async_tx_descriptor *txd = NULL; txd 878 drivers/rapidio/devices/tsi721_dma.c txd = &desc->txd; txd 879 drivers/rapidio/devices/tsi721_dma.c txd->flags = flags; txd 884 drivers/rapidio/devices/tsi721_dma.c if (!txd) { txd 890 drivers/rapidio/devices/tsi721_dma.c return txd; txd 78 drivers/spi/spi-altera.c unsigned int txd = 0; txd 83 drivers/spi/spi-altera.c txd = hw->tx[hw->count]; txd 86 drivers/spi/spi-altera.c txd = (hw->tx[hw->count * 2] txd 92 drivers/spi/spi-altera.c writel(txd, hw->base + ALTERA_SPI_TXDATA); txd 273 drivers/spi/spi-ep93xx.c struct dma_async_tx_descriptor *txd; txd 355 drivers/spi/spi-ep93xx.c txd = dmaengine_prep_slave_sg(chan, sgt->sgl, nents, conf.direction, txd 357 drivers/spi/spi-ep93xx.c if (!txd) { txd 361 drivers/spi/spi-ep93xx.c return txd; txd 403 drivers/spi/spi-ep93xx.c struct dma_async_tx_descriptor *rxd, *txd; txd 411 drivers/spi/spi-ep93xx.c txd = ep93xx_spi_dma_prepare(master, DMA_TO_DEVICE); txd 412 drivers/spi/spi-ep93xx.c if (IS_ERR(txd)) { txd 414 drivers/spi/spi-ep93xx.c dev_err(&master->dev, "DMA TX failed: %ld\n", PTR_ERR(txd)); txd 415 drivers/spi/spi-ep93xx.c return PTR_ERR(txd); txd 424 drivers/spi/spi-ep93xx.c dmaengine_submit(txd); txd 96 drivers/spi/spi-ppc4xx.c u8 txd; txd 159 drivers/spi/spi-ppc4xx.c out_8(&hw->regs->txd, data); txd 336 drivers/spi/spi-ppc4xx.c out_8(&hw->regs->txd, data); txd 266 drivers/video/fbdev/mx3fb.c struct dma_async_tx_descriptor *txd; txd 394 drivers/video/fbdev/mx3fb.c if (mx3_fbi->txd) txd 396 drivers/video/fbdev/mx3fb.c to_tx_desc(mx3_fbi->txd), to_tx_desc(mx3_fbi->txd)->sg); txd 402 drivers/video/fbdev/mx3fb.c mx3_fbi->txd = dmaengine_prep_slave_sg(dma_chan, txd 404 drivers/video/fbdev/mx3fb.c if (!mx3_fbi->txd) { txd 410 drivers/video/fbdev/mx3fb.c mx3_fbi->txd->callback_param = mx3_fbi->txd; txd 411 drivers/video/fbdev/mx3fb.c mx3_fbi->txd->callback = mx3fb_dma_done; txd 413 drivers/video/fbdev/mx3fb.c cookie = mx3_fbi->txd->tx_submit(mx3_fbi->txd); txd 415 drivers/video/fbdev/mx3fb.c mx3_fbi->txd, cookie, list_empty(&ichan->queue) ? '-' : '+'); txd 417 drivers/video/fbdev/mx3fb.c if (!mx3_fbi->txd || !mx3_fbi->txd->tx_submit) { txd 427 drivers/video/fbdev/mx3fb.c mx3_fbi->txd, cookie, list_empty(&ichan->queue) ? '-' : '+'); txd 451 drivers/video/fbdev/mx3fb.c if (mx3_fbi->txd == NULL) txd 460 drivers/video/fbdev/mx3fb.c dmaengine_terminate_all(mx3_fbi->txd->chan); txd 461 drivers/video/fbdev/mx3fb.c mx3_fbi->txd = NULL; txd 759 drivers/video/fbdev/mx3fb.c struct dma_chan *chan = tx_desc->txd.chan; txd 804 drivers/video/fbdev/mx3fb.c if (mx3_fbi->txd) txd 1143 drivers/video/fbdev/mx3fb.c struct dma_async_tx_descriptor *txd; txd 1199 drivers/video/fbdev/mx3fb.c if (mx3_fbi->txd) txd 1200 drivers/video/fbdev/mx3fb.c async_tx_ack(mx3_fbi->txd); txd 1202 drivers/video/fbdev/mx3fb.c txd = dmaengine_prep_slave_sg(dma_chan, sg + txd 1204 drivers/video/fbdev/mx3fb.c if (!txd) { txd 1211 drivers/video/fbdev/mx3fb.c txd->callback_param = txd; txd 1212 drivers/video/fbdev/mx3fb.c txd->callback = mx3fb_dma_done; txd 1218 drivers/video/fbdev/mx3fb.c cookie = txd->tx_submit(txd); txd 1219 drivers/video/fbdev/mx3fb.c dev_dbg(fbi->device, "%d: Submit %p #%d\n", __LINE__, txd, cookie); txd 1228 drivers/video/fbdev/mx3fb.c mx3_fbi->txd = txd; txd 145 include/linux/dma/ipu-dma.h struct dma_async_tx_descriptor txd; txd 171 include/linux/dma/ipu-dma.h #define to_tx_desc(tx) container_of(tx, struct idmac_tx_desc, txd) txd 549 include/linux/dmaengine.h static inline void txd_lock(struct dma_async_tx_descriptor *txd) txd 552 include/linux/dmaengine.h static inline void txd_unlock(struct dma_async_tx_descriptor *txd) txd 555 include/linux/dmaengine.h static inline void txd_chain(struct dma_async_tx_descriptor *txd, struct dma_async_tx_descriptor *next) txd 559 include/linux/dmaengine.h static inline void txd_clear_parent(struct dma_async_tx_descriptor *txd) txd 562 include/linux/dmaengine.h static inline void txd_clear_next(struct dma_async_tx_descriptor *txd) txd 565 include/linux/dmaengine.h static inline struct dma_async_tx_descriptor *txd_next(struct dma_async_tx_descriptor *txd) txd 569 include/linux/dmaengine.h static inline struct dma_async_tx_descriptor *txd_parent(struct dma_async_tx_descriptor *txd) txd 575 include/linux/dmaengine.h static inline void txd_lock(struct dma_async_tx_descriptor *txd) txd 577 include/linux/dmaengine.h spin_lock_bh(&txd->lock); txd 579 include/linux/dmaengine.h static inline void txd_unlock(struct dma_async_tx_descriptor *txd) txd 581 include/linux/dmaengine.h spin_unlock_bh(&txd->lock); txd 583 include/linux/dmaengine.h static inline void txd_chain(struct dma_async_tx_descriptor *txd, struct dma_async_tx_descriptor *next) txd 585 include/linux/dmaengine.h txd->next = next; txd 586 include/linux/dmaengine.h next->parent = txd; txd 588 include/linux/dmaengine.h static inline void txd_clear_parent(struct dma_async_tx_descriptor *txd) txd 590 include/linux/dmaengine.h txd->parent = NULL; txd 592 include/linux/dmaengine.h static inline void txd_clear_next(struct dma_async_tx_descriptor *txd) txd 594 include/linux/dmaengine.h txd->next = NULL; txd 596 include/linux/dmaengine.h static inline struct dma_async_tx_descriptor *txd_parent(struct dma_async_tx_descriptor *txd) txd 598 include/linux/dmaengine.h return txd->parent; txd 600 include/linux/dmaengine.h static inline struct dma_async_tx_descriptor *txd_next(struct dma_async_tx_descriptor *txd) txd 602 include/linux/dmaengine.h return txd->next;