txd               238 arch/arm/mach-imx/mach-pcm043.c 	iomux_v3_cfg_t txd = MX35_PAD_STXD4__AUDMUX_AUD4_TXD;
txd               266 arch/arm/mach-imx/mach-pcm043.c 	mxc_iomux_v3_setup_pad(txd);
txd               704 drivers/ata/pata_ep93xx.c 	struct dma_async_tx_descriptor *txd;
txd               712 drivers/ata/pata_ep93xx.c 	txd = dmaengine_prep_slave_sg(channel, qc->sg, qc->n_elem, qc->dma_dir,
txd               714 drivers/ata/pata_ep93xx.c 	if (!txd) {
txd               718 drivers/ata/pata_ep93xx.c 	txd->callback = NULL;
txd               719 drivers/ata/pata_ep93xx.c 	txd->callback_param = NULL;
txd               721 drivers/ata/pata_ep93xx.c 	if (dmaengine_submit(txd) < 0) {
txd               532 drivers/dma/amba-pl08x.c 	struct pl08x_txd *txd = to_pl08x_txd(&vd->tx);
txd               535 drivers/dma/amba-pl08x.c 	list_del(&txd->vd.node);
txd               537 drivers/dma/amba-pl08x.c 	plchan->at = txd;
txd               543 drivers/dma/amba-pl08x.c 	pl08x_write_lli(pl08x, phychan, &txd->llis_va[0], txd->ccfg);
txd               759 drivers/dma/amba-pl08x.c 	struct pl08x_txd *txd;
txd               765 drivers/dma/amba-pl08x.c 	txd = plchan->at;
txd               767 drivers/dma/amba-pl08x.c 	if (!ch || !txd)
txd               782 drivers/dma/amba-pl08x.c 	llis_va = txd->llis_va;
txd               783 drivers/dma/amba-pl08x.c 	llis_bus = txd->llis_bus;
txd              1107 drivers/dma/amba-pl08x.c 	struct pl08x_txd *txd;
txd              1173 drivers/dma/amba-pl08x.c 	u32 *llis_va = bd->txd->llis_va + offset;
txd              1174 drivers/dma/amba-pl08x.c 	dma_addr_t llis_bus = bd->txd->llis_bus;
txd              1258 drivers/dma/amba-pl08x.c 			      struct pl08x_txd *txd)
txd              1268 drivers/dma/amba-pl08x.c 	txd->llis_va = dma_pool_alloc(pl08x->pool, GFP_NOWAIT, &txd->llis_bus);
txd              1269 drivers/dma/amba-pl08x.c 	if (!txd->llis_va) {
txd              1274 drivers/dma/amba-pl08x.c 	bd.txd = txd;
txd              1276 drivers/dma/amba-pl08x.c 	cctl = txd->cctl;
txd              1284 drivers/dma/amba-pl08x.c 	list_for_each_entry(dsg, &txd->dsg_list, node) {
txd              1286 drivers/dma/amba-pl08x.c 		cctl = txd->cctl;
txd              1337 drivers/dma/amba-pl08x.c 				fc = (txd->ccfg & PL080_CONFIG_FLOW_CONTROL_MASK) >>
txd              1471 drivers/dma/amba-pl08x.c 	llis_va = txd->llis_va;
txd              1474 drivers/dma/amba-pl08x.c 	if (txd->cyclic) {
txd              1476 drivers/dma/amba-pl08x.c 		last_lli[PL080_LLI_LLI] = txd->llis_bus | bd.lli_bus;
txd              1493 drivers/dma/amba-pl08x.c 			   struct pl08x_txd *txd)
txd              1497 drivers/dma/amba-pl08x.c 	if (txd->llis_va)
txd              1498 drivers/dma/amba-pl08x.c 		dma_pool_free(pl08x->pool, txd->llis_va, txd->llis_bus);
txd              1500 drivers/dma/amba-pl08x.c 	list_for_each_entry_safe(dsg, _dsg, &txd->dsg_list, node) {
txd              1505 drivers/dma/amba-pl08x.c 	kfree(txd);
txd              1510 drivers/dma/amba-pl08x.c 	struct pl08x_txd *txd = to_pl08x_txd(&vd->tx);
txd              1514 drivers/dma/amba-pl08x.c 	if (!txd->done)
txd              1517 drivers/dma/amba-pl08x.c 	pl08x_free_txd(plchan->host, txd);
txd              1580 drivers/dma/amba-pl08x.c 			struct pl08x_txd *txd = to_pl08x_txd(&vd->tx);
txd              1583 drivers/dma/amba-pl08x.c 			list_for_each_entry(dsg, &txd->dsg_list, node)
txd              1754 drivers/dma/amba-pl08x.c 	struct pl08x_txd *txd = kzalloc(sizeof(*txd), GFP_NOWAIT);
txd              1756 drivers/dma/amba-pl08x.c 	if (txd)
txd              1757 drivers/dma/amba-pl08x.c 		INIT_LIST_HEAD(&txd->dsg_list);
txd              1758 drivers/dma/amba-pl08x.c 	return txd;
txd              1895 drivers/dma/amba-pl08x.c 	struct pl08x_txd *txd;
txd              1899 drivers/dma/amba-pl08x.c 	txd = pl08x_get_txd(plchan);
txd              1900 drivers/dma/amba-pl08x.c 	if (!txd) {
txd              1908 drivers/dma/amba-pl08x.c 		pl08x_free_txd(pl08x, txd);
txd              1911 drivers/dma/amba-pl08x.c 	list_add_tail(&dsg->node, &txd->dsg_list);
txd              1918 drivers/dma/amba-pl08x.c 		txd->ccfg = 0;
txd              1919 drivers/dma/amba-pl08x.c 		txd->cctl = pl08x_ftdmac020_memcpy_cctl(pl08x);
txd              1921 drivers/dma/amba-pl08x.c 		txd->ccfg = PL080_CONFIG_ERR_IRQ_MASK |
txd              1924 drivers/dma/amba-pl08x.c 		txd->cctl = pl08x_memcpy_cctl(pl08x);
txd              1927 drivers/dma/amba-pl08x.c 	ret = pl08x_fill_llis_for_desc(plchan->host, txd);
txd              1929 drivers/dma/amba-pl08x.c 		pl08x_free_txd(pl08x, txd);
txd              1933 drivers/dma/amba-pl08x.c 	return vchan_tx_prep(&plchan->vc, &txd->vd, flags);
txd              1943 drivers/dma/amba-pl08x.c 	struct pl08x_txd *txd;
txd              1949 drivers/dma/amba-pl08x.c 	txd = pl08x_get_txd(plchan);
txd              1950 drivers/dma/amba-pl08x.c 	if (!txd) {
txd              1975 drivers/dma/amba-pl08x.c 		pl08x_free_txd(pl08x, txd);
txd              1983 drivers/dma/amba-pl08x.c 		pl08x_free_txd(pl08x, txd);
txd              1989 drivers/dma/amba-pl08x.c 	txd->cctl = cctl | pl08x_select_bus(false, src_buses, dst_buses);
txd              1998 drivers/dma/amba-pl08x.c 	txd->ccfg = PL080_CONFIG_ERR_IRQ_MASK |
txd              2004 drivers/dma/amba-pl08x.c 		pl08x_free_txd(pl08x, txd);
txd              2016 drivers/dma/amba-pl08x.c 		txd->ccfg |= plchan->signal << PL080_CONFIG_DST_SEL_SHIFT;
txd              2018 drivers/dma/amba-pl08x.c 		txd->ccfg |= plchan->signal << PL080_CONFIG_SRC_SEL_SHIFT;
txd              2020 drivers/dma/amba-pl08x.c 	return txd;
txd              2023 drivers/dma/amba-pl08x.c static int pl08x_tx_add_sg(struct pl08x_txd *txd,
txd              2035 drivers/dma/amba-pl08x.c 	list_add_tail(&dsg->node, &txd->dsg_list);
txd              2056 drivers/dma/amba-pl08x.c 	struct pl08x_txd *txd;
txd              2064 drivers/dma/amba-pl08x.c 	txd = pl08x_init_txd(chan, direction, &slave_addr);
txd              2065 drivers/dma/amba-pl08x.c 	if (!txd)
txd              2069 drivers/dma/amba-pl08x.c 		ret = pl08x_tx_add_sg(txd, direction, slave_addr,
txd              2074 drivers/dma/amba-pl08x.c 			pl08x_free_txd(pl08x, txd);
txd              2081 drivers/dma/amba-pl08x.c 	ret = pl08x_fill_llis_for_desc(plchan->host, txd);
txd              2084 drivers/dma/amba-pl08x.c 		pl08x_free_txd(pl08x, txd);
txd              2088 drivers/dma/amba-pl08x.c 	return vchan_tx_prep(&plchan->vc, &txd->vd, flags);
txd              2098 drivers/dma/amba-pl08x.c 	struct pl08x_txd *txd;
txd              2108 drivers/dma/amba-pl08x.c 	txd = pl08x_init_txd(chan, direction, &slave_addr);
txd              2109 drivers/dma/amba-pl08x.c 	if (!txd)
txd              2112 drivers/dma/amba-pl08x.c 	txd->cyclic = true;
txd              2113 drivers/dma/amba-pl08x.c 	txd->cctl |= PL080_CONTROL_TC_IRQ_EN;
txd              2115 drivers/dma/amba-pl08x.c 		ret = pl08x_tx_add_sg(txd, direction, slave_addr,
txd              2119 drivers/dma/amba-pl08x.c 			pl08x_free_txd(pl08x, txd);
txd              2124 drivers/dma/amba-pl08x.c 	ret = pl08x_fill_llis_for_desc(plchan->host, txd);
txd              2127 drivers/dma/amba-pl08x.c 		pl08x_free_txd(pl08x, txd);
txd              2131 drivers/dma/amba-pl08x.c 	return vchan_tx_prep(&plchan->vc, &txd->vd, flags);
txd               112 drivers/dma/at_hdmac.c 		dma_async_tx_descriptor_init(&desc->txd, chan);
txd               114 drivers/dma/at_hdmac.c 		desc->txd.flags = DMA_CTRL_ACK;
txd               115 drivers/dma/at_hdmac.c 		desc->txd.tx_submit = atc_tx_submit;
txd               116 drivers/dma/at_hdmac.c 		desc->txd.phys = phys;
txd               136 drivers/dma/at_hdmac.c 		if (async_tx_test_ack(&desc->txd)) {
txd               203 drivers/dma/at_hdmac.c 		(*prev)->lli.dscr = desc->txd.phys;
txd               244 drivers/dma/at_hdmac.c 	channel_writel(atchan, DSCR, first->txd.phys);
txd               265 drivers/dma/at_hdmac.c 		if (desc->txd.cookie == cookie)
txd               270 drivers/dma/at_hdmac.c 		if (desc->txd.cookie == cookie)
txd               443 drivers/dma/at_hdmac.c 	struct dma_async_tx_descriptor	*txd = &desc->txd;
txd               447 drivers/dma/at_hdmac.c 		"descriptor %u complete\n", txd->cookie);
txd               451 drivers/dma/at_hdmac.c 		dma_cookie_complete(txd);
txd               465 drivers/dma/at_hdmac.c 	dma_descriptor_unmap(txd);
txd               473 drivers/dma/at_hdmac.c 		dmaengine_desc_get_callback_invoke(txd, NULL);
txd               476 drivers/dma/at_hdmac.c 	dma_run_dependencies(txd);
txd               571 drivers/dma/at_hdmac.c 			"  cookie: %d\n", bad_desc->txd.cookie);
txd               589 drivers/dma/at_hdmac.c 	struct dma_async_tx_descriptor	*txd = &first->txd;
txd               595 drivers/dma/at_hdmac.c 	dmaengine_desc_get_callback_invoke(txd, NULL);
txd               679 drivers/dma/at_hdmac.c 				desc->txd.cookie);
txd               684 drivers/dma/at_hdmac.c 				desc->txd.cookie);
txd               781 drivers/dma/at_hdmac.c 	desc->txd.cookie = -EBUSY;
txd               787 drivers/dma/at_hdmac.c 	desc->txd.flags = flags; /* client is in control of this ack */
txd               789 drivers/dma/at_hdmac.c 	return &desc->txd;
txd               850 drivers/dma/at_hdmac.c 		desc->txd.cookie = 0;
txd               857 drivers/dma/at_hdmac.c 	first->txd.cookie = -EBUSY;
txd               863 drivers/dma/at_hdmac.c 	first->txd.flags = flags; /* client is in control of this ack */
txd               865 drivers/dma/at_hdmac.c 	return &first->txd;
txd               906 drivers/dma/at_hdmac.c 	desc->txd.cookie = 0;
txd               962 drivers/dma/at_hdmac.c 	desc->txd.cookie = -EBUSY;
txd               968 drivers/dma/at_hdmac.c 	desc->txd.flags = flags;
txd               970 drivers/dma/at_hdmac.c 	return &desc->txd;
txd              1039 drivers/dma/at_hdmac.c 	first->txd.cookie = -EBUSY;
txd              1045 drivers/dma/at_hdmac.c 	first->txd.flags = flags;
txd              1047 drivers/dma/at_hdmac.c 	return &first->txd;
txd              1186 drivers/dma/at_hdmac.c 	first->txd.cookie = -EBUSY;
txd              1190 drivers/dma/at_hdmac.c 	first->txd.flags = flags; /* client is in control of this ack */
txd              1192 drivers/dma/at_hdmac.c 	return &first->txd;
txd              1342 drivers/dma/at_hdmac.c 	prev->lli.dscr = first->txd.phys;
txd              1345 drivers/dma/at_hdmac.c 	first->txd.cookie = -EBUSY;
txd              1348 drivers/dma/at_hdmac.c 	return &first->txd;
txd              1628 drivers/dma/at_hdmac.c 		dma_pool_free(atdma->dma_desc_pool, desc, desc->txd.phys);
txd               190 drivers/dma/at_hdmac_regs.h 	struct dma_async_tx_descriptor	txd;
txd               207 drivers/dma/at_hdmac_regs.h txd_to_at_desc(struct dma_async_tx_descriptor *txd)
txd               209 drivers/dma/at_hdmac_regs.h 	return container_of(txd, struct at_desc, txd);
txd               274 drivers/dma/at_xdmac.c static inline struct at_xdmac_desc *txd_to_at_desc(struct dma_async_tx_descriptor *txd)
txd               276 drivers/dma/at_xdmac.c 	return container_of(txd, struct at_xdmac_desc, tx_dma_desc);
txd              1569 drivers/dma/at_xdmac.c 	struct dma_async_tx_descriptor	*txd;
txd              1574 drivers/dma/at_xdmac.c 		txd = &desc->tx_dma_desc;
txd              1576 drivers/dma/at_xdmac.c 		if (txd->flags & DMA_PREP_INTERRUPT)
txd              1577 drivers/dma/at_xdmac.c 			dmaengine_desc_get_callback_invoke(txd, NULL);
txd              1638 drivers/dma/at_xdmac.c 		struct dma_async_tx_descriptor  *txd;
txd              1654 drivers/dma/at_xdmac.c 		txd = &desc->tx_dma_desc;
txd              1660 drivers/dma/at_xdmac.c 			dma_cookie_complete(txd);
txd              1661 drivers/dma/at_xdmac.c 			if (txd->flags & DMA_PREP_INTERRUPT)
txd              1662 drivers/dma/at_xdmac.c 				dmaengine_desc_get_callback_invoke(txd, NULL);
txd              1665 drivers/dma/at_xdmac.c 		dma_run_dependencies(txd);
txd                75 drivers/dma/dw/core.c 		 __func__, desc->txd.cookie);
txd                92 drivers/dma/dw/core.c 	dma_async_tx_descriptor_init(&desc->txd, &dwc->chan);
txd                93 drivers/dma/dw/core.c 	desc->txd.tx_submit = dwc_tx_submit;
txd                94 drivers/dma/dw/core.c 	desc->txd.flags = DMA_CTRL_ACK;
txd                95 drivers/dma/dw/core.c 	desc->txd.phys = phys;
txd               109 drivers/dma/dw/core.c 		dma_pool_free(dw->desc_pool, child, child->txd.phys);
txd               113 drivers/dma/dw/core.c 	dma_pool_free(dw->desc_pool, desc, desc->txd.phys);
txd               218 drivers/dma/dw/core.c 	channel_writel(dwc, LLP, first->txd.phys | lms);
txd               233 drivers/dma/dw/core.c 	dev_vdbg(chan2dev(&dwc->chan), "%s: started %u\n", __func__, desc->txd.cookie);
txd               243 drivers/dma/dw/core.c 	struct dma_async_tx_descriptor	*txd = &desc->txd;
txd               248 drivers/dma/dw/core.c 	dev_vdbg(chan2dev(&dwc->chan), "descriptor %u complete\n", txd->cookie);
txd               251 drivers/dma/dw/core.c 	dma_cookie_complete(txd);
txd               253 drivers/dma/dw/core.c 		dmaengine_desc_get_callback(txd, &cb);
txd               259 drivers/dma/dw/core.c 		async_tx_ack(&child->txd);
txd               260 drivers/dma/dw/core.c 	async_tx_ack(&desc->txd);
txd               375 drivers/dma/dw/core.c 		if (desc->txd.phys == DWC_LLP_LOC(llp)) {
txd               460 drivers/dma/dw/core.c 				       "  cookie: %d\n", bad_desc->txd.cookie);
txd               599 drivers/dma/dw/core.c 			lli_write(prev, llp, desc->txd.phys | lms);
txd               611 drivers/dma/dw/core.c 	first->txd.flags = flags;
txd               614 drivers/dma/dw/core.c 	return &first->txd;
txd               689 drivers/dma/dw/core.c 				lli_write(prev, llp, desc->txd.phys | lms);
txd               738 drivers/dma/dw/core.c 				lli_write(prev, llp, desc->txd.phys | lms);
txd               763 drivers/dma/dw/core.c 	return &first->txd;
txd               885 drivers/dma/dw/core.c 		if (desc->txd.cookie == c)
txd               394 drivers/dma/dw/regs.h 	struct dma_async_tx_descriptor	txd;
txd               403 drivers/dma/dw/regs.h txd_to_dw_desc(struct dma_async_tx_descriptor *txd)
txd               405 drivers/dma/dw/regs.h 	return container_of(txd, struct dw_desc, txd);
txd               127 drivers/dma/ep93xx_dma.c 	struct dma_async_tx_descriptor	txd;
txd               256 drivers/dma/ep93xx_dma.c 		d->txd.callback = desc->txd.callback;
txd               257 drivers/dma/ep93xx_dma.c 		d->txd.callback_param = desc->txd.callback_param;
txd               300 drivers/dma/ep93xx_dma.c 	return !desc->txd.cookie;
txd               428 drivers/dma/ep93xx_dma.c 			desc->txd.cookie, desc->src_addr, desc->dst_addr,
txd               619 drivers/dma/ep93xx_dma.c 	last_done = !desc || desc->txd.cookie;
txd               683 drivers/dma/ep93xx_dma.c 		if (async_tx_test_ack(&desc->txd)) {
txd               691 drivers/dma/ep93xx_dma.c 			desc->txd.cookie = 0;
txd               692 drivers/dma/ep93xx_dma.c 			desc->txd.callback = NULL;
txd               693 drivers/dma/ep93xx_dma.c 			desc->txd.callback_param = NULL;
txd               765 drivers/dma/ep93xx_dma.c 				dma_cookie_complete(&desc->txd);
txd               768 drivers/dma/ep93xx_dma.c 		dmaengine_desc_get_callback(&desc->txd, &cb);
txd               777 drivers/dma/ep93xx_dma.c 		dma_descriptor_unmap(&desc->txd);
txd               839 drivers/dma/ep93xx_dma.c 	desc = container_of(tx, struct ep93xx_dma_desc, txd);
txd               925 drivers/dma/ep93xx_dma.c 		dma_async_tx_descriptor_init(&desc->txd, chan);
txd               926 drivers/dma/ep93xx_dma.c 		desc->txd.flags = DMA_CTRL_ACK;
txd               927 drivers/dma/ep93xx_dma.c 		desc->txd.tx_submit = ep93xx_dma_tx_submit;
txd              1012 drivers/dma/ep93xx_dma.c 	first->txd.cookie = -EBUSY;
txd              1013 drivers/dma/ep93xx_dma.c 	first->txd.flags = flags;
txd              1015 drivers/dma/ep93xx_dma.c 	return &first->txd;
txd              1087 drivers/dma/ep93xx_dma.c 	first->txd.cookie = -EBUSY;
txd              1088 drivers/dma/ep93xx_dma.c 	first->txd.flags = flags;
txd              1090 drivers/dma/ep93xx_dma.c 	return &first->txd;
txd              1168 drivers/dma/ep93xx_dma.c 	first->txd.cookie = -EBUSY;
txd              1170 drivers/dma/ep93xx_dma.c 	return &first->txd;
txd               509 drivers/dma/fsldma.c 	struct dma_async_tx_descriptor *txd = &desc->async_tx;
txd               512 drivers/dma/fsldma.c 	BUG_ON(txd->cookie < 0);
txd               514 drivers/dma/fsldma.c 	if (txd->cookie > 0) {
txd               515 drivers/dma/fsldma.c 		ret = txd->cookie;
txd               517 drivers/dma/fsldma.c 		dma_descriptor_unmap(txd);
txd               519 drivers/dma/fsldma.c 		dmaengine_desc_get_callback_invoke(txd, NULL);
txd               523 drivers/dma/fsldma.c 	dma_run_dependencies(txd);
txd               220 drivers/dma/ioat/dma.c 	async_tx_ack(&desc->txd);
txd               221 drivers/dma/ioat/dma.c 	ioat_set_chainaddr(ioat_chan, desc->txd.phys);
txd               253 drivers/dma/ioat/dma.c 		ioat_set_chainaddr(ioat_chan, desc->txd.phys);
txd               347 drivers/dma/ioat/dma.c 	dma_async_tx_descriptor_init(&desc->txd, chan);
txd               348 drivers/dma/ioat/dma.c 	desc->txd.tx_submit = ioat_tx_submit_unlock;
txd               350 drivers/dma/ioat/dma.c 	desc->txd.phys = phys;
txd               426 drivers/dma/ioat/dma.c 		hw->next = next->txd.phys;
txd               428 drivers/dma/ioat/dma.c 	ring[i]->hw->next = ring[0]->txd.phys;
txd               609 drivers/dma/ioat/dma.c 		tx = &desc->txd;
txd               727 drivers/dma/ioat/dma.c 		tx = &desc->txd;
txd               756 drivers/dma/ioat/dma.c 	ioat_chan->last_completion = *ioat_chan->completion = desc->txd.phys;
txd               832 drivers/dma/ioat/dma.c 	tx = &desc->txd;
txd               842 drivers/dma/ioat/dma.c 	*ioat_chan->completion = desc->txd.phys;
txd               184 drivers/dma/ioat/dma.h 	struct dma_async_tx_descriptor txd;
txd               230 drivers/dma/ioat/dma.h 	({ if (d) __dump_desc_dbg(c, d->hw, &d->txd, desc_id(d)); 0; })
txd               139 drivers/dma/ioat/prep.c 	desc->txd.flags = flags;
txd               147 drivers/dma/ioat/prep.c 	return &desc->txd;
txd               223 drivers/dma/ioat/prep.c 	desc->txd.flags = flags;
txd               231 drivers/dma/ioat/prep.c 	compl_desc->txd.flags = flags & DMA_PREP_INTERRUPT;
txd               241 drivers/dma/ioat/prep.c 	return &compl_desc->txd;
txd               289 drivers/dma/ioat/prep.c 		desc_id(desc), (unsigned long long) desc->txd.phys,
txd               291 drivers/dma/ioat/prep.c 		desc->txd.flags, pq->size, pq->ctl, pq->ctl_f.op,
txd               322 drivers/dma/ioat/prep.c 		desc_id(desc), (unsigned long long) desc->txd.phys,
txd               324 drivers/dma/ioat/prep.c 		desc->txd.flags, pq->size, pq->ctl,
txd               433 drivers/dma/ioat/prep.c 	desc->txd.flags = flags;
txd               447 drivers/dma/ioat/prep.c 		compl_desc->txd.flags = flags & DMA_PREP_INTERRUPT;
txd               459 drivers/dma/ioat/prep.c 	return &compl_desc->txd;
txd               547 drivers/dma/ioat/prep.c 	desc->txd.flags = flags;
txd               560 drivers/dma/ioat/prep.c 	return &desc->txd;
txd               729 drivers/dma/ioat/prep.c 	desc->txd.flags = flags;
txd               735 drivers/dma/ioat/prep.c 	return &desc->txd;
txd               779 drivers/dma/ipu/ipu_idmac.c 	if (async_tx_test_ack(&desc->txd))
txd               926 drivers/dma/ipu/ipu_idmac.c 		struct dma_async_tx_descriptor *txd = &desc->txd;
txd               928 drivers/dma/ipu/ipu_idmac.c 		memset(txd, 0, sizeof(*txd));
txd               929 drivers/dma/ipu/ipu_idmac.c 		dma_async_tx_descriptor_init(txd, &ichan->dma_chan);
txd               930 drivers/dma/ipu/ipu_idmac.c 		txd->tx_submit		= idmac_tx_submit;
txd              1278 drivers/dma/ipu/ipu_idmac.c 		dmaengine_desc_get_callback(&descnew->txd, &cb);
txd              1290 drivers/dma/ipu/ipu_idmac.c 		dma_cookie_complete(&desc->txd);
txd              1292 drivers/dma/ipu/ipu_idmac.c 	dmaengine_desc_get_callback(&desc->txd, &cb);
txd              1296 drivers/dma/ipu/ipu_idmac.c 	if (done && (desc->txd.flags & DMA_PREP_INTERRUPT))
txd              1317 drivers/dma/ipu/ipu_idmac.c 			if (async_tx_test_ack(&desc->txd)) {
txd              1325 drivers/dma/ipu/ipu_idmac.c 				async_tx_clear_ack(&desc->txd);
txd              1340 drivers/dma/ipu/ipu_idmac.c 	struct dma_async_tx_descriptor *txd = NULL;
txd              1364 drivers/dma/ipu/ipu_idmac.c 		txd		= &desc->txd;
txd              1365 drivers/dma/ipu/ipu_idmac.c 		txd->flags	= tx_flags;
txd              1373 drivers/dma/ipu/ipu_idmac.c 	return txd;
txd              1450 drivers/dma/ipu/ipu_idmac.c 			async_tx_clear_ack(&desc->txd);
txd               166 drivers/dma/mediatek/mtk-hsdma.c 	struct mtk_hsdma_pdesc *txd;
txd               327 drivers/dma/mediatek/mtk-hsdma.c 	pc->sz_ring = 2 * MTK_DMA_SIZE * sizeof(*ring->txd);
txd               328 drivers/dma/mediatek/mtk-hsdma.c 	ring->txd = dma_alloc_coherent(hsdma2dev(hsdma), pc->sz_ring,
txd               330 drivers/dma/mediatek/mtk-hsdma.c 	if (!ring->txd)
txd               333 drivers/dma/mediatek/mtk-hsdma.c 	ring->rxd = &ring->txd[MTK_DMA_SIZE];
txd               334 drivers/dma/mediatek/mtk-hsdma.c 	ring->rphys = ring->tphys + MTK_DMA_SIZE * sizeof(*ring->txd);
txd               384 drivers/dma/mediatek/mtk-hsdma.c 			  pc->sz_ring, ring->txd, ring->tphys);
txd               409 drivers/dma/mediatek/mtk-hsdma.c 			  pc->sz_ring, ring->txd, ring->tphys);
txd               417 drivers/dma/mediatek/mtk-hsdma.c 	struct mtk_hsdma_pdesc *txd, *rxd;
txd               450 drivers/dma/mediatek/mtk-hsdma.c 		txd = &ring->txd[ring->cur_tptr];
txd               451 drivers/dma/mediatek/mtk-hsdma.c 		WRITE_ONCE(txd->desc1, hvd->src);
txd               452 drivers/dma/mediatek/mtk-hsdma.c 		WRITE_ONCE(txd->desc2,
txd               930 drivers/dma/mmp_pdma.c 		struct dma_async_tx_descriptor *txd = &desc->async_tx;
txd               935 drivers/dma/mmp_pdma.c 		dmaengine_desc_get_callback(txd, &cb);
txd               938 drivers/dma/mmp_pdma.c 		dma_pool_free(chan->desc_pool, desc, txd->phys);
txd               460 drivers/dma/mpc512x_dma.c static dma_cookie_t mpc_dma_tx_submit(struct dma_async_tx_descriptor *txd)
txd               462 drivers/dma/mpc512x_dma.c 	struct mpc_dma_chan *mchan = dma_chan_to_mpc_dma_chan(txd->chan);
txd               467 drivers/dma/mpc512x_dma.c 	mdesc = container_of(txd, struct mpc_dma_desc, desc);
txd               479 drivers/dma/mpc512x_dma.c 	cookie = dma_cookie_assign(txd);
txd               196 drivers/dma/owl-dma.c 	struct owl_dma_txd	*txd;
txd               343 drivers/dma/owl-dma.c static struct owl_dma_lli *owl_dma_add_lli(struct owl_dma_txd *txd,
txd               349 drivers/dma/owl-dma.c 		list_add_tail(&next->node, &txd->lli_list);
txd               500 drivers/dma/owl-dma.c 	struct owl_dma_txd *txd = to_owl_txd(&vd->tx);
txd               507 drivers/dma/owl-dma.c 	vchan->txd = txd;
txd               513 drivers/dma/owl-dma.c 	lli = list_first_entry(&txd->lli_list,
txd               516 drivers/dma/owl-dma.c 	if (txd->cyclic)
txd               601 drivers/dma/owl-dma.c 		struct owl_dma_txd *txd;
txd               614 drivers/dma/owl-dma.c 		txd = vchan->txd;
txd               615 drivers/dma/owl-dma.c 		if (txd) {
txd               616 drivers/dma/owl-dma.c 			vchan->txd = NULL;
txd               618 drivers/dma/owl-dma.c 			vchan_cookie_complete(&txd->vd);
txd               636 drivers/dma/owl-dma.c static void owl_dma_free_txd(struct owl_dma *od, struct owl_dma_txd *txd)
txd               640 drivers/dma/owl-dma.c 	if (unlikely(!txd))
txd               643 drivers/dma/owl-dma.c 	list_for_each_entry_safe(lli, _lli, &txd->lli_list, node)
txd               646 drivers/dma/owl-dma.c 	kfree(txd);
txd               652 drivers/dma/owl-dma.c 	struct owl_dma_txd *txd = to_owl_txd(&vd->tx);
txd               654 drivers/dma/owl-dma.c 	owl_dma_free_txd(od, txd);
txd               669 drivers/dma/owl-dma.c 	if (vchan->txd) {
txd               670 drivers/dma/owl-dma.c 		owl_dma_desc_free(&vchan->txd->vd);
txd               671 drivers/dma/owl-dma.c 		vchan->txd = NULL;
txd               716 drivers/dma/owl-dma.c 	if (!vchan->pchan && !vchan->txd)
txd               733 drivers/dma/owl-dma.c 	struct owl_dma_txd *txd;
txd               739 drivers/dma/owl-dma.c 	txd = vchan->txd;
txd               741 drivers/dma/owl-dma.c 	if (!pchan || !txd)
txd               750 drivers/dma/owl-dma.c 		list_for_each_entry(lli, &txd->lli_list, node) {
txd               753 drivers/dma/owl-dma.c 				list_for_each_entry(lli, &txd->lli_list, node)
txd               770 drivers/dma/owl-dma.c 	struct owl_dma_txd *txd;
txd               783 drivers/dma/owl-dma.c 		txd = to_owl_txd(&vd->tx);
txd               784 drivers/dma/owl-dma.c 		list_for_each_entry(lli, &txd->lli_list, node)
txd               832 drivers/dma/owl-dma.c 	struct owl_dma_txd *txd;
txd               840 drivers/dma/owl-dma.c 	txd = kzalloc(sizeof(*txd), GFP_NOWAIT);
txd               841 drivers/dma/owl-dma.c 	if (!txd)
txd               844 drivers/dma/owl-dma.c 	INIT_LIST_HEAD(&txd->lli_list);
txd               858 drivers/dma/owl-dma.c 				      &vchan->cfg, txd->cyclic);
txd               864 drivers/dma/owl-dma.c 		prev = owl_dma_add_lli(txd, prev, lli, false);
txd               867 drivers/dma/owl-dma.c 	return vchan_tx_prep(&vchan->vc, &txd->vd, flags);
txd               870 drivers/dma/owl-dma.c 	owl_dma_free_txd(od, txd);
txd               884 drivers/dma/owl-dma.c 	struct owl_dma_txd *txd;
txd               891 drivers/dma/owl-dma.c 	txd = kzalloc(sizeof(*txd), GFP_NOWAIT);
txd               892 drivers/dma/owl-dma.c 	if (!txd)
txd               895 drivers/dma/owl-dma.c 	INIT_LIST_HEAD(&txd->lli_list);
txd               922 drivers/dma/owl-dma.c 				      txd->cyclic);
txd               928 drivers/dma/owl-dma.c 		prev = owl_dma_add_lli(txd, prev, lli, false);
txd               931 drivers/dma/owl-dma.c 	return vchan_tx_prep(&vchan->vc, &txd->vd, flags);
txd               934 drivers/dma/owl-dma.c 	owl_dma_free_txd(od, txd);
txd               949 drivers/dma/owl-dma.c 	struct owl_dma_txd *txd;
txd               955 drivers/dma/owl-dma.c 	txd = kzalloc(sizeof(*txd), GFP_NOWAIT);
txd               956 drivers/dma/owl-dma.c 	if (!txd)
txd               959 drivers/dma/owl-dma.c 	INIT_LIST_HEAD(&txd->lli_list);
txd               960 drivers/dma/owl-dma.c 	txd->cyclic = true;
txd               978 drivers/dma/owl-dma.c 				      dir, sconfig, txd->cyclic);
txd               987 drivers/dma/owl-dma.c 		prev = owl_dma_add_lli(txd, prev, lli, false);
txd               991 drivers/dma/owl-dma.c 	owl_dma_add_lli(txd, prev, first, true);
txd               993 drivers/dma/owl-dma.c 	return vchan_tx_prep(&vchan->vc, &txd->vd, flags);
txd               996 drivers/dma/owl-dma.c 	owl_dma_free_txd(od, txd);
txd                85 drivers/dma/pch_dma.c 	struct dma_async_tx_descriptor txd;
txd               138 drivers/dma/pch_dma.c struct pch_dma_desc *to_pd_desc(struct dma_async_tx_descriptor *txd)
txd               140 drivers/dma/pch_dma.c 	return container_of(txd, struct pch_dma_desc, txd);
txd               343 drivers/dma/pch_dma.c 		channel_writel(pd_chan, NEXT, desc->txd.phys);
txd               351 drivers/dma/pch_dma.c 	struct dma_async_tx_descriptor *txd = &desc->txd;
txd               354 drivers/dma/pch_dma.c 	dmaengine_desc_get_callback(txd, &cb);
txd               392 drivers/dma/pch_dma.c 		 bad_desc->txd.cookie);
txd               408 drivers/dma/pch_dma.c static dma_cookie_t pd_tx_submit(struct dma_async_tx_descriptor *txd)
txd               410 drivers/dma/pch_dma.c 	struct pch_dma_desc *desc = to_pd_desc(txd);
txd               411 drivers/dma/pch_dma.c 	struct pch_dma_chan *pd_chan = to_pd_chan(txd->chan);
txd               435 drivers/dma/pch_dma.c 		dma_async_tx_descriptor_init(&desc->txd, chan);
txd               436 drivers/dma/pch_dma.c 		desc->txd.tx_submit = pd_tx_submit;
txd               437 drivers/dma/pch_dma.c 		desc->txd.flags = DMA_CTRL_ACK;
txd               438 drivers/dma/pch_dma.c 		desc->txd.phys = addr;
txd               453 drivers/dma/pch_dma.c 		if (async_tx_test_ack(&desc->txd)) {
txd               544 drivers/dma/pch_dma.c 		dma_pool_free(pd->pool, desc, desc->txd.phys);
txd               629 drivers/dma/pch_dma.c 			prev->regs.next |= desc->txd.phys;
txd               641 drivers/dma/pch_dma.c 	first->txd.cookie = -EBUSY;
txd               642 drivers/dma/pch_dma.c 	desc->txd.flags = flags;
txd               644 drivers/dma/pch_dma.c 	return &first->txd;
txd               520 drivers/dma/pl330.c 	struct dma_async_tx_descriptor txd;
txd              2027 drivers/dma/pl330.c 	return container_of(tx, struct dma_pl330_desc, txd);
txd              2051 drivers/dma/pl330.c 					__func__, __LINE__, desc->txd.cookie);
txd              2070 drivers/dma/pl330.c 				dma_cookie_complete(&desc->txd);
txd              2096 drivers/dma/pl330.c 		dmaengine_desc_get_callback(&desc->txd, &cb);
txd              2113 drivers/dma/pl330.c 		dma_descriptor_unmap(&desc->txd);
txd              2294 drivers/dma/pl330.c 		dma_cookie_complete(&desc->txd);
txd              2299 drivers/dma/pl330.c 		dma_cookie_complete(&desc->txd);
txd              2436 drivers/dma/pl330.c 		if (desc->txd.cookie == cookie) {
txd              2502 drivers/dma/pl330.c 			desc->txd.callback = last->txd.callback;
txd              2503 drivers/dma/pl330.c 			desc->txd.callback_param = last->txd.callback_param;
txd              2507 drivers/dma/pl330.c 		dma_cookie_assign(&desc->txd);
txd              2513 drivers/dma/pl330.c 	cookie = dma_cookie_assign(&last->txd);
txd              2525 drivers/dma/pl330.c 	desc->txd.tx_submit = pl330_tx_submit;
txd              2569 drivers/dma/pl330.c 		desc->txd.callback = NULL;
txd              2600 drivers/dma/pl330.c 	desc->txd.cookie = 0;
txd              2601 drivers/dma/pl330.c 	async_tx_ack(&desc->txd);
txd              2606 drivers/dma/pl330.c 	dma_async_tx_descriptor_init(&desc->txd, &pch->chan);
txd              2749 drivers/dma/pl330.c 	desc->txd.flags = flags;
txd              2751 drivers/dma/pl330.c 	return &desc->txd;
txd              2801 drivers/dma/pl330.c 	desc->txd.flags = flags;
txd              2803 drivers/dma/pl330.c 	return &desc->txd;
txd              2886 drivers/dma/pl330.c 	desc->txd.flags = flg;
txd              2887 drivers/dma/pl330.c 	return &desc->txd;
txd               317 drivers/dma/qcom/hidma.c static dma_cookie_t hidma_tx_submit(struct dma_async_tx_descriptor *txd)
txd               319 drivers/dma/qcom/hidma.c 	struct hidma_chan *mchan = to_hidma_chan(txd->chan);
txd               334 drivers/dma/qcom/hidma.c 	mdesc = container_of(txd, struct hidma_desc, desc);
txd               341 drivers/dma/qcom/hidma.c 	cookie = dma_cookie_assign(txd);
txd               495 drivers/dma/qcom/hidma.c 		struct dma_async_tx_descriptor *txd = &mdesc->desc;
txd               497 drivers/dma/qcom/hidma.c 		dma_descriptor_unmap(txd);
txd               498 drivers/dma/qcom/hidma.c 		dmaengine_desc_get_callback_invoke(txd, NULL);
txd               499 drivers/dma/qcom/hidma.c 		dma_run_dependencies(txd);
txd               373 drivers/dma/s3c24xx-dma.c 	struct s3c24xx_txd *txd = s3cchan->at;
txd               376 drivers/dma/s3c24xx-dma.c 	return tc * txd->width;
txd               417 drivers/dma/s3c24xx-dma.c 	struct s3c24xx_txd *txd = kzalloc(sizeof(*txd), GFP_NOWAIT);
txd               419 drivers/dma/s3c24xx-dma.c 	if (txd) {
txd               420 drivers/dma/s3c24xx-dma.c 		INIT_LIST_HEAD(&txd->dsg_list);
txd               421 drivers/dma/s3c24xx-dma.c 		txd->dcon = S3C24XX_DCON_INT | S3C24XX_DCON_NORELOAD;
txd               424 drivers/dma/s3c24xx-dma.c 	return txd;
txd               427 drivers/dma/s3c24xx-dma.c static void s3c24xx_dma_free_txd(struct s3c24xx_txd *txd)
txd               431 drivers/dma/s3c24xx-dma.c 	list_for_each_entry_safe(dsg, _dsg, &txd->dsg_list, node) {
txd               436 drivers/dma/s3c24xx-dma.c 	kfree(txd);
txd               440 drivers/dma/s3c24xx-dma.c 				       struct s3c24xx_txd *txd)
txd               445 drivers/dma/s3c24xx-dma.c 	struct s3c24xx_sg *dsg = list_entry(txd->at, struct s3c24xx_sg, node);
txd               446 drivers/dma/s3c24xx-dma.c 	u32 dcon = txd->dcon;
txd               450 drivers/dma/s3c24xx-dma.c 	switch (txd->width) {
txd               484 drivers/dma/s3c24xx-dma.c 	writel_relaxed(txd->disrcc, phy->base + S3C24XX_DISRCC);
txd               486 drivers/dma/s3c24xx-dma.c 	writel_relaxed(txd->didstc, phy->base + S3C24XX_DIDSTC);
txd               507 drivers/dma/s3c24xx-dma.c 	struct s3c24xx_txd *txd = to_s3c24xx_txd(&vd->tx);
txd               509 drivers/dma/s3c24xx-dma.c 	list_del(&txd->vd.node);
txd               511 drivers/dma/s3c24xx-dma.c 	s3cchan->at = txd;
txd               518 drivers/dma/s3c24xx-dma.c 	txd->at = txd->dsg_list.next;
txd               519 drivers/dma/s3c24xx-dma.c 	s3c24xx_dma_start_next_sg(s3cchan, txd);
txd               637 drivers/dma/s3c24xx-dma.c 	struct s3c24xx_txd *txd = to_s3c24xx_txd(&vd->tx);
txd               643 drivers/dma/s3c24xx-dma.c 	s3c24xx_dma_free_txd(txd);
txd               650 drivers/dma/s3c24xx-dma.c 	struct s3c24xx_txd *txd;
txd               671 drivers/dma/s3c24xx-dma.c 	txd = s3cchan->at;
txd               672 drivers/dma/s3c24xx-dma.c 	if (txd) {
txd               674 drivers/dma/s3c24xx-dma.c 		if (!list_is_last(txd->at, &txd->dsg_list)) {
txd               675 drivers/dma/s3c24xx-dma.c 			txd->at = txd->at->next;
txd               676 drivers/dma/s3c24xx-dma.c 			if (txd->cyclic)
txd               677 drivers/dma/s3c24xx-dma.c 				vchan_cyclic_callback(&txd->vd);
txd               678 drivers/dma/s3c24xx-dma.c 			s3c24xx_dma_start_next_sg(s3cchan, txd);
txd               679 drivers/dma/s3c24xx-dma.c 		} else if (!txd->cyclic) {
txd               681 drivers/dma/s3c24xx-dma.c 			vchan_cookie_complete(&txd->vd);
txd               692 drivers/dma/s3c24xx-dma.c 			vchan_cyclic_callback(&txd->vd);
txd               695 drivers/dma/s3c24xx-dma.c 			txd->at = txd->dsg_list.next;
txd               696 drivers/dma/s3c24xx-dma.c 			s3c24xx_dma_start_next_sg(s3cchan, txd);
txd               761 drivers/dma/s3c24xx-dma.c 	struct s3c24xx_txd *txd;
txd               783 drivers/dma/s3c24xx-dma.c 		txd = to_s3c24xx_txd(&vd->tx);
txd               785 drivers/dma/s3c24xx-dma.c 		list_for_each_entry(dsg, &txd->dsg_list, node)
txd               792 drivers/dma/s3c24xx-dma.c 		txd = s3cchan->at;
txd               794 drivers/dma/s3c24xx-dma.c 		dsg = list_entry(txd->at, struct s3c24xx_sg, node);
txd               795 drivers/dma/s3c24xx-dma.c 		list_for_each_entry_from(dsg, &txd->dsg_list, node)
txd               821 drivers/dma/s3c24xx-dma.c 	struct s3c24xx_txd *txd;
txd               833 drivers/dma/s3c24xx-dma.c 	txd = s3c24xx_dma_get_txd();
txd               834 drivers/dma/s3c24xx-dma.c 	if (!txd)
txd               839 drivers/dma/s3c24xx-dma.c 		s3c24xx_dma_free_txd(txd);
txd               842 drivers/dma/s3c24xx-dma.c 	list_add_tail(&dsg->node, &txd->dsg_list);
txd               858 drivers/dma/s3c24xx-dma.c 		txd->width = (src_mod == 0 && dest_mod == 0) ? 4 : 1;
txd               861 drivers/dma/s3c24xx-dma.c 		txd->width = ((src_mod == 2 || src_mod == 0) &&
txd               865 drivers/dma/s3c24xx-dma.c 		txd->width = 1;
txd               869 drivers/dma/s3c24xx-dma.c 	txd->disrcc = S3C24XX_DISRCC_LOC_AHB | S3C24XX_DISRCC_INC_INCREMENT;
txd               870 drivers/dma/s3c24xx-dma.c 	txd->didstc = S3C24XX_DIDSTC_LOC_AHB | S3C24XX_DIDSTC_INC_INCREMENT;
txd               871 drivers/dma/s3c24xx-dma.c 	txd->dcon |= S3C24XX_DCON_DEMAND | S3C24XX_DCON_SYNC_HCLK |
txd               874 drivers/dma/s3c24xx-dma.c 	return vchan_tx_prep(&s3cchan->vc, &txd->vd, flags);
txd               885 drivers/dma/s3c24xx-dma.c 	struct s3c24xx_txd *txd;
txd               902 drivers/dma/s3c24xx-dma.c 	txd = s3c24xx_dma_get_txd();
txd               903 drivers/dma/s3c24xx-dma.c 	if (!txd)
txd               906 drivers/dma/s3c24xx-dma.c 	txd->cyclic = 1;
txd               909 drivers/dma/s3c24xx-dma.c 		txd->dcon |= S3C24XX_DCON_HANDSHAKE;
txd               913 drivers/dma/s3c24xx-dma.c 		txd->dcon |= S3C24XX_DCON_SYNC_PCLK;
txd               917 drivers/dma/s3c24xx-dma.c 		txd->dcon |= S3C24XX_DCON_SYNC_HCLK;
txd               932 drivers/dma/s3c24xx-dma.c 	txd->dcon |= S3C24XX_DCON_SERV_SINGLE;
txd               935 drivers/dma/s3c24xx-dma.c 		txd->disrcc = S3C24XX_DISRCC_LOC_AHB |
txd               937 drivers/dma/s3c24xx-dma.c 		txd->didstc = hwcfg;
txd               939 drivers/dma/s3c24xx-dma.c 		txd->width = s3cchan->cfg.dst_addr_width;
txd               941 drivers/dma/s3c24xx-dma.c 		txd->disrcc = hwcfg;
txd               942 drivers/dma/s3c24xx-dma.c 		txd->didstc = S3C24XX_DIDSTC_LOC_AHB |
txd               945 drivers/dma/s3c24xx-dma.c 		txd->width = s3cchan->cfg.src_addr_width;
txd               953 drivers/dma/s3c24xx-dma.c 			s3c24xx_dma_free_txd(txd);
txd               956 drivers/dma/s3c24xx-dma.c 		list_add_tail(&dsg->node, &txd->dsg_list);
txd               971 drivers/dma/s3c24xx-dma.c 	return vchan_tx_prep(&s3cchan->vc, &txd->vd, flags);
txd               983 drivers/dma/s3c24xx-dma.c 	struct s3c24xx_txd *txd;
txd               993 drivers/dma/s3c24xx-dma.c 	txd = s3c24xx_dma_get_txd();
txd               994 drivers/dma/s3c24xx-dma.c 	if (!txd)
txd               998 drivers/dma/s3c24xx-dma.c 		txd->dcon |= S3C24XX_DCON_HANDSHAKE;
txd              1002 drivers/dma/s3c24xx-dma.c 		txd->dcon |= S3C24XX_DCON_SYNC_PCLK;
txd              1006 drivers/dma/s3c24xx-dma.c 		txd->dcon |= S3C24XX_DCON_SYNC_HCLK;
txd              1021 drivers/dma/s3c24xx-dma.c 	txd->dcon |= S3C24XX_DCON_SERV_SINGLE;
txd              1024 drivers/dma/s3c24xx-dma.c 		txd->disrcc = S3C24XX_DISRCC_LOC_AHB |
txd              1026 drivers/dma/s3c24xx-dma.c 		txd->didstc = hwcfg;
txd              1028 drivers/dma/s3c24xx-dma.c 		txd->width = s3cchan->cfg.dst_addr_width;
txd              1030 drivers/dma/s3c24xx-dma.c 		txd->disrcc = hwcfg;
txd              1031 drivers/dma/s3c24xx-dma.c 		txd->didstc = S3C24XX_DIDSTC_LOC_AHB |
txd              1034 drivers/dma/s3c24xx-dma.c 		txd->width = s3cchan->cfg.src_addr_width;
txd              1036 drivers/dma/s3c24xx-dma.c 		s3c24xx_dma_free_txd(txd);
txd              1045 drivers/dma/s3c24xx-dma.c 			s3c24xx_dma_free_txd(txd);
txd              1048 drivers/dma/s3c24xx-dma.c 		list_add_tail(&dsg->node, &txd->dsg_list);
txd              1060 drivers/dma/s3c24xx-dma.c 	return vchan_tx_prep(&s3cchan->vc, &txd->vd, flags);
txd               148 drivers/dma/sa11x0-dma.c static void sa11x0_dma_start_desc(struct sa11x0_dma_phy *p, struct sa11x0_dma_desc *txd)
txd               150 drivers/dma/sa11x0-dma.c 	list_del(&txd->vd.node);
txd               151 drivers/dma/sa11x0-dma.c 	p->txd_load = txd;
txd               155 drivers/dma/sa11x0-dma.c 		p->num, &txd->vd, txd->vd.tx.cookie, txd->ddar);
txd               161 drivers/dma/sa11x0-dma.c 	struct sa11x0_dma_desc *txd = p->txd_load;
txd               167 drivers/dma/sa11x0-dma.c 	if (!txd)
txd               176 drivers/dma/sa11x0-dma.c 	if (p->sg_load == txd->sglen) {
txd               177 drivers/dma/sa11x0-dma.c 		if (!txd->cyclic) {
txd               185 drivers/dma/sa11x0-dma.c 			if (txn && txn->ddar == txd->ddar) {
txd               186 drivers/dma/sa11x0-dma.c 				txd = txn;
txd               198 drivers/dma/sa11x0-dma.c 	sg = &txd->sg[p->sg_load++];
txd               225 drivers/dma/sa11x0-dma.c 	struct sa11x0_dma_desc *txd = p->txd_done;
txd               227 drivers/dma/sa11x0-dma.c 	if (++p->sg_done == txd->sglen) {
txd               228 drivers/dma/sa11x0-dma.c 		if (!txd->cyclic) {
txd               229 drivers/dma/sa11x0-dma.c 			vchan_cookie_complete(&txd->vd);
txd               237 drivers/dma/sa11x0-dma.c 			if ((p->sg_done % txd->period) == 0)
txd               238 drivers/dma/sa11x0-dma.c 				vchan_cyclic_callback(&txd->vd);
txd               301 drivers/dma/sa11x0-dma.c 	struct sa11x0_dma_desc *txd = sa11x0_dma_next_desc(c);
txd               304 drivers/dma/sa11x0-dma.c 	if (txd) {
txd               307 drivers/dma/sa11x0-dma.c 		sa11x0_dma_start_desc(p, txd);
txd               308 drivers/dma/sa11x0-dma.c 		p->txd_done = txd;
txd               318 drivers/dma/sa11x0-dma.c 		writel_relaxed(txd->ddar, p->base + DMA_DDAR);
txd               447 drivers/dma/sa11x0-dma.c 		struct sa11x0_dma_desc *txd;
txd               451 drivers/dma/sa11x0-dma.c 			txd = p->txd_done;
txd               453 drivers/dma/sa11x0-dma.c 			txd = p->txd_load;
txd               455 drivers/dma/sa11x0-dma.c 			txd = NULL;
txd               458 drivers/dma/sa11x0-dma.c 		if (txd) {
txd               464 drivers/dma/sa11x0-dma.c 			for (i = 0; i < txd->sglen; i++) {
txd               466 drivers/dma/sa11x0-dma.c 					i, txd->sg[i].addr, txd->sg[i].len);
txd               467 drivers/dma/sa11x0-dma.c 				if (addr >= txd->sg[i].addr &&
txd               468 drivers/dma/sa11x0-dma.c 				    addr < txd->sg[i].addr + txd->sg[i].len) {
txd               471 drivers/dma/sa11x0-dma.c 					len = txd->sg[i].len -
txd               472 drivers/dma/sa11x0-dma.c 						(addr - txd->sg[i].addr);
txd               480 drivers/dma/sa11x0-dma.c 			for (; i < txd->sglen; i++) {
txd               482 drivers/dma/sa11x0-dma.c 					i, txd->sg[i].addr, txd->sg[i].len);
txd               483 drivers/dma/sa11x0-dma.c 				bytes += txd->sg[i].len;
txd               527 drivers/dma/sa11x0-dma.c 	struct sa11x0_dma_desc *txd;
txd               556 drivers/dma/sa11x0-dma.c 	txd = kzalloc(struct_size(txd, sg, j), GFP_ATOMIC);
txd               557 drivers/dma/sa11x0-dma.c 	if (!txd) {
txd               585 drivers/dma/sa11x0-dma.c 			txd->sg[j].addr = addr;
txd               586 drivers/dma/sa11x0-dma.c 			txd->sg[j].len = tlen;
txd               594 drivers/dma/sa11x0-dma.c 	txd->ddar = c->ddar;
txd               595 drivers/dma/sa11x0-dma.c 	txd->size = size;
txd               596 drivers/dma/sa11x0-dma.c 	txd->sglen = j;
txd               599 drivers/dma/sa11x0-dma.c 		&c->vc, &txd->vd, txd->size, txd->sglen);
txd               601 drivers/dma/sa11x0-dma.c 	return vchan_tx_prep(&c->vc, &txd->vd, flags);
txd               609 drivers/dma/sa11x0-dma.c 	struct sa11x0_dma_desc *txd;
txd               626 drivers/dma/sa11x0-dma.c 	txd = kzalloc(struct_size(txd, sg, sglen), GFP_ATOMIC);
txd               627 drivers/dma/sa11x0-dma.c 	if (!txd) {
txd               643 drivers/dma/sa11x0-dma.c 			txd->sg[k].addr = addr;
txd               644 drivers/dma/sa11x0-dma.c 			txd->sg[k].len = tlen;
txd               654 drivers/dma/sa11x0-dma.c 	txd->ddar = c->ddar;
txd               655 drivers/dma/sa11x0-dma.c 	txd->size = size;
txd               656 drivers/dma/sa11x0-dma.c 	txd->sglen = sglen;
txd               657 drivers/dma/sa11x0-dma.c 	txd->cyclic = 1;
txd               658 drivers/dma/sa11x0-dma.c 	txd->period = sgperiod;
txd               660 drivers/dma/sa11x0-dma.c 	return vchan_tx_prep(&c->vc, &txd->vd, DMA_PREP_INTERRUPT | DMA_CTRL_ACK);
txd              1049 drivers/dma/sa11x0-dma.c 		struct sa11x0_dma_desc *txd = NULL;
txd              1055 drivers/dma/sa11x0-dma.c 			txd = p->txd_done;
txd              1057 drivers/dma/sa11x0-dma.c 			txd = p->txd_load;
txd              1059 drivers/dma/sa11x0-dma.c 		if (!txd)
txd              1062 drivers/dma/sa11x0-dma.c 		writel_relaxed(txd->ddar, p->base + DMA_DDAR);
txd               404 drivers/dma/sirf-dma.c static dma_cookie_t sirfsoc_dma_tx_submit(struct dma_async_tx_descriptor *txd)
txd               406 drivers/dma/sirf-dma.c 	struct sirfsoc_dma_chan *schan = dma_chan_to_sirfsoc_dma_chan(txd->chan);
txd               411 drivers/dma/sirf-dma.c 	sdesc = container_of(txd, struct sirfsoc_dma_desc, desc);
txd               418 drivers/dma/sirf-dma.c 	cookie = dma_cookie_assign(txd);
txd               373 drivers/dma/ste_dma40.c 	struct dma_async_tx_descriptor	 txd;
txd               775 drivers/dma/ste_dma40.c 			if (async_tx_test_ack(&d->txd)) {
txd              1470 drivers/dma/ste_dma40.c 	struct d40_desc *d40d = container_of(tx, struct d40_desc, txd);
txd              1594 drivers/dma/ste_dma40.c 		dma_cookie_complete(&d40d->txd);
txd              1606 drivers/dma/ste_dma40.c 	callback_active = !!(d40d->txd.flags & DMA_PREP_INTERRUPT);
txd              1607 drivers/dma/ste_dma40.c 	dmaengine_desc_get_callback(&d40d->txd, &cb);
txd              1610 drivers/dma/ste_dma40.c 		if (async_tx_test_ack(&d40d->txd)) {
txd              2197 drivers/dma/ste_dma40.c 	desc->txd.flags = dma_flags;
txd              2198 drivers/dma/ste_dma40.c 	desc->txd.tx_submit = d40_tx_submit;
txd              2200 drivers/dma/ste_dma40.c 	dma_async_tx_descriptor_init(&desc->txd, &chan->chan);
txd              2264 drivers/dma/ste_dma40.c 	return &desc->txd;
txd              2514 drivers/dma/ste_dma40.c 	struct dma_async_tx_descriptor *txd;
txd              2530 drivers/dma/ste_dma40.c 	txd = d40_prep_sg(chan, sg, sg, periods, direction,
txd              2535 drivers/dma/ste_dma40.c 	return txd;
txd               343 drivers/dma/sun6i-dma.c 	struct sun6i_desc *txd = pchan->desc;
txd               354 drivers/dma/sun6i-dma.c 	for (lli = txd->v_lli; lli; lli = lli->v_lli_next) {
txd               368 drivers/dma/sun6i-dma.c 			       struct sun6i_desc *txd)
txd               370 drivers/dma/sun6i-dma.c 	if ((!prev && !txd) || !next)
txd               374 drivers/dma/sun6i-dma.c 		txd->p_lli = next_phy;
txd               375 drivers/dma/sun6i-dma.c 		txd->v_lli = next;
txd               403 drivers/dma/sun6i-dma.c 	struct sun6i_desc *txd = to_sun6i_desc(&vd->tx);
txd               408 drivers/dma/sun6i-dma.c 	if (unlikely(!txd))
txd               411 drivers/dma/sun6i-dma.c 	p_lli = txd->p_lli;
txd               412 drivers/dma/sun6i-dma.c 	v_lli = txd->v_lli;
txd               424 drivers/dma/sun6i-dma.c 	kfree(txd);
txd               636 drivers/dma/sun6i-dma.c 	struct sun6i_desc *txd;
txd               647 drivers/dma/sun6i-dma.c 	txd = kzalloc(sizeof(*txd), GFP_NOWAIT);
txd               648 drivers/dma/sun6i-dma.c 	if (!txd)
txd               671 drivers/dma/sun6i-dma.c 	sun6i_dma_lli_add(NULL, v_lli, p_lli, txd);
txd               675 drivers/dma/sun6i-dma.c 	return vchan_tx_prep(&vchan->vc, &txd->vd, flags);
txd               678 drivers/dma/sun6i-dma.c 	kfree(txd);
txd               691 drivers/dma/sun6i-dma.c 	struct sun6i_desc *txd;
txd               706 drivers/dma/sun6i-dma.c 	txd = kzalloc(sizeof(*txd), GFP_NOWAIT);
txd               707 drivers/dma/sun6i-dma.c 	if (!txd)
txd               745 drivers/dma/sun6i-dma.c 		prev = sun6i_dma_lli_add(prev, v_lli, p_lli, txd);
txd               748 drivers/dma/sun6i-dma.c 	dev_dbg(chan2dev(chan), "First: %pad\n", &txd->p_lli);
txd               749 drivers/dma/sun6i-dma.c 	for (prev = txd->v_lli; prev; prev = prev->v_lli_next)
txd               752 drivers/dma/sun6i-dma.c 	return vchan_tx_prep(&vchan->vc, &txd->vd, flags);
txd               755 drivers/dma/sun6i-dma.c 	for (prev = txd->v_lli; prev; prev = prev->v_lli_next)
txd               757 drivers/dma/sun6i-dma.c 	kfree(txd);
txd               773 drivers/dma/sun6i-dma.c 	struct sun6i_desc *txd;
txd               785 drivers/dma/sun6i-dma.c 	txd = kzalloc(sizeof(*txd), GFP_NOWAIT);
txd               786 drivers/dma/sun6i-dma.c 	if (!txd)
txd               813 drivers/dma/sun6i-dma.c 		prev = sun6i_dma_lli_add(prev, v_lli, p_lli, txd);
txd               816 drivers/dma/sun6i-dma.c 	prev->p_lli_next = txd->p_lli;		/* cyclic list */
txd               820 drivers/dma/sun6i-dma.c 	return vchan_tx_prep(&vchan->vc, &txd->vd, flags);
txd               823 drivers/dma/sun6i-dma.c 	for (prev = txd->v_lli; prev; prev = prev->v_lli_next)
txd               825 drivers/dma/sun6i-dma.c 	kfree(txd);
txd               935 drivers/dma/sun6i-dma.c 	struct sun6i_desc *txd;
txd               947 drivers/dma/sun6i-dma.c 	txd = to_sun6i_desc(&vd->tx);
txd               950 drivers/dma/sun6i-dma.c 		for (lli = txd->v_lli; lli != NULL; lli = lli->v_lli_next)
txd               164 drivers/dma/tegra20-apb-dma.c 	struct dma_async_tx_descriptor	txd;
txd               261 drivers/dma/tegra20-apb-dma.c 	return container_of(td, struct tegra_dma_desc, txd);
txd               284 drivers/dma/tegra20-apb-dma.c 		if (async_tx_test_ack(&dma_desc->txd) && !dma_desc->cb_count) {
txd               287 drivers/dma/tegra20-apb-dma.c 			dma_desc->txd.flags = 0;
txd               299 drivers/dma/tegra20-apb-dma.c 	dma_async_tx_descriptor_init(&dma_desc->txd, &tdc->dma_chan);
txd               300 drivers/dma/tegra20-apb-dma.c 	dma_desc->txd.tx_submit = tegra_dma_tx_submit;
txd               301 drivers/dma/tegra20-apb-dma.c 	dma_desc->txd.flags = 0;
txd               610 drivers/dma/tegra20-apb-dma.c 		dma_cookie_complete(&dma_desc->txd);
txd               669 drivers/dma/tegra20-apb-dma.c 		dmaengine_desc_get_callback(&dma_desc->txd, &cb);
txd               706 drivers/dma/tegra20-apb-dma.c static dma_cookie_t tegra_dma_tx_submit(struct dma_async_tx_descriptor *txd)
txd               708 drivers/dma/tegra20-apb-dma.c 	struct tegra_dma_desc *dma_desc = txd_to_tegra_dma_desc(txd);
txd               709 drivers/dma/tegra20-apb-dma.c 	struct tegra_dma_channel *tdc = to_tegra_dma_chan(txd->chan);
txd               715 drivers/dma/tegra20-apb-dma.c 	cookie = dma_cookie_assign(&dma_desc->txd);
txd               879 drivers/dma/tegra20-apb-dma.c 		if (dma_desc->txd.cookie == cookie) {
txd               888 drivers/dma/tegra20-apb-dma.c 		if (dma_desc->txd.cookie == cookie) {
txd              1103 drivers/dma/tegra20-apb-dma.c 		dma_desc->txd.flags = DMA_CTRL_ACK;
txd              1120 drivers/dma/tegra20-apb-dma.c 	return &dma_desc->txd;
txd              1239 drivers/dma/tegra20-apb-dma.c 		dma_desc->txd.flags = DMA_CTRL_ACK;
txd              1256 drivers/dma/tegra20-apb-dma.c 	return &dma_desc->txd;
txd                87 drivers/dma/ti/cppi41.c 	struct dma_async_tx_descriptor txd;
txd               355 drivers/dma/ti/cppi41.c 			dma_cookie_complete(&c->txd);
txd               356 drivers/dma/ti/cppi41.c 			dmaengine_desc_get_callback_invoke(&c->txd, NULL);
txd               387 drivers/dma/ti/cppi41.c 	dma_async_tx_descriptor_init(&c->txd, chan);
txd               388 drivers/dma/ti/cppi41.c 	c->txd.tx_submit = cppi41_tx_submit;
txd               589 drivers/dma/ti/cppi41.c 	struct dma_async_tx_descriptor *txd = NULL;
txd               627 drivers/dma/ti/cppi41.c 	txd = &c->txd;
txd               633 drivers/dma/ti/cppi41.c 	return txd;
txd               730 drivers/dma/ti/cppi41.c 	dma_cookie_complete(&c->txd);
txd               731 drivers/dma/ti/cppi41.c 	dmaengine_desc_get_callback_invoke(&c->txd, &abort_result);
txd                64 drivers/dma/timb_dma.c 	struct dma_async_tx_descriptor	txd;
txd               199 drivers/dma/timb_dma.c 		iowrite32(td_desc->txd.phys, td_chan->membase +
txd               209 drivers/dma/timb_dma.c 		iowrite32(td_desc->txd.phys, td_chan->membase +
txd               222 drivers/dma/timb_dma.c 	struct dma_async_tx_descriptor	*txd;
txd               231 drivers/dma/timb_dma.c 	txd = &td_desc->txd;
txd               234 drivers/dma/timb_dma.c 		txd->cookie);
txd               243 drivers/dma/timb_dma.c 	dma_cookie_complete(txd);
txd               246 drivers/dma/timb_dma.c 	dmaengine_desc_get_callback(txd, &cb);
txd               250 drivers/dma/timb_dma.c 	dma_descriptor_unmap(txd);
txd               288 drivers/dma/timb_dma.c 		__func__, td_desc->txd.cookie);
txd               294 drivers/dma/timb_dma.c static dma_cookie_t td_tx_submit(struct dma_async_tx_descriptor *txd)
txd               296 drivers/dma/timb_dma.c 	struct timb_dma_desc *td_desc = container_of(txd, struct timb_dma_desc,
txd               297 drivers/dma/timb_dma.c 		txd);
txd               298 drivers/dma/timb_dma.c 	struct timb_dma_chan *td_chan = container_of(txd->chan,
txd               303 drivers/dma/timb_dma.c 	cookie = dma_cookie_assign(txd);
txd               306 drivers/dma/timb_dma.c 		dev_dbg(chan2dev(txd->chan), "%s: started %u\n", __func__,
txd               307 drivers/dma/timb_dma.c 			txd->cookie);
txd               311 drivers/dma/timb_dma.c 		dev_dbg(chan2dev(txd->chan), "tx_submit: queued %u\n",
txd               312 drivers/dma/timb_dma.c 			txd->cookie);
txd               338 drivers/dma/timb_dma.c 	dma_async_tx_descriptor_init(&td_desc->txd, chan);
txd               339 drivers/dma/timb_dma.c 	td_desc->txd.tx_submit = td_tx_submit;
txd               340 drivers/dma/timb_dma.c 	td_desc->txd.flags = DMA_CTRL_ACK;
txd               342 drivers/dma/timb_dma.c 	td_desc->txd.phys = dma_map_single(chan2dmadev(chan),
txd               345 drivers/dma/timb_dma.c 	err = dma_mapping_error(chan2dmadev(chan), td_desc->txd.phys);
txd               362 drivers/dma/timb_dma.c 	dev_dbg(chan2dev(td_desc->txd.chan), "Freeing desc: %p\n", td_desc);
txd               363 drivers/dma/timb_dma.c 	dma_unmap_single(chan2dmadev(td_desc->txd.chan), td_desc->txd.phys,
txd               388 drivers/dma/timb_dma.c 		if (async_tx_test_ack(&td_desc->txd)) {
txd               539 drivers/dma/timb_dma.c 	dma_sync_single_for_device(chan2dmadev(chan), td_desc->txd.phys,
txd               542 drivers/dma/timb_dma.c 	return &td_desc->txd;
txd               139 drivers/dma/txx9dmac.c txd_to_txx9dmac_desc(struct dma_async_tx_descriptor *txd)
txd               141 drivers/dma/txx9dmac.c 	return container_of(txd, struct txx9dmac_desc, txd);
txd               199 drivers/dma/txx9dmac.c 	dma_async_tx_descriptor_init(&desc->txd, &dc->chan);
txd               200 drivers/dma/txx9dmac.c 	desc->txd.tx_submit = txx9dmac_tx_submit;
txd               202 drivers/dma/txx9dmac.c 	desc->txd.flags = DMA_CTRL_ACK;
txd               203 drivers/dma/txx9dmac.c 	desc->txd.phys = dma_map_single(chan2parent(&dc->chan), &desc->hwdesc,
txd               216 drivers/dma/txx9dmac.c 		if (async_tx_test_ack(&desc->txd)) {
txd               249 drivers/dma/txx9dmac.c 				child->txd.phys, ddev->descsize,
txd               252 drivers/dma/txx9dmac.c 			desc->txd.phys, ddev->descsize,
txd               337 drivers/dma/txx9dmac.c 		 first->txd.cookie, first);
txd               367 drivers/dma/txx9dmac.c 		channel64_write_CHAR(dc, first->txd.phys);
txd               388 drivers/dma/txx9dmac.c 			channel32_writel(dc, CHAR, first->txd.phys);
txd               390 drivers/dma/txx9dmac.c 			channel32_writel(dc, CHAR, first->txd.phys);
txd               403 drivers/dma/txx9dmac.c 	struct dma_async_tx_descriptor *txd = &desc->txd;
txd               406 drivers/dma/txx9dmac.c 		 txd->cookie, desc);
txd               408 drivers/dma/txx9dmac.c 	dma_cookie_complete(txd);
txd               409 drivers/dma/txx9dmac.c 	dmaengine_desc_get_callback(txd, &cb);
txd               415 drivers/dma/txx9dmac.c 	dma_descriptor_unmap(txd);
txd               421 drivers/dma/txx9dmac.c 	dma_run_dependencies(txd);
txd               434 drivers/dma/txx9dmac.c 			desc_write_CHAR(dc, prev, desc->txd.phys);
txd               436 drivers/dma/txx9dmac.c 				prev->txd.phys, ddev->descsize,
txd               442 drivers/dma/txx9dmac.c 		if ((desc->txd.flags & DMA_PREP_INTERRUPT) &&
txd               528 drivers/dma/txx9dmac.c 		 bad_desc->txd.cookie);
txd               700 drivers/dma/txx9dmac.c 		 desc->txd.cookie, desc);
txd               778 drivers/dma/txx9dmac.c 			desc_write_CHAR(dc, prev, desc->txd.phys);
txd               780 drivers/dma/txx9dmac.c 					prev->txd.phys, ddev->descsize,
txd               793 drivers/dma/txx9dmac.c 			prev->txd.phys, ddev->descsize,
txd               796 drivers/dma/txx9dmac.c 	first->txd.flags = flags;
txd               799 drivers/dma/txx9dmac.c 	return &first->txd;
txd               872 drivers/dma/txx9dmac.c 			desc_write_CHAR(dc, prev, desc->txd.phys);
txd               874 drivers/dma/txx9dmac.c 					prev->txd.phys,
txd               888 drivers/dma/txx9dmac.c 			prev->txd.phys, ddev->descsize,
txd               891 drivers/dma/txx9dmac.c 	first->txd.flags = flags;
txd               894 drivers/dma/txx9dmac.c 	return &first->txd;
txd               949 drivers/dma/txx9dmac.c 	desc_write_CHAR(dc, prev, desc->txd.phys);
txd               951 drivers/dma/txx9dmac.c 				   prev->txd.phys, ddev->descsize,
txd               954 drivers/dma/txx9dmac.c 	    channel_read_CHAR(dc) == prev->txd.phys)
txd               956 drivers/dma/txx9dmac.c 		channel_write_CHAR(dc, desc->txd.phys);
txd               975 drivers/dma/txx9dmac.c 			if (!(prev->txd.flags & DMA_PREP_INTERRUPT) ||
txd              1065 drivers/dma/txx9dmac.c 		dma_unmap_single(chan2parent(chan), desc->txd.phys,
txd               231 drivers/dma/txx9dmac.h 	struct dma_async_tx_descriptor	txd;
txd               729 drivers/i2c/busses/i2c-qup.c 	struct dma_async_tx_descriptor *txd, *rxd = NULL;
txd               755 drivers/i2c/busses/i2c-qup.c 	txd = dmaengine_prep_slave_sg(qup->btx.dma, qup->btx.sg, tx_cnt,
txd               758 drivers/i2c/busses/i2c-qup.c 	if (!txd) {
txd               765 drivers/i2c/busses/i2c-qup.c 		txd->callback = qup_i2c_bam_cb;
txd               766 drivers/i2c/busses/i2c-qup.c 		txd->callback_param = qup;
txd               769 drivers/i2c/busses/i2c-qup.c 	cookie_tx = dmaengine_submit(txd);
txd               552 drivers/mailbox/bcm-pdc-mailbox.c 	struct dma64dd *txd = &pdcs->txd_64[pdcs->txout];
txd               558 drivers/mailbox/bcm-pdc-mailbox.c 	txd->addrlow = cpu_to_le32(lower_32_bits(dma_addr));
txd               559 drivers/mailbox/bcm-pdc-mailbox.c 	txd->addrhigh = cpu_to_le32(upper_32_bits(dma_addr));
txd               560 drivers/mailbox/bcm-pdc-mailbox.c 	txd->ctrl1 = cpu_to_le32(flags);
txd               561 drivers/mailbox/bcm-pdc-mailbox.c 	txd->ctrl2 = cpu_to_le32(buf_len);
txd                75 drivers/media/pci/mantis/mantis_i2c.c 	u32 txd = 0, stat, trials;
txd                82 drivers/media/pci/mantis/mantis_i2c.c 		txd = (msg->addr << 25) | (msg->buf[i] << 8)
txd                88 drivers/media/pci/mantis/mantis_i2c.c 			txd &= ~MANTIS_I2C_STOP;
txd                91 drivers/media/pci/mantis/mantis_i2c.c 		mmwrite(txd, MANTIS_I2CDATA_CTL);
txd               119 drivers/media/pci/mantis/mantis_i2c.c 	u32 stat, data, txd;
txd               142 drivers/media/pci/mantis/mantis_i2c.c 			txd = msgs[i].addr << 25 | (0x1 << 24)
txd               146 drivers/media/pci/mantis/mantis_i2c.c 			mmwrite(txd, MANTIS_I2CDATA_CTL);
txd               716 drivers/net/ethernet/3com/typhoon.c 	struct tx_desc *txd, *first_txd;
txd               786 drivers/net/ethernet/3com/typhoon.c 	txd = (struct tx_desc *) (txRing->ringBase + txRing->lastWrite);
txd               795 drivers/net/ethernet/3com/typhoon.c 		txd->flags = TYPHOON_FRAG_DESC | TYPHOON_DESC_VALID;
txd               796 drivers/net/ethernet/3com/typhoon.c 		txd->len = cpu_to_le16(skb->len);
txd               797 drivers/net/ethernet/3com/typhoon.c 		txd->frag.addr = cpu_to_le32(skb_dma);
txd               798 drivers/net/ethernet/3com/typhoon.c 		txd->frag.addrHi = 0;
txd               806 drivers/net/ethernet/3com/typhoon.c 		txd->flags = TYPHOON_FRAG_DESC | TYPHOON_DESC_VALID;
txd               807 drivers/net/ethernet/3com/typhoon.c 		txd->len = cpu_to_le16(len);
txd               808 drivers/net/ethernet/3com/typhoon.c 		txd->frag.addr = cpu_to_le32(skb_dma);
txd               809 drivers/net/ethernet/3com/typhoon.c 		txd->frag.addrHi = 0;
txd               816 drivers/net/ethernet/3com/typhoon.c 			txd = (struct tx_desc *) (txRing->ringBase +
txd               824 drivers/net/ethernet/3com/typhoon.c 			txd->flags = TYPHOON_FRAG_DESC | TYPHOON_DESC_VALID;
txd               825 drivers/net/ethernet/3com/typhoon.c 			txd->len = cpu_to_le16(len);
txd               826 drivers/net/ethernet/3com/typhoon.c 			txd->frag.addr = cpu_to_le32(skb_dma);
txd               827 drivers/net/ethernet/3com/typhoon.c 			txd->frag.addrHi = 0;
txd               436 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c 	struct hw_atl_txd_s *txd = NULL;
txd               446 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c 		txd = (struct hw_atl_txd_s *)&ring->dx_ring[ring->sw_tail *
txd               448 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c 		txd->ctl = 0;
txd               449 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c 		txd->ctl2 = 0;
txd               450 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c 		txd->buf_addr = 0;
txd               455 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c 			txd->ctl |= (buff->len_l3 << 31) |
txd               459 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c 			txd->ctl2 |= (buff->mss << 16) |
txd               469 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c 				txd->ctl |= HW_ATL_A0_TXD_CTL_CMD_IPV6;
txd               473 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c 			txd->buf_addr = buff->pa;
txd               474 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c 			txd->ctl |= (HW_ATL_A0_TXD_CTL_BLEN &
txd               476 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c 			txd->ctl |= HW_ATL_A0_TXD_CTL_DESC_TYPE_TXD;
txd               478 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c 			txd->ctl2 |= HW_ATL_A0_TXD_CTL2_LEN & (pkt_len << 14);
txd               481 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c 				txd->ctl |= HW_ATL_A0_TXD_CTL_CMD_LSO;
txd               482 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c 				txd->ctl2 |= HW_ATL_A0_TXD_CTL2_CTX_EN;
txd               487 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c 				txd->ctl |= HW_ATL_A0_TXD_CTL_CMD_IPCSO;
txd               490 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c 				txd->ctl |= HW_ATL_A0_TXD_CTL_CMD_TUCSO;
txd               493 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c 				txd->ctl |= HW_ATL_A0_TXD_CTL_EOP;
txd               494 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c 				txd->ctl |= HW_ATL_A0_TXD_CTL_CMD_WB;
txd               491 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c 	struct hw_atl_txd_s *txd = NULL;
txd               502 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c 		txd = (struct hw_atl_txd_s *)&ring->dx_ring[ring->sw_tail *
txd               504 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c 		txd->ctl = 0;
txd               505 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c 		txd->ctl2 = 0;
txd               506 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c 		txd->buf_addr = 0;
txd               511 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c 			txd->ctl |= HW_ATL_B0_TXD_CTL_CMD_TCP;
txd               512 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c 			txd->ctl |= HW_ATL_B0_TXD_CTL_DESC_TYPE_TXC;
txd               513 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c 			txd->ctl |= (buff->len_l3 << 31) |
txd               515 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c 			txd->ctl2 |= (buff->mss << 16);
txd               522 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c 				txd->ctl |= HW_ATL_B0_TXD_CTL_CMD_IPV6;
txd               523 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c 			txd->ctl2 |= (buff->len_l4 << 8) |
txd               527 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c 			txd->ctl |= HW_ATL_B0_TXD_CTL_DESC_TYPE_TXC;
txd               528 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c 			txd->ctl |= buff->vlan_tx_tag << 4;
txd               534 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c 			txd->buf_addr = buff->pa;
txd               535 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c 			txd->ctl |= (HW_ATL_B0_TXD_CTL_BLEN &
txd               537 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c 			txd->ctl |= HW_ATL_B0_TXD_CTL_DESC_TYPE_TXD;
txd               540 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c 			txd->ctl2 |= HW_ATL_B0_TXD_CTL2_LEN & (pkt_len << 14);
txd               544 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c 				txd->ctl2 |= HW_ATL_B0_TXD_CTL2_CTX_EN;
txd               547 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c 				txd->ctl |= HW_ATL_B0_TXD_CTL_CMD_LSO;
txd               551 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c 				txd->ctl |= HW_ATL_B0_TXD_CTL_CMD_IPCSO;
txd               554 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c 				txd->ctl |= HW_ATL_B0_TXD_CTL_CMD_TUCSO;
txd               557 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c 				txd->ctl |= HW_ATL_B0_TXD_CTL_CMD_VLAN;
txd               560 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c 				txd->ctl |= HW_ATL_B0_TXD_CTL_EOP;
txd               561 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c 				txd->ctl |= HW_ATL_B0_TXD_CTL_CMD_WB;
txd               390 drivers/net/ethernet/aurora/nb8800.c 	struct nb8800_tx_desc *txd;
txd               425 drivers/net/ethernet/aurora/nb8800.c 	txd = &priv->tx_descs[next];
txd               426 drivers/net/ethernet/aurora/nb8800.c 	desc = &txd->desc[0];
txd               431 drivers/net/ethernet/aurora/nb8800.c 		memcpy(txd->buf, skb->data, align);
txd               435 drivers/net/ethernet/aurora/nb8800.c 		desc->n_addr = txb->dma_desc + sizeof(txd->desc[0]);
txd               496 drivers/net/ethernet/aurora/nb8800.c 		struct nb8800_tx_desc *txd = &priv->tx_descs[done];
txd               500 drivers/net/ethernet/aurora/nb8800.c 		if (!txd->report)
txd               509 drivers/net/ethernet/aurora/nb8800.c 		if (IS_TX_ERROR(txd->report)) {
txd               510 drivers/net/ethernet/aurora/nb8800.c 			nb8800_tx_error(dev, txd->report);
txd               517 drivers/net/ethernet/aurora/nb8800.c 		dev->stats.tx_bytes += TX_BYTES_TRANSFERRED(txd->report);
txd               518 drivers/net/ethernet/aurora/nb8800.c 		dev->stats.collisions += TX_EARLY_COLLISIONS(txd->report);
txd               522 drivers/net/ethernet/aurora/nb8800.c 		txd->report = 0;
txd               785 drivers/net/ethernet/aurora/nb8800.c 	struct nb8800_tx_desc *txd;
txd               809 drivers/net/ethernet/aurora/nb8800.c 		txd = &priv->tx_descs[i];
txd               810 drivers/net/ethernet/aurora/nb8800.c 		txd->desc[0].r_addr = r_dma;
txd               811 drivers/net/ethernet/aurora/nb8800.c 		txd->desc[1].r_addr = r_dma;
txd               812 drivers/net/ethernet/aurora/nb8800.c 		txd->report = 0;
txd               875 drivers/net/ethernet/aurora/nb8800.c 	struct nb8800_tx_desc *txd = &priv->tx_descs[0];
txd               901 drivers/net/ethernet/aurora/nb8800.c 	txd->desc[0].s_addr =
txd               903 drivers/net/ethernet/aurora/nb8800.c 	txd->desc[0].config = DESC_BTS(2) | DESC_DS | DESC_EOF | DESC_EOC | 8;
txd               904 drivers/net/ethernet/aurora/nb8800.c 	memset(txd->buf, 0, sizeof(txd->buf));
txd               334 drivers/net/ethernet/cirrus/ep93xx_eth.c 	struct ep93xx_tdesc *txd;
txd               346 drivers/net/ethernet/cirrus/ep93xx_eth.c 	txd = &ep->descs->tdesc[entry];
txd               348 drivers/net/ethernet/cirrus/ep93xx_eth.c 	txd->tdesc1 = TDESC1_EOF | (entry << 16) | (skb->len & 0xfff);
txd               349 drivers/net/ethernet/cirrus/ep93xx_eth.c 	dma_sync_single_for_cpu(dev->dev.parent, txd->buf_addr, skb->len,
txd               352 drivers/net/ethernet/cirrus/ep93xx_eth.c 	dma_sync_single_for_device(dev->dev.parent, txd->buf_addr, skb->len,
txd               618 drivers/net/ethernet/cortina/gemini.c 	struct gmac_txdesc *txd;
txd               629 drivers/net/ethernet/cortina/gemini.c 		txd = txq->ring + c;
txd               630 drivers/net/ethernet/cortina/gemini.c 		word0 = txd->word0;
txd               631 drivers/net/ethernet/cortina/gemini.c 		word1 = txd->word1;
txd               632 drivers/net/ethernet/cortina/gemini.c 		mapping = txd->word2.buf_adr;
txd               633 drivers/net/ethernet/cortina/gemini.c 		word3 = txd->word3.bits32;
txd               653 drivers/net/ethernet/cortina/gemini.c 		bytes += txd->word1.bits.byte_count;
txd              1147 drivers/net/ethernet/cortina/gemini.c 	struct gmac_txdesc *txd;
txd              1201 drivers/net/ethernet/cortina/gemini.c 		txd = txq->ring + w;
txd              1202 drivers/net/ethernet/cortina/gemini.c 		txd->word0.bits32 = buflen;
txd              1203 drivers/net/ethernet/cortina/gemini.c 		txd->word1.bits32 = word1;
txd              1204 drivers/net/ethernet/cortina/gemini.c 		txd->word2.buf_adr = mapping;
txd              1205 drivers/net/ethernet/cortina/gemini.c 		txd->word3.bits32 = word3;
txd               609 drivers/net/ethernet/dec/tulip/de2104x.c 	struct de_desc *txd;
txd               623 drivers/net/ethernet/dec/tulip/de2104x.c 	txd = &de->tx_ring[entry];
txd               632 drivers/net/ethernet/dec/tulip/de2104x.c 	txd->opts2 = cpu_to_le32(flags);
txd               633 drivers/net/ethernet/dec/tulip/de2104x.c 	txd->addr1 = cpu_to_le32(mapping);
txd               639 drivers/net/ethernet/dec/tulip/de2104x.c 	txd->opts1 = cpu_to_le32(DescOwn);
txd               724 drivers/net/ethernet/dec/tulip/de2104x.c 	struct de_desc *txd;
txd               773 drivers/net/ethernet/dec/tulip/de2104x.c 	txd = &de->tx_ring[entry];
txd               775 drivers/net/ethernet/dec/tulip/de2104x.c 		txd->opts2 = cpu_to_le32(SetupFrame | RingEnd | sizeof (de->setup_frame));
txd               777 drivers/net/ethernet/dec/tulip/de2104x.c 		txd->opts2 = cpu_to_le32(SetupFrame | sizeof (de->setup_frame));
txd               778 drivers/net/ethernet/dec/tulip/de2104x.c 	txd->addr1 = cpu_to_le32(mapping);
txd               781 drivers/net/ethernet/dec/tulip/de2104x.c 	txd->opts1 = cpu_to_le32(DescOwn);
txd                60 drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c 		struct fm10k_tx_desc *txd = FM10K_TX_DESC(ring, i);
txd                63 drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c 			   i, txd->buffer_addr, txd->buflen, txd->vlan,
txd                64 drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c 			   txd->mss, txd->hdrlen, txd->flags);
txd               497 drivers/net/ethernet/intel/i40e/i40e_debugfs.c 	struct i40e_tx_desc *txd;
txd               530 drivers/net/ethernet/intel/i40e/i40e_debugfs.c 				txd = I40E_TX_DESC(ring, i);
txd               533 drivers/net/ethernet/intel/i40e/i40e_debugfs.c 					 i, txd->buffer_addr,
txd               534 drivers/net/ethernet/intel/i40e/i40e_debugfs.c 					 txd->cmd_type_offset_bsz);
txd               551 drivers/net/ethernet/intel/i40e/i40e_debugfs.c 			txd = I40E_TX_DESC(ring, desc_n);
txd               555 drivers/net/ethernet/intel/i40e/i40e_debugfs.c 				 txd->buffer_addr, txd->cmd_type_offset_bsz);
txd              2074 drivers/net/ethernet/marvell/mv643xx_eth.c 		struct tx_desc *txd = tx_desc + i;
txd              2081 drivers/net/ethernet/marvell/mv643xx_eth.c 		txd->cmd_sts = 0;
txd              2082 drivers/net/ethernet/marvell/mv643xx_eth.c 		txd->next_desc_ptr = txq->tx_desc_dma +
txd               829 drivers/net/ethernet/mediatek/mtk_eth_soc.c 						    struct mtk_tx_dma *txd)
txd               831 drivers/net/ethernet/mediatek/mtk_eth_soc.c 	int idx = txd - ring->dma;
txd               885 drivers/net/ethernet/mediatek/mtk_eth_soc.c 			 struct mtk_tx_dma *txd, dma_addr_t mapped_addr,
txd               893 drivers/net/ethernet/mediatek/mtk_eth_soc.c 			txd->txd3 = mapped_addr;
txd               894 drivers/net/ethernet/mediatek/mtk_eth_soc.c 			txd->txd2 |= TX_DMA_PLEN1(size);
txd               899 drivers/net/ethernet/mediatek/mtk_eth_soc.c 			txd->txd1 = mapped_addr;
txd               900 drivers/net/ethernet/mediatek/mtk_eth_soc.c 			txd->txd2 = TX_DMA_PLEN0(size);
txd               912 drivers/net/ethernet/mediatek/mtk_eth_soc.c 	struct mtk_tx_dma *itxd, *txd;
txd               957 drivers/net/ethernet/mediatek/mtk_eth_soc.c 	txd = itxd;
txd               958 drivers/net/ethernet/mediatek/mtk_eth_soc.c 	txd_pdma = qdma_to_pdma(ring, txd);
txd               973 drivers/net/ethernet/mediatek/mtk_eth_soc.c 				txd = mtk_qdma_phys_to_virt(ring, txd->txd2);
txd               974 drivers/net/ethernet/mediatek/mtk_eth_soc.c 				txd_pdma = qdma_to_pdma(ring, txd);
txd               975 drivers/net/ethernet/mediatek/mtk_eth_soc.c 				if (txd == ring->last_free)
txd               995 drivers/net/ethernet/mediatek/mtk_eth_soc.c 			WRITE_ONCE(txd->txd1, mapped_addr);
txd               996 drivers/net/ethernet/mediatek/mtk_eth_soc.c 			WRITE_ONCE(txd->txd3, (TX_DMA_SWC |
txd               999 drivers/net/ethernet/mediatek/mtk_eth_soc.c 			WRITE_ONCE(txd->txd4, fport);
txd              1001 drivers/net/ethernet/mediatek/mtk_eth_soc.c 			tx_buf = mtk_desc_to_tx_buf(ring, txd);
txd              1033 drivers/net/ethernet/mediatek/mtk_eth_soc.c 	ring->next_free = mtk_qdma_phys_to_virt(ring, txd->txd2);
txd              1044 drivers/net/ethernet/mediatek/mtk_eth_soc.c 			mtk_w32(eth, txd->txd2, MTK_QTX_CTX_PTR);
txd              1046 drivers/net/ethernet/mediatek/mtk_eth_soc.c 		int next_idx = NEXT_DESP_IDX(txd_to_idx(ring, txd),
txd              1066 drivers/net/ethernet/mediatek/mtk_eth_soc.c 	} while (itxd != txd);
txd               111 drivers/net/ethernet/micrel/ks8851.c 	u8			txd[8];
txd               233 drivers/net/ethernet/micrel/ks8851.c 	__le16 *txb = (__le16 *)ks->txd;
txd               727 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 			   struct nfp_net_tx_desc *txd, struct sk_buff *skb,
txd               751 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	txd->l3_offset = l3_offset - md_bytes;
txd               752 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	txd->l4_offset = l4_offset - md_bytes;
txd               753 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	txd->lso_hdrlen = hdrlen - md_bytes;
txd               754 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	txd->mss = cpu_to_le16(mss);
txd               755 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	txd->flags |= PCIE_DESC_TX_LSO;
txd               776 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 			    struct nfp_net_tx_desc *txd, struct sk_buff *skb)
txd               788 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	txd->flags |= PCIE_DESC_TX_CSUM;
txd               790 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 		txd->flags |= PCIE_DESC_TX_ENCAP;
txd               796 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 		txd->flags |= PCIE_DESC_TX_IP4_CSUM;
txd               807 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 		txd->flags |= PCIE_DESC_TX_TCP_CSUM;
txd               810 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 		txd->flags |= PCIE_DESC_TX_UDP_CSUM;
txd               983 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	struct nfp_net_tx_desc *txd;
txd              1037 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	txd = &tx_ring->txds[wr_idx];
txd              1038 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	txd->offset_eop = (nr_frags ? 0 : PCIE_DESC_TX_EOP) | md_bytes;
txd              1039 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	txd->dma_len = cpu_to_le16(skb_headlen(skb));
txd              1040 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	nfp_desc_set_dma_addr(txd, dma_addr);
txd              1041 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	txd->data_len = cpu_to_le16(skb->len);
txd              1043 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	txd->flags = 0;
txd              1044 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	txd->mss = 0;
txd              1045 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	txd->lso_hdrlen = 0;
txd              1048 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	nfp_net_tx_tso(r_vec, txbuf, txd, skb, md_bytes);
txd              1049 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	nfp_net_tx_csum(dp, r_vec, txbuf, txd, skb);
txd              1051 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 		txd->flags |= PCIE_DESC_TX_VLAN;
txd              1052 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 		txd->vlan = cpu_to_le16(skb_vlan_tag_get(skb));
txd              1060 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 		second_half = txd->vals8[1];
txd              1076 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 			txd = &tx_ring->txds[wr_idx];
txd              1077 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 			txd->dma_len = cpu_to_le16(fsize);
txd              1078 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 			nfp_desc_set_dma_addr(txd, dma_addr);
txd              1079 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 			txd->offset_eop = md_bytes |
txd              1081 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 			txd->vals8[1] = second_half;
txd              1743 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	struct nfp_net_tx_desc *txd;
txd              1776 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	txd = &tx_ring->txds[wr_idx];
txd              1777 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	txd->offset_eop = PCIE_DESC_TX_EOP;
txd              1778 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	txd->dma_len = cpu_to_le16(pkt_len);
txd              1779 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	nfp_desc_set_dma_addr(txd, rxbuf->dma_addr + dma_off);
txd              1780 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	txd->data_len = cpu_to_le16(pkt_len);
txd              1782 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	txd->flags = 0;
txd              1783 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	txd->mss = 0;
txd              1784 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	txd->lso_hdrlen = 0;
txd              2073 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	struct nfp_net_tx_desc *txd;
txd              2124 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	txd = &tx_ring->txds[wr_idx];
txd              2125 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	txd->offset_eop = meta_len | PCIE_DESC_TX_EOP;
txd              2126 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	txd->dma_len = cpu_to_le16(skb_headlen(skb));
txd              2127 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	nfp_desc_set_dma_addr(txd, dma_addr);
txd              2128 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	txd->data_len = cpu_to_le16(skb->len);
txd              2130 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	txd->flags = 0;
txd              2131 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	txd->mss = 0;
txd              2132 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	txd->lso_hdrlen = 0;
txd                77 drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c 	struct nfp_net_tx_desc *txd;
txd               106 drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c 		txd = &tx_ring->txds[i];
txd               108 drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c 			   txd->vals[0], txd->vals[1],
txd               109 drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c 			   txd->vals[2], txd->vals[3]);
txd               655 drivers/net/ethernet/realtek/8139cp.c 		struct cp_desc *txd = cp->tx_ring + tx_tail;
txd               660 drivers/net/ethernet/realtek/8139cp.c 		status = le32_to_cpu(txd->opts1);
txd               667 drivers/net/ethernet/realtek/8139cp.c 		dma_unmap_single(&cp->pdev->dev, le64_to_cpu(txd->addr),
txd               719 drivers/net/ethernet/realtek/8139cp.c 	struct cp_desc *txd;
txd               724 drivers/net/ethernet/realtek/8139cp.c 		txd = &cp->tx_ring[index];
txd               726 drivers/net/ethernet/realtek/8139cp.c 		dma_unmap_single(&cp->pdev->dev, le64_to_cpu(txd->addr),
txd               779 drivers/net/ethernet/realtek/8139cp.c 		struct cp_desc *txd = &cp->tx_ring[entry];
txd               788 drivers/net/ethernet/realtek/8139cp.c 		txd->opts2 = opts2;
txd               789 drivers/net/ethernet/realtek/8139cp.c 		txd->addr = cpu_to_le64(mapping);
txd               794 drivers/net/ethernet/realtek/8139cp.c 		txd->opts1 = cpu_to_le32(opts1);
txd               802 drivers/net/ethernet/realtek/8139cp.c 		struct cp_desc *txd;
txd               842 drivers/net/ethernet/realtek/8139cp.c 			txd = &cp->tx_ring[entry];
txd               843 drivers/net/ethernet/realtek/8139cp.c 			txd->opts2 = opts2;
txd               844 drivers/net/ethernet/realtek/8139cp.c 			txd->addr = cpu_to_le64(mapping);
txd               847 drivers/net/ethernet/realtek/8139cp.c 			txd->opts1 = cpu_to_le32(ctrl);
txd               854 drivers/net/ethernet/realtek/8139cp.c 		txd = &cp->tx_ring[first_entry];
txd               855 drivers/net/ethernet/realtek/8139cp.c 		txd->opts2 = opts2;
txd               856 drivers/net/ethernet/realtek/8139cp.c 		txd->addr = cpu_to_le64(first_mapping);
txd               860 drivers/net/ethernet/realtek/8139cp.c 		txd->opts1 = cpu_to_le32(ctrl);
txd              5701 drivers/net/ethernet/realtek/r8169_main.c 	struct TxDesc *uninitialized_var(txd);
txd              5713 drivers/net/ethernet/realtek/r8169_main.c 		txd = tp->TxDescArray + entry;
txd              5724 drivers/net/ethernet/realtek/r8169_main.c 		txd->opts1 = rtl8169_get_txd_opts1(opts[0], len, entry);
txd              5725 drivers/net/ethernet/realtek/r8169_main.c 		txd->opts2 = cpu_to_le32(opts[1]);
txd              5726 drivers/net/ethernet/realtek/r8169_main.c 		txd->addr = cpu_to_le64(mapping);
txd              5733 drivers/net/ethernet/realtek/r8169_main.c 		txd->opts1 |= cpu_to_le32(LastFrag);
txd              5885 drivers/net/ethernet/realtek/r8169_main.c 	struct TxDesc *txd = tp->TxDescArray + entry;
txd              5898 drivers/net/ethernet/realtek/r8169_main.c 	if (unlikely(le32_to_cpu(txd->opts1) & DescOwn))
txd              5920 drivers/net/ethernet/realtek/r8169_main.c 	txd->addr = cpu_to_le64(mapping);
txd              5932 drivers/net/ethernet/realtek/r8169_main.c 	txd->opts2 = cpu_to_le32(opts[1]);
txd              5941 drivers/net/ethernet/realtek/r8169_main.c 	txd->opts1 = rtl8169_get_txd_opts1(opts[0], len, entry);
txd              5977 drivers/net/ethernet/realtek/r8169_main.c 	rtl8169_unmap_tx_skb(d, tp->tx_skb + entry, txd);
txd              2300 drivers/net/ethernet/sfc/ef10.c 	return efx_nic_alloc_buffer(tx_queue->efx, &tx_queue->txd.buf,
txd              2308 drivers/net/ethernet/sfc/ef10.c 					 const efx_qword_t *txd)
txd              2315 drivers/net/ethernet/sfc/ef10.c 	reg.qword[0] = *txd;
txd              2411 drivers/net/ethernet/sfc/ef10.c 	size_t entries = tx_queue->txd.buf.len / EFX_BUF_SIZE;
txd              2418 drivers/net/ethernet/sfc/ef10.c 	efx_qword_t *txd;
txd              2454 drivers/net/ethernet/sfc/ef10.c 	dma_addr = tx_queue->txd.buf.dma_addr;
txd              2501 drivers/net/ethernet/sfc/ef10.c 	txd = efx_tx_desc(tx_queue, 0);
txd              2502 drivers/net/ethernet/sfc/ef10.c 	EFX_POPULATE_QWORD_5(*txd,
txd              2520 drivers/net/ethernet/sfc/ef10.c 	efx_ef10_push_tx_desc(tx_queue, txd);
txd              2555 drivers/net/ethernet/sfc/ef10.c 	efx_nic_free_buffer(tx_queue->efx, &tx_queue->txd.buf);
txd              2594 drivers/net/ethernet/sfc/ef10.c 	efx_qword_t *txd;
txd              2603 drivers/net/ethernet/sfc/ef10.c 		txd = efx_tx_desc(tx_queue, write_ptr);
txd              2608 drivers/net/ethernet/sfc/ef10.c 			*txd = buffer->option;
txd              2609 drivers/net/ethernet/sfc/ef10.c 			if (EFX_QWORD_FIELD(*txd, ESF_DZ_TX_OPTION_TYPE) == 1)
txd              2616 drivers/net/ethernet/sfc/ef10.c 				*txd,
txd              2627 drivers/net/ethernet/sfc/ef10.c 		txd = efx_tx_desc(tx_queue,
txd              2629 drivers/net/ethernet/sfc/ef10.c 		efx_ef10_push_tx_desc(tx_queue, txd);
txd               523 drivers/net/ethernet/sfc/efx.c 		memset(&tx_queue->txd, 0, sizeof(tx_queue->txd));
txd               834 drivers/net/ethernet/sfc/efx.c 						tx_queue->txd.index +
txd               835 drivers/net/ethernet/sfc/efx.c 						tx_queue->txd.entries);
txd               482 drivers/net/ethernet/sfc/falcon/efx.c 		memset(&tx_queue->txd, 0, sizeof(tx_queue->txd));
txd               793 drivers/net/ethernet/sfc/falcon/efx.c 						tx_queue->txd.index +
txd               794 drivers/net/ethernet/sfc/falcon/efx.c 						tx_queue->txd.entries);
txd               285 drivers/net/ethernet/sfc/falcon/farch.c 					  const ef4_qword_t *txd)
txd               296 drivers/net/ethernet/sfc/falcon/farch.c 	reg.qword[0] = *txd;
txd               309 drivers/net/ethernet/sfc/falcon/farch.c 	ef4_qword_t *txd;
txd               320 drivers/net/ethernet/sfc/falcon/farch.c 		txd = ef4_tx_desc(tx_queue, write_ptr);
txd               327 drivers/net/ethernet/sfc/falcon/farch.c 		EF4_POPULATE_QWORD_4(*txd,
txd               338 drivers/net/ethernet/sfc/falcon/farch.c 		txd = ef4_tx_desc(tx_queue,
txd               340 drivers/net/ethernet/sfc/falcon/farch.c 		ef4_farch_push_tx_desc(tx_queue, txd);
txd               369 drivers/net/ethernet/sfc/falcon/farch.c 	return ef4_alloc_special_buffer(efx, &tx_queue->txd,
txd               379 drivers/net/ethernet/sfc/falcon/farch.c 	ef4_init_special_buffer(efx, &tx_queue->txd);
txd               386 drivers/net/ethernet/sfc/falcon/farch.c 			      FRF_AZ_TX_DESCQ_BUF_BASE_ID, tx_queue->txd.index,
txd               392 drivers/net/ethernet/sfc/falcon/farch.c 			      __ffs(tx_queue->txd.entries),
txd               454 drivers/net/ethernet/sfc/falcon/farch.c 	ef4_fini_special_buffer(efx, &tx_queue->txd);
txd               460 drivers/net/ethernet/sfc/falcon/farch.c 	ef4_free_special_buffer(tx_queue->efx, &tx_queue->txd);
txd               220 drivers/net/ethernet/sfc/falcon/net_driver.h 	struct ef4_special_buffer txd;
txd                65 drivers/net/ethernet/sfc/falcon/nic.h 	return ((ef4_qword_t *) (tx_queue->txd.buf.addr)) + index;
txd               294 drivers/net/ethernet/sfc/farch.c 					  const efx_qword_t *txd)
txd               305 drivers/net/ethernet/sfc/farch.c 	reg.qword[0] = *txd;
txd               318 drivers/net/ethernet/sfc/farch.c 	efx_qword_t *txd;
txd               329 drivers/net/ethernet/sfc/farch.c 		txd = efx_tx_desc(tx_queue, write_ptr);
txd               336 drivers/net/ethernet/sfc/farch.c 		EFX_POPULATE_QWORD_4(*txd,
txd               347 drivers/net/ethernet/sfc/farch.c 		txd = efx_tx_desc(tx_queue,
txd               349 drivers/net/ethernet/sfc/farch.c 		efx_farch_push_tx_desc(tx_queue, txd);
txd               375 drivers/net/ethernet/sfc/farch.c 	return efx_alloc_special_buffer(efx, &tx_queue->txd,
txd               386 drivers/net/ethernet/sfc/farch.c 	efx_init_special_buffer(efx, &tx_queue->txd);
txd               393 drivers/net/ethernet/sfc/farch.c 			      FRF_AZ_TX_DESCQ_BUF_BASE_ID, tx_queue->txd.index,
txd               399 drivers/net/ethernet/sfc/farch.c 			      __ffs(tx_queue->txd.entries),
txd               442 drivers/net/ethernet/sfc/farch.c 	efx_fini_special_buffer(efx, &tx_queue->txd);
txd               448 drivers/net/ethernet/sfc/farch.c 	efx_free_special_buffer(tx_queue->efx, &tx_queue->txd);
txd               247 drivers/net/ethernet/sfc/net_driver.h 	struct efx_special_buffer txd;
txd                63 drivers/net/ethernet/sfc/nic.h 	return ((efx_qword_t *) (tx_queue->txd.buf.addr)) + index;
txd                88 drivers/net/ethernet/sfc/tx_tso.c 	ptr = (char *) (((efx_qword_t *)tx_queue->txd.buf.addr) + insert_ptr);
txd               700 drivers/net/ethernet/sis/sis190.c 		struct TxDesc *txd = tp->TxDescRing + entry;
txd               701 drivers/net/ethernet/sis/sis190.c 		u32 status = le32_to_cpu(txd->status);
txd               715 drivers/net/ethernet/sis/sis190.c 		sis190_unmap_tx_skb(tp->pci_dev, skb, txd);
txd              1881 drivers/net/ethernet/sun/cassini.c 			struct cas_tx_desc *txd = txds + entry;
txd              1883 drivers/net/ethernet/sun/cassini.c 			daddr = le64_to_cpu(txd->buffer);
txd              1885 drivers/net/ethernet/sun/cassini.c 				       le64_to_cpu(txd->control));
txd              2728 drivers/net/ethernet/sun/cassini.c 	struct cas_tx_desc *txd = cp->init_txds[ring] + entry;
txd              2735 drivers/net/ethernet/sun/cassini.c 	txd->control = cpu_to_le64(ctrl);
txd              2736 drivers/net/ethernet/sun/cassini.c 	txd->buffer = cpu_to_le64(mapping);
txd              3881 drivers/net/ethernet/sun/cassini.c 	struct cas_tx_desc *txd = cp->init_txds[ring];
txd              3902 drivers/net/ethernet/sun/cassini.c 			daddr = le64_to_cpu(txd[ent].buffer);
txd              3904 drivers/net/ethernet/sun/cassini.c 					 le64_to_cpu(txd[ent].control));
txd               651 drivers/net/ethernet/sun/sungem.c 		struct gem_txd *txd;
txd               680 drivers/net/ethernet/sun/sungem.c 			txd = &gp->init_block->txd[entry];
txd               682 drivers/net/ethernet/sun/sungem.c 			dma_addr = le64_to_cpu(txd->buffer);
txd               683 drivers/net/ethernet/sun/sungem.c 			dma_len = le64_to_cpu(txd->control_word) & TXDCTRL_BUFSZ;
txd              1030 drivers/net/ethernet/sun/sungem.c 		struct gem_txd *txd = &gp->init_block->txd[entry];
txd              1042 drivers/net/ethernet/sun/sungem.c 		txd->buffer = cpu_to_le64(mapping);
txd              1044 drivers/net/ethernet/sun/sungem.c 		txd->control_word = cpu_to_le64(ctrl);
txd              1047 drivers/net/ethernet/sun/sungem.c 		struct gem_txd *txd;
txd              1079 drivers/net/ethernet/sun/sungem.c 			txd = &gp->init_block->txd[entry];
txd              1080 drivers/net/ethernet/sun/sungem.c 			txd->buffer = cpu_to_le64(mapping);
txd              1082 drivers/net/ethernet/sun/sungem.c 			txd->control_word = cpu_to_le64(this_ctrl | len);
txd              1089 drivers/net/ethernet/sun/sungem.c 		txd = &gp->init_block->txd[first_entry];
txd              1090 drivers/net/ethernet/sun/sungem.c 		txd->buffer = cpu_to_le64(first_mapping);
txd              1092 drivers/net/ethernet/sun/sungem.c 		txd->control_word =
txd              1602 drivers/net/ethernet/sun/sungem.c 			struct gem_txd *txd;
txd              1611 drivers/net/ethernet/sun/sungem.c 				txd = &gb->txd[ent];
txd              1612 drivers/net/ethernet/sun/sungem.c 				dma_addr = le64_to_cpu(txd->buffer);
txd              1614 drivers/net/ethernet/sun/sungem.c 					       le64_to_cpu(txd->control_word) &
txd              1664 drivers/net/ethernet/sun/sungem.c 		struct gem_txd *txd = &gb->txd[i];
txd              1666 drivers/net/ethernet/sun/sungem.c 		txd->control_word = 0;
txd              1668 drivers/net/ethernet/sun/sungem.c 		txd->buffer = 0;
txd               956 drivers/net/ethernet/sun/sungem.h 	struct gem_txd	txd[INIT_BLOCK_TX_RING_SIZE];
txd               204 drivers/net/ethernet/sun/sunhme.c static void sbus_hme_write_txd(struct happy_meal_txd *txd, u32 flags, u32 addr)
txd               206 drivers/net/ethernet/sun/sunhme.c 	txd->tx_addr = (__force hme32)addr;
txd               208 drivers/net/ethernet/sun/sunhme.c 	txd->tx_flags = (__force hme32)flags;
txd               233 drivers/net/ethernet/sun/sunhme.c static void pci_hme_write_txd(struct happy_meal_txd *txd, u32 flags, u32 addr)
txd               235 drivers/net/ethernet/sun/sunhme.c 	txd->tx_addr = (__force hme32)cpu_to_le32(addr);
txd               237 drivers/net/ethernet/sun/sunhme.c 	txd->tx_flags = (__force hme32)cpu_to_le32(flags);
txd              1220 drivers/net/ethernet/sun/sunhme.c 			struct happy_meal_txd *txd;
txd              1227 drivers/net/ethernet/sun/sunhme.c 				txd = &hp->happy_block->happy_meal_txd[i];
txd              1228 drivers/net/ethernet/sun/sunhme.c 				dma_addr = hme_read_desc32(hp, &txd->tx_addr);
txd              1231 drivers/net/ethernet/sun/sunhme.c 							 (hme_read_desc32(hp, &txd->tx_flags)
txd              1236 drivers/net/ethernet/sun/sunhme.c 							 (hme_read_desc32(hp, &txd->tx_flags)
txd               209 drivers/net/fddi/skfp/h/hwmtm.h #define	HWM_GET_TX_PHYS(txd)		(u_long)AIX_REVERSE((txd)->txd_tbadr)
txd               225 drivers/net/fddi/skfp/h/hwmtm.h #define	HWM_GET_TX_LEN(txd)	((int)AIX_REVERSE((txd)->txd_tbctrl)& RD_LENGTH)
txd                98 drivers/net/fddi/skfp/hwmtm.c 				volatile struct s_smt_fp_txd *txd);
txd               121 drivers/net/fddi/skfp/skfddi.c void mac_drv_tx_complete(struct s_smc *smc, volatile struct s_smt_fp_txd *txd);
txd              1106 drivers/net/fddi/skfp/skfddi.c 	struct s_smt_fp_txd *txd;	// Current TxD.
txd              1170 drivers/net/fddi/skfp/skfddi.c 		txd = (struct s_smt_fp_txd *) HWM_GET_CURR_TXD(smc, queue);
txd              1175 drivers/net/fddi/skfp/skfddi.c 			txd->txd_os.skb = skb;			// save skb
txd              1176 drivers/net/fddi/skfp/skfddi.c 			txd->txd_os.dma_addr = dma_address;	// save dma mapping
txd              1487 drivers/net/fddi/skfp/skfddi.c void mac_drv_tx_complete(struct s_smc *smc, volatile struct s_smt_fp_txd *txd)
txd              1494 drivers/net/fddi/skfp/skfddi.c 	if (!(skb = txd->txd_os.skb)) {
txd              1498 drivers/net/fddi/skfp/skfddi.c 	txd->txd_os.skb = NULL;
txd              1501 drivers/net/fddi/skfp/skfddi.c 	pci_unmap_single(&smc->os.pdev, txd->txd_os.dma_addr,
txd              1503 drivers/net/fddi/skfp/skfddi.c 	txd->txd_os.dma_addr = 0;
txd               111 drivers/net/hamradio/yam.c 	int txd;				/* tx delay */
txd               595 drivers/net/hamradio/yam.c 	if ((yp->tx_state == TX_TAIL) || (yp->txd == 0))
txd               598 drivers/net/hamradio/yam.c 		yp->tx_count = (yp->bitrate * yp->txd) / 8000;
txd               810 drivers/net/hamradio/yam.c 	seq_printf(seq, "  TxDelay  %u\n", yp->txd);
txd              1015 drivers/net/hamradio/yam.c 			yp->txd = yi.cfg.txdelay;
txd              1043 drivers/net/hamradio/yam.c 		yi.cfg.txdelay = yp->txd;
txd              1092 drivers/net/hamradio/yam.c 	yp->txd = DEFAULT_TXD;
txd               353 drivers/net/vmxnet3/vmxnet3_defs.h 	struct Vmxnet3_TxDesc		txd;
txd               337 drivers/net/vmxnet3/vmxnet3_drv.c 	BUG_ON(VMXNET3_TXDESC_GET_EOP(&(tq->tx_ring.base[eop_idx].txd)) != 1);
txd               694 drivers/net/vmxnet3/vmxnet3_drv.c 		ctx->sop_txd->txd.addr = cpu_to_le64(tq->data_ring.basePA +
txd               706 drivers/net/vmxnet3/vmxnet3_drv.c 			le64_to_cpu(ctx->sop_txd->txd.addr),
txd               739 drivers/net/vmxnet3/vmxnet3_drv.c 		BUG_ON(gdesc->txd.gen == tq->tx_ring.gen);
txd               741 drivers/net/vmxnet3/vmxnet3_drv.c 		gdesc->txd.addr = cpu_to_le64(tbi->dma_addr);
txd               747 drivers/net/vmxnet3/vmxnet3_drv.c 			tq->tx_ring.next2fill, le64_to_cpu(gdesc->txd.addr),
txd               781 drivers/net/vmxnet3/vmxnet3_drv.c 			BUG_ON(gdesc->txd.gen == tq->tx_ring.gen);
txd               783 drivers/net/vmxnet3/vmxnet3_drv.c 			gdesc->txd.addr = cpu_to_le64(tbi->dma_addr);
txd               789 drivers/net/vmxnet3/vmxnet3_drv.c 				tq->tx_ring.next2fill, le64_to_cpu(gdesc->txd.addr),
txd              1086 drivers/net/vmxnet3/vmxnet3_drv.c 		gdesc->txd.hlen = ctx.eth_ip_hdr_size + ctx.l4_hdr_size;
txd              1087 drivers/net/vmxnet3/vmxnet3_drv.c 		gdesc->txd.om = VMXNET3_OM_TSO;
txd              1088 drivers/net/vmxnet3/vmxnet3_drv.c 		gdesc->txd.msscof = ctx.mss;
txd              1089 drivers/net/vmxnet3/vmxnet3_drv.c 		num_pkts = (skb->len - gdesc->txd.hlen + ctx.mss - 1) / ctx.mss;
txd              1092 drivers/net/vmxnet3/vmxnet3_drv.c 			gdesc->txd.hlen = ctx.eth_ip_hdr_size;
txd              1093 drivers/net/vmxnet3/vmxnet3_drv.c 			gdesc->txd.om = VMXNET3_OM_CSUM;
txd              1094 drivers/net/vmxnet3/vmxnet3_drv.c 			gdesc->txd.msscof = ctx.eth_ip_hdr_size +
txd              1097 drivers/net/vmxnet3/vmxnet3_drv.c 			gdesc->txd.om = 0;
txd              1098 drivers/net/vmxnet3/vmxnet3_drv.c 			gdesc->txd.msscof = 0;
txd              1106 drivers/net/vmxnet3/vmxnet3_drv.c 		gdesc->txd.ti = 1;
txd              1107 drivers/net/vmxnet3/vmxnet3_drv.c 		gdesc->txd.tci = skb_vlan_tag_get(skb);
txd              1129 drivers/net/vmxnet3/vmxnet3_drv.c 		tq->tx_ring.base), le64_to_cpu(gdesc->txd.addr),
txd               325 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 	return txd(di, i + 1);
txd               330 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 	return txd(di, i - 1);
txd               340 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 	return txd(di, t-h);
txd               162 drivers/net/wireless/marvell/mwl8k.c 	struct mwl8k_tx_desc *txd;
txd              1463 drivers/net/wireless/marvell/mwl8k.c 	txq->txd = pci_zalloc_consistent(priv->pdev, size, &txq->txd_dma);
txd              1464 drivers/net/wireless/marvell/mwl8k.c 	if (txq->txd == NULL) {
txd              1471 drivers/net/wireless/marvell/mwl8k.c 		pci_free_consistent(priv->pdev, size, txq->txd, txq->txd_dma);
txd              1479 drivers/net/wireless/marvell/mwl8k.c 		tx_desc = txq->txd + i;
txd              1512 drivers/net/wireless/marvell/mwl8k.c 			struct mwl8k_tx_desc *tx_desc = txq->txd + desc;
txd              1688 drivers/net/wireless/marvell/mwl8k.c 		tx_desc = txq->txd + tx;
txd              1769 drivers/net/wireless/marvell/mwl8k.c 	if (txq->txd == NULL)
txd              1779 drivers/net/wireless/marvell/mwl8k.c 			    txq->txd, txq->txd_dma);
txd              1780 drivers/net/wireless/marvell/mwl8k.c 	txq->txd = NULL;
txd              2090 drivers/net/wireless/marvell/mwl8k.c 	tx = txq->txd + txq->tail;
txd                33 drivers/net/wireless/mediatek/mt76/mt7603/dma.c 	__le32 *txd = (__le32 *)skb->data;
txd                46 drivers/net/wireless/mediatek/mt76/mt7603/dma.c 	val = le32_to_cpu(txd[1]);
txd                58 drivers/net/wireless/mediatek/mt76/mt7603/dma.c 	val = le32_to_cpu(txd[0]);
txd                63 drivers/net/wireless/mediatek/mt76/mt7603/dma.c 	txd[0] = cpu_to_le32(val);
txd                22 drivers/net/wireless/mediatek/mt76/mt7603/mcu.c 	struct mt7603_mcu_txd *txd;
txd                29 drivers/net/wireless/mediatek/mt76/mt7603/mcu.c 	txd = (struct mt7603_mcu_txd *)skb_push(skb, hdrlen);
txd                30 drivers/net/wireless/mediatek/mt76/mt7603/mcu.c 	memset(txd, 0, hdrlen);
txd                32 drivers/net/wireless/mediatek/mt76/mt7603/mcu.c 	txd->len = cpu_to_le16(skb->len);
txd                34 drivers/net/wireless/mediatek/mt76/mt7603/mcu.c 		txd->pq_id = cpu_to_le16(MCU_PORT_QUEUE_FW);
txd                36 drivers/net/wireless/mediatek/mt76/mt7603/mcu.c 		txd->pq_id = cpu_to_le16(MCU_PORT_QUEUE);
txd                37 drivers/net/wireless/mediatek/mt76/mt7603/mcu.c 	txd->pkt_type = MCU_PKT_ID;
txd                38 drivers/net/wireless/mediatek/mt76/mt7603/mcu.c 	txd->seq = seq;
txd                41 drivers/net/wireless/mediatek/mt76/mt7603/mcu.c 		txd->cid = -cmd;
txd                42 drivers/net/wireless/mediatek/mt76/mt7603/mcu.c 		txd->set_query = MCU_Q_NA;
txd                44 drivers/net/wireless/mediatek/mt76/mt7603/mcu.c 		txd->cid = MCU_CMD_EXT_CID;
txd                45 drivers/net/wireless/mediatek/mt76/mt7603/mcu.c 		txd->ext_cid = cmd;
txd                46 drivers/net/wireless/mediatek/mt76/mt7603/mcu.c 		txd->set_query = MCU_Q_SET;
txd                47 drivers/net/wireless/mediatek/mt76/mt7603/mcu.c 		txd->ext_cid_ack = 1;
txd                58 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c 	__le32 *txd;
txd                76 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c 	txd = mcu_txd->txd;
txd                81 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c 	txd[0] = cpu_to_le32(val);
txd                86 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c 	txd[1] = cpu_to_le32(val);
txd                88 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c 	mcu_txd->len = cpu_to_le16(skb->len - sizeof(mcu_txd->txd));
txd                 8 drivers/net/wireless/mediatek/mt76/mt7615/mcu.h 	__le32 txd[8];
txd              1099 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	__le32 *txd = entry_priv->desc;
txd              1105 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	word = rt2x00_desc_read(txd, 1);
txd              1107 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	rt2x00_desc_write(txd, 1, word);
txd              1109 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	word = rt2x00_desc_read(txd, 2);
txd              1112 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	rt2x00_desc_write(txd, 2, word);
txd              1114 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	word = rt2x00_desc_read(txd, 3);
txd              1121 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	rt2x00_desc_write(txd, 3, word);
txd              1123 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	word = rt2x00_desc_read(txd, 4);
txd              1132 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	rt2x00_desc_write(txd, 4, word);
txd              1139 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	word = rt2x00_desc_read(txd, 0);
txd              1153 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	rt2x00_desc_write(txd, 0, word);
txd              1158 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	skbdesc->desc = txd;
txd              1252 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	__le32 *txd = entry_priv->desc;
txd              1258 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	word = rt2x00_desc_read(txd, 1);
txd              1260 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	rt2x00_desc_write(txd, 1, word);
txd              1262 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	word = rt2x00_desc_read(txd, 2);
txd              1267 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	rt2x00_desc_write(txd, 2, word);
txd              1269 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	word = rt2x00_desc_read(txd, 3);
txd              1276 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	rt2x00_desc_write(txd, 3, word);
txd              1278 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	word = rt2x00_desc_read(txd, 10);
txd              1281 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	rt2x00_desc_write(txd, 10, word);
txd              1288 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	word = rt2x00_desc_read(txd, 0);
txd              1305 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	rt2x00_desc_write(txd, 0, word);
txd              1310 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	skbdesc->desc = txd;
txd              1060 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	__le32 *txd = (__le32 *) entry->skb->data;
txd              1066 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	word = rt2x00_desc_read(txd, 0);
txd              1082 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	rt2x00_desc_write(txd, 0, word);
txd              1084 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	word = rt2x00_desc_read(txd, 1);
txd              1089 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	rt2x00_desc_write(txd, 1, word);
txd              1091 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	word = rt2x00_desc_read(txd, 2);
txd              1098 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	rt2x00_desc_write(txd, 2, word);
txd              1101 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 		_rt2x00_desc_write(txd, 3, skbdesc->iv[0]);
txd              1102 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 		_rt2x00_desc_write(txd, 4, skbdesc->iv[1]);
txd              1109 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	skbdesc->desc = txd;
txd                72 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 	__le32 *txd = entry_priv->desc;
txd                89 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 	rt2x00_desc_write(txd, 0, word);
txd               100 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 	rt2x00_desc_write(txd, 1, word);
txd               105 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 	rt2x00_desc_write(txd, 2, word);
txd               111 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 	rt2x00_desc_write(txd, 3, word);
txd               116 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 	skbdesc->desc = txd;
txd              1780 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	__le32 *txd = entry_priv->desc;
txd              1786 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	word = rt2x00_desc_read(txd, 1);
txd              1795 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	rt2x00_desc_write(txd, 1, word);
txd              1797 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	word = rt2x00_desc_read(txd, 2);
txd              1804 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	rt2x00_desc_write(txd, 2, word);
txd              1807 drivers/net/wireless/ralink/rt2x00/rt61pci.c 		_rt2x00_desc_write(txd, 3, skbdesc->iv[0]);
txd              1808 drivers/net/wireless/ralink/rt2x00/rt61pci.c 		_rt2x00_desc_write(txd, 4, skbdesc->iv[1]);
txd              1811 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	word = rt2x00_desc_read(txd, 5);
txd              1817 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	rt2x00_desc_write(txd, 5, word);
txd              1820 drivers/net/wireless/ralink/rt2x00/rt61pci.c 		word = rt2x00_desc_read(txd, 6);
txd              1823 drivers/net/wireless/ralink/rt2x00/rt61pci.c 		rt2x00_desc_write(txd, 6, word);
txd              1825 drivers/net/wireless/ralink/rt2x00/rt61pci.c 		word = rt2x00_desc_read(txd, 11);
txd              1828 drivers/net/wireless/ralink/rt2x00/rt61pci.c 		rt2x00_desc_write(txd, 11, word);
txd              1836 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	word = rt2x00_desc_read(txd, 0);
txd              1859 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	rt2x00_desc_write(txd, 0, word);
txd              1864 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	skbdesc->desc = txd;
txd              1448 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	__le32 *txd = (__le32 *) entry->skb->data;
txd              1454 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	word = rt2x00_desc_read(txd, 0);
txd              1478 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	rt2x00_desc_write(txd, 0, word);
txd              1480 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	word = rt2x00_desc_read(txd, 1);
txd              1488 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	rt2x00_desc_write(txd, 1, word);
txd              1490 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	word = rt2x00_desc_read(txd, 2);
txd              1497 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	rt2x00_desc_write(txd, 2, word);
txd              1500 drivers/net/wireless/ralink/rt2x00/rt73usb.c 		_rt2x00_desc_write(txd, 3, skbdesc->iv[0]);
txd              1501 drivers/net/wireless/ralink/rt2x00/rt73usb.c 		_rt2x00_desc_write(txd, 4, skbdesc->iv[1]);
txd              1504 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	word = rt2x00_desc_read(txd, 5);
txd              1508 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	rt2x00_desc_write(txd, 5, word);
txd              1514 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	skbdesc->desc = txd;
txd              1524 drivers/ntb/ntb_transport.c 	struct dma_async_tx_descriptor *txd;
txd              1560 drivers/ntb/ntb_transport.c 	txd = device->device_prep_dma_memcpy(chan, unmap->addr[1],
txd              1563 drivers/ntb/ntb_transport.c 	if (!txd)
txd              1566 drivers/ntb/ntb_transport.c 	txd->callback_result = ntb_rx_copy_callback;
txd              1567 drivers/ntb/ntb_transport.c 	txd->callback_param = entry;
txd              1568 drivers/ntb/ntb_transport.c 	dma_set_unmap(txd, unmap);
txd              1570 drivers/ntb/ntb_transport.c 	cookie = dmaengine_submit(txd);
txd              1804 drivers/ntb/ntb_transport.c 	struct dma_async_tx_descriptor *txd;
txd              1834 drivers/ntb/ntb_transport.c 	txd = device->device_prep_dma_memcpy(chan, dest, unmap->addr[0], len,
txd              1836 drivers/ntb/ntb_transport.c 	if (!txd)
txd              1839 drivers/ntb/ntb_transport.c 	txd->callback_result = ntb_tx_copy_callback;
txd              1840 drivers/ntb/ntb_transport.c 	txd->callback_param = entry;
txd              1841 drivers/ntb/ntb_transport.c 	dma_set_unmap(txd, unmap);
txd              1843 drivers/ntb/ntb_transport.c 	cookie = dmaengine_submit(txd);
txd               675 drivers/rapidio/devices/tsi721.h 	struct dma_async_tx_descriptor	txd;
txd                58 drivers/rapidio/devices/tsi721_dma.c struct tsi721_tx_desc *to_tsi721_desc(struct dma_async_tx_descriptor *txd)
txd                60 drivers/rapidio/devices/tsi721_dma.c 	return container_of(txd, struct tsi721_tx_desc, txd);
txd               375 drivers/rapidio/devices/tsi721_dma.c 	struct dma_async_tx_descriptor *txd = &desc->txd;
txd               376 drivers/rapidio/devices/tsi721_dma.c 	dma_async_tx_callback callback = txd->callback;
txd               377 drivers/rapidio/devices/tsi721_dma.c 	void *param = txd->callback_param;
txd               411 drivers/rapidio/devices/tsi721_dma.c 	struct dma_chan *dchan = desc->txd.chan;
txd               641 drivers/rapidio/devices/tsi721_dma.c 		dma_cookie_complete(&desc->txd);
txd               667 drivers/rapidio/devices/tsi721_dma.c 			dma_cookie_complete(&desc->txd);
txd               668 drivers/rapidio/devices/tsi721_dma.c 			if (desc->txd.flags & DMA_PREP_INTERRUPT) {
txd               669 drivers/rapidio/devices/tsi721_dma.c 				callback = desc->txd.callback;
txd               670 drivers/rapidio/devices/tsi721_dma.c 				param = desc->txd.callback_param;
txd               691 drivers/rapidio/devices/tsi721_dma.c static dma_cookie_t tsi721_tx_submit(struct dma_async_tx_descriptor *txd)
txd               693 drivers/rapidio/devices/tsi721_dma.c 	struct tsi721_tx_desc *desc = to_tsi721_desc(txd);
txd               694 drivers/rapidio/devices/tsi721_dma.c 	struct tsi721_bdma_chan *bdma_chan = to_tsi721_chan(txd->chan);
txd               701 drivers/rapidio/devices/tsi721_dma.c 			bdma_chan->id, txd);
txd               712 drivers/rapidio/devices/tsi721_dma.c 	cookie = dma_cookie_assign(txd);
txd               750 drivers/rapidio/devices/tsi721_dma.c 		dma_async_tx_descriptor_init(&desc[i].txd, dchan);
txd               751 drivers/rapidio/devices/tsi721_dma.c 		desc[i].txd.tx_submit = tsi721_tx_submit;
txd               752 drivers/rapidio/devices/tsi721_dma.c 		desc[i].txd.flags = DMA_CTRL_ACK;
txd               833 drivers/rapidio/devices/tsi721_dma.c 	struct dma_async_tx_descriptor *txd = NULL;
txd               878 drivers/rapidio/devices/tsi721_dma.c 		txd		= &desc->txd;
txd               879 drivers/rapidio/devices/tsi721_dma.c 		txd->flags	= flags;
txd               884 drivers/rapidio/devices/tsi721_dma.c 	if (!txd) {
txd               890 drivers/rapidio/devices/tsi721_dma.c 	return txd;
txd                78 drivers/spi/spi-altera.c 	unsigned int txd = 0;
txd                83 drivers/spi/spi-altera.c 			txd = hw->tx[hw->count];
txd                86 drivers/spi/spi-altera.c 			txd = (hw->tx[hw->count * 2]
txd                92 drivers/spi/spi-altera.c 	writel(txd, hw->base + ALTERA_SPI_TXDATA);
txd               273 drivers/spi/spi-ep93xx.c 	struct dma_async_tx_descriptor *txd;
txd               355 drivers/spi/spi-ep93xx.c 	txd = dmaengine_prep_slave_sg(chan, sgt->sgl, nents, conf.direction,
txd               357 drivers/spi/spi-ep93xx.c 	if (!txd) {
txd               361 drivers/spi/spi-ep93xx.c 	return txd;
txd               403 drivers/spi/spi-ep93xx.c 	struct dma_async_tx_descriptor *rxd, *txd;
txd               411 drivers/spi/spi-ep93xx.c 	txd = ep93xx_spi_dma_prepare(master, DMA_TO_DEVICE);
txd               412 drivers/spi/spi-ep93xx.c 	if (IS_ERR(txd)) {
txd               414 drivers/spi/spi-ep93xx.c 		dev_err(&master->dev, "DMA TX failed: %ld\n", PTR_ERR(txd));
txd               415 drivers/spi/spi-ep93xx.c 		return PTR_ERR(txd);
txd               424 drivers/spi/spi-ep93xx.c 	dmaengine_submit(txd);
txd                96 drivers/spi/spi-ppc4xx.c 	u8 txd;
txd               159 drivers/spi/spi-ppc4xx.c 	out_8(&hw->regs->txd, data);
txd               336 drivers/spi/spi-ppc4xx.c 		out_8(&hw->regs->txd, data);
txd               266 drivers/video/fbdev/mx3fb.c 	struct dma_async_tx_descriptor	*txd;
txd               394 drivers/video/fbdev/mx3fb.c 	if (mx3_fbi->txd)
txd               396 drivers/video/fbdev/mx3fb.c 			to_tx_desc(mx3_fbi->txd), to_tx_desc(mx3_fbi->txd)->sg);
txd               402 drivers/video/fbdev/mx3fb.c 		mx3_fbi->txd = dmaengine_prep_slave_sg(dma_chan,
txd               404 drivers/video/fbdev/mx3fb.c 		if (!mx3_fbi->txd) {
txd               410 drivers/video/fbdev/mx3fb.c 		mx3_fbi->txd->callback_param	= mx3_fbi->txd;
txd               411 drivers/video/fbdev/mx3fb.c 		mx3_fbi->txd->callback		= mx3fb_dma_done;
txd               413 drivers/video/fbdev/mx3fb.c 		cookie = mx3_fbi->txd->tx_submit(mx3_fbi->txd);
txd               415 drivers/video/fbdev/mx3fb.c 		       mx3_fbi->txd, cookie, list_empty(&ichan->queue) ? '-' : '+');
txd               417 drivers/video/fbdev/mx3fb.c 		if (!mx3_fbi->txd || !mx3_fbi->txd->tx_submit) {
txd               427 drivers/video/fbdev/mx3fb.c 		       mx3_fbi->txd, cookie, list_empty(&ichan->queue) ? '-' : '+');
txd               451 drivers/video/fbdev/mx3fb.c 	if (mx3_fbi->txd == NULL)
txd               460 drivers/video/fbdev/mx3fb.c 	dmaengine_terminate_all(mx3_fbi->txd->chan);
txd               461 drivers/video/fbdev/mx3fb.c 	mx3_fbi->txd = NULL;
txd               759 drivers/video/fbdev/mx3fb.c 	struct dma_chan *chan = tx_desc->txd.chan;
txd               804 drivers/video/fbdev/mx3fb.c 	if (mx3_fbi->txd)
txd              1143 drivers/video/fbdev/mx3fb.c 	struct dma_async_tx_descriptor *txd;
txd              1199 drivers/video/fbdev/mx3fb.c 	if (mx3_fbi->txd)
txd              1200 drivers/video/fbdev/mx3fb.c 		async_tx_ack(mx3_fbi->txd);
txd              1202 drivers/video/fbdev/mx3fb.c 	txd = dmaengine_prep_slave_sg(dma_chan, sg +
txd              1204 drivers/video/fbdev/mx3fb.c 	if (!txd) {
txd              1211 drivers/video/fbdev/mx3fb.c 	txd->callback_param	= txd;
txd              1212 drivers/video/fbdev/mx3fb.c 	txd->callback		= mx3fb_dma_done;
txd              1218 drivers/video/fbdev/mx3fb.c 	cookie = txd->tx_submit(txd);
txd              1219 drivers/video/fbdev/mx3fb.c 	dev_dbg(fbi->device, "%d: Submit %p #%d\n", __LINE__, txd, cookie);
txd              1228 drivers/video/fbdev/mx3fb.c 	mx3_fbi->txd = txd;
txd               145 include/linux/dma/ipu-dma.h 	struct dma_async_tx_descriptor	txd;
txd               171 include/linux/dma/ipu-dma.h #define to_tx_desc(tx) container_of(tx, struct idmac_tx_desc, txd)
txd               549 include/linux/dmaengine.h static inline void txd_lock(struct dma_async_tx_descriptor *txd)
txd               552 include/linux/dmaengine.h static inline void txd_unlock(struct dma_async_tx_descriptor *txd)
txd               555 include/linux/dmaengine.h static inline void txd_chain(struct dma_async_tx_descriptor *txd, struct dma_async_tx_descriptor *next)
txd               559 include/linux/dmaengine.h static inline void txd_clear_parent(struct dma_async_tx_descriptor *txd)
txd               562 include/linux/dmaengine.h static inline void txd_clear_next(struct dma_async_tx_descriptor *txd)
txd               565 include/linux/dmaengine.h static inline struct dma_async_tx_descriptor *txd_next(struct dma_async_tx_descriptor *txd)
txd               569 include/linux/dmaengine.h static inline struct dma_async_tx_descriptor *txd_parent(struct dma_async_tx_descriptor *txd)
txd               575 include/linux/dmaengine.h static inline void txd_lock(struct dma_async_tx_descriptor *txd)
txd               577 include/linux/dmaengine.h 	spin_lock_bh(&txd->lock);
txd               579 include/linux/dmaengine.h static inline void txd_unlock(struct dma_async_tx_descriptor *txd)
txd               581 include/linux/dmaengine.h 	spin_unlock_bh(&txd->lock);
txd               583 include/linux/dmaengine.h static inline void txd_chain(struct dma_async_tx_descriptor *txd, struct dma_async_tx_descriptor *next)
txd               585 include/linux/dmaengine.h 	txd->next = next;
txd               586 include/linux/dmaengine.h 	next->parent = txd;
txd               588 include/linux/dmaengine.h static inline void txd_clear_parent(struct dma_async_tx_descriptor *txd)
txd               590 include/linux/dmaengine.h 	txd->parent = NULL;
txd               592 include/linux/dmaengine.h static inline void txd_clear_next(struct dma_async_tx_descriptor *txd)
txd               594 include/linux/dmaengine.h 	txd->next = NULL;
txd               596 include/linux/dmaengine.h static inline struct dma_async_tx_descriptor *txd_parent(struct dma_async_tx_descriptor *txd)
txd               598 include/linux/dmaengine.h 	return txd->parent;
txd               600 include/linux/dmaengine.h static inline struct dma_async_tx_descriptor *txd_next(struct dma_async_tx_descriptor *txd)
txd               602 include/linux/dmaengine.h 	return txd->next;