Home
last modified time | relevance | path

Searched refs:txd (Results 1 – 87 of 87) sorted by relevance

/linux-4.4.14/drivers/dma/
Ds3c24xx-dma.c381 struct s3c24xx_txd *txd = s3cchan->at; in s3c24xx_dma_getbytes_chan() local
384 return tc * txd->width; in s3c24xx_dma_getbytes_chan()
425 struct s3c24xx_txd *txd = kzalloc(sizeof(*txd), GFP_NOWAIT); in s3c24xx_dma_get_txd() local
427 if (txd) { in s3c24xx_dma_get_txd()
428 INIT_LIST_HEAD(&txd->dsg_list); in s3c24xx_dma_get_txd()
429 txd->dcon = S3C24XX_DCON_INT | S3C24XX_DCON_NORELOAD; in s3c24xx_dma_get_txd()
432 return txd; in s3c24xx_dma_get_txd()
435 static void s3c24xx_dma_free_txd(struct s3c24xx_txd *txd) in s3c24xx_dma_free_txd() argument
439 list_for_each_entry_safe(dsg, _dsg, &txd->dsg_list, node) { in s3c24xx_dma_free_txd()
444 kfree(txd); in s3c24xx_dma_free_txd()
[all …]
Dsa11x0-dma.c152 static void sa11x0_dma_start_desc(struct sa11x0_dma_phy *p, struct sa11x0_dma_desc *txd) in sa11x0_dma_start_desc() argument
154 list_del(&txd->vd.node); in sa11x0_dma_start_desc()
155 p->txd_load = txd; in sa11x0_dma_start_desc()
159 p->num, &txd->vd, txd->vd.tx.cookie, txd->ddar); in sa11x0_dma_start_desc()
165 struct sa11x0_dma_desc *txd = p->txd_load; in sa11x0_dma_start_sg() local
171 if (!txd) in sa11x0_dma_start_sg()
180 if (p->sg_load == txd->sglen) { in sa11x0_dma_start_sg()
181 if (!txd->cyclic) { in sa11x0_dma_start_sg()
189 if (txn && txn->ddar == txd->ddar) { in sa11x0_dma_start_sg()
190 txd = txn; in sa11x0_dma_start_sg()
[all …]
Dtimb_dma.c72 struct dma_async_tx_descriptor txd; member
207 iowrite32(td_desc->txd.phys, td_chan->membase + in __td_start_dma()
217 iowrite32(td_desc->txd.phys, td_chan->membase + in __td_start_dma()
231 struct dma_async_tx_descriptor *txd; in __td_finish() local
240 txd = &td_desc->txd; in __td_finish()
243 txd->cookie); in __td_finish()
252 dma_cookie_complete(txd); in __td_finish()
255 callback = txd->callback; in __td_finish()
256 param = txd->callback_param; in __td_finish()
260 dma_descriptor_unmap(txd); in __td_finish()
[all …]
Damba-pl08x.c401 struct pl08x_txd *txd = to_pl08x_txd(&vd->tx); in pl08x_start_next_txd() local
404 list_del(&txd->vd.node); in pl08x_start_next_txd()
406 plchan->at = txd; in pl08x_start_next_txd()
412 pl08x_write_lli(pl08x, phychan, &txd->llis_va[0], txd->ccfg); in pl08x_start_next_txd()
534 struct pl08x_txd *txd; in pl08x_getbytes_chan() local
540 txd = plchan->at; in pl08x_getbytes_chan()
542 if (!ch || !txd) in pl08x_getbytes_chan()
562 llis_va = txd->llis_va; in pl08x_getbytes_chan()
563 llis_bus = txd->llis_bus; in pl08x_getbytes_chan()
806 struct pl08x_txd *txd; member
[all …]
Dtxx9dmac.c142 txd_to_txx9dmac_desc(struct dma_async_tx_descriptor *txd) in txd_to_txx9dmac_desc() argument
144 return container_of(txd, struct txx9dmac_desc, txd); in txd_to_txx9dmac_desc()
202 dma_async_tx_descriptor_init(&desc->txd, &dc->chan); in txx9dmac_desc_alloc()
203 desc->txd.tx_submit = txx9dmac_tx_submit; in txx9dmac_desc_alloc()
205 desc->txd.flags = DMA_CTRL_ACK; in txx9dmac_desc_alloc()
206 desc->txd.phys = dma_map_single(chan2parent(&dc->chan), &desc->hwdesc, in txx9dmac_desc_alloc()
219 if (async_tx_test_ack(&desc->txd)) { in txx9dmac_desc_get()
252 child->txd.phys, ddev->descsize, in txx9dmac_sync_desc_for_cpu()
255 desc->txd.phys, ddev->descsize, in txx9dmac_sync_desc_for_cpu()
341 first->txd.cookie, first); in txx9dmac_dostart()
[all …]
Dsun6i-dma.c260 struct sun6i_desc *txd) in sun6i_dma_lli_add() argument
262 if ((!prev && !txd) || !next) in sun6i_dma_lli_add()
266 txd->p_lli = next_phy; in sun6i_dma_lli_add()
267 txd->v_lli = next; in sun6i_dma_lli_add()
334 struct sun6i_desc *txd = to_sun6i_desc(&vd->tx); in sun6i_dma_free_desc() local
339 if (unlikely(!txd)) in sun6i_dma_free_desc()
342 p_lli = txd->p_lli; in sun6i_dma_free_desc()
343 v_lli = txd->v_lli; in sun6i_dma_free_desc()
355 kfree(txd); in sun6i_dma_free_desc()
512 struct sun6i_desc *txd; in sun6i_dma_prep_dma_memcpy() local
[all …]
Dep93xx_dma.c127 struct dma_async_tx_descriptor txd; member
254 d->txd.callback = desc->txd.callback; in ep93xx_dma_set_active()
255 d->txd.callback_param = desc->txd.callback_param; in ep93xx_dma_set_active()
300 return !desc->txd.cookie; in ep93xx_dma_advance_active()
420 desc->txd.cookie, desc->src_addr, desc->dst_addr, in m2p_hw_interrupt()
609 last_done = !desc || desc->txd.cookie; in m2m_hw_interrupt()
673 if (async_tx_test_ack(&desc->txd)) { in ep93xx_dma_desc_get()
681 desc->txd.cookie = 0; in ep93xx_dma_desc_get()
682 desc->txd.callback = NULL; in ep93xx_dma_desc_get()
683 desc->txd.callback_param = NULL; in ep93xx_dma_desc_get()
[all …]
Dat_hdmac.c118 dma_async_tx_descriptor_init(&desc->txd, chan); in atc_alloc_descriptor()
120 desc->txd.flags = DMA_CTRL_ACK; in atc_alloc_descriptor()
121 desc->txd.tx_submit = atc_tx_submit; in atc_alloc_descriptor()
122 desc->txd.phys = phys; in atc_alloc_descriptor()
143 if (async_tx_test_ack(&desc->txd)) { in atc_desc_get()
210 (*prev)->lli.dscr = desc->txd.phys; in atc_desc_chain()
251 channel_writel(atchan, DSCR, first->txd.phys); in atc_dostart()
272 if (desc->txd.cookie == cookie) in atc_get_desc_by_cookie()
277 if (desc->txd.cookie == cookie) in atc_get_desc_by_cookie()
450 struct dma_async_tx_descriptor *txd = &desc->txd; in atc_chain_complete() local
[all …]
Dpch_dma.c93 struct dma_async_tx_descriptor txd; member
146 struct pch_dma_desc *to_pd_desc(struct dma_async_tx_descriptor *txd) in to_pd_desc() argument
148 return container_of(txd, struct pch_dma_desc, txd); in to_pd_desc()
351 channel_writel(pd_chan, NEXT, desc->txd.phys); in pdc_dostart()
359 struct dma_async_tx_descriptor *txd = &desc->txd; in pdc_chain_complete() local
360 dma_async_tx_callback callback = txd->callback; in pdc_chain_complete()
361 void *param = txd->callback_param; in pdc_chain_complete()
401 bad_desc->txd.cookie); in pdc_handle_error()
417 static dma_cookie_t pd_tx_submit(struct dma_async_tx_descriptor *txd) in pd_tx_submit() argument
419 struct pch_dma_desc *desc = to_pd_desc(txd); in pd_tx_submit()
[all …]
Dtegra20-apb-dma.c168 struct dma_async_tx_descriptor txd; member
265 return container_of(td, struct tegra_dma_desc, txd); in txd_to_tegra_dma_desc()
288 if (async_tx_test_ack(&dma_desc->txd)) { in tegra_dma_desc_get()
291 dma_desc->txd.flags = 0; in tegra_dma_desc_get()
305 dma_async_tx_descriptor_init(&dma_desc->txd, &tdc->dma_chan); in tegra_dma_desc_get()
306 dma_desc->txd.tx_submit = tegra_dma_tx_submit; in tegra_dma_desc_get()
307 dma_desc->txd.flags = 0; in tegra_dma_desc_get()
611 dma_cookie_complete(&dma_desc->txd); in handle_once_dma_done()
666 callback = dma_desc->txd.callback; in tegra_dma_tasklet()
667 callback_param = dma_desc->txd.callback_param; in tegra_dma_tasklet()
[all …]
Dpl330.c498 struct dma_async_tx_descriptor txd; member
1942 return container_of(tx, struct dma_pl330_desc, txd); in to_desc()
1966 __func__, __LINE__, desc->txd.cookie); in fill_queue()
1985 dma_cookie_complete(&desc->txd); in pl330_tasklet()
2011 callback = desc->txd.callback; in pl330_tasklet()
2012 callback_param = desc->txd.callback_param; in pl330_tasklet()
2028 dma_descriptor_unmap(&desc->txd); in pl330_tasklet()
2146 dma_cookie_complete(&desc->txd); in pl330_terminate_all()
2151 dma_cookie_complete(&desc->txd); in pl330_terminate_all()
2267 if (desc->txd.cookie == cookie) { in pl330_tx_status()
[all …]
Dat_hdmac_regs.h194 struct dma_async_tx_descriptor txd; member
211 txd_to_at_desc(struct dma_async_tx_descriptor *txd) in txd_to_at_desc() argument
213 return container_of(txd, struct at_desc, txd); in txd_to_at_desc()
Dat_xdmac.c285 static inline struct at_xdmac_desc *txd_to_at_desc(struct dma_async_tx_descriptor *txd) in txd_to_at_desc() argument
287 return container_of(txd, struct at_xdmac_desc, tx_dma_desc); in txd_to_at_desc()
1526 struct dma_async_tx_descriptor *txd; in at_xdmac_handle_cyclic() local
1529 txd = &desc->tx_dma_desc; in at_xdmac_handle_cyclic()
1531 if (txd->callback && (txd->flags & DMA_PREP_INTERRUPT)) in at_xdmac_handle_cyclic()
1532 txd->callback(txd->callback_param); in at_xdmac_handle_cyclic()
1552 struct dma_async_tx_descriptor *txd; in at_xdmac_tasklet() local
1568 txd = &desc->tx_dma_desc; in at_xdmac_tasklet()
1574 dma_cookie_complete(txd); in at_xdmac_tasklet()
1575 if (txd->callback && (txd->flags & DMA_PREP_INTERRUPT)) in at_xdmac_tasklet()
[all …]
Dfsldma.c512 struct dma_async_tx_descriptor *txd = &desc->async_tx; in fsldma_run_tx_complete_actions() local
515 BUG_ON(txd->cookie < 0); in fsldma_run_tx_complete_actions()
517 if (txd->cookie > 0) { in fsldma_run_tx_complete_actions()
518 ret = txd->cookie; in fsldma_run_tx_complete_actions()
521 if (txd->callback) { in fsldma_run_tx_complete_actions()
523 txd->callback(txd->callback_param); in fsldma_run_tx_complete_actions()
528 dma_run_dependencies(txd); in fsldma_run_tx_complete_actions()
Dcppi41.c95 struct dma_async_tx_descriptor txd; member
333 dma_cookie_complete(&c->txd); in cppi41_irq()
334 c->txd.callback(c->txd.callback_param); in cppi41_irq()
354 dma_async_tx_descriptor_init(&c->txd, chan); in cppi41_dma_alloc_chan_resources()
355 c->txd.tx_submit = cppi41_tx_submit; in cppi41_dma_alloc_chan_resources()
525 return &c->txd; in cppi41_dma_prep_slave_sg()
Dste_dma40.c373 struct dma_async_tx_descriptor txd; member
767 if (async_tx_test_ack(&d->txd)) { in d40_desc_get()
1489 struct d40_desc *d40d = container_of(tx, struct d40_desc, txd); in d40_tx_submit()
1614 dma_cookie_complete(&d40d->txd); in dma_tasklet()
1626 callback_active = !!(d40d->txd.flags & DMA_PREP_INTERRUPT); in dma_tasklet()
1627 callback = d40d->txd.callback; in dma_tasklet()
1628 callback_param = d40d->txd.callback_param; in dma_tasklet()
1631 if (async_tx_test_ack(&d40d->txd)) { in dma_tasklet()
2223 desc->txd.flags = dma_flags; in d40_prep_desc()
2224 desc->txd.tx_submit = d40_tx_submit; in d40_prep_desc()
[all …]
Dmmp_pdma.c920 struct dma_async_tx_descriptor *txd = &desc->async_tx; in dma_do_tasklet() local
925 if (txd->callback) in dma_do_tasklet()
926 txd->callback(txd->callback_param); in dma_do_tasklet()
928 dma_pool_free(chan->desc_pool, desc, txd->phys); in dma_do_tasklet()
Dmpc512x_dma.c473 static dma_cookie_t mpc_dma_tx_submit(struct dma_async_tx_descriptor *txd) in mpc_dma_tx_submit() argument
475 struct mpc_dma_chan *mchan = dma_chan_to_mpc_dma_chan(txd->chan); in mpc_dma_tx_submit()
480 mdesc = container_of(txd, struct mpc_dma_desc, desc); in mpc_dma_tx_submit()
492 cookie = dma_cookie_assign(txd); in mpc_dma_tx_submit()
Dtxx9dmac.h234 struct dma_async_tx_descriptor txd; member
Dsirf-dma.c408 static dma_cookie_t sirfsoc_dma_tx_submit(struct dma_async_tx_descriptor *txd) in sirfsoc_dma_tx_submit() argument
410 struct sirfsoc_dma_chan *schan = dma_chan_to_sirfsoc_dma_chan(txd->chan); in sirfsoc_dma_tx_submit()
415 sdesc = container_of(txd, struct sirfsoc_dma_desc, desc); in sirfsoc_dma_tx_submit()
422 cookie = dma_cookie_assign(txd); in sirfsoc_dma_tx_submit()
/linux-4.4.14/Documentation/devicetree/bindings/pinctrl/
Dmarvell,armada-39x-pinctrl.txt19 mpp1 1 gpio, ua0(txd)
22 mpp4 4 gpio, ua1(txd), ua0(rts), smi(mdc)
32 mpp14 14 gpio, dram(vttctrl), dev(we1), ua1(txd)
36 mpp18 18 gpio, ua1(txd), spi0(cs0), i2c2(sck)
38 mpp20 20 gpio, sata0(prsnt) [1], ua0(rts), ua1(txd), smi(mdc)
44 mpp25 25 gpio, spi0(cs0), ua0(rts), ua1(txd), sd0(d5), dev(cs0)
61 mpp42 42 gpio, ua1(txd), ua0(rts), dev(ad7)
66 mpp46 46 gpio, ref(clk), pcie0(rstout), ua1(txd), led(stb)
69 …, sata0(prsnt) [1], dram(vttctrl), tdm(pclk) [2], audio(mclk) [2], sd0(d4), pcie0(clkreq), ua1(txd)
73 mpp51 51 gpio, tdm(dtx) [2], audio(sdo) [2], dram(deccerr), ua2(txd)
[all …]
Dmarvell,kirkwood-pinctrl.txt29 mpp5 5 gpo, nand(io7), uart0(txd), ptp(trig)
36 mpp10 10 gpo, spi(sck), uart0(txd), ptp(trig)
40 mpp13 13 gpio, sdio(cmd), uart1(txd)
42 mpp15 15 gpio, sdio(d1), uart0(rts), uart1(txd)
67 mpp5 5 gpo, nand(io7), uart0(txd), ptp(trig), sata0(act)
74 mpp10 10 gpo, spi(sck), uart0(txd), ptp(trig)
78 mpp13 13 gpio, sdio(cmd), uart1(txd)
80 mpp15 15 gpio, sdio(d1), uart0(rts), uart1(txd), sata0(act)
111 mpp5 5 gpo, nand(io7), uart0(txd), ptp(trig), sata0(act)
118 mpp10 10 gpo, spi(sck), uart0(txd), ptp(trig), sata1(act)
[all …]
Dmarvell,armada-370-pinctrl.txt17 mpp1 1 gpo, uart0(txd)
18 mpp2 2 gpio, i2c0(sck), uart0(txd)
21 mpp5 5 gpo, ge0(txclkout), uart1(txd), spi1(sck), audio(mclk)
25 mpp9 9 gpo, ge0(txd3), uart1(txd), sd0(clk), audio(spdifo)
42 mpp21 21 gpo, ge0(txd5), ge1(txd1), uart1(txd)
63 mpp42 42 gpo, dev(ad3), uart1(txd)
80 mpp55 55 gpio, dev(cs1), uart1(txd), tdm(rst), sata1(prsnt),
91 mpp61 61 gpo, dev(we1), uart1(txd), audio(lrclk)
Dmarvell,armada-38x-pinctrl.txt19 mpp1 1 gpio, ua0(txd)
22 mpp4 4 gpio, ge(mdc), ua1(txd), ua0(rts)
36 mpp18 18 gpio, ge0(rxerr), ptp(trig), ua1(txd), spi0(cs0)
43 mpp25 25 gpio, spi0(cs0), ua0(rts), ua1(txd), sd0(d5), dev(cs0)
60 mpp42 42 gpio, ua1(txd), ua0(rts), dev(ad7)
64 mpp46 46 gpio, ref(clk_out1), pcie0(rstout), ua1(txd)
74 mpp56 56 gpio, ua1(rts), ge(mdc), dram(deccerr), spi1(mosi), ua1(txd)
75 mpp57 57 gpio, spi1(sck), sd0(clk), ua1(txd)
Dmarvell,dove-pinctrl.txt20 mpp2 2 gpio, pmu, uart2(txd), sdio0(buspwr), sata(prsnt),
26 mpp6 6 gpio, pmu, uart3(txd), sdio1(buspwr), spi1(mosi), pmu*
37 mpp14 14 gpio, pmu, uart2(txd), sdio1(buspwr), ssp(rxd), pmu*
42 mpp18 18 gpio, uart3(txd), sdio0(buspwr), ac97(sdi3), lcd0(pwm)
49 lcd-spi(mosi), uart1(cts), ssp(txd)
Dmarvell,armada-375-pinctrl.txt30 mpp14 14 gpio, i2c0(sda), uart1(txd)
32 mpp16 16 gpio, uart0(txd)
56 mpp40 40 gpio, uart1(txd)
76 mpp60 60 gpio, uart1(txd), led(c2)
Dmarvell,orion-pinctrl.txt64 mpp17 17 uart1(txd), ge(rxd5), gpio
89 mpp17 17 uart1(txd), ge(rxd5)
Dmarvell,armada-xp-pinctrl.txt66 mpp43 43 gpio, uart2(txd), uart0(rts), spi0(cs3), pcie(rstout),
70 mpp45 45 gpio, uart2(rts), uart3(txd), spi0(cs5), sata1(prsnt),
Dimg,pistachio-pinctrl.txt188 uart0-txd {
/linux-4.4.14/drivers/rapidio/devices/
Dtsi721_dma.c61 struct tsi721_tx_desc *to_tsi721_desc(struct dma_async_tx_descriptor *txd) in to_tsi721_desc() argument
63 return container_of(txd, struct tsi721_tx_desc, txd); in to_tsi721_desc()
380 struct dma_async_tx_descriptor *txd = &desc->txd; in tsi721_dma_tx_err() local
381 dma_async_tx_callback callback = txd->callback; in tsi721_dma_tx_err()
382 void *param = txd->callback_param; in tsi721_dma_tx_err()
416 struct dma_chan *dchan = desc->txd.chan; in tsi721_submit_sg()
604 dma_cookie_complete(&desc->txd); in tsi721_dma_tasklet()
605 if (desc->txd.flags & DMA_PREP_INTERRUPT) { in tsi721_dma_tasklet()
606 callback = desc->txd.callback; in tsi721_dma_tasklet()
607 param = desc->txd.callback_param; in tsi721_dma_tasklet()
[all …]
Dtsi721.h648 struct dma_async_tx_descriptor txd; member
/linux-4.4.14/include/linux/
Ddmaengine.h523 static inline void txd_lock(struct dma_async_tx_descriptor *txd) in txd_lock() argument
526 static inline void txd_unlock(struct dma_async_tx_descriptor *txd) in txd_unlock() argument
529 static inline void txd_chain(struct dma_async_tx_descriptor *txd, struct dma_async_tx_descriptor *n… in txd_chain() argument
533 static inline void txd_clear_parent(struct dma_async_tx_descriptor *txd) in txd_clear_parent() argument
536 static inline void txd_clear_next(struct dma_async_tx_descriptor *txd) in txd_clear_next() argument
539 static inline struct dma_async_tx_descriptor *txd_next(struct dma_async_tx_descriptor *txd) in txd_next() argument
543 static inline struct dma_async_tx_descriptor *txd_parent(struct dma_async_tx_descriptor *txd) in txd_parent() argument
549 static inline void txd_lock(struct dma_async_tx_descriptor *txd) in txd_lock() argument
551 spin_lock_bh(&txd->lock); in txd_lock()
553 static inline void txd_unlock(struct dma_async_tx_descriptor *txd) in txd_unlock() argument
[all …]
/linux-4.4.14/drivers/net/ethernet/intel/fm10k/
Dfm10k_debugfs.c79 struct fm10k_tx_desc *txd = FM10K_TX_DESC(ring, i); in fm10k_dbg_tx_desc_seq_show() local
82 i, txd->buffer_addr, txd->buflen, txd->vlan, in fm10k_dbg_tx_desc_seq_show()
83 txd->mss, txd->hdrlen, txd->flags); in fm10k_dbg_tx_desc_seq_show()
/linux-4.4.14/drivers/media/pci/mantis/
Dmantis_i2c.c87 u32 txd = 0, stat, trials; in mantis_i2c_write() local
94 txd = (msg->addr << 25) | (msg->buf[i] << 8) in mantis_i2c_write()
100 txd &= ~MANTIS_I2C_STOP; in mantis_i2c_write()
103 mmwrite(txd, MANTIS_I2CDATA_CTL); in mantis_i2c_write()
131 u32 stat, data, txd; in mantis_i2c_xfer() local
154 txd = msgs[i].addr << 25 | (0x1 << 24) in mantis_i2c_xfer()
158 mmwrite(txd, MANTIS_I2CDATA_CTL); in mantis_i2c_xfer()
/linux-4.4.14/drivers/video/fbdev/
Dmx3fb.c269 struct dma_async_tx_descriptor *txd; member
397 if (mx3_fbi->txd) in sdc_enable_channel()
399 to_tx_desc(mx3_fbi->txd), to_tx_desc(mx3_fbi->txd)->sg); in sdc_enable_channel()
405 mx3_fbi->txd = dmaengine_prep_slave_sg(dma_chan, in sdc_enable_channel()
407 if (!mx3_fbi->txd) { in sdc_enable_channel()
413 mx3_fbi->txd->callback_param = mx3_fbi->txd; in sdc_enable_channel()
414 mx3_fbi->txd->callback = mx3fb_dma_done; in sdc_enable_channel()
416 cookie = mx3_fbi->txd->tx_submit(mx3_fbi->txd); in sdc_enable_channel()
418 mx3_fbi->txd, cookie, list_empty(&ichan->queue) ? '-' : '+'); in sdc_enable_channel()
420 if (!mx3_fbi->txd || !mx3_fbi->txd->tx_submit) { in sdc_enable_channel()
[all …]
/linux-4.4.14/drivers/media/platform/soc_camera/
Dmx3_camera.c70 struct dma_async_tx_descriptor *txd; member
145 struct dma_chan *chan = desc->txd.chan; in mx3_cam_dma_done()
150 desc->txd.cookie, mx3_cam->active ? sg_dma_address(&mx3_cam->active->sg) : 0); in mx3_cam_dma_done()
267 struct dma_async_tx_descriptor *txd; in mx3_videobuf_queue() local
282 if (!buf->txd) { in mx3_videobuf_queue()
286 txd = dmaengine_prep_slave_sg( in mx3_videobuf_queue()
289 if (!txd) in mx3_videobuf_queue()
292 txd->callback_param = txd; in mx3_videobuf_queue()
293 txd->callback = mx3_cam_dma_done; in mx3_videobuf_queue()
295 buf->txd = txd; in mx3_videobuf_queue()
[all …]
/linux-4.4.14/drivers/dma/ioat/
Dprep.c151 desc->txd.flags = flags; in ioat_dma_prep_memcpy_lock()
159 return &desc->txd; in ioat_dma_prep_memcpy_lock()
235 desc->txd.flags = flags; in __ioat_prep_xor_lock()
243 compl_desc->txd.flags = flags & DMA_PREP_INTERRUPT; in __ioat_prep_xor_lock()
253 return &compl_desc->txd; in __ioat_prep_xor_lock()
301 desc_id(desc), (unsigned long long) desc->txd.phys, in dump_pq_desc_dbg()
303 desc->txd.flags, pq->size, pq->ctl, pq->ctl_f.op, in dump_pq_desc_dbg()
334 desc_id(desc), (unsigned long long) desc->txd.phys, in dump_pq16_desc_dbg()
336 desc->txd.flags, pq->size, pq->ctl, in dump_pq16_desc_dbg()
445 desc->txd.flags = flags; in __ioat_prep_pq_lock()
[all …]
Ddma.c188 async_tx_ack(&desc->txd); in __ioat_start_null_desc()
189 ioat_set_chainaddr(ioat_chan, desc->txd.phys); in __ioat_start_null_desc()
221 ioat_set_chainaddr(ioat_chan, desc->txd.phys); in __ioat_restart_chan()
312 dma_async_tx_descriptor_init(&desc->txd, chan); in ioat_alloc_ring_ent()
313 desc->txd.tx_submit = ioat_tx_submit_unlock; in ioat_alloc_ring_ent()
315 desc->txd.phys = phys; in ioat_alloc_ring_ent()
324 pci_pool_free(ioat_dma->dma_pool, desc->hw, desc->txd.phys); in ioat_free_ring_ent()
358 hw->next = next->txd.phys; in ioat_alloc_ring()
360 ring[i]->hw->next = ring[0]->txd.phys; in ioat_alloc_ring()
432 hw->next = next->txd.phys; in reshape_ring()
[all …]
Ddma.h184 struct dma_async_tx_descriptor txd; member
230 ({ if (d) __dump_desc_dbg(c, d->hw, &d->txd, desc_id(d)); 0; })
/linux-4.4.14/drivers/dma/dw/
Dcore.c93 if (async_tx_test_ack(&desc->txd)) { in dwc_desc_get()
251 channel_writel(dwc, LLP, first->txd.phys); in dwc_dostart()
267 dev_vdbg(chan2dev(&dwc->chan), "%s: started %u\n", __func__, desc->txd.cookie); in dwc_dostart_first_queued()
279 struct dma_async_tx_descriptor *txd = &desc->txd; in dwc_descriptor_complete() local
283 dev_vdbg(chan2dev(&dwc->chan), "descriptor %u complete\n", txd->cookie); in dwc_descriptor_complete()
286 dma_cookie_complete(txd); in dwc_descriptor_complete()
288 callback = txd->callback; in dwc_descriptor_complete()
289 param = txd->callback_param; in dwc_descriptor_complete()
294 async_tx_ack(&child->txd); in dwc_descriptor_complete()
295 async_tx_ack(&desc->txd); in dwc_descriptor_complete()
[all …]
Dregs.h334 struct dma_async_tx_descriptor txd; member
342 txd_to_dw_desc(struct dma_async_tx_descriptor *txd) in txd_to_dw_desc() argument
344 return container_of(txd, struct dw_desc, txd); in txd_to_dw_desc()
/linux-4.4.14/drivers/net/fddi/skfp/h/
Dhwmtm.h201 #define HWM_GET_TX_PHYS(txd) (u_long)AIX_REVERSE((txd)->txd_tbadr) argument
217 #define HWM_GET_TX_LEN(txd) ((int)AIX_REVERSE((txd)->txd_tbctrl)& RD_LENGTH) argument
/linux-4.4.14/drivers/net/ethernet/aurora/
Dnb8800.c401 struct nb8800_tx_desc *txd; in nb8800_xmit() local
434 txd = &priv->tx_descs[next]; in nb8800_xmit()
435 desc = &txd->desc[0]; in nb8800_xmit()
440 memcpy(txd->buf, skb->data, align); in nb8800_xmit()
444 desc->n_addr = txb->dma_desc + sizeof(txd->desc[0]); in nb8800_xmit()
505 struct nb8800_tx_desc *txd = &priv->tx_descs[done]; in nb8800_tx_done() local
509 if (!txd->report) in nb8800_tx_done()
518 if (IS_TX_ERROR(txd->report)) { in nb8800_tx_done()
519 nb8800_tx_error(dev, txd->report); in nb8800_tx_done()
526 dev->stats.tx_bytes += TX_BYTES_TRANSFERRED(txd->report); in nb8800_tx_done()
[all …]
/linux-4.4.14/drivers/dma/ipu/
Dipu_idmac.c782 if (async_tx_test_ack(&desc->txd)) in ipu_submit_buffer()
928 struct dma_async_tx_descriptor *txd = &desc->txd; in idmac_desc_alloc() local
930 memset(txd, 0, sizeof(*txd)); in idmac_desc_alloc()
931 dma_async_tx_descriptor_init(txd, &ichan->dma_chan); in idmac_desc_alloc()
932 txd->tx_submit = idmac_tx_submit; in idmac_desc_alloc()
1281 callback = descnew->txd.callback; in idmac_interrupt()
1282 callback_param = descnew->txd.callback_param; in idmac_interrupt()
1293 dma_cookie_complete(&desc->txd); in idmac_interrupt()
1295 callback = desc->txd.callback; in idmac_interrupt()
1296 callback_param = desc->txd.callback_param; in idmac_interrupt()
[all …]
/linux-4.4.14/drivers/spi/
Dspi-ep93xx.c435 struct dma_async_tx_descriptor *txd; in ep93xx_spi_dma_prepare() local
517 txd = dmaengine_prep_slave_sg(chan, sgt->sgl, nents, dir, DMA_CTRL_ACK); in ep93xx_spi_dma_prepare()
518 if (!txd) { in ep93xx_spi_dma_prepare()
522 return txd; in ep93xx_spi_dma_prepare()
558 struct dma_async_tx_descriptor *rxd, *txd; in ep93xx_spi_dma_transfer() local
567 txd = ep93xx_spi_dma_prepare(espi, DMA_MEM_TO_DEV); in ep93xx_spi_dma_transfer()
568 if (IS_ERR(txd)) { in ep93xx_spi_dma_transfer()
571 msg->status = PTR_ERR(txd); in ep93xx_spi_dma_transfer()
581 dmaengine_submit(txd); in ep93xx_spi_dma_transfer()
Dspi-ppc4xx.c99 u8 txd; member
162 out_8(&hw->regs->txd, data); in spi_ppc4xx_txrx()
339 out_8(&hw->regs->txd, data); in spi_ppc4xx_int()
/linux-4.4.14/include/linux/dma/
Dipu-dma.h148 struct dma_async_tx_descriptor txd; member
174 #define to_tx_desc(tx) container_of(tx, struct idmac_tx_desc, txd)
/linux-4.4.14/drivers/net/ethernet/sun/
Dsungem.c652 struct gem_txd *txd; in gem_tx() local
681 txd = &gp->init_block->txd[entry]; in gem_tx()
683 dma_addr = le64_to_cpu(txd->buffer); in gem_tx()
684 dma_len = le64_to_cpu(txd->control_word) & TXDCTRL_BUFSZ; in gem_tx()
1028 struct gem_txd *txd = &gp->init_block->txd[entry]; in gem_start_xmit() local
1040 txd->buffer = cpu_to_le64(mapping); in gem_start_xmit()
1042 txd->control_word = cpu_to_le64(ctrl); in gem_start_xmit()
1045 struct gem_txd *txd; in gem_start_xmit() local
1077 txd = &gp->init_block->txd[entry]; in gem_start_xmit()
1078 txd->buffer = cpu_to_le64(mapping); in gem_start_xmit()
[all …]
Dsunhme.c203 static void sbus_hme_write_txd(struct happy_meal_txd *txd, u32 flags, u32 addr) in sbus_hme_write_txd() argument
205 txd->tx_addr = (__force hme32)addr; in sbus_hme_write_txd()
207 txd->tx_flags = (__force hme32)flags; in sbus_hme_write_txd()
232 static void pci_hme_write_txd(struct happy_meal_txd *txd, u32 flags, u32 addr) in pci_hme_write_txd() argument
234 txd->tx_addr = (__force hme32)cpu_to_le32(addr); in pci_hme_write_txd()
236 txd->tx_flags = (__force hme32)cpu_to_le32(flags); in pci_hme_write_txd()
1219 struct happy_meal_txd *txd; in happy_meal_clean_rings() local
1226 txd = &hp->happy_block->happy_meal_txd[i]; in happy_meal_clean_rings()
1227 dma_addr = hme_read_desc32(hp, &txd->tx_addr); in happy_meal_clean_rings()
1230 (hme_read_desc32(hp, &txd->tx_flags) in happy_meal_clean_rings()
[all …]
Dsungem.h955 struct gem_txd txd[INIT_BLOCK_TX_RING_SIZE]; member
Dcassini.c1892 struct cas_tx_desc *txd = txds + entry; in cas_tx_ringN() local
1894 daddr = le64_to_cpu(txd->buffer); in cas_tx_ringN()
1896 le64_to_cpu(txd->control)); in cas_tx_ringN()
2739 struct cas_tx_desc *txd = cp->init_txds[ring] + entry; in cas_write_txd() local
2746 txd->control = cpu_to_le64(ctrl); in cas_write_txd()
2747 txd->buffer = cpu_to_le64(mapping); in cas_write_txd()
3895 struct cas_tx_desc *txd = cp->init_txds[ring]; in cas_clean_txd() local
3916 daddr = le64_to_cpu(txd[ent].buffer); in cas_clean_txd()
3918 le64_to_cpu(txd[ent].control)); in cas_clean_txd()
/linux-4.4.14/arch/cris/include/arch-v32/arch/hwregs/
Dser_defs.h99 unsigned int txd : 1; member
191 unsigned int txd : 1; member
212 unsigned int txd : 1; member
/linux-4.4.14/drivers/net/ethernet/realtek/
D8139cp.c657 struct cp_desc *txd = cp->tx_ring + tx_tail; in cp_tx() local
662 status = le32_to_cpu(txd->opts1); in cp_tx()
669 dma_unmap_single(&cp->pdev->dev, le64_to_cpu(txd->addr), in cp_tx()
721 struct cp_desc *txd; in unwind_tx_frag_mapping() local
726 txd = &cp->tx_ring[index]; in unwind_tx_frag_mapping()
728 dma_unmap_single(&cp->pdev->dev, le64_to_cpu(txd->addr), in unwind_tx_frag_mapping()
781 struct cp_desc *txd = &cp->tx_ring[entry]; in cp_start_xmit() local
790 txd->opts2 = opts2; in cp_start_xmit()
791 txd->addr = cpu_to_le64(mapping); in cp_start_xmit()
796 txd->opts1 = cpu_to_le32(opts1); in cp_start_xmit()
[all …]
Dr8169.c6881 struct TxDesc *uninitialized_var(txd); in rtl8169_xmit_frags()
6893 txd = tp->TxDescArray + entry; in rtl8169_xmit_frags()
6908 txd->opts1 = cpu_to_le32(status); in rtl8169_xmit_frags()
6909 txd->opts2 = cpu_to_le32(opts[1]); in rtl8169_xmit_frags()
6910 txd->addr = cpu_to_le64(mapping); in rtl8169_xmit_frags()
6917 txd->opts1 |= cpu_to_le32(LastFrag); in rtl8169_xmit_frags()
7113 struct TxDesc *txd = tp->TxDescArray + entry; in rtl8169_start_xmit() local
7126 if (unlikely(le32_to_cpu(txd->opts1) & DescOwn)) in rtl8169_start_xmit()
7146 txd->addr = cpu_to_le64(mapping); in rtl8169_start_xmit()
7158 txd->opts2 = cpu_to_le32(opts[1]); in rtl8169_start_xmit()
[all …]
/linux-4.4.14/drivers/net/wireless/rt2x00/
Drt2800mmio.c53 __le32 *txd = entry_priv->desc; in rt2800mmio_write_tx_desc() local
70 rt2x00_desc_write(txd, 0, word); in rt2800mmio_write_tx_desc()
81 rt2x00_desc_write(txd, 1, word); in rt2800mmio_write_tx_desc()
86 rt2x00_desc_write(txd, 2, word); in rt2800mmio_write_tx_desc()
92 rt2x00_desc_write(txd, 3, word); in rt2800mmio_write_tx_desc()
97 skbdesc->desc = txd; in rt2800mmio_write_tx_desc()
Drt2400pci.c1105 __le32 *txd = entry_priv->desc; in rt2400pci_write_tx_desc() local
1111 rt2x00_desc_read(txd, 1, &word); in rt2400pci_write_tx_desc()
1113 rt2x00_desc_write(txd, 1, word); in rt2400pci_write_tx_desc()
1115 rt2x00_desc_read(txd, 2, &word); in rt2400pci_write_tx_desc()
1118 rt2x00_desc_write(txd, 2, word); in rt2400pci_write_tx_desc()
1120 rt2x00_desc_read(txd, 3, &word); in rt2400pci_write_tx_desc()
1127 rt2x00_desc_write(txd, 3, word); in rt2400pci_write_tx_desc()
1129 rt2x00_desc_read(txd, 4, &word); in rt2400pci_write_tx_desc()
1138 rt2x00_desc_write(txd, 4, word); in rt2400pci_write_tx_desc()
1145 rt2x00_desc_read(txd, 0, &word); in rt2400pci_write_tx_desc()
[all …]
Drt2500pci.c1258 __le32 *txd = entry_priv->desc; in rt2500pci_write_tx_desc() local
1264 rt2x00_desc_read(txd, 1, &word); in rt2500pci_write_tx_desc()
1266 rt2x00_desc_write(txd, 1, word); in rt2500pci_write_tx_desc()
1268 rt2x00_desc_read(txd, 2, &word); in rt2500pci_write_tx_desc()
1273 rt2x00_desc_write(txd, 2, word); in rt2500pci_write_tx_desc()
1275 rt2x00_desc_read(txd, 3, &word); in rt2500pci_write_tx_desc()
1282 rt2x00_desc_write(txd, 3, word); in rt2500pci_write_tx_desc()
1284 rt2x00_desc_read(txd, 10, &word); in rt2500pci_write_tx_desc()
1287 rt2x00_desc_write(txd, 10, word); in rt2500pci_write_tx_desc()
1294 rt2x00_desc_read(txd, 0, &word); in rt2500pci_write_tx_desc()
[all …]
Drt61pci.c1871 __le32 *txd = entry_priv->desc; in rt61pci_write_tx_desc() local
1877 rt2x00_desc_read(txd, 1, &word); in rt61pci_write_tx_desc()
1886 rt2x00_desc_write(txd, 1, word); in rt61pci_write_tx_desc()
1888 rt2x00_desc_read(txd, 2, &word); in rt61pci_write_tx_desc()
1895 rt2x00_desc_write(txd, 2, word); in rt61pci_write_tx_desc()
1898 _rt2x00_desc_write(txd, 3, skbdesc->iv[0]); in rt61pci_write_tx_desc()
1899 _rt2x00_desc_write(txd, 4, skbdesc->iv[1]); in rt61pci_write_tx_desc()
1902 rt2x00_desc_read(txd, 5, &word); in rt61pci_write_tx_desc()
1909 rt2x00_desc_write(txd, 5, word); in rt61pci_write_tx_desc()
1912 rt2x00_desc_read(txd, 6, &word); in rt61pci_write_tx_desc()
[all …]
Drt2500usb.c1078 __le32 *txd = (__le32 *) entry->skb->data; in rt2500usb_write_tx_desc() local
1084 rt2x00_desc_read(txd, 0, &word); in rt2500usb_write_tx_desc()
1100 rt2x00_desc_write(txd, 0, word); in rt2500usb_write_tx_desc()
1102 rt2x00_desc_read(txd, 1, &word); in rt2500usb_write_tx_desc()
1107 rt2x00_desc_write(txd, 1, word); in rt2500usb_write_tx_desc()
1109 rt2x00_desc_read(txd, 2, &word); in rt2500usb_write_tx_desc()
1116 rt2x00_desc_write(txd, 2, word); in rt2500usb_write_tx_desc()
1119 _rt2x00_desc_write(txd, 3, skbdesc->iv[0]); in rt2500usb_write_tx_desc()
1120 _rt2x00_desc_write(txd, 4, skbdesc->iv[1]); in rt2500usb_write_tx_desc()
1127 skbdesc->desc = txd; in rt2500usb_write_tx_desc()
Drt73usb.c1454 __le32 *txd = (__le32 *) entry->skb->data; in rt73usb_write_tx_desc() local
1460 rt2x00_desc_read(txd, 0, &word); in rt73usb_write_tx_desc()
1484 rt2x00_desc_write(txd, 0, word); in rt73usb_write_tx_desc()
1486 rt2x00_desc_read(txd, 1, &word); in rt73usb_write_tx_desc()
1494 rt2x00_desc_write(txd, 1, word); in rt73usb_write_tx_desc()
1496 rt2x00_desc_read(txd, 2, &word); in rt73usb_write_tx_desc()
1503 rt2x00_desc_write(txd, 2, word); in rt73usb_write_tx_desc()
1506 _rt2x00_desc_write(txd, 3, skbdesc->iv[0]); in rt73usb_write_tx_desc()
1507 _rt2x00_desc_write(txd, 4, skbdesc->iv[1]); in rt73usb_write_tx_desc()
1510 rt2x00_desc_read(txd, 5, &word); in rt73usb_write_tx_desc()
[all …]
/linux-4.4.14/drivers/ntb/
Dntb_transport.c1223 struct dma_async_tx_descriptor *txd; in ntb_async_rx() local
1266 txd = device->device_prep_dma_memcpy(chan, unmap->addr[1], in ntb_async_rx()
1269 if (!txd) in ntb_async_rx()
1272 txd->callback = ntb_rx_copy_callback; in ntb_async_rx()
1273 txd->callback_param = entry; in ntb_async_rx()
1274 dma_set_unmap(txd, unmap); in ntb_async_rx()
1276 cookie = dmaengine_submit(txd); in ntb_async_rx()
1453 struct dma_async_tx_descriptor *txd; in ntb_async_tx() local
1497 txd = device->device_prep_dma_memcpy(chan, dest, unmap->addr[0], len, in ntb_async_tx()
1499 if (!txd) in ntb_async_tx()
[all …]
/linux-4.4.14/drivers/net/ethernet/3com/
Dtyphoon.c717 struct tx_desc *txd, *first_txd; in typhoon_start_tx() local
787 txd = (struct tx_desc *) (txRing->ringBase + txRing->lastWrite); in typhoon_start_tx()
796 txd->flags = TYPHOON_FRAG_DESC | TYPHOON_DESC_VALID; in typhoon_start_tx()
797 txd->len = cpu_to_le16(skb->len); in typhoon_start_tx()
798 txd->frag.addr = cpu_to_le32(skb_dma); in typhoon_start_tx()
799 txd->frag.addrHi = 0; in typhoon_start_tx()
807 txd->flags = TYPHOON_FRAG_DESC | TYPHOON_DESC_VALID; in typhoon_start_tx()
808 txd->len = cpu_to_le16(len); in typhoon_start_tx()
809 txd->frag.addr = cpu_to_le32(skb_dma); in typhoon_start_tx()
810 txd->frag.addrHi = 0; in typhoon_start_tx()
[all …]
/linux-4.4.14/drivers/ata/
Dpata_ep93xx.c703 struct dma_async_tx_descriptor *txd; in ep93xx_pata_dma_start() local
711 txd = dmaengine_prep_slave_sg(channel, qc->sg, qc->n_elem, qc->dma_dir, in ep93xx_pata_dma_start()
713 if (!txd) { in ep93xx_pata_dma_start()
717 txd->callback = NULL; in ep93xx_pata_dma_start()
718 txd->callback_param = NULL; in ep93xx_pata_dma_start()
720 if (dmaengine_submit(txd) < 0) { in ep93xx_pata_dma_start()
/linux-4.4.14/drivers/net/fddi/skfp/
Dskfddi.c125 void mac_drv_tx_complete(struct s_smc *smc, volatile struct s_smt_fp_txd *txd);
1110 struct s_smt_fp_txd *txd; // Current TxD. in send_queued_packets() local
1174 txd = (struct s_smt_fp_txd *) HWM_GET_CURR_TXD(smc, queue); in send_queued_packets()
1179 txd->txd_os.skb = skb; // save skb in send_queued_packets()
1180 txd->txd_os.dma_addr = dma_address; // save dma mapping in send_queued_packets()
1491 void mac_drv_tx_complete(struct s_smc *smc, volatile struct s_smt_fp_txd *txd) in mac_drv_tx_complete() argument
1498 if (!(skb = txd->txd_os.skb)) { in mac_drv_tx_complete()
1502 txd->txd_os.skb = NULL; in mac_drv_tx_complete()
1505 pci_unmap_single(&smc->os.pdev, txd->txd_os.dma_addr, in mac_drv_tx_complete()
1507 txd->txd_os.dma_addr = 0; in mac_drv_tx_complete()
Dhwmtm.c102 volatile struct s_smt_fp_txd *txd);
/linux-4.4.14/drivers/net/hamradio/
Dyam.c126 int txd; /* tx delay */ member
610 if ((yp->tx_state == TX_TAIL) || (yp->txd == 0)) in yam_start_tx()
613 yp->tx_count = (yp->bitrate * yp->txd) / 8000; in yam_start_tx()
825 seq_printf(seq, " TxDelay %u\n", yp->txd); in yam_seq_show()
1042 yp->txd = yi.cfg.txdelay; in yam_ioctl()
1070 yi.cfg.txdelay = yp->txd; in yam_ioctl()
1119 yp->txd = DEFAULT_TXD; in yam_setup()
/linux-4.4.14/drivers/net/vmxnet3/
Dvmxnet3_drv.c337 BUG_ON(VMXNET3_TXDESC_GET_EOP(&(tq->tx_ring.base[eop_idx].txd)) != 1); in vmxnet3_unmap_pkt()
690 ctx->sop_txd->txd.addr = cpu_to_le64(tq->data_ring.basePA + in vmxnet3_map_pkt()
702 le64_to_cpu(ctx->sop_txd->txd.addr), in vmxnet3_map_pkt()
735 BUG_ON(gdesc->txd.gen == tq->tx_ring.gen); in vmxnet3_map_pkt()
737 gdesc->txd.addr = cpu_to_le64(tbi->dma_addr); in vmxnet3_map_pkt()
743 tq->tx_ring.next2fill, le64_to_cpu(gdesc->txd.addr), in vmxnet3_map_pkt()
777 BUG_ON(gdesc->txd.gen == tq->tx_ring.gen); in vmxnet3_map_pkt()
779 gdesc->txd.addr = cpu_to_le64(tbi->dma_addr); in vmxnet3_map_pkt()
785 tq->tx_ring.next2fill, le64_to_cpu(gdesc->txd.addr), in vmxnet3_map_pkt()
1055 gdesc->txd.hlen = ctx.eth_ip_hdr_size + ctx.l4_hdr_size; in vmxnet3_tq_xmit()
[all …]
Dvmxnet3_defs.h343 struct Vmxnet3_TxDesc txd; member
/linux-4.4.14/arch/arm/mach-imx/
Dmach-pcm043.c249 iomux_v3_cfg_t txd = MX35_PAD_STXD4__AUDMUX_AUD4_TXD; in pcm043_ac97_cold_reset() local
277 mxc_iomux_v3_setup_pad(txd); in pcm043_ac97_cold_reset()
/linux-4.4.14/drivers/net/ethernet/dec/tulip/
Dde2104x.c611 struct de_desc *txd; in de_start_xmit() local
625 txd = &de->tx_ring[entry]; in de_start_xmit()
634 txd->opts2 = cpu_to_le32(flags); in de_start_xmit()
635 txd->addr1 = cpu_to_le32(mapping); in de_start_xmit()
641 txd->opts1 = cpu_to_le32(DescOwn); in de_start_xmit()
726 struct de_desc *txd; in __de_set_rx_mode() local
775 txd = &de->tx_ring[entry]; in __de_set_rx_mode()
777 txd->opts2 = cpu_to_le32(SetupFrame | RingEnd | sizeof (de->setup_frame)); in __de_set_rx_mode()
779 txd->opts2 = cpu_to_le32(SetupFrame | sizeof (de->setup_frame)); in __de_set_rx_mode()
780 txd->addr1 = cpu_to_le32(mapping); in __de_set_rx_mode()
[all …]
/linux-4.4.14/drivers/net/ethernet/cirrus/
Dep93xx_eth.c355 struct ep93xx_tdesc *txd; in ep93xx_xmit() local
367 txd = &ep->descs->tdesc[entry]; in ep93xx_xmit()
369 txd->tdesc1 = TDESC1_EOF | (entry << 16) | (skb->len & 0xfff); in ep93xx_xmit()
370 dma_sync_single_for_cpu(dev->dev.parent, txd->buf_addr, skb->len, in ep93xx_xmit()
373 dma_sync_single_for_device(dev->dev.parent, txd->buf_addr, skb->len, in ep93xx_xmit()
/linux-4.4.14/drivers/net/ethernet/sfc/
Dfarch.c296 const efx_qword_t *txd) in efx_farch_push_tx_desc() argument
307 reg.qword[0] = *txd; in efx_farch_push_tx_desc()
320 efx_qword_t *txd; in efx_farch_tx_write() local
331 txd = efx_tx_desc(tx_queue, write_ptr); in efx_farch_tx_write()
338 EFX_POPULATE_QWORD_4(*txd, in efx_farch_tx_write()
349 txd = efx_tx_desc(tx_queue, in efx_farch_tx_write()
351 efx_farch_push_tx_desc(tx_queue, txd); in efx_farch_tx_write()
365 return efx_alloc_special_buffer(efx, &tx_queue->txd, in efx_farch_tx_probe()
375 efx_init_special_buffer(efx, &tx_queue->txd); in efx_farch_tx_init()
382 FRF_AZ_TX_DESCQ_BUF_BASE_ID, tx_queue->txd.index, in efx_farch_tx_init()
[all …]
Def10.c1737 return efx_nic_alloc_buffer(tx_queue->efx, &tx_queue->txd.buf, in efx_ef10_tx_probe()
1745 const efx_qword_t *txd) in efx_ef10_push_tx_desc() argument
1752 reg.qword[0] = *txd; in efx_ef10_push_tx_desc()
1762 size_t entries = tx_queue->txd.buf.len / EFX_BUF_SIZE; in efx_ef10_tx_init()
1768 efx_qword_t *txd; in efx_ef10_tx_init() local
1783 dma_addr = tx_queue->txd.buf.dma_addr; in efx_ef10_tx_init()
1808 txd = efx_tx_desc(tx_queue, 0); in efx_ef10_tx_init()
1809 EFX_POPULATE_QWORD_4(*txd, in efx_ef10_tx_init()
1817 efx_ef10_push_tx_desc(tx_queue, txd); in efx_ef10_tx_init()
1852 efx_nic_free_buffer(tx_queue->efx, &tx_queue->txd.buf); in efx_ef10_tx_remove()
[all …]
Dnic.h71 return ((efx_qword_t *) (tx_queue->txd.buf.addr)) + index; in efx_tx_desc()
Defx.c489 memset(&tx_queue->txd, 0, sizeof(tx_queue->txd)); in efx_copy_channel()
795 tx_queue->txd.index + in efx_realloc_channels()
796 tx_queue->txd.entries); in efx_realloc_channels()
Dnet_driver.h235 struct efx_special_buffer txd; member
/linux-4.4.14/arch/cris/arch-v10/kernel/
Dhead.S479 move.b IO_FIELD (R_SERIAL0_TR_CTRL, txd, 0) \
513 move.b IO_FIELD (R_SERIAL1_TR_CTRL, txd, 0) \
548 move.b IO_FIELD (R_SERIAL2_TR_CTRL, txd, 0) \
584 move.b IO_FIELD (R_SERIAL3_TR_CTRL, txd, 0) \
Ddebugport.c270 IO_FIELD(R_SERIAL0_TR_CTRL, txd, 0) | in start_port()
/linux-4.4.14/drivers/tty/serial/
Detraxfs-uart.c432 tr_ctrl.txd = 0; in etraxfs_uart_break_ctl()
438 tr_ctrl.txd = 1; in etraxfs_uart_break_ctl()
673 tx_ctrl.txd = 1; in etraxfs_uart_set_termios()
/linux-4.4.14/drivers/net/ethernet/intel/i40e/
Di40e_debugfs.c783 struct i40e_tx_desc *txd; in i40e_dbg_dump_desc() local
816 txd = I40E_TX_DESC(ring, i); in i40e_dbg_dump_desc()
819 i, txd->buffer_addr, in i40e_dbg_dump_desc()
820 txd->cmd_type_offset_bsz); in i40e_dbg_dump_desc()
844 txd = I40E_TX_DESC(ring, desc_n); in i40e_dbg_dump_desc()
848 txd->buffer_addr, txd->cmd_type_offset_bsz); in i40e_dbg_dump_desc()
/linux-4.4.14/drivers/net/wireless/brcm80211/brcmsmac/
Ddma.c313 static uint txd(struct dma_info *di, uint x) in txd() function
325 return txd(di, i + 1); in nexttxd()
330 return txd(di, i - 1); in prevtxd()
340 return txd(di, t-h); in ntxdactive()
/linux-4.4.14/drivers/net/wireless/
Dmwl8k.c162 struct mwl8k_tx_desc *txd; member
1453 txq->txd = pci_zalloc_consistent(priv->pdev, size, &txq->txd_dma); in mwl8k_txq_init()
1454 if (txq->txd == NULL) { in mwl8k_txq_init()
1461 pci_free_consistent(priv->pdev, size, txq->txd, txq->txd_dma); in mwl8k_txq_init()
1469 tx_desc = txq->txd + i; in mwl8k_txq_init()
1502 struct mwl8k_tx_desc *tx_desc = txq->txd + desc; in mwl8k_dump_tx_rings()
1678 tx_desc = txq->txd + tx; in mwl8k_txq_reclaim()
1759 if (txq->txd == NULL) in mwl8k_txq_deinit()
1769 txq->txd, txq->txd_dma); in mwl8k_txq_deinit()
1770 txq->txd = NULL; in mwl8k_txq_deinit()
[all …]
/linux-4.4.14/drivers/net/ethernet/sis/
Dsis190.c700 struct TxDesc *txd = tp->TxDescRing + entry; in sis190_tx_interrupt() local
701 u32 status = le32_to_cpu(txd->status); in sis190_tx_interrupt()
715 sis190_unmap_tx_skb(tp->pci_dev, skb, txd); in sis190_tx_interrupt()
/linux-4.4.14/drivers/net/ethernet/micrel/
Dks8851.c114 u8 txd[8]; member
250 __le16 *txb = (__le16 *)ks->txd; in ks8851_rdreg()
/linux-4.4.14/arch/arm/boot/dts/
Dexynos5250-pinctrl.dtsi789 c2c_txd: c2c-txd {
/linux-4.4.14/drivers/net/ethernet/marvell/
Dmv643xx_eth.c2089 struct tx_desc *txd = tx_desc + i; in txq_init() local
2096 txd->cmd_sts = 0; in txq_init()
2097 txd->next_desc_ptr = txq->tx_desc_dma + in txq_init()
/linux-4.4.14/Documentation/networking/
Dz8530drv.txt459 Example: sccparam /dev/scc0 txd 20