Home
last modified time | relevance | path

Searched refs:vchan (Results 1 – 14 of 14) sorted by relevance

/linux-4.1.27/drivers/dma/
Dsun6i-dma.c138 struct sun6i_vchan *vchan; member
318 static inline void sun6i_dma_dump_lli(struct sun6i_vchan *vchan, in sun6i_dma_dump_lli() argument
323 dev_dbg(chan2dev(&vchan->vc.chan), in sun6i_dma_dump_lli()
358 static int sun6i_dma_start_desc(struct sun6i_vchan *vchan) in sun6i_dma_start_desc() argument
360 struct sun6i_dma_dev *sdev = to_sun6i_dma_dev(vchan->vc.chan.device); in sun6i_dma_start_desc()
361 struct virt_dma_desc *desc = vchan_next_desc(&vchan->vc); in sun6i_dma_start_desc()
362 struct sun6i_pchan *pchan = vchan->phy; in sun6i_dma_start_desc()
379 sun6i_dma_dump_lli(vchan, pchan->desc->v_lli); in sun6i_dma_start_desc()
401 struct sun6i_vchan *vchan; in sun6i_dma_tasklet() local
406 list_for_each_entry(vchan, &sdev->slave.channels, vc.chan.device_node) { in sun6i_dma_tasklet()
[all …]
Ddma-jz4740.c116 struct virt_dma_chan vchan; member
137 return container_of(chan->vchan.chan.device, struct jz4740_dma_dev, in jz4740_dma_chan_get_dev()
143 return container_of(c, struct jz4740_dmaengine_chan, vchan.chan); in to_jz4740_dma_chan()
277 spin_lock_irqsave(&chan->vchan.lock, flags); in jz4740_dma_terminate_all()
281 vchan_get_all_descriptors(&chan->vchan, &head); in jz4740_dma_terminate_all()
282 spin_unlock_irqrestore(&chan->vchan.lock, flags); in jz4740_dma_terminate_all()
284 vchan_dma_desc_free_list(&chan->vchan, &head); in jz4740_dma_terminate_all()
300 vdesc = vchan_next_desc(&chan->vchan); in jz4740_dma_start_transfer()
340 spin_lock(&chan->vchan.lock); in jz4740_dma_chan_irq()
353 spin_unlock(&chan->vchan.lock); in jz4740_dma_chan_irq()
[all …]
Ddma-jz4780.c126 struct virt_dma_chan vchan; member
155 return container_of(chan, struct jz4780_dma_chan, vchan.chan); in to_jz4780_dma_chan()
167 return container_of(jzchan->vchan.chan.device, struct jz4780_dma_dev, in jz4780_dma_chan_parent()
333 return vchan_tx_prep(&jzchan->vchan, &desc->vdesc, flags); in jz4780_dma_prep_slave_sg()
383 return vchan_tx_prep(&jzchan->vchan, &desc->vdesc, flags); in jz4780_dma_prep_dma_cyclic()
412 return vchan_tx_prep(&jzchan->vchan, &desc->vdesc, flags); in jz4780_dma_prep_dma_memcpy()
423 vdesc = vchan_next_desc(&jzchan->vchan); in jz4780_dma_begin()
479 spin_lock_irqsave(&jzchan->vchan.lock, flags); in jz4780_dma_issue_pending()
481 if (vchan_issue_pending(&jzchan->vchan) && !jzchan->desc) in jz4780_dma_issue_pending()
484 spin_unlock_irqrestore(&jzchan->vchan.lock, flags); in jz4780_dma_issue_pending()
[all …]
Dfsl-edma.c148 struct virt_dma_chan vchan; member
220 return container_of(chan, struct fsl_edma_chan, vchan.chan); in to_fsl_edma_chan()
231 u32 ch = fsl_chan->vchan.chan.chan_id; in fsl_edma_enable_request()
240 u32 ch = fsl_chan->vchan.chan.chan_id; in fsl_edma_disable_request()
249 u32 ch = fsl_chan->vchan.chan.chan_id; in fsl_edma_chan_mux()
254 ch_off = fsl_chan->vchan.chan.chan_id % chans_per_mux; in fsl_edma_chan_mux()
298 spin_lock_irqsave(&fsl_chan->vchan.lock, flags); in fsl_edma_terminate_all()
301 vchan_get_all_descriptors(&fsl_chan->vchan, &head); in fsl_edma_terminate_all()
302 spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags); in fsl_edma_terminate_all()
303 vchan_dma_desc_free_list(&fsl_chan->vchan, &head); in fsl_edma_terminate_all()
[all …]
Dedma.c109 struct virt_dma_chan vchan; member
135 return container_of(c, struct edma_chan, vchan.chan); in to_edma_chan()
154 struct device *dev = echan->vchan.chan.device->dev; in edma_execute()
161 vdesc = vchan_next_desc(&echan->vchan); in edma_execute()
182 dev_vdbg(echan->vchan.chan.device->dev, in edma_execute()
254 spin_lock_irqsave(&echan->vchan.lock, flags); in edma_terminate_all()
278 vchan_get_all_descriptors(&echan->vchan, &head); in edma_terminate_all()
279 spin_unlock_irqrestore(&echan->vchan.lock, flags); in edma_terminate_all()
280 vchan_dma_desc_free_list(&echan->vchan, &head); in edma_terminate_all()
542 return vchan_tx_prep(&echan->vchan, &edesc->vdesc, tx_flags); in edma_prep_slave_sg()
[all …]
Dmmp_pdma.c123 struct mmp_pdma_chan *vchan; member
155 if (!phy->vchan) in enable_chan()
158 reg = DRCMR(phy->vchan->drcmr); in enable_chan()
162 if (phy->vchan->byte_align) in enable_chan()
195 if ((dcsr & DCSR_BUSERR) && (phy->vchan)) in clear_chan_irq()
196 dev_warn(phy->vchan->dev, "DCSR_BUSERR\n"); in clear_chan_irq()
208 tasklet_schedule(&phy->vchan->tasklet); in mmp_pdma_chan_handler()
260 if (!phy->vchan) { in lookup_phy()
261 phy->vchan = pchan; in lookup_phy()
287 pchan->phy->vchan = NULL; in mmp_pdma_free_phy()
Dk3dma.c90 struct k3_dma_chan *vchan; member
205 c = p->vchan; in k3_dma_int_handler()
278 p->vchan = NULL; in k3_dma_tasklet()
289 if (p->vchan == NULL && !list_empty(&d->chan_pending)) { in k3_dma_tasklet()
296 p->vchan = c; in k3_dma_tasklet()
306 c = p->vchan; in k3_dma_tasklet()
590 p->vchan = NULL; in k3_dma_terminate_all()
Dsa11x0-dma.c109 struct sa11x0_dma_chan *vchan; member
279 c = p->vchan; in sa11x0_dma_irq()
350 p->vchan = NULL; in sa11x0_dma_tasklet()
360 if (p->vchan == NULL && !list_empty(&d->chan_pending)) { in sa11x0_dma_tasklet()
368 p->vchan = c; in sa11x0_dma_tasklet()
378 c = p->vchan; in sa11x0_dma_tasklet()
791 p->vchan = NULL; in sa11x0_dma_device_terminate_all()
/linux-4.1.27/drivers/dma/hsu/
Dhsu.c124 vdesc = vchan_next_desc(&hsuc->vchan); in hsu_dma_start_transfer()
178 spin_lock_irqsave(&hsuc->vchan.lock, flags); in hsu_dma_irq()
191 spin_unlock_irqrestore(&hsuc->vchan.lock, flags); in hsu_dma_irq()
246 return vchan_tx_prep(&hsuc->vchan, &desc->vdesc, flags); in hsu_dma_prep_slave_sg()
254 spin_lock_irqsave(&hsuc->vchan.lock, flags); in hsu_dma_issue_pending()
255 if (vchan_issue_pending(&hsuc->vchan) && !hsuc->desc) in hsu_dma_issue_pending()
257 spin_unlock_irqrestore(&hsuc->vchan.lock, flags); in hsu_dma_issue_pending()
301 spin_lock_irqsave(&hsuc->vchan.lock, flags); in hsu_dma_tx_status()
302 vdesc = vchan_find_desc(&hsuc->vchan, cookie); in hsu_dma_tx_status()
311 spin_unlock_irqrestore(&hsuc->vchan.lock, flags); in hsu_dma_tx_status()
[all …]
Dhsu.h81 struct virt_dma_chan vchan; member
95 return container_of(chan, struct hsu_dma_chan, vchan.chan); in to_hsu_dma_chan()
/linux-4.1.27/arch/ia64/include/asm/sn/
Dio.h253 sn_pci_set_vchan(struct pci_dev *pci_dev, unsigned long *addr, int vchan) in sn_pci_set_vchan() argument
256 if (vchan > 1) { in sn_pci_set_vchan()
263 if (vchan == 1) { in sn_pci_set_vchan()
/linux-4.1.27/arch/sh/drivers/dma/
Ddma-sysfs.c87 dma_configure_channel(channel->vchan, config); in dma_store_config()
130 dev->id = chan->vchan; in dma_create_sysfs_files()
Ddma-api.c87 if (channel->vchan == chan) in get_dma_channel()
391 chan->vchan = info->first_channel_nr + i + total_channels; in register_dmac()
/linux-4.1.27/arch/sh/include/asm/
Ddma.h70 unsigned int vchan; /* Virtual channel number */ member