/linux-4.1.27/drivers/dma/ |
D | sun6i-dma.c | 138 struct sun6i_vchan *vchan; member 318 static inline void sun6i_dma_dump_lli(struct sun6i_vchan *vchan, in sun6i_dma_dump_lli() argument 323 dev_dbg(chan2dev(&vchan->vc.chan), in sun6i_dma_dump_lli() 358 static int sun6i_dma_start_desc(struct sun6i_vchan *vchan) in sun6i_dma_start_desc() argument 360 struct sun6i_dma_dev *sdev = to_sun6i_dma_dev(vchan->vc.chan.device); in sun6i_dma_start_desc() 361 struct virt_dma_desc *desc = vchan_next_desc(&vchan->vc); in sun6i_dma_start_desc() 362 struct sun6i_pchan *pchan = vchan->phy; in sun6i_dma_start_desc() 379 sun6i_dma_dump_lli(vchan, pchan->desc->v_lli); in sun6i_dma_start_desc() 401 struct sun6i_vchan *vchan; in sun6i_dma_tasklet() local 406 list_for_each_entry(vchan, &sdev->slave.channels, vc.chan.device_node) { in sun6i_dma_tasklet() [all …]
|
D | dma-jz4740.c | 116 struct virt_dma_chan vchan; member 137 return container_of(chan->vchan.chan.device, struct jz4740_dma_dev, in jz4740_dma_chan_get_dev() 143 return container_of(c, struct jz4740_dmaengine_chan, vchan.chan); in to_jz4740_dma_chan() 277 spin_lock_irqsave(&chan->vchan.lock, flags); in jz4740_dma_terminate_all() 281 vchan_get_all_descriptors(&chan->vchan, &head); in jz4740_dma_terminate_all() 282 spin_unlock_irqrestore(&chan->vchan.lock, flags); in jz4740_dma_terminate_all() 284 vchan_dma_desc_free_list(&chan->vchan, &head); in jz4740_dma_terminate_all() 300 vdesc = vchan_next_desc(&chan->vchan); in jz4740_dma_start_transfer() 340 spin_lock(&chan->vchan.lock); in jz4740_dma_chan_irq() 353 spin_unlock(&chan->vchan.lock); in jz4740_dma_chan_irq() [all …]
|
D | dma-jz4780.c | 126 struct virt_dma_chan vchan; member 155 return container_of(chan, struct jz4780_dma_chan, vchan.chan); in to_jz4780_dma_chan() 167 return container_of(jzchan->vchan.chan.device, struct jz4780_dma_dev, in jz4780_dma_chan_parent() 333 return vchan_tx_prep(&jzchan->vchan, &desc->vdesc, flags); in jz4780_dma_prep_slave_sg() 383 return vchan_tx_prep(&jzchan->vchan, &desc->vdesc, flags); in jz4780_dma_prep_dma_cyclic() 412 return vchan_tx_prep(&jzchan->vchan, &desc->vdesc, flags); in jz4780_dma_prep_dma_memcpy() 423 vdesc = vchan_next_desc(&jzchan->vchan); in jz4780_dma_begin() 479 spin_lock_irqsave(&jzchan->vchan.lock, flags); in jz4780_dma_issue_pending() 481 if (vchan_issue_pending(&jzchan->vchan) && !jzchan->desc) in jz4780_dma_issue_pending() 484 spin_unlock_irqrestore(&jzchan->vchan.lock, flags); in jz4780_dma_issue_pending() [all …]
|
D | fsl-edma.c | 148 struct virt_dma_chan vchan; member 220 return container_of(chan, struct fsl_edma_chan, vchan.chan); in to_fsl_edma_chan() 231 u32 ch = fsl_chan->vchan.chan.chan_id; in fsl_edma_enable_request() 240 u32 ch = fsl_chan->vchan.chan.chan_id; in fsl_edma_disable_request() 249 u32 ch = fsl_chan->vchan.chan.chan_id; in fsl_edma_chan_mux() 254 ch_off = fsl_chan->vchan.chan.chan_id % chans_per_mux; in fsl_edma_chan_mux() 298 spin_lock_irqsave(&fsl_chan->vchan.lock, flags); in fsl_edma_terminate_all() 301 vchan_get_all_descriptors(&fsl_chan->vchan, &head); in fsl_edma_terminate_all() 302 spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags); in fsl_edma_terminate_all() 303 vchan_dma_desc_free_list(&fsl_chan->vchan, &head); in fsl_edma_terminate_all() [all …]
|
D | edma.c | 109 struct virt_dma_chan vchan; member 135 return container_of(c, struct edma_chan, vchan.chan); in to_edma_chan() 154 struct device *dev = echan->vchan.chan.device->dev; in edma_execute() 161 vdesc = vchan_next_desc(&echan->vchan); in edma_execute() 182 dev_vdbg(echan->vchan.chan.device->dev, in edma_execute() 254 spin_lock_irqsave(&echan->vchan.lock, flags); in edma_terminate_all() 278 vchan_get_all_descriptors(&echan->vchan, &head); in edma_terminate_all() 279 spin_unlock_irqrestore(&echan->vchan.lock, flags); in edma_terminate_all() 280 vchan_dma_desc_free_list(&echan->vchan, &head); in edma_terminate_all() 542 return vchan_tx_prep(&echan->vchan, &edesc->vdesc, tx_flags); in edma_prep_slave_sg() [all …]
|
D | mmp_pdma.c | 123 struct mmp_pdma_chan *vchan; member 155 if (!phy->vchan) in enable_chan() 158 reg = DRCMR(phy->vchan->drcmr); in enable_chan() 162 if (phy->vchan->byte_align) in enable_chan() 195 if ((dcsr & DCSR_BUSERR) && (phy->vchan)) in clear_chan_irq() 196 dev_warn(phy->vchan->dev, "DCSR_BUSERR\n"); in clear_chan_irq() 208 tasklet_schedule(&phy->vchan->tasklet); in mmp_pdma_chan_handler() 260 if (!phy->vchan) { in lookup_phy() 261 phy->vchan = pchan; in lookup_phy() 287 pchan->phy->vchan = NULL; in mmp_pdma_free_phy()
|
D | k3dma.c | 90 struct k3_dma_chan *vchan; member 205 c = p->vchan; in k3_dma_int_handler() 278 p->vchan = NULL; in k3_dma_tasklet() 289 if (p->vchan == NULL && !list_empty(&d->chan_pending)) { in k3_dma_tasklet() 296 p->vchan = c; in k3_dma_tasklet() 306 c = p->vchan; in k3_dma_tasklet() 590 p->vchan = NULL; in k3_dma_terminate_all()
|
D | sa11x0-dma.c | 109 struct sa11x0_dma_chan *vchan; member 279 c = p->vchan; in sa11x0_dma_irq() 350 p->vchan = NULL; in sa11x0_dma_tasklet() 360 if (p->vchan == NULL && !list_empty(&d->chan_pending)) { in sa11x0_dma_tasklet() 368 p->vchan = c; in sa11x0_dma_tasklet() 378 c = p->vchan; in sa11x0_dma_tasklet() 791 p->vchan = NULL; in sa11x0_dma_device_terminate_all()
|
/linux-4.1.27/drivers/dma/hsu/ |
D | hsu.c | 124 vdesc = vchan_next_desc(&hsuc->vchan); in hsu_dma_start_transfer() 178 spin_lock_irqsave(&hsuc->vchan.lock, flags); in hsu_dma_irq() 191 spin_unlock_irqrestore(&hsuc->vchan.lock, flags); in hsu_dma_irq() 246 return vchan_tx_prep(&hsuc->vchan, &desc->vdesc, flags); in hsu_dma_prep_slave_sg() 254 spin_lock_irqsave(&hsuc->vchan.lock, flags); in hsu_dma_issue_pending() 255 if (vchan_issue_pending(&hsuc->vchan) && !hsuc->desc) in hsu_dma_issue_pending() 257 spin_unlock_irqrestore(&hsuc->vchan.lock, flags); in hsu_dma_issue_pending() 301 spin_lock_irqsave(&hsuc->vchan.lock, flags); in hsu_dma_tx_status() 302 vdesc = vchan_find_desc(&hsuc->vchan, cookie); in hsu_dma_tx_status() 311 spin_unlock_irqrestore(&hsuc->vchan.lock, flags); in hsu_dma_tx_status() [all …]
|
D | hsu.h | 81 struct virt_dma_chan vchan; member 95 return container_of(chan, struct hsu_dma_chan, vchan.chan); in to_hsu_dma_chan()
|
/linux-4.1.27/arch/ia64/include/asm/sn/ |
D | io.h | 253 sn_pci_set_vchan(struct pci_dev *pci_dev, unsigned long *addr, int vchan) in sn_pci_set_vchan() argument 256 if (vchan > 1) { in sn_pci_set_vchan() 263 if (vchan == 1) { in sn_pci_set_vchan()
|
/linux-4.1.27/arch/sh/drivers/dma/ |
D | dma-sysfs.c | 87 dma_configure_channel(channel->vchan, config); in dma_store_config() 130 dev->id = chan->vchan; in dma_create_sysfs_files()
|
D | dma-api.c | 87 if (channel->vchan == chan) in get_dma_channel() 391 chan->vchan = info->first_channel_nr + i + total_channels; in register_dmac()
|
/linux-4.1.27/arch/sh/include/asm/ |
D | dma.h | 70 unsigned int vchan; /* Virtual channel number */ member
|