vchan              84 arch/sh/drivers/dma/dma-api.c 		if (channel->vchan == chan)
vchan             376 arch/sh/drivers/dma/dma-api.c 		chan->vchan = info->first_channel_nr + i + total_channels;
vchan              84 arch/sh/drivers/dma/dma-sysfs.c 	dma_configure_channel(channel->vchan, config);
vchan             127 arch/sh/drivers/dma/dma-sysfs.c 	dev->id  = chan->vchan;
vchan              67 arch/sh/include/asm/dma.h 	unsigned int vchan;		/* Virtual channel number */
vchan             109 drivers/dma/dma-axi-dmac.c 	struct virt_dma_chan vchan;
vchan             143 drivers/dma/dma-axi-dmac.c 	return container_of(chan->vchan.chan.device, struct axi_dmac,
vchan             149 drivers/dma/dma-axi-dmac.c 	return container_of(c, struct axi_dmac_chan, vchan.chan);
vchan             210 drivers/dma/dma-axi-dmac.c 		vdesc = vchan_next_desc(&chan->vchan);
vchan             410 drivers/dma/dma-axi-dmac.c 	spin_lock(&dmac->chan.vchan.lock);
vchan             421 drivers/dma/dma-axi-dmac.c 	spin_unlock(&dmac->chan.vchan.lock);
vchan             433 drivers/dma/dma-axi-dmac.c 	spin_lock_irqsave(&chan->vchan.lock, flags);
vchan             436 drivers/dma/dma-axi-dmac.c 	vchan_get_all_descriptors(&chan->vchan, &head);
vchan             438 drivers/dma/dma-axi-dmac.c 	spin_unlock_irqrestore(&chan->vchan.lock, flags);
vchan             440 drivers/dma/dma-axi-dmac.c 	vchan_dma_desc_free_list(&chan->vchan, &head);
vchan             449 drivers/dma/dma-axi-dmac.c 	vchan_synchronize(&chan->vchan);
vchan             460 drivers/dma/dma-axi-dmac.c 	spin_lock_irqsave(&chan->vchan.lock, flags);
vchan             461 drivers/dma/dma-axi-dmac.c 	if (vchan_issue_pending(&chan->vchan))
vchan             463 drivers/dma/dma-axi-dmac.c 	spin_unlock_irqrestore(&chan->vchan.lock, flags);
vchan             564 drivers/dma/dma-axi-dmac.c 	return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags);
vchan             598 drivers/dma/dma-axi-dmac.c 	return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags);
vchan             669 drivers/dma/dma-axi-dmac.c 	return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags);
vchan             894 drivers/dma/dma-axi-dmac.c 	dmac->chan.vchan.desc_free = axi_dmac_desc_free;
vchan             895 drivers/dma/dma-axi-dmac.c 	vchan_init(&dmac->chan.vchan, dma_dev);
vchan             952 drivers/dma/dma-axi-dmac.c 	tasklet_kill(&dmac->chan.vchan.task);
vchan             129 drivers/dma/dma-jz4780.c 	struct virt_dma_chan vchan;
vchan             166 drivers/dma/dma-jz4780.c 	return container_of(chan, struct jz4780_dma_chan, vchan.chan);
vchan             178 drivers/dma/dma-jz4780.c 	return container_of(jzchan->vchan.chan.device, struct jz4780_dma_dev,
vchan             395 drivers/dma/dma-jz4780.c 	return vchan_tx_prep(&jzchan->vchan, &desc->vdesc, flags);
vchan             447 drivers/dma/dma-jz4780.c 	return vchan_tx_prep(&jzchan->vchan, &desc->vdesc, flags);
vchan             475 drivers/dma/dma-jz4780.c 	return vchan_tx_prep(&jzchan->vchan, &desc->vdesc, flags);
vchan             486 drivers/dma/dma-jz4780.c 		vdesc = vchan_next_desc(&jzchan->vchan);
vchan             558 drivers/dma/dma-jz4780.c 	spin_lock_irqsave(&jzchan->vchan.lock, flags);
vchan             560 drivers/dma/dma-jz4780.c 	if (vchan_issue_pending(&jzchan->vchan) && !jzchan->desc)
vchan             563 drivers/dma/dma-jz4780.c 	spin_unlock_irqrestore(&jzchan->vchan.lock, flags);
vchan             573 drivers/dma/dma-jz4780.c 	spin_lock_irqsave(&jzchan->vchan.lock, flags);
vchan             584 drivers/dma/dma-jz4780.c 	vchan_get_all_descriptors(&jzchan->vchan, &head);
vchan             586 drivers/dma/dma-jz4780.c 	spin_unlock_irqrestore(&jzchan->vchan.lock, flags);
vchan             588 drivers/dma/dma-jz4780.c 	vchan_dma_desc_free_list(&jzchan->vchan, &head);
vchan             597 drivers/dma/dma-jz4780.c 	vchan_synchronize(&jzchan->vchan);
vchan             646 drivers/dma/dma-jz4780.c 	spin_lock_irqsave(&jzchan->vchan.lock, flags);
vchan             648 drivers/dma/dma-jz4780.c 	vdesc = vchan_find_desc(&jzchan->vchan, cookie);
vchan             663 drivers/dma/dma-jz4780.c 	spin_unlock_irqrestore(&jzchan->vchan.lock, flags);
vchan             675 drivers/dma/dma-jz4780.c 	spin_lock(&jzchan->vchan.lock);
vchan             681 drivers/dma/dma-jz4780.c 		dev_warn(&jzchan->vchan.chan.dev->device,
vchan             686 drivers/dma/dma-jz4780.c 		dev_warn(&jzchan->vchan.chan.dev->device,
vchan             715 drivers/dma/dma-jz4780.c 		dev_err(&jzchan->vchan.chan.dev->device,
vchan             719 drivers/dma/dma-jz4780.c 	spin_unlock(&jzchan->vchan.lock);
vchan             771 drivers/dma/dma-jz4780.c 	vchan_free_chan_resources(&jzchan->vchan);
vchan             827 drivers/dma/dma-jz4780.c 			&jzdma->chan[data.channel].vchan.chan);
vchan             954 drivers/dma/dma-jz4780.c 		vchan_init(&jzchan->vchan, dd);
vchan             955 drivers/dma/dma-jz4780.c 		jzchan->vchan.desc_free = jz4780_dma_desc_free;
vchan             993 drivers/dma/dma-jz4780.c 		tasklet_kill(&jzdma->chan[i].vchan.task);
vchan              48 drivers/dma/fsl-edma-common.c 	u32 ch = fsl_chan->vchan.chan.chan_id;
vchan              65 drivers/dma/fsl-edma-common.c 	u32 ch = fsl_chan->vchan.chan.chan_id;
vchan             109 drivers/dma/fsl-edma-common.c 	u32 ch = fsl_chan->vchan.chan.chan_id;
vchan             115 drivers/dma/fsl-edma-common.c 	ch_off = fsl_chan->vchan.chan.chan_id % chans_per_mux;
vchan             161 drivers/dma/fsl-edma-common.c 	spin_lock_irqsave(&fsl_chan->vchan.lock, flags);
vchan             165 drivers/dma/fsl-edma-common.c 	vchan_get_all_descriptors(&fsl_chan->vchan, &head);
vchan             166 drivers/dma/fsl-edma-common.c 	spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags);
vchan             167 drivers/dma/fsl-edma-common.c 	vchan_dma_desc_free_list(&fsl_chan->vchan, &head);
vchan             177 drivers/dma/fsl-edma-common.c 	spin_lock_irqsave(&fsl_chan->vchan.lock, flags);
vchan             183 drivers/dma/fsl-edma-common.c 	spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags);
vchan             193 drivers/dma/fsl-edma-common.c 	spin_lock_irqsave(&fsl_chan->vchan.lock, flags);
vchan             199 drivers/dma/fsl-edma-common.c 	spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags);
vchan             207 drivers/dma/fsl-edma-common.c 		dma_unmap_resource(fsl_chan->vchan.chan.device->dev,
vchan             217 drivers/dma/fsl-edma-common.c 	struct device *dev = fsl_chan->vchan.chan.device->dev;
vchan             270 drivers/dma/fsl-edma-common.c 	u32 ch = fsl_chan->vchan.chan.chan_id;
vchan             323 drivers/dma/fsl-edma-common.c 	spin_lock_irqsave(&fsl_chan->vchan.lock, flags);
vchan             324 drivers/dma/fsl-edma-common.c 	vdesc = vchan_find_desc(&fsl_chan->vchan, cookie);
vchan             334 drivers/dma/fsl-edma-common.c 	spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags);
vchan             345 drivers/dma/fsl-edma-common.c 	u32 ch = fsl_chan->vchan.chan.chan_id;
vchan             507 drivers/dma/fsl-edma-common.c 	return vchan_tx_prep(&fsl_chan->vchan, &fsl_desc->vdesc, flags);
vchan             579 drivers/dma/fsl-edma-common.c 	return vchan_tx_prep(&fsl_chan->vchan, &fsl_desc->vdesc, flags);
vchan             587 drivers/dma/fsl-edma-common.c 	vdesc = vchan_next_desc(&fsl_chan->vchan);
vchan             603 drivers/dma/fsl-edma-common.c 	spin_lock_irqsave(&fsl_chan->vchan.lock, flags);
vchan             606 drivers/dma/fsl-edma-common.c 		spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags);
vchan             611 drivers/dma/fsl-edma-common.c 	if (vchan_issue_pending(&fsl_chan->vchan) && !fsl_chan->edesc)
vchan             614 drivers/dma/fsl-edma-common.c 	spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags);
vchan             635 drivers/dma/fsl-edma-common.c 	spin_lock_irqsave(&fsl_chan->vchan.lock, flags);
vchan             639 drivers/dma/fsl-edma-common.c 	vchan_get_all_descriptors(&fsl_chan->vchan, &head);
vchan             641 drivers/dma/fsl-edma-common.c 	spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags);
vchan             643 drivers/dma/fsl-edma-common.c 	vchan_dma_desc_free_list(&fsl_chan->vchan, &head);
vchan             654 drivers/dma/fsl-edma-common.c 				&dmadev->channels, vchan.chan.device_node) {
vchan             655 drivers/dma/fsl-edma-common.c 		list_del(&chan->vchan.chan.device_node);
vchan             656 drivers/dma/fsl-edma-common.c 		tasklet_kill(&chan->vchan.task);
vchan             115 drivers/dma/fsl-edma-common.h 	struct virt_dma_chan		vchan;
vchan             215 drivers/dma/fsl-edma-common.h 	return container_of(chan, struct fsl_edma_chan, vchan.chan);
vchan              27 drivers/dma/fsl-edma.c 	vchan_synchronize(&fsl_chan->vchan);
vchan              47 drivers/dma/fsl-edma.c 			spin_lock(&fsl_chan->vchan.lock);
vchan              61 drivers/dma/fsl-edma.c 			spin_unlock(&fsl_chan->vchan.lock);
vchan             345 drivers/dma/fsl-edma.c 		fsl_chan->vchan.desc_free = fsl_edma_free_desc;
vchan             346 drivers/dma/fsl-edma.c 		vchan_init(&fsl_chan->vchan, &fsl_edma->dma_dev);
vchan             428 drivers/dma/fsl-edma.c 		spin_lock_irqsave(&fsl_chan->vchan.lock, flags);
vchan             437 drivers/dma/fsl-edma.c 		spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags);
vchan             165 drivers/dma/fsl-qdma.c 	struct virt_dma_chan		vchan;
vchan             284 drivers/dma/fsl-qdma.c 	return container_of(chan, struct fsl_qdma_chan, vchan.chan);
vchan             301 drivers/dma/fsl-qdma.c 	spin_lock_irqsave(&fsl_chan->vchan.lock, flags);
vchan             302 drivers/dma/fsl-qdma.c 	vchan_get_all_descriptors(&fsl_chan->vchan, &head);
vchan             303 drivers/dma/fsl-qdma.c 	spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags);
vchan             305 drivers/dma/fsl-qdma.c 	vchan_dma_desc_free_list(&fsl_chan->vchan, &head);
vchan             689 drivers/dma/fsl-qdma.c 		spin_lock(&fsl_comp->qchan->vchan.lock);
vchan             692 drivers/dma/fsl-qdma.c 		spin_unlock(&fsl_comp->qchan->vchan.lock);
vchan             934 drivers/dma/fsl-qdma.c 	return vchan_tx_prep(&fsl_chan->vchan, &fsl_comp->vdesc, flags);
vchan             948 drivers/dma/fsl-qdma.c 	vdesc = vchan_next_desc(&fsl_chan->vchan);
vchan             988 drivers/dma/fsl-qdma.c 	spin_lock(&fsl_chan->vchan.lock);
vchan             989 drivers/dma/fsl-qdma.c 	if (vchan_issue_pending(&fsl_chan->vchan))
vchan             991 drivers/dma/fsl-qdma.c 	spin_unlock(&fsl_chan->vchan.lock);
vchan             999 drivers/dma/fsl-qdma.c 	vchan_synchronize(&fsl_chan->vchan);
vchan            1008 drivers/dma/fsl-qdma.c 	spin_lock_irqsave(&fsl_chan->vchan.lock, flags);
vchan            1009 drivers/dma/fsl-qdma.c 	vchan_get_all_descriptors(&fsl_chan->vchan, &head);
vchan            1010 drivers/dma/fsl-qdma.c 	spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags);
vchan            1011 drivers/dma/fsl-qdma.c 	vchan_dma_desc_free_list(&fsl_chan->vchan, &head);
vchan            1170 drivers/dma/fsl-qdma.c 		fsl_chan->vchan.desc_free = fsl_qdma_free_desc;
vchan            1171 drivers/dma/fsl-qdma.c 		vchan_init(&fsl_chan->vchan, &fsl_qdma->dma_dev);
vchan            1212 drivers/dma/fsl-qdma.c 				 &dmadev->channels, vchan.chan.device_node) {
vchan            1213 drivers/dma/fsl-qdma.c 		list_del(&chan->vchan.chan.device_node);
vchan            1214 drivers/dma/fsl-qdma.c 		tasklet_kill(&chan->vchan.task);
vchan             113 drivers/dma/hsu/hsu.c 	vdesc = vchan_next_desc(&hsuc->vchan);
vchan             160 drivers/dma/hsu/hsu.c 	spin_lock_irqsave(&hsuc->vchan.lock, flags);
vchan             162 drivers/dma/hsu/hsu.c 	spin_unlock_irqrestore(&hsuc->vchan.lock, flags);
vchan             214 drivers/dma/hsu/hsu.c 	spin_lock_irqsave(&hsuc->vchan.lock, flags);
vchan             227 drivers/dma/hsu/hsu.c 	spin_unlock_irqrestore(&hsuc->vchan.lock, flags);
vchan             284 drivers/dma/hsu/hsu.c 	return vchan_tx_prep(&hsuc->vchan, &desc->vdesc, flags);
vchan             292 drivers/dma/hsu/hsu.c 	spin_lock_irqsave(&hsuc->vchan.lock, flags);
vchan             293 drivers/dma/hsu/hsu.c 	if (vchan_issue_pending(&hsuc->vchan) && !hsuc->desc)
vchan             295 drivers/dma/hsu/hsu.c 	spin_unlock_irqrestore(&hsuc->vchan.lock, flags);
vchan             328 drivers/dma/hsu/hsu.c 	spin_lock_irqsave(&hsuc->vchan.lock, flags);
vchan             329 drivers/dma/hsu/hsu.c 	vdesc = vchan_find_desc(&hsuc->vchan, cookie);
vchan             338 drivers/dma/hsu/hsu.c 	spin_unlock_irqrestore(&hsuc->vchan.lock, flags);
vchan             358 drivers/dma/hsu/hsu.c 	spin_lock_irqsave(&hsuc->vchan.lock, flags);
vchan             363 drivers/dma/hsu/hsu.c 	spin_unlock_irqrestore(&hsuc->vchan.lock, flags);
vchan             373 drivers/dma/hsu/hsu.c 	spin_lock_irqsave(&hsuc->vchan.lock, flags);
vchan             378 drivers/dma/hsu/hsu.c 	spin_unlock_irqrestore(&hsuc->vchan.lock, flags);
vchan             389 drivers/dma/hsu/hsu.c 	spin_lock_irqsave(&hsuc->vchan.lock, flags);
vchan             397 drivers/dma/hsu/hsu.c 	vchan_get_all_descriptors(&hsuc->vchan, &head);
vchan             398 drivers/dma/hsu/hsu.c 	spin_unlock_irqrestore(&hsuc->vchan.lock, flags);
vchan             399 drivers/dma/hsu/hsu.c 	vchan_dma_desc_free_list(&hsuc->vchan, &head);
vchan             413 drivers/dma/hsu/hsu.c 	vchan_synchronize(&hsuc->vchan);
vchan             441 drivers/dma/hsu/hsu.c 		hsuc->vchan.desc_free = hsu_dma_desc_free;
vchan             442 drivers/dma/hsu/hsu.c 		vchan_init(&hsuc->vchan, &hsu->dma);
vchan             492 drivers/dma/hsu/hsu.c 		tasklet_kill(&hsuc->vchan.task);
vchan              83 drivers/dma/hsu/hsu.h 	struct virt_dma_chan vchan;
vchan              96 drivers/dma/hsu/hsu.h 	return container_of(chan, struct hsu_dma_chan, vchan.chan);
vchan             107 drivers/dma/idma64.c 	struct idma64 *idma64 = to_idma64(idma64c->vchan.chan.device);
vchan             114 drivers/dma/idma64.c 	struct idma64 *idma64 = to_idma64(idma64c->vchan.chan.device);
vchan             118 drivers/dma/idma64.c 	vdesc = vchan_next_desc(&idma64c->vchan);
vchan             142 drivers/dma/idma64.c 	spin_lock(&idma64c->vchan.lock);
vchan             159 drivers/dma/idma64.c 	spin_unlock(&idma64c->vchan.lock);
vchan             326 drivers/dma/idma64.c 	return vchan_tx_prep(&idma64c->vchan, &desc->vdesc, flags);
vchan             334 drivers/dma/idma64.c 	spin_lock_irqsave(&idma64c->vchan.lock, flags);
vchan             335 drivers/dma/idma64.c 	if (vchan_issue_pending(&idma64c->vchan) && !idma64c->desc)
vchan             337 drivers/dma/idma64.c 	spin_unlock_irqrestore(&idma64c->vchan.lock, flags);
vchan             378 drivers/dma/idma64.c 	spin_lock_irqsave(&idma64c->vchan.lock, flags);
vchan             379 drivers/dma/idma64.c 	vdesc = vchan_find_desc(&idma64c->vchan, cookie);
vchan             388 drivers/dma/idma64.c 	spin_unlock_irqrestore(&idma64c->vchan.lock, flags);
vchan             445 drivers/dma/idma64.c 	spin_lock_irqsave(&idma64c->vchan.lock, flags);
vchan             450 drivers/dma/idma64.c 	spin_unlock_irqrestore(&idma64c->vchan.lock, flags);
vchan             460 drivers/dma/idma64.c 	spin_lock_irqsave(&idma64c->vchan.lock, flags);
vchan             465 drivers/dma/idma64.c 	spin_unlock_irqrestore(&idma64c->vchan.lock, flags);
vchan             476 drivers/dma/idma64.c 	spin_lock_irqsave(&idma64c->vchan.lock, flags);
vchan             483 drivers/dma/idma64.c 	vchan_get_all_descriptors(&idma64c->vchan, &head);
vchan             484 drivers/dma/idma64.c 	spin_unlock_irqrestore(&idma64c->vchan.lock, flags);
vchan             486 drivers/dma/idma64.c 	vchan_dma_desc_free_list(&idma64c->vchan, &head);
vchan             494 drivers/dma/idma64.c 	vchan_synchronize(&idma64c->vchan);
vchan             562 drivers/dma/idma64.c 		idma64c->vchan.desc_free = idma64_vdesc_free;
vchan             563 drivers/dma/idma64.c 		vchan_init(&idma64c->vchan, &idma64->dma);
vchan             619 drivers/dma/idma64.c 		tasklet_kill(&idma64c->vchan.task);
vchan             128 drivers/dma/idma64.h 	struct virt_dma_chan vchan;
vchan             143 drivers/dma/idma64.h 	return container_of(chan, struct idma64_chan, vchan.chan);
vchan              94 drivers/dma/k3dma.c 	struct k3_dma_chan	*vchan;
vchan             229 drivers/dma/k3dma.c 			c = p->vchan;
vchan             317 drivers/dma/k3dma.c 				p->vchan = NULL;
vchan             331 drivers/dma/k3dma.c 		if (p->vchan == NULL && !list_empty(&d->chan_pending)) {
vchan             338 drivers/dma/k3dma.c 			p->vchan = c;
vchan             351 drivers/dma/k3dma.c 			c = p->vchan;
vchan             743 drivers/dma/k3dma.c 		p->vchan = NULL;
vchan              37 drivers/dma/mcf-edma.c 			spin_lock(&mcf_chan->vchan.lock);
vchan              51 drivers/dma/mcf-edma.c 			spin_unlock(&mcf_chan->vchan.lock);
vchan             223 drivers/dma/mcf-edma.c 		mcf_chan->vchan.desc_free = fsl_edma_free_desc;
vchan             224 drivers/dma/mcf-edma.c 		vchan_init(&mcf_chan->vchan, &mcf_edma->dma_dev);
vchan             120 drivers/dma/mmp_pdma.c 	struct mmp_pdma_chan *vchan;
vchan             156 drivers/dma/mmp_pdma.c 	if (!phy->vchan)
vchan             159 drivers/dma/mmp_pdma.c 	reg = DRCMR(phy->vchan->drcmr);
vchan             163 drivers/dma/mmp_pdma.c 	if (phy->vchan->byte_align)
vchan             196 drivers/dma/mmp_pdma.c 	if ((dcsr & DCSR_BUSERR) && (phy->vchan))
vchan             197 drivers/dma/mmp_pdma.c 		dev_warn(phy->vchan->dev, "DCSR_BUSERR\n");
vchan             209 drivers/dma/mmp_pdma.c 	tasklet_schedule(&phy->vchan->tasklet);
vchan             261 drivers/dma/mmp_pdma.c 			if (!phy->vchan) {
vchan             262 drivers/dma/mmp_pdma.c 				phy->vchan = pchan;
vchan             288 drivers/dma/mmp_pdma.c 	pchan->phy->vchan = NULL;
vchan             182 drivers/dma/owl-dma.c 	struct owl_dma_vchan	*vchan;
vchan             359 drivers/dma/owl-dma.c static inline int owl_dma_cfg_lli(struct owl_dma_vchan *vchan,
vchan             379 drivers/dma/owl-dma.c 		mode |= OWL_DMA_MODE_TS(vchan->drq)
vchan             392 drivers/dma/owl-dma.c 		 mode |= OWL_DMA_MODE_TS(vchan->drq)
vchan             429 drivers/dma/owl-dma.c 					       struct owl_dma_vchan *vchan)
vchan             439 drivers/dma/owl-dma.c 		if (!pchan->vchan) {
vchan             440 drivers/dma/owl-dma.c 			pchan->vchan = vchan;
vchan             480 drivers/dma/owl-dma.c 	pchan->vchan = NULL;
vchan             495 drivers/dma/owl-dma.c static int owl_dma_start_next_txd(struct owl_dma_vchan *vchan)
vchan             497 drivers/dma/owl-dma.c 	struct owl_dma *od = to_owl_dma(vchan->vc.chan.device);
vchan             498 drivers/dma/owl-dma.c 	struct virt_dma_desc *vd = vchan_next_desc(&vchan->vc);
vchan             499 drivers/dma/owl-dma.c 	struct owl_dma_pchan *pchan = vchan->pchan;
vchan             507 drivers/dma/owl-dma.c 	vchan->txd = txd;
vchan             536 drivers/dma/owl-dma.c 	dev_dbg(chan2dev(&vchan->vc.chan), "starting pchan %d\n", pchan->id);
vchan             544 drivers/dma/owl-dma.c static void owl_dma_phy_free(struct owl_dma *od, struct owl_dma_vchan *vchan)
vchan             547 drivers/dma/owl-dma.c 	owl_dma_terminate_pchan(od, vchan->pchan);
vchan             549 drivers/dma/owl-dma.c 	vchan->pchan = NULL;
vchan             555 drivers/dma/owl-dma.c 	struct owl_dma_vchan *vchan;
vchan             605 drivers/dma/owl-dma.c 		vchan = pchan->vchan;
vchan             606 drivers/dma/owl-dma.c 		if (!vchan) {
vchan             612 drivers/dma/owl-dma.c 		spin_lock(&vchan->vc.lock);
vchan             614 drivers/dma/owl-dma.c 		txd = vchan->txd;
vchan             616 drivers/dma/owl-dma.c 			vchan->txd = NULL;
vchan             624 drivers/dma/owl-dma.c 			if (vchan_next_desc(&vchan->vc))
vchan             625 drivers/dma/owl-dma.c 				owl_dma_start_next_txd(vchan);
vchan             627 drivers/dma/owl-dma.c 				owl_dma_phy_free(od, vchan);
vchan             630 drivers/dma/owl-dma.c 		spin_unlock(&vchan->vc.lock);
vchan             660 drivers/dma/owl-dma.c 	struct owl_dma_vchan *vchan = to_owl_vchan(chan);
vchan             664 drivers/dma/owl-dma.c 	spin_lock_irqsave(&vchan->vc.lock, flags);
vchan             666 drivers/dma/owl-dma.c 	if (vchan->pchan)
vchan             667 drivers/dma/owl-dma.c 		owl_dma_phy_free(od, vchan);
vchan             669 drivers/dma/owl-dma.c 	if (vchan->txd) {
vchan             670 drivers/dma/owl-dma.c 		owl_dma_desc_free(&vchan->txd->vd);
vchan             671 drivers/dma/owl-dma.c 		vchan->txd = NULL;
vchan             674 drivers/dma/owl-dma.c 	vchan_get_all_descriptors(&vchan->vc, &head);
vchan             675 drivers/dma/owl-dma.c 	vchan_dma_desc_free_list(&vchan->vc, &head);
vchan             677 drivers/dma/owl-dma.c 	spin_unlock_irqrestore(&vchan->vc.lock, flags);
vchan             685 drivers/dma/owl-dma.c 	struct owl_dma_vchan *vchan = to_owl_vchan(chan);
vchan             692 drivers/dma/owl-dma.c 	memcpy(&vchan->cfg, config, sizeof(struct dma_slave_config));
vchan             699 drivers/dma/owl-dma.c 	struct owl_dma_vchan *vchan = to_owl_vchan(chan);
vchan             702 drivers/dma/owl-dma.c 	spin_lock_irqsave(&vchan->vc.lock, flags);
vchan             704 drivers/dma/owl-dma.c 	owl_dma_pause_pchan(vchan->pchan);
vchan             706 drivers/dma/owl-dma.c 	spin_unlock_irqrestore(&vchan->vc.lock, flags);
vchan             713 drivers/dma/owl-dma.c 	struct owl_dma_vchan *vchan = to_owl_vchan(chan);
vchan             716 drivers/dma/owl-dma.c 	if (!vchan->pchan && !vchan->txd)
vchan             719 drivers/dma/owl-dma.c 	dev_dbg(chan2dev(chan), "vchan %p: resume\n", &vchan->vc);
vchan             721 drivers/dma/owl-dma.c 	spin_lock_irqsave(&vchan->vc.lock, flags);
vchan             723 drivers/dma/owl-dma.c 	owl_dma_resume_pchan(vchan->pchan);
vchan             725 drivers/dma/owl-dma.c 	spin_unlock_irqrestore(&vchan->vc.lock, flags);
vchan             730 drivers/dma/owl-dma.c static u32 owl_dma_getbytes_chan(struct owl_dma_vchan *vchan)
vchan             738 drivers/dma/owl-dma.c 	pchan = vchan->pchan;
vchan             739 drivers/dma/owl-dma.c 	txd = vchan->txd;
vchan             767 drivers/dma/owl-dma.c 	struct owl_dma_vchan *vchan = to_owl_vchan(chan);
vchan             779 drivers/dma/owl-dma.c 	spin_lock_irqsave(&vchan->vc.lock, flags);
vchan             781 drivers/dma/owl-dma.c 	vd = vchan_find_desc(&vchan->vc, cookie);
vchan             787 drivers/dma/owl-dma.c 		bytes = owl_dma_getbytes_chan(vchan);
vchan             790 drivers/dma/owl-dma.c 	spin_unlock_irqrestore(&vchan->vc.lock, flags);
vchan             797 drivers/dma/owl-dma.c static void owl_dma_phy_alloc_and_start(struct owl_dma_vchan *vchan)
vchan             799 drivers/dma/owl-dma.c 	struct owl_dma *od = to_owl_dma(vchan->vc.chan.device);
vchan             802 drivers/dma/owl-dma.c 	pchan = owl_dma_get_pchan(od, vchan);
vchan             808 drivers/dma/owl-dma.c 	vchan->pchan = pchan;
vchan             809 drivers/dma/owl-dma.c 	owl_dma_start_next_txd(vchan);
vchan             814 drivers/dma/owl-dma.c 	struct owl_dma_vchan *vchan = to_owl_vchan(chan);
vchan             817 drivers/dma/owl-dma.c 	spin_lock_irqsave(&vchan->vc.lock, flags);
vchan             818 drivers/dma/owl-dma.c 	if (vchan_issue_pending(&vchan->vc)) {
vchan             819 drivers/dma/owl-dma.c 		if (!vchan->pchan)
vchan             820 drivers/dma/owl-dma.c 			owl_dma_phy_alloc_and_start(vchan);
vchan             822 drivers/dma/owl-dma.c 	spin_unlock_irqrestore(&vchan->vc.lock, flags);
vchan             831 drivers/dma/owl-dma.c 	struct owl_dma_vchan *vchan = to_owl_vchan(chan);
vchan             856 drivers/dma/owl-dma.c 		ret = owl_dma_cfg_lli(vchan, lli, src + offset, dst + offset,
vchan             858 drivers/dma/owl-dma.c 				      &vchan->cfg, txd->cyclic);
vchan             867 drivers/dma/owl-dma.c 	return vchan_tx_prep(&vchan->vc, &txd->vd, flags);
vchan             882 drivers/dma/owl-dma.c 	struct owl_dma_vchan *vchan = to_owl_vchan(chan);
vchan             883 drivers/dma/owl-dma.c 	struct dma_slave_config *sconfig = &vchan->cfg;
vchan             921 drivers/dma/owl-dma.c 		ret = owl_dma_cfg_lli(vchan, lli, src, dst, len, dir, sconfig,
vchan             931 drivers/dma/owl-dma.c 	return vchan_tx_prep(&vchan->vc, &txd->vd, flags);
vchan             947 drivers/dma/owl-dma.c 	struct owl_dma_vchan *vchan = to_owl_vchan(chan);
vchan             948 drivers/dma/owl-dma.c 	struct dma_slave_config *sconfig = &vchan->cfg;
vchan             977 drivers/dma/owl-dma.c 		ret = owl_dma_cfg_lli(vchan, lli, src, dst, period_len,
vchan             993 drivers/dma/owl-dma.c 	return vchan_tx_prep(&vchan->vc, &txd->vd, flags);
vchan            1003 drivers/dma/owl-dma.c 	struct owl_dma_vchan *vchan = to_owl_vchan(chan);
vchan            1006 drivers/dma/owl-dma.c 	vchan_free_chan_resources(&vchan->vc);
vchan            1011 drivers/dma/owl-dma.c 	struct owl_dma_vchan *vchan = NULL;
vchan            1014 drivers/dma/owl-dma.c 	list_for_each_entry_safe(vchan,
vchan            1016 drivers/dma/owl-dma.c 		list_del(&vchan->vc.chan.device_node);
vchan            1017 drivers/dma/owl-dma.c 		tasklet_kill(&vchan->vc.task);
vchan            1025 drivers/dma/owl-dma.c 	struct owl_dma_vchan *vchan;
vchan            1036 drivers/dma/owl-dma.c 	vchan = to_owl_vchan(chan);
vchan            1037 drivers/dma/owl-dma.c 	vchan->drq = drq;
vchan            1145 drivers/dma/owl-dma.c 		struct owl_dma_vchan *vchan = &od->vchans[i];
vchan            1147 drivers/dma/owl-dma.c 		vchan->vc.desc_free = owl_dma_desc_free;
vchan            1148 drivers/dma/owl-dma.c 		vchan_init(&vchan->vc, &od->dma);
vchan             100 drivers/dma/pxa_dma.c 	struct pxad_chan	*vchan;
vchan             151 drivers/dma/pxa_dma.c 		dev_vdbg(&phy->vchan->vc.chan.dev->device,		\
vchan             159 drivers/dma/pxa_dma.c 		dev_vdbg(&phy->vchan->vc.chan.dev->device,		\
vchan             166 drivers/dma/pxa_dma.c 		dev_vdbg(&phy->vchan->vc.chan.dev->device,		\
vchan             390 drivers/dma/pxa_dma.c 			if (!phy->vchan) {
vchan             391 drivers/dma/pxa_dma.c 				phy->vchan = pchan;
vchan             425 drivers/dma/pxa_dma.c 	chan->phy->vchan = NULL;
vchan             455 drivers/dma/pxa_dma.c 	if (!phy->vchan)
vchan             458 drivers/dma/pxa_dma.c 	dev_dbg(&phy->vchan->vc.chan.dev->device,
vchan             462 drivers/dma/pxa_dma.c 	pdev = to_pxad_dev(phy->vchan->vc.chan.device);
vchan             463 drivers/dma/pxa_dma.c 	if (phy->vchan->drcmr <= pdev->nr_requestors) {
vchan             464 drivers/dma/pxa_dma.c 		reg = pxad_drcmr(phy->vchan->drcmr);
vchan             487 drivers/dma/pxa_dma.c 	dev_dbg(&phy->vchan->vc.chan.dev->device,
vchan             595 drivers/dma/pxa_dma.c 	if ((dcsr & PXA_DCSR_BUSERR) && (phy->vchan))
vchan             596 drivers/dma/pxa_dma.c 		dev_warn(&phy->vchan->vc.chan.dev->device,
vchan             598 drivers/dma/pxa_dma.c 			 __func__, &phy->vchan);
vchan             606 drivers/dma/pxa_dma.c 	struct pxad_chan *chan = phy->vchan;
vchan            1112 drivers/dma/pxa_dma.c 		phy->vchan = NULL;
vchan             105 drivers/dma/sa11x0-dma.c 	struct sa11x0_dma_chan	*vchan;
vchan             275 drivers/dma/sa11x0-dma.c 	c = p->vchan;
vchan             346 drivers/dma/sa11x0-dma.c 				p->vchan = NULL;
vchan             356 drivers/dma/sa11x0-dma.c 		if (p->vchan == NULL && !list_empty(&d->chan_pending)) {
vchan             364 drivers/dma/sa11x0-dma.c 			p->vchan = c;
vchan             374 drivers/dma/sa11x0-dma.c 			c = p->vchan;
vchan             785 drivers/dma/sa11x0-dma.c 		p->vchan = NULL;
vchan              23 drivers/dma/st_fdma.c 	return container_of(c, struct st_fdma_chan, vchan.chan);
vchan              78 drivers/dma/st_fdma.c 	vdesc = vchan_next_desc(&fchan->vchan);
vchan              84 drivers/dma/st_fdma.c 	cmd = FDMA_CMD_START(fchan->vchan.chan.chan_id);
vchan              93 drivers/dma/st_fdma.c 	dev_dbg(fchan->fdev->dev, "start chan:%d\n", fchan->vchan.chan.chan_id);
vchan             100 drivers/dma/st_fdma.c 	int ch_id = fchan->vchan.chan.chan_id;
vchan             138 drivers/dma/st_fdma.c 		spin_lock(&fchan->vchan.lock);
vchan             156 drivers/dma/st_fdma.c 		spin_unlock(&fchan->vchan.lock);
vchan             281 drivers/dma/st_fdma.c 		fchan->vchan.chan.chan_id, fchan->cfg.type);
vchan             293 drivers/dma/st_fdma.c 		__func__, fchan->vchan.chan.chan_id);
vchan             298 drivers/dma/st_fdma.c 	spin_lock_irqsave(&fchan->vchan.lock, flags);
vchan             300 drivers/dma/st_fdma.c 	spin_unlock_irqrestore(&fchan->vchan.lock, flags);
vchan             342 drivers/dma/st_fdma.c 	return vchan_tx_prep(&fchan->vchan, &fdesc->vdesc, flags);
vchan             350 drivers/dma/st_fdma.c 	int ch_id = fchan->vchan.chan.chan_id;
vchan             502 drivers/dma/st_fdma.c 	return vchan_tx_prep(&fchan->vchan, &fdesc->vdesc, flags);
vchan             551 drivers/dma/st_fdma.c 	return vchan_tx_prep(&fchan->vchan, &fdesc->vdesc, flags);
vchan             592 drivers/dma/st_fdma.c 	spin_lock_irqsave(&fchan->vchan.lock, flags);
vchan             593 drivers/dma/st_fdma.c 	vd = vchan_find_desc(&fchan->vchan, cookie);
vchan             601 drivers/dma/st_fdma.c 	spin_unlock_irqrestore(&fchan->vchan.lock, flags);
vchan             611 drivers/dma/st_fdma.c 	spin_lock_irqsave(&fchan->vchan.lock, flags);
vchan             613 drivers/dma/st_fdma.c 	if (vchan_issue_pending(&fchan->vchan) && !fchan->fdesc)
vchan             616 drivers/dma/st_fdma.c 	spin_unlock_irqrestore(&fchan->vchan.lock, flags);
vchan             623 drivers/dma/st_fdma.c 	int ch_id = fchan->vchan.chan.chan_id;
vchan             628 drivers/dma/st_fdma.c 	spin_lock_irqsave(&fchan->vchan.lock, flags);
vchan             631 drivers/dma/st_fdma.c 	spin_unlock_irqrestore(&fchan->vchan.lock, flags);
vchan             641 drivers/dma/st_fdma.c 	int ch_id = fchan->vchan.chan.chan_id;
vchan             645 drivers/dma/st_fdma.c 	spin_lock_irqsave(&fchan->vchan.lock, flags);
vchan             651 drivers/dma/st_fdma.c 	spin_unlock_irqrestore(&fchan->vchan.lock, flags);
vchan             661 drivers/dma/st_fdma.c 	int ch_id = fchan->vchan.chan.chan_id;
vchan             666 drivers/dma/st_fdma.c 	spin_lock_irqsave(&fchan->vchan.lock, flags);
vchan             669 drivers/dma/st_fdma.c 	vchan_get_all_descriptors(&fchan->vchan, &head);
vchan             670 drivers/dma/st_fdma.c 	spin_unlock_irqrestore(&fchan->vchan.lock, flags);
vchan             671 drivers/dma/st_fdma.c 	vchan_dma_desc_free_list(&fchan->vchan, &head);
vchan             733 drivers/dma/st_fdma.c 		list_del(&fchan->vchan.chan.device_node);
vchan             734 drivers/dma/st_fdma.c 		tasklet_kill(&fchan->vchan.task);
vchan             797 drivers/dma/st_fdma.c 		fchan->vchan.desc_free = st_fdma_free_desc;
vchan             798 drivers/dma/st_fdma.c 		vchan_init(&fchan->vchan, &fdev->dma_device);
vchan             125 drivers/dma/st_fdma.h 	struct virt_dma_chan vchan;
vchan             186 drivers/dma/st_fdma.h 			+ (fchan)->vchan.chan.chan_id * 0x4 \
vchan             191 drivers/dma/st_fdma.h 			+ (fchan)->vchan.chan.chan_id * 0x4 \
vchan             208 drivers/dma/st_fdma.h 			+ (fchan)->vchan.chan.chan_id * FDMA_NODE_SZ \
vchan             213 drivers/dma/st_fdma.h 			+ (fchan)->vchan.chan.chan_id * FDMA_NODE_SZ \
vchan             192 drivers/dma/stm32-dma.c 	struct virt_dma_chan vchan;
vchan             217 drivers/dma/stm32-dma.c 	return container_of(chan->vchan.chan.device, struct stm32_dma_device,
vchan             223 drivers/dma/stm32-dma.c 	return container_of(c, struct stm32_dma_chan, vchan.chan);
vchan             233 drivers/dma/stm32-dma.c 	return &chan->vchan.chan.dev->device;
vchan             489 drivers/dma/stm32-dma.c 	spin_lock_irqsave(&chan->vchan.lock, flags);
vchan             496 drivers/dma/stm32-dma.c 	vchan_get_all_descriptors(&chan->vchan, &head);
vchan             497 drivers/dma/stm32-dma.c 	spin_unlock_irqrestore(&chan->vchan.lock, flags);
vchan             498 drivers/dma/stm32-dma.c 	vchan_dma_desc_free_list(&chan->vchan, &head);
vchan             507 drivers/dma/stm32-dma.c 	vchan_synchronize(&chan->vchan);
vchan             544 drivers/dma/stm32-dma.c 		vdesc = vchan_next_desc(&chan->vchan);
vchan             583 drivers/dma/stm32-dma.c 	dev_dbg(chan2dev(chan), "vchan %pK: started\n", &chan->vchan);
vchan             640 drivers/dma/stm32-dma.c 	spin_lock(&chan->vchan.lock);
vchan             673 drivers/dma/stm32-dma.c 	spin_unlock(&chan->vchan.lock);
vchan             683 drivers/dma/stm32-dma.c 	spin_lock_irqsave(&chan->vchan.lock, flags);
vchan             684 drivers/dma/stm32-dma.c 	if (vchan_issue_pending(&chan->vchan) && !chan->desc && !chan->busy) {
vchan             685 drivers/dma/stm32-dma.c 		dev_dbg(chan2dev(chan), "vchan %pK: issued\n", &chan->vchan);
vchan             689 drivers/dma/stm32-dma.c 	spin_unlock_irqrestore(&chan->vchan.lock, flags);
vchan             886 drivers/dma/stm32-dma.c 	return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags);
vchan             971 drivers/dma/stm32-dma.c 	return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags);
vchan            1023 drivers/dma/stm32-dma.c 	return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags);
vchan            1154 drivers/dma/stm32-dma.c 	spin_lock_irqsave(&chan->vchan.lock, flags);
vchan            1155 drivers/dma/stm32-dma.c 	vdesc = vchan_find_desc(&chan->vchan, cookie);
vchan            1164 drivers/dma/stm32-dma.c 	spin_unlock_irqrestore(&chan->vchan.lock, flags);
vchan            1197 drivers/dma/stm32-dma.c 		spin_lock_irqsave(&chan->vchan.lock, flags);
vchan            1200 drivers/dma/stm32-dma.c 		spin_unlock_irqrestore(&chan->vchan.lock, flags);
vchan            1254 drivers/dma/stm32-dma.c 	c = dma_get_slave_channel(&chan->vchan.chan);
vchan            1352 drivers/dma/stm32-dma.c 		chan->vchan.desc_free = stm32_dma_desc_free;
vchan            1353 drivers/dma/stm32-dma.c 		vchan_init(&chan->vchan, dd);
vchan             259 drivers/dma/stm32-mdma.c 	struct virt_dma_chan vchan;
vchan             287 drivers/dma/stm32-mdma.c 	return container_of(chan->vchan.chan.device, struct stm32_mdma_device,
vchan             293 drivers/dma/stm32-mdma.c 	return container_of(c, struct stm32_mdma_chan, vchan.chan);
vchan             303 drivers/dma/stm32-mdma.c 	return &chan->vchan.chan.dev->device;
vchan             810 drivers/dma/stm32-mdma.c 	return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags);
vchan             901 drivers/dma/stm32-mdma.c 	return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags);
vchan            1087 drivers/dma/stm32-mdma.c 	return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags);
vchan            1124 drivers/dma/stm32-mdma.c 	vdesc = vchan_next_desc(&chan->vchan);
vchan            1163 drivers/dma/stm32-mdma.c 	dev_dbg(chan2dev(chan), "vchan %pK: started\n", &chan->vchan);
vchan            1171 drivers/dma/stm32-mdma.c 	spin_lock_irqsave(&chan->vchan.lock, flags);
vchan            1173 drivers/dma/stm32-mdma.c 	if (!vchan_issue_pending(&chan->vchan))
vchan            1176 drivers/dma/stm32-mdma.c 	dev_dbg(chan2dev(chan), "vchan %pK: issued\n", &chan->vchan);
vchan            1182 drivers/dma/stm32-mdma.c 	spin_unlock_irqrestore(&chan->vchan.lock, flags);
vchan            1191 drivers/dma/stm32-mdma.c 	spin_lock_irqsave(&chan->vchan.lock, flags);
vchan            1193 drivers/dma/stm32-mdma.c 	spin_unlock_irqrestore(&chan->vchan.lock, flags);
vchan            1196 drivers/dma/stm32-mdma.c 		dev_dbg(chan2dev(chan), "vchan %pK: pause\n", &chan->vchan);
vchan            1211 drivers/dma/stm32-mdma.c 	spin_lock_irqsave(&chan->vchan.lock, flags);
vchan            1231 drivers/dma/stm32-mdma.c 	spin_unlock_irqrestore(&chan->vchan.lock, flags);
vchan            1233 drivers/dma/stm32-mdma.c 	dev_dbg(chan2dev(chan), "vchan %pK: resume\n", &chan->vchan);
vchan            1244 drivers/dma/stm32-mdma.c 	spin_lock_irqsave(&chan->vchan.lock, flags);
vchan            1249 drivers/dma/stm32-mdma.c 	vchan_get_all_descriptors(&chan->vchan, &head);
vchan            1250 drivers/dma/stm32-mdma.c 	spin_unlock_irqrestore(&chan->vchan.lock, flags);
vchan            1252 drivers/dma/stm32-mdma.c 	vchan_dma_desc_free_list(&chan->vchan, &head);
vchan            1261 drivers/dma/stm32-mdma.c 	vchan_synchronize(&chan->vchan);
vchan            1316 drivers/dma/stm32-mdma.c 	spin_lock_irqsave(&chan->vchan.lock, flags);
vchan            1318 drivers/dma/stm32-mdma.c 	vdesc = vchan_find_desc(&chan->vchan, cookie);
vchan            1327 drivers/dma/stm32-mdma.c 	spin_unlock_irqrestore(&chan->vchan.lock, flags);
vchan            1374 drivers/dma/stm32-mdma.c 	spin_lock(&chan->vchan.lock);
vchan            1381 drivers/dma/stm32-mdma.c 		spin_unlock(&chan->vchan.lock);
vchan            1427 drivers/dma/stm32-mdma.c 	spin_unlock(&chan->vchan.lock);
vchan            1469 drivers/dma/stm32-mdma.c 		spin_lock_irqsave(&chan->vchan.lock, flags);
vchan            1472 drivers/dma/stm32-mdma.c 		spin_unlock_irqrestore(&chan->vchan.lock, flags);
vchan            1635 drivers/dma/stm32-mdma.c 		chan->vchan.desc_free = stm32_mdma_desc_free;
vchan            1636 drivers/dma/stm32-mdma.c 		vchan_init(&chan->vchan, dd);
vchan             129 drivers/dma/sun4i-dma.c 	struct sun4i_dma_vchan		*vchan;
vchan             212 drivers/dma/sun4i-dma.c 	struct sun4i_dma_vchan *vchan = to_sun4i_dma_vchan(chan);
vchan             214 drivers/dma/sun4i-dma.c 	vchan_free_chan_resources(&vchan->vc);
vchan             218 drivers/dma/sun4i-dma.c 						  struct sun4i_dma_vchan *vchan)
vchan             228 drivers/dma/sun4i-dma.c 	if (vchan->is_dedicated) {
vchan             239 drivers/dma/sun4i-dma.c 		pchan->vchan = vchan;
vchan             256 drivers/dma/sun4i-dma.c 	pchan->vchan = NULL;
vchan             320 drivers/dma/sun4i-dma.c 				   struct sun4i_dma_vchan *vchan)
vchan             328 drivers/dma/sun4i-dma.c 	lockdep_assert_held(&vchan->vc.lock);
vchan             331 drivers/dma/sun4i-dma.c 	pchan = find_and_use_pchan(priv, vchan);
vchan             339 drivers/dma/sun4i-dma.c 	if (vchan->processing) {
vchan             340 drivers/dma/sun4i-dma.c 		dev_dbg(chan2dev(&vchan->vc.chan),
vchan             348 drivers/dma/sun4i-dma.c 		vd = vchan_next_desc(&vchan->vc);
vchan             350 drivers/dma/sun4i-dma.c 			dev_dbg(chan2dev(&vchan->vc.chan),
vchan             361 drivers/dma/sun4i-dma.c 			dev_dbg(chan2dev(&vchan->vc.chan),
vchan             369 drivers/dma/sun4i-dma.c 	vchan->processing = promise;
vchan             373 drivers/dma/sun4i-dma.c 		vchan->contract = contract;
vchan             374 drivers/dma/sun4i-dma.c 		vchan->pchan = pchan;
vchan             614 drivers/dma/sun4i-dma.c 	struct sun4i_dma_vchan *vchan = to_sun4i_dma_vchan(chan);
vchan             615 drivers/dma/sun4i-dma.c 	struct dma_slave_config *sconfig = &vchan->cfg;
vchan             633 drivers/dma/sun4i-dma.c 	if (vchan->is_dedicated)
vchan             645 drivers/dma/sun4i-dma.c 	if (vchan->is_dedicated) {
vchan             657 drivers/dma/sun4i-dma.c 	return vchan_tx_prep(&vchan->vc, &contract->vd, flags);
vchan             665 drivers/dma/sun4i-dma.c 	struct sun4i_dma_vchan *vchan = to_sun4i_dma_vchan(chan);
vchan             666 drivers/dma/sun4i-dma.c 	struct dma_slave_config *sconfig = &vchan->cfg;
vchan             678 drivers/dma/sun4i-dma.c 	if (vchan->is_dedicated) {
vchan             701 drivers/dma/sun4i-dma.c 			    SUN4I_DMA_CFG_DST_DRQ_TYPE(vchan->endpoint) |
vchan             706 drivers/dma/sun4i-dma.c 		endpoints = SUN4I_DMA_CFG_SRC_DRQ_TYPE(vchan->endpoint) |
vchan             763 drivers/dma/sun4i-dma.c 	return vchan_tx_prep(&vchan->vc, &contract->vd, flags);
vchan             771 drivers/dma/sun4i-dma.c 	struct sun4i_dma_vchan *vchan = to_sun4i_dma_vchan(chan);
vchan             772 drivers/dma/sun4i-dma.c 	struct dma_slave_config *sconfig = &vchan->cfg;
vchan             793 drivers/dma/sun4i-dma.c 	if (vchan->is_dedicated) {
vchan             804 drivers/dma/sun4i-dma.c 		endpoints = SUN4I_DMA_CFG_DST_DRQ_TYPE(vchan->endpoint) |
vchan             811 drivers/dma/sun4i-dma.c 			    SUN4I_DMA_CFG_SRC_DRQ_TYPE(vchan->endpoint) |
vchan             836 drivers/dma/sun4i-dma.c 		if (vchan->is_dedicated)
vchan             859 drivers/dma/sun4i-dma.c 	return vchan_tx_prep(&vchan->vc, &contract->vd, flags);
vchan             865 drivers/dma/sun4i-dma.c 	struct sun4i_dma_vchan *vchan = to_sun4i_dma_vchan(chan);
vchan             866 drivers/dma/sun4i-dma.c 	struct sun4i_dma_pchan *pchan = vchan->pchan;
vchan             870 drivers/dma/sun4i-dma.c 	spin_lock_irqsave(&vchan->vc.lock, flags);
vchan             871 drivers/dma/sun4i-dma.c 	vchan_get_all_descriptors(&vchan->vc, &head);
vchan             872 drivers/dma/sun4i-dma.c 	spin_unlock_irqrestore(&vchan->vc.lock, flags);
vchan             887 drivers/dma/sun4i-dma.c 	spin_lock_irqsave(&vchan->vc.lock, flags);
vchan             888 drivers/dma/sun4i-dma.c 	vchan_dma_desc_free_list(&vchan->vc, &head);
vchan             890 drivers/dma/sun4i-dma.c 	vchan->processing = NULL;
vchan             891 drivers/dma/sun4i-dma.c 	vchan->pchan = NULL;
vchan             892 drivers/dma/sun4i-dma.c 	spin_unlock_irqrestore(&vchan->vc.lock, flags);
vchan             900 drivers/dma/sun4i-dma.c 	struct sun4i_dma_vchan *vchan = to_sun4i_dma_vchan(chan);
vchan             902 drivers/dma/sun4i-dma.c 	memcpy(&vchan->cfg, config, sizeof(*config));
vchan             911 drivers/dma/sun4i-dma.c 	struct sun4i_dma_vchan *vchan;
vchan             930 drivers/dma/sun4i-dma.c 	vchan = to_sun4i_dma_vchan(chan);
vchan             931 drivers/dma/sun4i-dma.c 	vchan->is_dedicated = is_dedicated;
vchan             932 drivers/dma/sun4i-dma.c 	vchan->endpoint = endpoint;
vchan             941 drivers/dma/sun4i-dma.c 	struct sun4i_dma_vchan *vchan = to_sun4i_dma_vchan(chan);
vchan             942 drivers/dma/sun4i-dma.c 	struct sun4i_dma_pchan *pchan = vchan->pchan;
vchan             954 drivers/dma/sun4i-dma.c 	spin_lock_irqsave(&vchan->vc.lock, flags);
vchan             955 drivers/dma/sun4i-dma.c 	vd = vchan_find_desc(&vchan->vc, cookie);
vchan             981 drivers/dma/sun4i-dma.c 	spin_unlock_irqrestore(&vchan->vc.lock, flags);
vchan             989 drivers/dma/sun4i-dma.c 	struct sun4i_dma_vchan *vchan = to_sun4i_dma_vchan(chan);
vchan             992 drivers/dma/sun4i-dma.c 	spin_lock_irqsave(&vchan->vc.lock, flags);
vchan             998 drivers/dma/sun4i-dma.c 	if (vchan_issue_pending(&vchan->vc))
vchan             999 drivers/dma/sun4i-dma.c 		__execute_vchan_pending(priv, vchan);
vchan            1001 drivers/dma/sun4i-dma.c 	spin_unlock_irqrestore(&vchan->vc.lock, flags);
vchan            1008 drivers/dma/sun4i-dma.c 	struct sun4i_dma_vchan *vchan;
vchan            1023 drivers/dma/sun4i-dma.c 		vchan = pchan->vchan;
vchan            1024 drivers/dma/sun4i-dma.c 		if (!vchan) /* a terminated channel may still interrupt */
vchan            1026 drivers/dma/sun4i-dma.c 		contract = vchan->contract;
vchan            1033 drivers/dma/sun4i-dma.c 			spin_lock(&vchan->vc.lock);
vchan            1039 drivers/dma/sun4i-dma.c 			list_del(&vchan->processing->list);
vchan            1040 drivers/dma/sun4i-dma.c 			list_add_tail(&vchan->processing->list,
vchan            1058 drivers/dma/sun4i-dma.c 				vchan->processing = promise;
vchan            1062 drivers/dma/sun4i-dma.c 				vchan->processing = NULL;
vchan            1063 drivers/dma/sun4i-dma.c 				vchan->pchan = NULL;
vchan            1070 drivers/dma/sun4i-dma.c 			spin_unlock(&vchan->vc.lock);
vchan            1096 drivers/dma/sun4i-dma.c 			vchan = &priv->vchans[i];
vchan            1097 drivers/dma/sun4i-dma.c 			spin_lock(&vchan->vc.lock);
vchan            1098 drivers/dma/sun4i-dma.c 			__execute_vchan_pending(priv, vchan);
vchan            1099 drivers/dma/sun4i-dma.c 			spin_unlock(&vchan->vc.lock);
vchan            1198 drivers/dma/sun4i-dma.c 		struct sun4i_dma_vchan *vchan = &priv->vchans[i];
vchan            1200 drivers/dma/sun4i-dma.c 		spin_lock_init(&vchan->vc.lock);
vchan            1201 drivers/dma/sun4i-dma.c 		vchan->vc.desc_free = sun4i_dma_free_contract;
vchan            1202 drivers/dma/sun4i-dma.c 		vchan_init(&vchan->vc, &priv->slave);
vchan             170 drivers/dma/sun6i-dma.c 	struct sun6i_vchan	*vchan;
vchan             387 drivers/dma/sun6i-dma.c static inline void sun6i_dma_dump_lli(struct sun6i_vchan *vchan,
vchan             392 drivers/dma/sun6i-dma.c 	dev_dbg(chan2dev(&vchan->vc.chan),
vchan             427 drivers/dma/sun6i-dma.c static int sun6i_dma_start_desc(struct sun6i_vchan *vchan)
vchan             429 drivers/dma/sun6i-dma.c 	struct sun6i_dma_dev *sdev = to_sun6i_dma_dev(vchan->vc.chan.device);
vchan             430 drivers/dma/sun6i-dma.c 	struct virt_dma_desc *desc = vchan_next_desc(&vchan->vc);
vchan             431 drivers/dma/sun6i-dma.c 	struct sun6i_pchan *pchan = vchan->phy;
vchan             448 drivers/dma/sun6i-dma.c 	sun6i_dma_dump_lli(vchan, pchan->desc->v_lli);
vchan             453 drivers/dma/sun6i-dma.c 	vchan->irq_type = vchan->cyclic ? DMA_IRQ_PKG : DMA_IRQ_QUEUE;
vchan             458 drivers/dma/sun6i-dma.c 	irq_val |= vchan->irq_type << (irq_offset * DMA_IRQ_CHAN_WIDTH);
vchan             473 drivers/dma/sun6i-dma.c 	struct sun6i_vchan *vchan;
vchan             478 drivers/dma/sun6i-dma.c 	list_for_each_entry(vchan, &sdev->slave.channels, vc.chan.device_node) {
vchan             479 drivers/dma/sun6i-dma.c 		spin_lock_irq(&vchan->vc.lock);
vchan             481 drivers/dma/sun6i-dma.c 		pchan = vchan->phy;
vchan             484 drivers/dma/sun6i-dma.c 			if (sun6i_dma_start_desc(vchan)) {
vchan             492 drivers/dma/sun6i-dma.c 				vchan->phy = NULL;
vchan             493 drivers/dma/sun6i-dma.c 				pchan->vchan = NULL;
vchan             496 drivers/dma/sun6i-dma.c 		spin_unlock_irq(&vchan->vc.lock);
vchan             503 drivers/dma/sun6i-dma.c 		if (pchan->vchan || list_empty(&sdev->pending))
vchan             506 drivers/dma/sun6i-dma.c 		vchan = list_first_entry(&sdev->pending,
vchan             510 drivers/dma/sun6i-dma.c 		list_del_init(&vchan->node);
vchan             514 drivers/dma/sun6i-dma.c 		pchan->vchan = vchan;
vchan             515 drivers/dma/sun6i-dma.c 		vchan->phy = pchan;
vchan             517 drivers/dma/sun6i-dma.c 			pchan->idx, &vchan->vc);
vchan             526 drivers/dma/sun6i-dma.c 		vchan = pchan->vchan;
vchan             527 drivers/dma/sun6i-dma.c 		if (vchan) {
vchan             528 drivers/dma/sun6i-dma.c 			spin_lock_irq(&vchan->vc.lock);
vchan             529 drivers/dma/sun6i-dma.c 			sun6i_dma_start_desc(vchan);
vchan             530 drivers/dma/sun6i-dma.c 			spin_unlock_irq(&vchan->vc.lock);
vchan             538 drivers/dma/sun6i-dma.c 	struct sun6i_vchan *vchan;
vchan             555 drivers/dma/sun6i-dma.c 			vchan = pchan->vchan;
vchan             556 drivers/dma/sun6i-dma.c 			if (vchan && (status & vchan->irq_type)) {
vchan             557 drivers/dma/sun6i-dma.c 				if (vchan->cyclic) {
vchan             560 drivers/dma/sun6i-dma.c 					spin_lock(&vchan->vc.lock);
vchan             563 drivers/dma/sun6i-dma.c 					spin_unlock(&vchan->vc.lock);
vchan             634 drivers/dma/sun6i-dma.c 	struct sun6i_vchan *vchan = to_sun6i_vchan(chan);
vchan             642 drivers/dma/sun6i-dma.c 		__func__, vchan->vc.chan.chan_id, &dest, &src, len, flags);
vchan             673 drivers/dma/sun6i-dma.c 	sun6i_dma_dump_lli(vchan, v_lli);
vchan             675 drivers/dma/sun6i-dma.c 	return vchan_tx_prep(&vchan->vc, &txd->vd, flags);
vchan             688 drivers/dma/sun6i-dma.c 	struct sun6i_vchan *vchan = to_sun6i_vchan(chan);
vchan             689 drivers/dma/sun6i-dma.c 	struct dma_slave_config *sconfig = &vchan->cfg;
vchan             722 drivers/dma/sun6i-dma.c 			sdev->cfg->set_drq(&v_lli->cfg, DRQ_SDRAM, vchan->port);
vchan             727 drivers/dma/sun6i-dma.c 				__func__, vchan->vc.chan.chan_id,
vchan             735 drivers/dma/sun6i-dma.c 			sdev->cfg->set_drq(&v_lli->cfg, vchan->port, DRQ_SDRAM);
vchan             740 drivers/dma/sun6i-dma.c 				__func__, vchan->vc.chan.chan_id,
vchan             750 drivers/dma/sun6i-dma.c 		sun6i_dma_dump_lli(vchan, prev);
vchan             752 drivers/dma/sun6i-dma.c 	return vchan_tx_prep(&vchan->vc, &txd->vd, flags);
vchan             770 drivers/dma/sun6i-dma.c 	struct sun6i_vchan *vchan = to_sun6i_vchan(chan);
vchan             771 drivers/dma/sun6i-dma.c 	struct dma_slave_config *sconfig = &vchan->cfg;
vchan             803 drivers/dma/sun6i-dma.c 			sdev->cfg->set_drq(&v_lli->cfg, DRQ_SDRAM, vchan->port);
vchan             809 drivers/dma/sun6i-dma.c 			sdev->cfg->set_drq(&v_lli->cfg, vchan->port, DRQ_SDRAM);
vchan             818 drivers/dma/sun6i-dma.c 	vchan->cyclic = true;
vchan             820 drivers/dma/sun6i-dma.c 	return vchan_tx_prep(&vchan->vc, &txd->vd, flags);
vchan             832 drivers/dma/sun6i-dma.c 	struct sun6i_vchan *vchan = to_sun6i_vchan(chan);
vchan             834 drivers/dma/sun6i-dma.c 	memcpy(&vchan->cfg, config, sizeof(*config));
vchan             842 drivers/dma/sun6i-dma.c 	struct sun6i_vchan *vchan = to_sun6i_vchan(chan);
vchan             843 drivers/dma/sun6i-dma.c 	struct sun6i_pchan *pchan = vchan->phy;
vchan             845 drivers/dma/sun6i-dma.c 	dev_dbg(chan2dev(chan), "vchan %p: pause\n", &vchan->vc);
vchan             852 drivers/dma/sun6i-dma.c 		list_del_init(&vchan->node);
vchan             862 drivers/dma/sun6i-dma.c 	struct sun6i_vchan *vchan = to_sun6i_vchan(chan);
vchan             863 drivers/dma/sun6i-dma.c 	struct sun6i_pchan *pchan = vchan->phy;
vchan             866 drivers/dma/sun6i-dma.c 	dev_dbg(chan2dev(chan), "vchan %p: resume\n", &vchan->vc);
vchan             868 drivers/dma/sun6i-dma.c 	spin_lock_irqsave(&vchan->vc.lock, flags);
vchan             873 drivers/dma/sun6i-dma.c 	} else if (!list_empty(&vchan->vc.desc_issued)) {
vchan             875 drivers/dma/sun6i-dma.c 		list_add_tail(&vchan->node, &sdev->pending);
vchan             879 drivers/dma/sun6i-dma.c 	spin_unlock_irqrestore(&vchan->vc.lock, flags);
vchan             887 drivers/dma/sun6i-dma.c 	struct sun6i_vchan *vchan = to_sun6i_vchan(chan);
vchan             888 drivers/dma/sun6i-dma.c 	struct sun6i_pchan *pchan = vchan->phy;
vchan             893 drivers/dma/sun6i-dma.c 	list_del_init(&vchan->node);
vchan             896 drivers/dma/sun6i-dma.c 	spin_lock_irqsave(&vchan->vc.lock, flags);
vchan             898 drivers/dma/sun6i-dma.c 	if (vchan->cyclic) {
vchan             899 drivers/dma/sun6i-dma.c 		vchan->cyclic = false;
vchan             902 drivers/dma/sun6i-dma.c 			struct virt_dma_chan *vc = &vchan->vc;
vchan             908 drivers/dma/sun6i-dma.c 	vchan_get_all_descriptors(&vchan->vc, &head);
vchan             914 drivers/dma/sun6i-dma.c 		vchan->phy = NULL;
vchan             915 drivers/dma/sun6i-dma.c 		pchan->vchan = NULL;
vchan             920 drivers/dma/sun6i-dma.c 	spin_unlock_irqrestore(&vchan->vc.lock, flags);
vchan             922 drivers/dma/sun6i-dma.c 	vchan_dma_desc_free_list(&vchan->vc, &head);
vchan             931 drivers/dma/sun6i-dma.c 	struct sun6i_vchan *vchan = to_sun6i_vchan(chan);
vchan             932 drivers/dma/sun6i-dma.c 	struct sun6i_pchan *pchan = vchan->phy;
vchan             944 drivers/dma/sun6i-dma.c 	spin_lock_irqsave(&vchan->vc.lock, flags);
vchan             946 drivers/dma/sun6i-dma.c 	vd = vchan_find_desc(&vchan->vc, cookie);
vchan             958 drivers/dma/sun6i-dma.c 	spin_unlock_irqrestore(&vchan->vc.lock, flags);
vchan             968 drivers/dma/sun6i-dma.c 	struct sun6i_vchan *vchan = to_sun6i_vchan(chan);
vchan             971 drivers/dma/sun6i-dma.c 	spin_lock_irqsave(&vchan->vc.lock, flags);
vchan             973 drivers/dma/sun6i-dma.c 	if (vchan_issue_pending(&vchan->vc)) {
vchan             976 drivers/dma/sun6i-dma.c 		if (!vchan->phy && list_empty(&vchan->node)) {
vchan             977 drivers/dma/sun6i-dma.c 			list_add_tail(&vchan->node, &sdev->pending);
vchan             980 drivers/dma/sun6i-dma.c 				&vchan->vc);
vchan             986 drivers/dma/sun6i-dma.c 			&vchan->vc);
vchan             989 drivers/dma/sun6i-dma.c 	spin_unlock_irqrestore(&vchan->vc.lock, flags);
vchan             995 drivers/dma/sun6i-dma.c 	struct sun6i_vchan *vchan = to_sun6i_vchan(chan);
vchan             999 drivers/dma/sun6i-dma.c 	list_del_init(&vchan->node);
vchan            1002 drivers/dma/sun6i-dma.c 	vchan_free_chan_resources(&vchan->vc);
vchan            1009 drivers/dma/sun6i-dma.c 	struct sun6i_vchan *vchan;
vchan            1020 drivers/dma/sun6i-dma.c 	vchan = to_sun6i_vchan(chan);
vchan            1021 drivers/dma/sun6i-dma.c 	vchan->port = port;
vchan            1047 drivers/dma/sun6i-dma.c 		struct sun6i_vchan *vchan = &sdev->vchans[i];
vchan            1049 drivers/dma/sun6i-dma.c 		list_del(&vchan->vc.chan.device_node);
vchan            1050 drivers/dma/sun6i-dma.c 		tasklet_kill(&vchan->vc.task);
vchan            1356 drivers/dma/sun6i-dma.c 		struct sun6i_vchan *vchan = &sdc->vchans[i];
vchan            1358 drivers/dma/sun6i-dma.c 		INIT_LIST_HEAD(&vchan->node);
vchan            1359 drivers/dma/sun6i-dma.c 		vchan->vc.desc_free = sun6i_dma_free_desc;
vchan            1360 drivers/dma/sun6i-dma.c 		vchan_init(&vchan->vc, &sdc->slave);
vchan             225 drivers/dma/ti/edma.c 	struct virt_dma_chan		vchan;
vchan             748 drivers/dma/ti/edma.c 	return container_of(c, struct edma_chan, vchan.chan);
vchan             767 drivers/dma/ti/edma.c 	struct device *dev = echan->vchan.chan.device->dev;
vchan             772 drivers/dma/ti/edma.c 		vdesc = vchan_next_desc(&echan->vchan);
vchan             861 drivers/dma/ti/edma.c 	spin_lock_irqsave(&echan->vchan.lock, flags);
vchan             878 drivers/dma/ti/edma.c 	vchan_get_all_descriptors(&echan->vchan, &head);
vchan             879 drivers/dma/ti/edma.c 	spin_unlock_irqrestore(&echan->vchan.lock, flags);
vchan             880 drivers/dma/ti/edma.c 	vchan_dma_desc_free_list(&echan->vchan, &head);
vchan             889 drivers/dma/ti/edma.c 	vchan_synchronize(&echan->vchan);
vchan            1146 drivers/dma/ti/edma.c 	return vchan_tx_prep(&echan->vchan, &edesc->vdesc, tx_flags);
vchan            1262 drivers/dma/ti/edma.c 	return vchan_tx_prep(&echan->vchan, &edesc->vdesc, tx_flags);
vchan            1419 drivers/dma/ti/edma.c 	return vchan_tx_prep(&echan->vchan, &edesc->vdesc, tx_flags);
vchan            1424 drivers/dma/ti/edma.c 	struct device *dev = echan->vchan.chan.device->dev;
vchan            1427 drivers/dma/ti/edma.c 	spin_lock(&echan->vchan.lock);
vchan            1432 drivers/dma/ti/edma.c 			spin_unlock(&echan->vchan.lock);
vchan            1456 drivers/dma/ti/edma.c 	spin_unlock(&echan->vchan.lock);
vchan            1508 drivers/dma/ti/edma.c 	struct device *dev = echan->vchan.chan.device->dev;
vchan            1515 drivers/dma/ti/edma.c 	spin_lock(&echan->vchan.lock);
vchan            1545 drivers/dma/ti/edma.c 	spin_unlock(&echan->vchan.lock);
vchan            1688 drivers/dma/ti/edma.c 	vchan_free_chan_resources(&echan->vchan);
vchan            1720 drivers/dma/ti/edma.c 	spin_lock_irqsave(&echan->vchan.lock, flags);
vchan            1721 drivers/dma/ti/edma.c 	if (vchan_issue_pending(&echan->vchan) && !echan->edesc)
vchan            1723 drivers/dma/ti/edma.c 	spin_unlock_irqrestore(&echan->vchan.lock, flags);
vchan            1774 drivers/dma/ti/edma.c 			dev_dbg_ratelimited(echan->vchan.chan.device->dev,
vchan            1843 drivers/dma/ti/edma.c 	spin_lock_irqsave(&echan->vchan.lock, flags);
vchan            1847 drivers/dma/ti/edma.c 		struct virt_dma_desc *vdesc = vchan_find_desc(&echan->vchan,
vchan            1870 drivers/dma/ti/edma.c 	spin_unlock_irqrestore(&echan->vchan.lock, flags);
vchan            1971 drivers/dma/ti/edma.c 		echan->vchan.desc_free = edma_desc_free;
vchan            1974 drivers/dma/ti/edma.c 			vchan_init(&echan->vchan, m_ddev);
vchan            1976 drivers/dma/ti/edma.c 			vchan_init(&echan->vchan, s_ddev);
vchan            2209 drivers/dma/ti/edma.c 			chan = &echan->vchan.chan;
vchan            2486 drivers/dma/ti/edma.c 			&dmadev->channels, vchan.chan.device_node) {
vchan            2487 drivers/dma/ti/edma.c 		list_del(&echan->vchan.chan.device_node);
vchan            2488 drivers/dma/ti/edma.c 		tasklet_kill(&echan->vchan.task);
vchan             112 drivers/dma/zx_dma.c 	struct zx_dma_chan	*vchan;
vchan             234 drivers/dma/zx_dma.c 			p->vchan = NULL;
vchan             245 drivers/dma/zx_dma.c 		if (!p->vchan) {
vchan             250 drivers/dma/zx_dma.c 			p->vchan = c;
vchan             261 drivers/dma/zx_dma.c 			c = p->vchan;
vchan             286 drivers/dma/zx_dma.c 		c = p->vchan;
vchan             681 drivers/dma/zx_dma.c 		p->vchan = NULL;
vchan             144 drivers/staging/mt7621-dma/mtk-hsdma.c 	struct virt_dma_chan vchan;
vchan             168 drivers/staging/mt7621-dma/mtk-hsdma.c 	return container_of(chan->vchan.chan.device, struct mtk_hsdam_engine,
vchan             174 drivers/staging/mt7621-dma/mtk-hsdma.c 	return container_of(c, struct mtk_hsdma_chan, vchan.chan);
vchan             291 drivers/staging/mt7621-dma/mtk-hsdma.c 	spin_lock_bh(&chan->vchan.lock);
vchan             294 drivers/staging/mt7621-dma/mtk-hsdma.c 	vchan_get_all_descriptors(&chan->vchan, &head);
vchan             295 drivers/staging/mt7621-dma/mtk-hsdma.c 	spin_unlock_bh(&chan->vchan.lock);
vchan             297 drivers/staging/mt7621-dma/mtk-hsdma.c 	vchan_dma_desc_free_list(&chan->vchan, &head);
vchan             390 drivers/staging/mt7621-dma/mtk-hsdma.c 	vdesc = vchan_next_desc(&chan->vchan);
vchan             408 drivers/staging/mt7621-dma/mtk-hsdma.c 	spin_lock_bh(&chan->vchan.lock);
vchan             422 drivers/staging/mt7621-dma/mtk-hsdma.c 	spin_unlock_bh(&chan->vchan.lock);
vchan             449 drivers/staging/mt7621-dma/mtk-hsdma.c 	spin_lock_bh(&chan->vchan.lock);
vchan             450 drivers/staging/mt7621-dma/mtk-hsdma.c 	if (vchan_issue_pending(&chan->vchan) && !chan->desc) {
vchan             458 drivers/staging/mt7621-dma/mtk-hsdma.c 	spin_unlock_bh(&chan->vchan.lock);
vchan             481 drivers/staging/mt7621-dma/mtk-hsdma.c 	return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags);
vchan             704 drivers/staging/mt7621-dma/mtk-hsdma.c 	chan->vchan.desc_free = mtk_hsdma_desc_free;
vchan             705 drivers/staging/mt7621-dma/mtk-hsdma.c 	vchan_init(&chan->vchan, dd);
vchan             108 drivers/staging/ralink-gdma/ralink-gdma.c 	struct virt_dma_chan vchan;
vchan             141 drivers/staging/ralink-gdma/ralink-gdma.c 	return container_of(chan->vchan.chan.device, struct gdma_dma_dev,
vchan             147 drivers/staging/ralink-gdma/ralink-gdma.c 	return container_of(c, struct gdma_dmaengine_chan, vchan.chan);
vchan             228 drivers/staging/ralink-gdma/ralink-gdma.c 	spin_lock_irqsave(&chan->vchan.lock, flags);
vchan             231 drivers/staging/ralink-gdma/ralink-gdma.c 	vchan_get_all_descriptors(&chan->vchan, &head);
vchan             232 drivers/staging/ralink-gdma/ralink-gdma.c 	spin_unlock_irqrestore(&chan->vchan.lock, flags);
vchan             234 drivers/staging/ralink-gdma/ralink-gdma.c 	vchan_dma_desc_free_list(&chan->vchan, &head);
vchan             417 drivers/staging/ralink-gdma/ralink-gdma.c 	vdesc = vchan_next_desc(&chan->vchan);
vchan             436 drivers/staging/ralink-gdma/ralink-gdma.c 	spin_lock_irqsave(&chan->vchan.lock, flags);
vchan             460 drivers/staging/ralink-gdma/ralink-gdma.c 	spin_unlock_irqrestore(&chan->vchan.lock, flags);
vchan             500 drivers/staging/ralink-gdma/ralink-gdma.c 	spin_lock_irqsave(&chan->vchan.lock, flags);
vchan             501 drivers/staging/ralink-gdma/ralink-gdma.c 	if (vchan_issue_pending(&chan->vchan) && !chan->desc) {
vchan             510 drivers/staging/ralink-gdma/ralink-gdma.c 	spin_unlock_irqrestore(&chan->vchan.lock, flags);
vchan             554 drivers/staging/ralink-gdma/ralink-gdma.c 	return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags);
vchan             601 drivers/staging/ralink-gdma/ralink-gdma.c 	return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags);
vchan             648 drivers/staging/ralink-gdma/ralink-gdma.c 	return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags);
vchan             669 drivers/staging/ralink-gdma/ralink-gdma.c 	spin_lock_irqsave(&chan->vchan.lock, flags);
vchan             683 drivers/staging/ralink-gdma/ralink-gdma.c 		vdesc = vchan_find_desc(&chan->vchan, cookie);
vchan             687 drivers/staging/ralink-gdma/ralink-gdma.c 	spin_unlock_irqrestore(&chan->vchan.lock, flags);
vchan             866 drivers/staging/ralink-gdma/ralink-gdma.c 		chan->vchan.desc_free = gdma_dma_desc_free;
vchan             867 drivers/staging/ralink-gdma/ralink-gdma.c 		vchan_init(&chan->vchan, dd);