vchan 84 arch/sh/drivers/dma/dma-api.c if (channel->vchan == chan) vchan 376 arch/sh/drivers/dma/dma-api.c chan->vchan = info->first_channel_nr + i + total_channels; vchan 84 arch/sh/drivers/dma/dma-sysfs.c dma_configure_channel(channel->vchan, config); vchan 127 arch/sh/drivers/dma/dma-sysfs.c dev->id = chan->vchan; vchan 67 arch/sh/include/asm/dma.h unsigned int vchan; /* Virtual channel number */ vchan 109 drivers/dma/dma-axi-dmac.c struct virt_dma_chan vchan; vchan 143 drivers/dma/dma-axi-dmac.c return container_of(chan->vchan.chan.device, struct axi_dmac, vchan 149 drivers/dma/dma-axi-dmac.c return container_of(c, struct axi_dmac_chan, vchan.chan); vchan 210 drivers/dma/dma-axi-dmac.c vdesc = vchan_next_desc(&chan->vchan); vchan 410 drivers/dma/dma-axi-dmac.c spin_lock(&dmac->chan.vchan.lock); vchan 421 drivers/dma/dma-axi-dmac.c spin_unlock(&dmac->chan.vchan.lock); vchan 433 drivers/dma/dma-axi-dmac.c spin_lock_irqsave(&chan->vchan.lock, flags); vchan 436 drivers/dma/dma-axi-dmac.c vchan_get_all_descriptors(&chan->vchan, &head); vchan 438 drivers/dma/dma-axi-dmac.c spin_unlock_irqrestore(&chan->vchan.lock, flags); vchan 440 drivers/dma/dma-axi-dmac.c vchan_dma_desc_free_list(&chan->vchan, &head); vchan 449 drivers/dma/dma-axi-dmac.c vchan_synchronize(&chan->vchan); vchan 460 drivers/dma/dma-axi-dmac.c spin_lock_irqsave(&chan->vchan.lock, flags); vchan 461 drivers/dma/dma-axi-dmac.c if (vchan_issue_pending(&chan->vchan)) vchan 463 drivers/dma/dma-axi-dmac.c spin_unlock_irqrestore(&chan->vchan.lock, flags); vchan 564 drivers/dma/dma-axi-dmac.c return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); vchan 598 drivers/dma/dma-axi-dmac.c return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); vchan 669 drivers/dma/dma-axi-dmac.c return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); vchan 894 drivers/dma/dma-axi-dmac.c dmac->chan.vchan.desc_free = axi_dmac_desc_free; vchan 895 drivers/dma/dma-axi-dmac.c vchan_init(&dmac->chan.vchan, dma_dev); vchan 952 drivers/dma/dma-axi-dmac.c tasklet_kill(&dmac->chan.vchan.task); vchan 129 drivers/dma/dma-jz4780.c struct virt_dma_chan vchan; vchan 166 drivers/dma/dma-jz4780.c return container_of(chan, struct jz4780_dma_chan, vchan.chan); vchan 178 drivers/dma/dma-jz4780.c return container_of(jzchan->vchan.chan.device, struct jz4780_dma_dev, vchan 395 drivers/dma/dma-jz4780.c return vchan_tx_prep(&jzchan->vchan, &desc->vdesc, flags); vchan 447 drivers/dma/dma-jz4780.c return vchan_tx_prep(&jzchan->vchan, &desc->vdesc, flags); vchan 475 drivers/dma/dma-jz4780.c return vchan_tx_prep(&jzchan->vchan, &desc->vdesc, flags); vchan 486 drivers/dma/dma-jz4780.c vdesc = vchan_next_desc(&jzchan->vchan); vchan 558 drivers/dma/dma-jz4780.c spin_lock_irqsave(&jzchan->vchan.lock, flags); vchan 560 drivers/dma/dma-jz4780.c if (vchan_issue_pending(&jzchan->vchan) && !jzchan->desc) vchan 563 drivers/dma/dma-jz4780.c spin_unlock_irqrestore(&jzchan->vchan.lock, flags); vchan 573 drivers/dma/dma-jz4780.c spin_lock_irqsave(&jzchan->vchan.lock, flags); vchan 584 drivers/dma/dma-jz4780.c vchan_get_all_descriptors(&jzchan->vchan, &head); vchan 586 drivers/dma/dma-jz4780.c spin_unlock_irqrestore(&jzchan->vchan.lock, flags); vchan 588 drivers/dma/dma-jz4780.c vchan_dma_desc_free_list(&jzchan->vchan, &head); vchan 597 drivers/dma/dma-jz4780.c vchan_synchronize(&jzchan->vchan); vchan 646 drivers/dma/dma-jz4780.c spin_lock_irqsave(&jzchan->vchan.lock, flags); vchan 648 drivers/dma/dma-jz4780.c vdesc = vchan_find_desc(&jzchan->vchan, cookie); vchan 663 drivers/dma/dma-jz4780.c spin_unlock_irqrestore(&jzchan->vchan.lock, flags); vchan 675 drivers/dma/dma-jz4780.c spin_lock(&jzchan->vchan.lock); vchan 681 drivers/dma/dma-jz4780.c dev_warn(&jzchan->vchan.chan.dev->device, vchan 686 drivers/dma/dma-jz4780.c dev_warn(&jzchan->vchan.chan.dev->device, vchan 715 drivers/dma/dma-jz4780.c dev_err(&jzchan->vchan.chan.dev->device, vchan 719 drivers/dma/dma-jz4780.c spin_unlock(&jzchan->vchan.lock); vchan 771 drivers/dma/dma-jz4780.c vchan_free_chan_resources(&jzchan->vchan); vchan 827 drivers/dma/dma-jz4780.c &jzdma->chan[data.channel].vchan.chan); vchan 954 drivers/dma/dma-jz4780.c vchan_init(&jzchan->vchan, dd); vchan 955 drivers/dma/dma-jz4780.c jzchan->vchan.desc_free = jz4780_dma_desc_free; vchan 993 drivers/dma/dma-jz4780.c tasklet_kill(&jzdma->chan[i].vchan.task); vchan 48 drivers/dma/fsl-edma-common.c u32 ch = fsl_chan->vchan.chan.chan_id; vchan 65 drivers/dma/fsl-edma-common.c u32 ch = fsl_chan->vchan.chan.chan_id; vchan 109 drivers/dma/fsl-edma-common.c u32 ch = fsl_chan->vchan.chan.chan_id; vchan 115 drivers/dma/fsl-edma-common.c ch_off = fsl_chan->vchan.chan.chan_id % chans_per_mux; vchan 161 drivers/dma/fsl-edma-common.c spin_lock_irqsave(&fsl_chan->vchan.lock, flags); vchan 165 drivers/dma/fsl-edma-common.c vchan_get_all_descriptors(&fsl_chan->vchan, &head); vchan 166 drivers/dma/fsl-edma-common.c spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags); vchan 167 drivers/dma/fsl-edma-common.c vchan_dma_desc_free_list(&fsl_chan->vchan, &head); vchan 177 drivers/dma/fsl-edma-common.c spin_lock_irqsave(&fsl_chan->vchan.lock, flags); vchan 183 drivers/dma/fsl-edma-common.c spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags); vchan 193 drivers/dma/fsl-edma-common.c spin_lock_irqsave(&fsl_chan->vchan.lock, flags); vchan 199 drivers/dma/fsl-edma-common.c spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags); vchan 207 drivers/dma/fsl-edma-common.c dma_unmap_resource(fsl_chan->vchan.chan.device->dev, vchan 217 drivers/dma/fsl-edma-common.c struct device *dev = fsl_chan->vchan.chan.device->dev; vchan 270 drivers/dma/fsl-edma-common.c u32 ch = fsl_chan->vchan.chan.chan_id; vchan 323 drivers/dma/fsl-edma-common.c spin_lock_irqsave(&fsl_chan->vchan.lock, flags); vchan 324 drivers/dma/fsl-edma-common.c vdesc = vchan_find_desc(&fsl_chan->vchan, cookie); vchan 334 drivers/dma/fsl-edma-common.c spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags); vchan 345 drivers/dma/fsl-edma-common.c u32 ch = fsl_chan->vchan.chan.chan_id; vchan 507 drivers/dma/fsl-edma-common.c return vchan_tx_prep(&fsl_chan->vchan, &fsl_desc->vdesc, flags); vchan 579 drivers/dma/fsl-edma-common.c return vchan_tx_prep(&fsl_chan->vchan, &fsl_desc->vdesc, flags); vchan 587 drivers/dma/fsl-edma-common.c vdesc = vchan_next_desc(&fsl_chan->vchan); vchan 603 drivers/dma/fsl-edma-common.c spin_lock_irqsave(&fsl_chan->vchan.lock, flags); vchan 606 drivers/dma/fsl-edma-common.c spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags); vchan 611 drivers/dma/fsl-edma-common.c if (vchan_issue_pending(&fsl_chan->vchan) && !fsl_chan->edesc) vchan 614 drivers/dma/fsl-edma-common.c spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags); vchan 635 drivers/dma/fsl-edma-common.c spin_lock_irqsave(&fsl_chan->vchan.lock, flags); vchan 639 drivers/dma/fsl-edma-common.c vchan_get_all_descriptors(&fsl_chan->vchan, &head); vchan 641 drivers/dma/fsl-edma-common.c spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags); vchan 643 drivers/dma/fsl-edma-common.c vchan_dma_desc_free_list(&fsl_chan->vchan, &head); vchan 654 drivers/dma/fsl-edma-common.c &dmadev->channels, vchan.chan.device_node) { vchan 655 drivers/dma/fsl-edma-common.c list_del(&chan->vchan.chan.device_node); vchan 656 drivers/dma/fsl-edma-common.c tasklet_kill(&chan->vchan.task); vchan 115 drivers/dma/fsl-edma-common.h struct virt_dma_chan vchan; vchan 215 drivers/dma/fsl-edma-common.h return container_of(chan, struct fsl_edma_chan, vchan.chan); vchan 27 drivers/dma/fsl-edma.c vchan_synchronize(&fsl_chan->vchan); vchan 47 drivers/dma/fsl-edma.c spin_lock(&fsl_chan->vchan.lock); vchan 61 drivers/dma/fsl-edma.c spin_unlock(&fsl_chan->vchan.lock); vchan 345 drivers/dma/fsl-edma.c fsl_chan->vchan.desc_free = fsl_edma_free_desc; vchan 346 drivers/dma/fsl-edma.c vchan_init(&fsl_chan->vchan, &fsl_edma->dma_dev); vchan 428 drivers/dma/fsl-edma.c spin_lock_irqsave(&fsl_chan->vchan.lock, flags); vchan 437 drivers/dma/fsl-edma.c spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags); vchan 165 drivers/dma/fsl-qdma.c struct virt_dma_chan vchan; vchan 284 drivers/dma/fsl-qdma.c return container_of(chan, struct fsl_qdma_chan, vchan.chan); vchan 301 drivers/dma/fsl-qdma.c spin_lock_irqsave(&fsl_chan->vchan.lock, flags); vchan 302 drivers/dma/fsl-qdma.c vchan_get_all_descriptors(&fsl_chan->vchan, &head); vchan 303 drivers/dma/fsl-qdma.c spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags); vchan 305 drivers/dma/fsl-qdma.c vchan_dma_desc_free_list(&fsl_chan->vchan, &head); vchan 689 drivers/dma/fsl-qdma.c spin_lock(&fsl_comp->qchan->vchan.lock); vchan 692 drivers/dma/fsl-qdma.c spin_unlock(&fsl_comp->qchan->vchan.lock); vchan 934 drivers/dma/fsl-qdma.c return vchan_tx_prep(&fsl_chan->vchan, &fsl_comp->vdesc, flags); vchan 948 drivers/dma/fsl-qdma.c vdesc = vchan_next_desc(&fsl_chan->vchan); vchan 988 drivers/dma/fsl-qdma.c spin_lock(&fsl_chan->vchan.lock); vchan 989 drivers/dma/fsl-qdma.c if (vchan_issue_pending(&fsl_chan->vchan)) vchan 991 drivers/dma/fsl-qdma.c spin_unlock(&fsl_chan->vchan.lock); vchan 999 drivers/dma/fsl-qdma.c vchan_synchronize(&fsl_chan->vchan); vchan 1008 drivers/dma/fsl-qdma.c spin_lock_irqsave(&fsl_chan->vchan.lock, flags); vchan 1009 drivers/dma/fsl-qdma.c vchan_get_all_descriptors(&fsl_chan->vchan, &head); vchan 1010 drivers/dma/fsl-qdma.c spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags); vchan 1011 drivers/dma/fsl-qdma.c vchan_dma_desc_free_list(&fsl_chan->vchan, &head); vchan 1170 drivers/dma/fsl-qdma.c fsl_chan->vchan.desc_free = fsl_qdma_free_desc; vchan 1171 drivers/dma/fsl-qdma.c vchan_init(&fsl_chan->vchan, &fsl_qdma->dma_dev); vchan 1212 drivers/dma/fsl-qdma.c &dmadev->channels, vchan.chan.device_node) { vchan 1213 drivers/dma/fsl-qdma.c list_del(&chan->vchan.chan.device_node); vchan 1214 drivers/dma/fsl-qdma.c tasklet_kill(&chan->vchan.task); vchan 113 drivers/dma/hsu/hsu.c vdesc = vchan_next_desc(&hsuc->vchan); vchan 160 drivers/dma/hsu/hsu.c spin_lock_irqsave(&hsuc->vchan.lock, flags); vchan 162 drivers/dma/hsu/hsu.c spin_unlock_irqrestore(&hsuc->vchan.lock, flags); vchan 214 drivers/dma/hsu/hsu.c spin_lock_irqsave(&hsuc->vchan.lock, flags); vchan 227 drivers/dma/hsu/hsu.c spin_unlock_irqrestore(&hsuc->vchan.lock, flags); vchan 284 drivers/dma/hsu/hsu.c return vchan_tx_prep(&hsuc->vchan, &desc->vdesc, flags); vchan 292 drivers/dma/hsu/hsu.c spin_lock_irqsave(&hsuc->vchan.lock, flags); vchan 293 drivers/dma/hsu/hsu.c if (vchan_issue_pending(&hsuc->vchan) && !hsuc->desc) vchan 295 drivers/dma/hsu/hsu.c spin_unlock_irqrestore(&hsuc->vchan.lock, flags); vchan 328 drivers/dma/hsu/hsu.c spin_lock_irqsave(&hsuc->vchan.lock, flags); vchan 329 drivers/dma/hsu/hsu.c vdesc = vchan_find_desc(&hsuc->vchan, cookie); vchan 338 drivers/dma/hsu/hsu.c spin_unlock_irqrestore(&hsuc->vchan.lock, flags); vchan 358 drivers/dma/hsu/hsu.c spin_lock_irqsave(&hsuc->vchan.lock, flags); vchan 363 drivers/dma/hsu/hsu.c spin_unlock_irqrestore(&hsuc->vchan.lock, flags); vchan 373 drivers/dma/hsu/hsu.c spin_lock_irqsave(&hsuc->vchan.lock, flags); vchan 378 drivers/dma/hsu/hsu.c spin_unlock_irqrestore(&hsuc->vchan.lock, flags); vchan 389 drivers/dma/hsu/hsu.c spin_lock_irqsave(&hsuc->vchan.lock, flags); vchan 397 drivers/dma/hsu/hsu.c vchan_get_all_descriptors(&hsuc->vchan, &head); vchan 398 drivers/dma/hsu/hsu.c spin_unlock_irqrestore(&hsuc->vchan.lock, flags); vchan 399 drivers/dma/hsu/hsu.c vchan_dma_desc_free_list(&hsuc->vchan, &head); vchan 413 drivers/dma/hsu/hsu.c vchan_synchronize(&hsuc->vchan); vchan 441 drivers/dma/hsu/hsu.c hsuc->vchan.desc_free = hsu_dma_desc_free; vchan 442 drivers/dma/hsu/hsu.c vchan_init(&hsuc->vchan, &hsu->dma); vchan 492 drivers/dma/hsu/hsu.c tasklet_kill(&hsuc->vchan.task); vchan 83 drivers/dma/hsu/hsu.h struct virt_dma_chan vchan; vchan 96 drivers/dma/hsu/hsu.h return container_of(chan, struct hsu_dma_chan, vchan.chan); vchan 107 drivers/dma/idma64.c struct idma64 *idma64 = to_idma64(idma64c->vchan.chan.device); vchan 114 drivers/dma/idma64.c struct idma64 *idma64 = to_idma64(idma64c->vchan.chan.device); vchan 118 drivers/dma/idma64.c vdesc = vchan_next_desc(&idma64c->vchan); vchan 142 drivers/dma/idma64.c spin_lock(&idma64c->vchan.lock); vchan 159 drivers/dma/idma64.c spin_unlock(&idma64c->vchan.lock); vchan 326 drivers/dma/idma64.c return vchan_tx_prep(&idma64c->vchan, &desc->vdesc, flags); vchan 334 drivers/dma/idma64.c spin_lock_irqsave(&idma64c->vchan.lock, flags); vchan 335 drivers/dma/idma64.c if (vchan_issue_pending(&idma64c->vchan) && !idma64c->desc) vchan 337 drivers/dma/idma64.c spin_unlock_irqrestore(&idma64c->vchan.lock, flags); vchan 378 drivers/dma/idma64.c spin_lock_irqsave(&idma64c->vchan.lock, flags); vchan 379 drivers/dma/idma64.c vdesc = vchan_find_desc(&idma64c->vchan, cookie); vchan 388 drivers/dma/idma64.c spin_unlock_irqrestore(&idma64c->vchan.lock, flags); vchan 445 drivers/dma/idma64.c spin_lock_irqsave(&idma64c->vchan.lock, flags); vchan 450 drivers/dma/idma64.c spin_unlock_irqrestore(&idma64c->vchan.lock, flags); vchan 460 drivers/dma/idma64.c spin_lock_irqsave(&idma64c->vchan.lock, flags); vchan 465 drivers/dma/idma64.c spin_unlock_irqrestore(&idma64c->vchan.lock, flags); vchan 476 drivers/dma/idma64.c spin_lock_irqsave(&idma64c->vchan.lock, flags); vchan 483 drivers/dma/idma64.c vchan_get_all_descriptors(&idma64c->vchan, &head); vchan 484 drivers/dma/idma64.c spin_unlock_irqrestore(&idma64c->vchan.lock, flags); vchan 486 drivers/dma/idma64.c vchan_dma_desc_free_list(&idma64c->vchan, &head); vchan 494 drivers/dma/idma64.c vchan_synchronize(&idma64c->vchan); vchan 562 drivers/dma/idma64.c idma64c->vchan.desc_free = idma64_vdesc_free; vchan 563 drivers/dma/idma64.c vchan_init(&idma64c->vchan, &idma64->dma); vchan 619 drivers/dma/idma64.c tasklet_kill(&idma64c->vchan.task); vchan 128 drivers/dma/idma64.h struct virt_dma_chan vchan; vchan 143 drivers/dma/idma64.h return container_of(chan, struct idma64_chan, vchan.chan); vchan 94 drivers/dma/k3dma.c struct k3_dma_chan *vchan; vchan 229 drivers/dma/k3dma.c c = p->vchan; vchan 317 drivers/dma/k3dma.c p->vchan = NULL; vchan 331 drivers/dma/k3dma.c if (p->vchan == NULL && !list_empty(&d->chan_pending)) { vchan 338 drivers/dma/k3dma.c p->vchan = c; vchan 351 drivers/dma/k3dma.c c = p->vchan; vchan 743 drivers/dma/k3dma.c p->vchan = NULL; vchan 37 drivers/dma/mcf-edma.c spin_lock(&mcf_chan->vchan.lock); vchan 51 drivers/dma/mcf-edma.c spin_unlock(&mcf_chan->vchan.lock); vchan 223 drivers/dma/mcf-edma.c mcf_chan->vchan.desc_free = fsl_edma_free_desc; vchan 224 drivers/dma/mcf-edma.c vchan_init(&mcf_chan->vchan, &mcf_edma->dma_dev); vchan 120 drivers/dma/mmp_pdma.c struct mmp_pdma_chan *vchan; vchan 156 drivers/dma/mmp_pdma.c if (!phy->vchan) vchan 159 drivers/dma/mmp_pdma.c reg = DRCMR(phy->vchan->drcmr); vchan 163 drivers/dma/mmp_pdma.c if (phy->vchan->byte_align) vchan 196 drivers/dma/mmp_pdma.c if ((dcsr & DCSR_BUSERR) && (phy->vchan)) vchan 197 drivers/dma/mmp_pdma.c dev_warn(phy->vchan->dev, "DCSR_BUSERR\n"); vchan 209 drivers/dma/mmp_pdma.c tasklet_schedule(&phy->vchan->tasklet); vchan 261 drivers/dma/mmp_pdma.c if (!phy->vchan) { vchan 262 drivers/dma/mmp_pdma.c phy->vchan = pchan; vchan 288 drivers/dma/mmp_pdma.c pchan->phy->vchan = NULL; vchan 182 drivers/dma/owl-dma.c struct owl_dma_vchan *vchan; vchan 359 drivers/dma/owl-dma.c static inline int owl_dma_cfg_lli(struct owl_dma_vchan *vchan, vchan 379 drivers/dma/owl-dma.c mode |= OWL_DMA_MODE_TS(vchan->drq) vchan 392 drivers/dma/owl-dma.c mode |= OWL_DMA_MODE_TS(vchan->drq) vchan 429 drivers/dma/owl-dma.c struct owl_dma_vchan *vchan) vchan 439 drivers/dma/owl-dma.c if (!pchan->vchan) { vchan 440 drivers/dma/owl-dma.c pchan->vchan = vchan; vchan 480 drivers/dma/owl-dma.c pchan->vchan = NULL; vchan 495 drivers/dma/owl-dma.c static int owl_dma_start_next_txd(struct owl_dma_vchan *vchan) vchan 497 drivers/dma/owl-dma.c struct owl_dma *od = to_owl_dma(vchan->vc.chan.device); vchan 498 drivers/dma/owl-dma.c struct virt_dma_desc *vd = vchan_next_desc(&vchan->vc); vchan 499 drivers/dma/owl-dma.c struct owl_dma_pchan *pchan = vchan->pchan; vchan 507 drivers/dma/owl-dma.c vchan->txd = txd; vchan 536 drivers/dma/owl-dma.c dev_dbg(chan2dev(&vchan->vc.chan), "starting pchan %d\n", pchan->id); vchan 544 drivers/dma/owl-dma.c static void owl_dma_phy_free(struct owl_dma *od, struct owl_dma_vchan *vchan) vchan 547 drivers/dma/owl-dma.c owl_dma_terminate_pchan(od, vchan->pchan); vchan 549 drivers/dma/owl-dma.c vchan->pchan = NULL; vchan 555 drivers/dma/owl-dma.c struct owl_dma_vchan *vchan; vchan 605 drivers/dma/owl-dma.c vchan = pchan->vchan; vchan 606 drivers/dma/owl-dma.c if (!vchan) { vchan 612 drivers/dma/owl-dma.c spin_lock(&vchan->vc.lock); vchan 614 drivers/dma/owl-dma.c txd = vchan->txd; vchan 616 drivers/dma/owl-dma.c vchan->txd = NULL; vchan 624 drivers/dma/owl-dma.c if (vchan_next_desc(&vchan->vc)) vchan 625 drivers/dma/owl-dma.c owl_dma_start_next_txd(vchan); vchan 627 drivers/dma/owl-dma.c owl_dma_phy_free(od, vchan); vchan 630 drivers/dma/owl-dma.c spin_unlock(&vchan->vc.lock); vchan 660 drivers/dma/owl-dma.c struct owl_dma_vchan *vchan = to_owl_vchan(chan); vchan 664 drivers/dma/owl-dma.c spin_lock_irqsave(&vchan->vc.lock, flags); vchan 666 drivers/dma/owl-dma.c if (vchan->pchan) vchan 667 drivers/dma/owl-dma.c owl_dma_phy_free(od, vchan); vchan 669 drivers/dma/owl-dma.c if (vchan->txd) { vchan 670 drivers/dma/owl-dma.c owl_dma_desc_free(&vchan->txd->vd); vchan 671 drivers/dma/owl-dma.c vchan->txd = NULL; vchan 674 drivers/dma/owl-dma.c vchan_get_all_descriptors(&vchan->vc, &head); vchan 675 drivers/dma/owl-dma.c vchan_dma_desc_free_list(&vchan->vc, &head); vchan 677 drivers/dma/owl-dma.c spin_unlock_irqrestore(&vchan->vc.lock, flags); vchan 685 drivers/dma/owl-dma.c struct owl_dma_vchan *vchan = to_owl_vchan(chan); vchan 692 drivers/dma/owl-dma.c memcpy(&vchan->cfg, config, sizeof(struct dma_slave_config)); vchan 699 drivers/dma/owl-dma.c struct owl_dma_vchan *vchan = to_owl_vchan(chan); vchan 702 drivers/dma/owl-dma.c spin_lock_irqsave(&vchan->vc.lock, flags); vchan 704 drivers/dma/owl-dma.c owl_dma_pause_pchan(vchan->pchan); vchan 706 drivers/dma/owl-dma.c spin_unlock_irqrestore(&vchan->vc.lock, flags); vchan 713 drivers/dma/owl-dma.c struct owl_dma_vchan *vchan = to_owl_vchan(chan); vchan 716 drivers/dma/owl-dma.c if (!vchan->pchan && !vchan->txd) vchan 719 drivers/dma/owl-dma.c dev_dbg(chan2dev(chan), "vchan %p: resume\n", &vchan->vc); vchan 721 drivers/dma/owl-dma.c spin_lock_irqsave(&vchan->vc.lock, flags); vchan 723 drivers/dma/owl-dma.c owl_dma_resume_pchan(vchan->pchan); vchan 725 drivers/dma/owl-dma.c spin_unlock_irqrestore(&vchan->vc.lock, flags); vchan 730 drivers/dma/owl-dma.c static u32 owl_dma_getbytes_chan(struct owl_dma_vchan *vchan) vchan 738 drivers/dma/owl-dma.c pchan = vchan->pchan; vchan 739 drivers/dma/owl-dma.c txd = vchan->txd; vchan 767 drivers/dma/owl-dma.c struct owl_dma_vchan *vchan = to_owl_vchan(chan); vchan 779 drivers/dma/owl-dma.c spin_lock_irqsave(&vchan->vc.lock, flags); vchan 781 drivers/dma/owl-dma.c vd = vchan_find_desc(&vchan->vc, cookie); vchan 787 drivers/dma/owl-dma.c bytes = owl_dma_getbytes_chan(vchan); vchan 790 drivers/dma/owl-dma.c spin_unlock_irqrestore(&vchan->vc.lock, flags); vchan 797 drivers/dma/owl-dma.c static void owl_dma_phy_alloc_and_start(struct owl_dma_vchan *vchan) vchan 799 drivers/dma/owl-dma.c struct owl_dma *od = to_owl_dma(vchan->vc.chan.device); vchan 802 drivers/dma/owl-dma.c pchan = owl_dma_get_pchan(od, vchan); vchan 808 drivers/dma/owl-dma.c vchan->pchan = pchan; vchan 809 drivers/dma/owl-dma.c owl_dma_start_next_txd(vchan); vchan 814 drivers/dma/owl-dma.c struct owl_dma_vchan *vchan = to_owl_vchan(chan); vchan 817 drivers/dma/owl-dma.c spin_lock_irqsave(&vchan->vc.lock, flags); vchan 818 drivers/dma/owl-dma.c if (vchan_issue_pending(&vchan->vc)) { vchan 819 drivers/dma/owl-dma.c if (!vchan->pchan) vchan 820 drivers/dma/owl-dma.c owl_dma_phy_alloc_and_start(vchan); vchan 822 drivers/dma/owl-dma.c spin_unlock_irqrestore(&vchan->vc.lock, flags); vchan 831 drivers/dma/owl-dma.c struct owl_dma_vchan *vchan = to_owl_vchan(chan); vchan 856 drivers/dma/owl-dma.c ret = owl_dma_cfg_lli(vchan, lli, src + offset, dst + offset, vchan 858 drivers/dma/owl-dma.c &vchan->cfg, txd->cyclic); vchan 867 drivers/dma/owl-dma.c return vchan_tx_prep(&vchan->vc, &txd->vd, flags); vchan 882 drivers/dma/owl-dma.c struct owl_dma_vchan *vchan = to_owl_vchan(chan); vchan 883 drivers/dma/owl-dma.c struct dma_slave_config *sconfig = &vchan->cfg; vchan 921 drivers/dma/owl-dma.c ret = owl_dma_cfg_lli(vchan, lli, src, dst, len, dir, sconfig, vchan 931 drivers/dma/owl-dma.c return vchan_tx_prep(&vchan->vc, &txd->vd, flags); vchan 947 drivers/dma/owl-dma.c struct owl_dma_vchan *vchan = to_owl_vchan(chan); vchan 948 drivers/dma/owl-dma.c struct dma_slave_config *sconfig = &vchan->cfg; vchan 977 drivers/dma/owl-dma.c ret = owl_dma_cfg_lli(vchan, lli, src, dst, period_len, vchan 993 drivers/dma/owl-dma.c return vchan_tx_prep(&vchan->vc, &txd->vd, flags); vchan 1003 drivers/dma/owl-dma.c struct owl_dma_vchan *vchan = to_owl_vchan(chan); vchan 1006 drivers/dma/owl-dma.c vchan_free_chan_resources(&vchan->vc); vchan 1011 drivers/dma/owl-dma.c struct owl_dma_vchan *vchan = NULL; vchan 1014 drivers/dma/owl-dma.c list_for_each_entry_safe(vchan, vchan 1016 drivers/dma/owl-dma.c list_del(&vchan->vc.chan.device_node); vchan 1017 drivers/dma/owl-dma.c tasklet_kill(&vchan->vc.task); vchan 1025 drivers/dma/owl-dma.c struct owl_dma_vchan *vchan; vchan 1036 drivers/dma/owl-dma.c vchan = to_owl_vchan(chan); vchan 1037 drivers/dma/owl-dma.c vchan->drq = drq; vchan 1145 drivers/dma/owl-dma.c struct owl_dma_vchan *vchan = &od->vchans[i]; vchan 1147 drivers/dma/owl-dma.c vchan->vc.desc_free = owl_dma_desc_free; vchan 1148 drivers/dma/owl-dma.c vchan_init(&vchan->vc, &od->dma); vchan 100 drivers/dma/pxa_dma.c struct pxad_chan *vchan; vchan 151 drivers/dma/pxa_dma.c dev_vdbg(&phy->vchan->vc.chan.dev->device, \ vchan 159 drivers/dma/pxa_dma.c dev_vdbg(&phy->vchan->vc.chan.dev->device, \ vchan 166 drivers/dma/pxa_dma.c dev_vdbg(&phy->vchan->vc.chan.dev->device, \ vchan 390 drivers/dma/pxa_dma.c if (!phy->vchan) { vchan 391 drivers/dma/pxa_dma.c phy->vchan = pchan; vchan 425 drivers/dma/pxa_dma.c chan->phy->vchan = NULL; vchan 455 drivers/dma/pxa_dma.c if (!phy->vchan) vchan 458 drivers/dma/pxa_dma.c dev_dbg(&phy->vchan->vc.chan.dev->device, vchan 462 drivers/dma/pxa_dma.c pdev = to_pxad_dev(phy->vchan->vc.chan.device); vchan 463 drivers/dma/pxa_dma.c if (phy->vchan->drcmr <= pdev->nr_requestors) { vchan 464 drivers/dma/pxa_dma.c reg = pxad_drcmr(phy->vchan->drcmr); vchan 487 drivers/dma/pxa_dma.c dev_dbg(&phy->vchan->vc.chan.dev->device, vchan 595 drivers/dma/pxa_dma.c if ((dcsr & PXA_DCSR_BUSERR) && (phy->vchan)) vchan 596 drivers/dma/pxa_dma.c dev_warn(&phy->vchan->vc.chan.dev->device, vchan 598 drivers/dma/pxa_dma.c __func__, &phy->vchan); vchan 606 drivers/dma/pxa_dma.c struct pxad_chan *chan = phy->vchan; vchan 1112 drivers/dma/pxa_dma.c phy->vchan = NULL; vchan 105 drivers/dma/sa11x0-dma.c struct sa11x0_dma_chan *vchan; vchan 275 drivers/dma/sa11x0-dma.c c = p->vchan; vchan 346 drivers/dma/sa11x0-dma.c p->vchan = NULL; vchan 356 drivers/dma/sa11x0-dma.c if (p->vchan == NULL && !list_empty(&d->chan_pending)) { vchan 364 drivers/dma/sa11x0-dma.c p->vchan = c; vchan 374 drivers/dma/sa11x0-dma.c c = p->vchan; vchan 785 drivers/dma/sa11x0-dma.c p->vchan = NULL; vchan 23 drivers/dma/st_fdma.c return container_of(c, struct st_fdma_chan, vchan.chan); vchan 78 drivers/dma/st_fdma.c vdesc = vchan_next_desc(&fchan->vchan); vchan 84 drivers/dma/st_fdma.c cmd = FDMA_CMD_START(fchan->vchan.chan.chan_id); vchan 93 drivers/dma/st_fdma.c dev_dbg(fchan->fdev->dev, "start chan:%d\n", fchan->vchan.chan.chan_id); vchan 100 drivers/dma/st_fdma.c int ch_id = fchan->vchan.chan.chan_id; vchan 138 drivers/dma/st_fdma.c spin_lock(&fchan->vchan.lock); vchan 156 drivers/dma/st_fdma.c spin_unlock(&fchan->vchan.lock); vchan 281 drivers/dma/st_fdma.c fchan->vchan.chan.chan_id, fchan->cfg.type); vchan 293 drivers/dma/st_fdma.c __func__, fchan->vchan.chan.chan_id); vchan 298 drivers/dma/st_fdma.c spin_lock_irqsave(&fchan->vchan.lock, flags); vchan 300 drivers/dma/st_fdma.c spin_unlock_irqrestore(&fchan->vchan.lock, flags); vchan 342 drivers/dma/st_fdma.c return vchan_tx_prep(&fchan->vchan, &fdesc->vdesc, flags); vchan 350 drivers/dma/st_fdma.c int ch_id = fchan->vchan.chan.chan_id; vchan 502 drivers/dma/st_fdma.c return vchan_tx_prep(&fchan->vchan, &fdesc->vdesc, flags); vchan 551 drivers/dma/st_fdma.c return vchan_tx_prep(&fchan->vchan, &fdesc->vdesc, flags); vchan 592 drivers/dma/st_fdma.c spin_lock_irqsave(&fchan->vchan.lock, flags); vchan 593 drivers/dma/st_fdma.c vd = vchan_find_desc(&fchan->vchan, cookie); vchan 601 drivers/dma/st_fdma.c spin_unlock_irqrestore(&fchan->vchan.lock, flags); vchan 611 drivers/dma/st_fdma.c spin_lock_irqsave(&fchan->vchan.lock, flags); vchan 613 drivers/dma/st_fdma.c if (vchan_issue_pending(&fchan->vchan) && !fchan->fdesc) vchan 616 drivers/dma/st_fdma.c spin_unlock_irqrestore(&fchan->vchan.lock, flags); vchan 623 drivers/dma/st_fdma.c int ch_id = fchan->vchan.chan.chan_id; vchan 628 drivers/dma/st_fdma.c spin_lock_irqsave(&fchan->vchan.lock, flags); vchan 631 drivers/dma/st_fdma.c spin_unlock_irqrestore(&fchan->vchan.lock, flags); vchan 641 drivers/dma/st_fdma.c int ch_id = fchan->vchan.chan.chan_id; vchan 645 drivers/dma/st_fdma.c spin_lock_irqsave(&fchan->vchan.lock, flags); vchan 651 drivers/dma/st_fdma.c spin_unlock_irqrestore(&fchan->vchan.lock, flags); vchan 661 drivers/dma/st_fdma.c int ch_id = fchan->vchan.chan.chan_id; vchan 666 drivers/dma/st_fdma.c spin_lock_irqsave(&fchan->vchan.lock, flags); vchan 669 drivers/dma/st_fdma.c vchan_get_all_descriptors(&fchan->vchan, &head); vchan 670 drivers/dma/st_fdma.c spin_unlock_irqrestore(&fchan->vchan.lock, flags); vchan 671 drivers/dma/st_fdma.c vchan_dma_desc_free_list(&fchan->vchan, &head); vchan 733 drivers/dma/st_fdma.c list_del(&fchan->vchan.chan.device_node); vchan 734 drivers/dma/st_fdma.c tasklet_kill(&fchan->vchan.task); vchan 797 drivers/dma/st_fdma.c fchan->vchan.desc_free = st_fdma_free_desc; vchan 798 drivers/dma/st_fdma.c vchan_init(&fchan->vchan, &fdev->dma_device); vchan 125 drivers/dma/st_fdma.h struct virt_dma_chan vchan; vchan 186 drivers/dma/st_fdma.h + (fchan)->vchan.chan.chan_id * 0x4 \ vchan 191 drivers/dma/st_fdma.h + (fchan)->vchan.chan.chan_id * 0x4 \ vchan 208 drivers/dma/st_fdma.h + (fchan)->vchan.chan.chan_id * FDMA_NODE_SZ \ vchan 213 drivers/dma/st_fdma.h + (fchan)->vchan.chan.chan_id * FDMA_NODE_SZ \ vchan 192 drivers/dma/stm32-dma.c struct virt_dma_chan vchan; vchan 217 drivers/dma/stm32-dma.c return container_of(chan->vchan.chan.device, struct stm32_dma_device, vchan 223 drivers/dma/stm32-dma.c return container_of(c, struct stm32_dma_chan, vchan.chan); vchan 233 drivers/dma/stm32-dma.c return &chan->vchan.chan.dev->device; vchan 489 drivers/dma/stm32-dma.c spin_lock_irqsave(&chan->vchan.lock, flags); vchan 496 drivers/dma/stm32-dma.c vchan_get_all_descriptors(&chan->vchan, &head); vchan 497 drivers/dma/stm32-dma.c spin_unlock_irqrestore(&chan->vchan.lock, flags); vchan 498 drivers/dma/stm32-dma.c vchan_dma_desc_free_list(&chan->vchan, &head); vchan 507 drivers/dma/stm32-dma.c vchan_synchronize(&chan->vchan); vchan 544 drivers/dma/stm32-dma.c vdesc = vchan_next_desc(&chan->vchan); vchan 583 drivers/dma/stm32-dma.c dev_dbg(chan2dev(chan), "vchan %pK: started\n", &chan->vchan); vchan 640 drivers/dma/stm32-dma.c spin_lock(&chan->vchan.lock); vchan 673 drivers/dma/stm32-dma.c spin_unlock(&chan->vchan.lock); vchan 683 drivers/dma/stm32-dma.c spin_lock_irqsave(&chan->vchan.lock, flags); vchan 684 drivers/dma/stm32-dma.c if (vchan_issue_pending(&chan->vchan) && !chan->desc && !chan->busy) { vchan 685 drivers/dma/stm32-dma.c dev_dbg(chan2dev(chan), "vchan %pK: issued\n", &chan->vchan); vchan 689 drivers/dma/stm32-dma.c spin_unlock_irqrestore(&chan->vchan.lock, flags); vchan 886 drivers/dma/stm32-dma.c return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); vchan 971 drivers/dma/stm32-dma.c return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); vchan 1023 drivers/dma/stm32-dma.c return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); vchan 1154 drivers/dma/stm32-dma.c spin_lock_irqsave(&chan->vchan.lock, flags); vchan 1155 drivers/dma/stm32-dma.c vdesc = vchan_find_desc(&chan->vchan, cookie); vchan 1164 drivers/dma/stm32-dma.c spin_unlock_irqrestore(&chan->vchan.lock, flags); vchan 1197 drivers/dma/stm32-dma.c spin_lock_irqsave(&chan->vchan.lock, flags); vchan 1200 drivers/dma/stm32-dma.c spin_unlock_irqrestore(&chan->vchan.lock, flags); vchan 1254 drivers/dma/stm32-dma.c c = dma_get_slave_channel(&chan->vchan.chan); vchan 1352 drivers/dma/stm32-dma.c chan->vchan.desc_free = stm32_dma_desc_free; vchan 1353 drivers/dma/stm32-dma.c vchan_init(&chan->vchan, dd); vchan 259 drivers/dma/stm32-mdma.c struct virt_dma_chan vchan; vchan 287 drivers/dma/stm32-mdma.c return container_of(chan->vchan.chan.device, struct stm32_mdma_device, vchan 293 drivers/dma/stm32-mdma.c return container_of(c, struct stm32_mdma_chan, vchan.chan); vchan 303 drivers/dma/stm32-mdma.c return &chan->vchan.chan.dev->device; vchan 810 drivers/dma/stm32-mdma.c return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); vchan 901 drivers/dma/stm32-mdma.c return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); vchan 1087 drivers/dma/stm32-mdma.c return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); vchan 1124 drivers/dma/stm32-mdma.c vdesc = vchan_next_desc(&chan->vchan); vchan 1163 drivers/dma/stm32-mdma.c dev_dbg(chan2dev(chan), "vchan %pK: started\n", &chan->vchan); vchan 1171 drivers/dma/stm32-mdma.c spin_lock_irqsave(&chan->vchan.lock, flags); vchan 1173 drivers/dma/stm32-mdma.c if (!vchan_issue_pending(&chan->vchan)) vchan 1176 drivers/dma/stm32-mdma.c dev_dbg(chan2dev(chan), "vchan %pK: issued\n", &chan->vchan); vchan 1182 drivers/dma/stm32-mdma.c spin_unlock_irqrestore(&chan->vchan.lock, flags); vchan 1191 drivers/dma/stm32-mdma.c spin_lock_irqsave(&chan->vchan.lock, flags); vchan 1193 drivers/dma/stm32-mdma.c spin_unlock_irqrestore(&chan->vchan.lock, flags); vchan 1196 drivers/dma/stm32-mdma.c dev_dbg(chan2dev(chan), "vchan %pK: pause\n", &chan->vchan); vchan 1211 drivers/dma/stm32-mdma.c spin_lock_irqsave(&chan->vchan.lock, flags); vchan 1231 drivers/dma/stm32-mdma.c spin_unlock_irqrestore(&chan->vchan.lock, flags); vchan 1233 drivers/dma/stm32-mdma.c dev_dbg(chan2dev(chan), "vchan %pK: resume\n", &chan->vchan); vchan 1244 drivers/dma/stm32-mdma.c spin_lock_irqsave(&chan->vchan.lock, flags); vchan 1249 drivers/dma/stm32-mdma.c vchan_get_all_descriptors(&chan->vchan, &head); vchan 1250 drivers/dma/stm32-mdma.c spin_unlock_irqrestore(&chan->vchan.lock, flags); vchan 1252 drivers/dma/stm32-mdma.c vchan_dma_desc_free_list(&chan->vchan, &head); vchan 1261 drivers/dma/stm32-mdma.c vchan_synchronize(&chan->vchan); vchan 1316 drivers/dma/stm32-mdma.c spin_lock_irqsave(&chan->vchan.lock, flags); vchan 1318 drivers/dma/stm32-mdma.c vdesc = vchan_find_desc(&chan->vchan, cookie); vchan 1327 drivers/dma/stm32-mdma.c spin_unlock_irqrestore(&chan->vchan.lock, flags); vchan 1374 drivers/dma/stm32-mdma.c spin_lock(&chan->vchan.lock); vchan 1381 drivers/dma/stm32-mdma.c spin_unlock(&chan->vchan.lock); vchan 1427 drivers/dma/stm32-mdma.c spin_unlock(&chan->vchan.lock); vchan 1469 drivers/dma/stm32-mdma.c spin_lock_irqsave(&chan->vchan.lock, flags); vchan 1472 drivers/dma/stm32-mdma.c spin_unlock_irqrestore(&chan->vchan.lock, flags); vchan 1635 drivers/dma/stm32-mdma.c chan->vchan.desc_free = stm32_mdma_desc_free; vchan 1636 drivers/dma/stm32-mdma.c vchan_init(&chan->vchan, dd); vchan 129 drivers/dma/sun4i-dma.c struct sun4i_dma_vchan *vchan; vchan 212 drivers/dma/sun4i-dma.c struct sun4i_dma_vchan *vchan = to_sun4i_dma_vchan(chan); vchan 214 drivers/dma/sun4i-dma.c vchan_free_chan_resources(&vchan->vc); vchan 218 drivers/dma/sun4i-dma.c struct sun4i_dma_vchan *vchan) vchan 228 drivers/dma/sun4i-dma.c if (vchan->is_dedicated) { vchan 239 drivers/dma/sun4i-dma.c pchan->vchan = vchan; vchan 256 drivers/dma/sun4i-dma.c pchan->vchan = NULL; vchan 320 drivers/dma/sun4i-dma.c struct sun4i_dma_vchan *vchan) vchan 328 drivers/dma/sun4i-dma.c lockdep_assert_held(&vchan->vc.lock); vchan 331 drivers/dma/sun4i-dma.c pchan = find_and_use_pchan(priv, vchan); vchan 339 drivers/dma/sun4i-dma.c if (vchan->processing) { vchan 340 drivers/dma/sun4i-dma.c dev_dbg(chan2dev(&vchan->vc.chan), vchan 348 drivers/dma/sun4i-dma.c vd = vchan_next_desc(&vchan->vc); vchan 350 drivers/dma/sun4i-dma.c dev_dbg(chan2dev(&vchan->vc.chan), vchan 361 drivers/dma/sun4i-dma.c dev_dbg(chan2dev(&vchan->vc.chan), vchan 369 drivers/dma/sun4i-dma.c vchan->processing = promise; vchan 373 drivers/dma/sun4i-dma.c vchan->contract = contract; vchan 374 drivers/dma/sun4i-dma.c vchan->pchan = pchan; vchan 614 drivers/dma/sun4i-dma.c struct sun4i_dma_vchan *vchan = to_sun4i_dma_vchan(chan); vchan 615 drivers/dma/sun4i-dma.c struct dma_slave_config *sconfig = &vchan->cfg; vchan 633 drivers/dma/sun4i-dma.c if (vchan->is_dedicated) vchan 645 drivers/dma/sun4i-dma.c if (vchan->is_dedicated) { vchan 657 drivers/dma/sun4i-dma.c return vchan_tx_prep(&vchan->vc, &contract->vd, flags); vchan 665 drivers/dma/sun4i-dma.c struct sun4i_dma_vchan *vchan = to_sun4i_dma_vchan(chan); vchan 666 drivers/dma/sun4i-dma.c struct dma_slave_config *sconfig = &vchan->cfg; vchan 678 drivers/dma/sun4i-dma.c if (vchan->is_dedicated) { vchan 701 drivers/dma/sun4i-dma.c SUN4I_DMA_CFG_DST_DRQ_TYPE(vchan->endpoint) | vchan 706 drivers/dma/sun4i-dma.c endpoints = SUN4I_DMA_CFG_SRC_DRQ_TYPE(vchan->endpoint) | vchan 763 drivers/dma/sun4i-dma.c return vchan_tx_prep(&vchan->vc, &contract->vd, flags); vchan 771 drivers/dma/sun4i-dma.c struct sun4i_dma_vchan *vchan = to_sun4i_dma_vchan(chan); vchan 772 drivers/dma/sun4i-dma.c struct dma_slave_config *sconfig = &vchan->cfg; vchan 793 drivers/dma/sun4i-dma.c if (vchan->is_dedicated) { vchan 804 drivers/dma/sun4i-dma.c endpoints = SUN4I_DMA_CFG_DST_DRQ_TYPE(vchan->endpoint) | vchan 811 drivers/dma/sun4i-dma.c SUN4I_DMA_CFG_SRC_DRQ_TYPE(vchan->endpoint) | vchan 836 drivers/dma/sun4i-dma.c if (vchan->is_dedicated) vchan 859 drivers/dma/sun4i-dma.c return vchan_tx_prep(&vchan->vc, &contract->vd, flags); vchan 865 drivers/dma/sun4i-dma.c struct sun4i_dma_vchan *vchan = to_sun4i_dma_vchan(chan); vchan 866 drivers/dma/sun4i-dma.c struct sun4i_dma_pchan *pchan = vchan->pchan; vchan 870 drivers/dma/sun4i-dma.c spin_lock_irqsave(&vchan->vc.lock, flags); vchan 871 drivers/dma/sun4i-dma.c vchan_get_all_descriptors(&vchan->vc, &head); vchan 872 drivers/dma/sun4i-dma.c spin_unlock_irqrestore(&vchan->vc.lock, flags); vchan 887 drivers/dma/sun4i-dma.c spin_lock_irqsave(&vchan->vc.lock, flags); vchan 888 drivers/dma/sun4i-dma.c vchan_dma_desc_free_list(&vchan->vc, &head); vchan 890 drivers/dma/sun4i-dma.c vchan->processing = NULL; vchan 891 drivers/dma/sun4i-dma.c vchan->pchan = NULL; vchan 892 drivers/dma/sun4i-dma.c spin_unlock_irqrestore(&vchan->vc.lock, flags); vchan 900 drivers/dma/sun4i-dma.c struct sun4i_dma_vchan *vchan = to_sun4i_dma_vchan(chan); vchan 902 drivers/dma/sun4i-dma.c memcpy(&vchan->cfg, config, sizeof(*config)); vchan 911 drivers/dma/sun4i-dma.c struct sun4i_dma_vchan *vchan; vchan 930 drivers/dma/sun4i-dma.c vchan = to_sun4i_dma_vchan(chan); vchan 931 drivers/dma/sun4i-dma.c vchan->is_dedicated = is_dedicated; vchan 932 drivers/dma/sun4i-dma.c vchan->endpoint = endpoint; vchan 941 drivers/dma/sun4i-dma.c struct sun4i_dma_vchan *vchan = to_sun4i_dma_vchan(chan); vchan 942 drivers/dma/sun4i-dma.c struct sun4i_dma_pchan *pchan = vchan->pchan; vchan 954 drivers/dma/sun4i-dma.c spin_lock_irqsave(&vchan->vc.lock, flags); vchan 955 drivers/dma/sun4i-dma.c vd = vchan_find_desc(&vchan->vc, cookie); vchan 981 drivers/dma/sun4i-dma.c spin_unlock_irqrestore(&vchan->vc.lock, flags); vchan 989 drivers/dma/sun4i-dma.c struct sun4i_dma_vchan *vchan = to_sun4i_dma_vchan(chan); vchan 992 drivers/dma/sun4i-dma.c spin_lock_irqsave(&vchan->vc.lock, flags); vchan 998 drivers/dma/sun4i-dma.c if (vchan_issue_pending(&vchan->vc)) vchan 999 drivers/dma/sun4i-dma.c __execute_vchan_pending(priv, vchan); vchan 1001 drivers/dma/sun4i-dma.c spin_unlock_irqrestore(&vchan->vc.lock, flags); vchan 1008 drivers/dma/sun4i-dma.c struct sun4i_dma_vchan *vchan; vchan 1023 drivers/dma/sun4i-dma.c vchan = pchan->vchan; vchan 1024 drivers/dma/sun4i-dma.c if (!vchan) /* a terminated channel may still interrupt */ vchan 1026 drivers/dma/sun4i-dma.c contract = vchan->contract; vchan 1033 drivers/dma/sun4i-dma.c spin_lock(&vchan->vc.lock); vchan 1039 drivers/dma/sun4i-dma.c list_del(&vchan->processing->list); vchan 1040 drivers/dma/sun4i-dma.c list_add_tail(&vchan->processing->list, vchan 1058 drivers/dma/sun4i-dma.c vchan->processing = promise; vchan 1062 drivers/dma/sun4i-dma.c vchan->processing = NULL; vchan 1063 drivers/dma/sun4i-dma.c vchan->pchan = NULL; vchan 1070 drivers/dma/sun4i-dma.c spin_unlock(&vchan->vc.lock); vchan 1096 drivers/dma/sun4i-dma.c vchan = &priv->vchans[i]; vchan 1097 drivers/dma/sun4i-dma.c spin_lock(&vchan->vc.lock); vchan 1098 drivers/dma/sun4i-dma.c __execute_vchan_pending(priv, vchan); vchan 1099 drivers/dma/sun4i-dma.c spin_unlock(&vchan->vc.lock); vchan 1198 drivers/dma/sun4i-dma.c struct sun4i_dma_vchan *vchan = &priv->vchans[i]; vchan 1200 drivers/dma/sun4i-dma.c spin_lock_init(&vchan->vc.lock); vchan 1201 drivers/dma/sun4i-dma.c vchan->vc.desc_free = sun4i_dma_free_contract; vchan 1202 drivers/dma/sun4i-dma.c vchan_init(&vchan->vc, &priv->slave); vchan 170 drivers/dma/sun6i-dma.c struct sun6i_vchan *vchan; vchan 387 drivers/dma/sun6i-dma.c static inline void sun6i_dma_dump_lli(struct sun6i_vchan *vchan, vchan 392 drivers/dma/sun6i-dma.c dev_dbg(chan2dev(&vchan->vc.chan), vchan 427 drivers/dma/sun6i-dma.c static int sun6i_dma_start_desc(struct sun6i_vchan *vchan) vchan 429 drivers/dma/sun6i-dma.c struct sun6i_dma_dev *sdev = to_sun6i_dma_dev(vchan->vc.chan.device); vchan 430 drivers/dma/sun6i-dma.c struct virt_dma_desc *desc = vchan_next_desc(&vchan->vc); vchan 431 drivers/dma/sun6i-dma.c struct sun6i_pchan *pchan = vchan->phy; vchan 448 drivers/dma/sun6i-dma.c sun6i_dma_dump_lli(vchan, pchan->desc->v_lli); vchan 453 drivers/dma/sun6i-dma.c vchan->irq_type = vchan->cyclic ? DMA_IRQ_PKG : DMA_IRQ_QUEUE; vchan 458 drivers/dma/sun6i-dma.c irq_val |= vchan->irq_type << (irq_offset * DMA_IRQ_CHAN_WIDTH); vchan 473 drivers/dma/sun6i-dma.c struct sun6i_vchan *vchan; vchan 478 drivers/dma/sun6i-dma.c list_for_each_entry(vchan, &sdev->slave.channels, vc.chan.device_node) { vchan 479 drivers/dma/sun6i-dma.c spin_lock_irq(&vchan->vc.lock); vchan 481 drivers/dma/sun6i-dma.c pchan = vchan->phy; vchan 484 drivers/dma/sun6i-dma.c if (sun6i_dma_start_desc(vchan)) { vchan 492 drivers/dma/sun6i-dma.c vchan->phy = NULL; vchan 493 drivers/dma/sun6i-dma.c pchan->vchan = NULL; vchan 496 drivers/dma/sun6i-dma.c spin_unlock_irq(&vchan->vc.lock); vchan 503 drivers/dma/sun6i-dma.c if (pchan->vchan || list_empty(&sdev->pending)) vchan 506 drivers/dma/sun6i-dma.c vchan = list_first_entry(&sdev->pending, vchan 510 drivers/dma/sun6i-dma.c list_del_init(&vchan->node); vchan 514 drivers/dma/sun6i-dma.c pchan->vchan = vchan; vchan 515 drivers/dma/sun6i-dma.c vchan->phy = pchan; vchan 517 drivers/dma/sun6i-dma.c pchan->idx, &vchan->vc); vchan 526 drivers/dma/sun6i-dma.c vchan = pchan->vchan; vchan 527 drivers/dma/sun6i-dma.c if (vchan) { vchan 528 drivers/dma/sun6i-dma.c spin_lock_irq(&vchan->vc.lock); vchan 529 drivers/dma/sun6i-dma.c sun6i_dma_start_desc(vchan); vchan 530 drivers/dma/sun6i-dma.c spin_unlock_irq(&vchan->vc.lock); vchan 538 drivers/dma/sun6i-dma.c struct sun6i_vchan *vchan; vchan 555 drivers/dma/sun6i-dma.c vchan = pchan->vchan; vchan 556 drivers/dma/sun6i-dma.c if (vchan && (status & vchan->irq_type)) { vchan 557 drivers/dma/sun6i-dma.c if (vchan->cyclic) { vchan 560 drivers/dma/sun6i-dma.c spin_lock(&vchan->vc.lock); vchan 563 drivers/dma/sun6i-dma.c spin_unlock(&vchan->vc.lock); vchan 634 drivers/dma/sun6i-dma.c struct sun6i_vchan *vchan = to_sun6i_vchan(chan); vchan 642 drivers/dma/sun6i-dma.c __func__, vchan->vc.chan.chan_id, &dest, &src, len, flags); vchan 673 drivers/dma/sun6i-dma.c sun6i_dma_dump_lli(vchan, v_lli); vchan 675 drivers/dma/sun6i-dma.c return vchan_tx_prep(&vchan->vc, &txd->vd, flags); vchan 688 drivers/dma/sun6i-dma.c struct sun6i_vchan *vchan = to_sun6i_vchan(chan); vchan 689 drivers/dma/sun6i-dma.c struct dma_slave_config *sconfig = &vchan->cfg; vchan 722 drivers/dma/sun6i-dma.c sdev->cfg->set_drq(&v_lli->cfg, DRQ_SDRAM, vchan->port); vchan 727 drivers/dma/sun6i-dma.c __func__, vchan->vc.chan.chan_id, vchan 735 drivers/dma/sun6i-dma.c sdev->cfg->set_drq(&v_lli->cfg, vchan->port, DRQ_SDRAM); vchan 740 drivers/dma/sun6i-dma.c __func__, vchan->vc.chan.chan_id, vchan 750 drivers/dma/sun6i-dma.c sun6i_dma_dump_lli(vchan, prev); vchan 752 drivers/dma/sun6i-dma.c return vchan_tx_prep(&vchan->vc, &txd->vd, flags); vchan 770 drivers/dma/sun6i-dma.c struct sun6i_vchan *vchan = to_sun6i_vchan(chan); vchan 771 drivers/dma/sun6i-dma.c struct dma_slave_config *sconfig = &vchan->cfg; vchan 803 drivers/dma/sun6i-dma.c sdev->cfg->set_drq(&v_lli->cfg, DRQ_SDRAM, vchan->port); vchan 809 drivers/dma/sun6i-dma.c sdev->cfg->set_drq(&v_lli->cfg, vchan->port, DRQ_SDRAM); vchan 818 drivers/dma/sun6i-dma.c vchan->cyclic = true; vchan 820 drivers/dma/sun6i-dma.c return vchan_tx_prep(&vchan->vc, &txd->vd, flags); vchan 832 drivers/dma/sun6i-dma.c struct sun6i_vchan *vchan = to_sun6i_vchan(chan); vchan 834 drivers/dma/sun6i-dma.c memcpy(&vchan->cfg, config, sizeof(*config)); vchan 842 drivers/dma/sun6i-dma.c struct sun6i_vchan *vchan = to_sun6i_vchan(chan); vchan 843 drivers/dma/sun6i-dma.c struct sun6i_pchan *pchan = vchan->phy; vchan 845 drivers/dma/sun6i-dma.c dev_dbg(chan2dev(chan), "vchan %p: pause\n", &vchan->vc); vchan 852 drivers/dma/sun6i-dma.c list_del_init(&vchan->node); vchan 862 drivers/dma/sun6i-dma.c struct sun6i_vchan *vchan = to_sun6i_vchan(chan); vchan 863 drivers/dma/sun6i-dma.c struct sun6i_pchan *pchan = vchan->phy; vchan 866 drivers/dma/sun6i-dma.c dev_dbg(chan2dev(chan), "vchan %p: resume\n", &vchan->vc); vchan 868 drivers/dma/sun6i-dma.c spin_lock_irqsave(&vchan->vc.lock, flags); vchan 873 drivers/dma/sun6i-dma.c } else if (!list_empty(&vchan->vc.desc_issued)) { vchan 875 drivers/dma/sun6i-dma.c list_add_tail(&vchan->node, &sdev->pending); vchan 879 drivers/dma/sun6i-dma.c spin_unlock_irqrestore(&vchan->vc.lock, flags); vchan 887 drivers/dma/sun6i-dma.c struct sun6i_vchan *vchan = to_sun6i_vchan(chan); vchan 888 drivers/dma/sun6i-dma.c struct sun6i_pchan *pchan = vchan->phy; vchan 893 drivers/dma/sun6i-dma.c list_del_init(&vchan->node); vchan 896 drivers/dma/sun6i-dma.c spin_lock_irqsave(&vchan->vc.lock, flags); vchan 898 drivers/dma/sun6i-dma.c if (vchan->cyclic) { vchan 899 drivers/dma/sun6i-dma.c vchan->cyclic = false; vchan 902 drivers/dma/sun6i-dma.c struct virt_dma_chan *vc = &vchan->vc; vchan 908 drivers/dma/sun6i-dma.c vchan_get_all_descriptors(&vchan->vc, &head); vchan 914 drivers/dma/sun6i-dma.c vchan->phy = NULL; vchan 915 drivers/dma/sun6i-dma.c pchan->vchan = NULL; vchan 920 drivers/dma/sun6i-dma.c spin_unlock_irqrestore(&vchan->vc.lock, flags); vchan 922 drivers/dma/sun6i-dma.c vchan_dma_desc_free_list(&vchan->vc, &head); vchan 931 drivers/dma/sun6i-dma.c struct sun6i_vchan *vchan = to_sun6i_vchan(chan); vchan 932 drivers/dma/sun6i-dma.c struct sun6i_pchan *pchan = vchan->phy; vchan 944 drivers/dma/sun6i-dma.c spin_lock_irqsave(&vchan->vc.lock, flags); vchan 946 drivers/dma/sun6i-dma.c vd = vchan_find_desc(&vchan->vc, cookie); vchan 958 drivers/dma/sun6i-dma.c spin_unlock_irqrestore(&vchan->vc.lock, flags); vchan 968 drivers/dma/sun6i-dma.c struct sun6i_vchan *vchan = to_sun6i_vchan(chan); vchan 971 drivers/dma/sun6i-dma.c spin_lock_irqsave(&vchan->vc.lock, flags); vchan 973 drivers/dma/sun6i-dma.c if (vchan_issue_pending(&vchan->vc)) { vchan 976 drivers/dma/sun6i-dma.c if (!vchan->phy && list_empty(&vchan->node)) { vchan 977 drivers/dma/sun6i-dma.c list_add_tail(&vchan->node, &sdev->pending); vchan 980 drivers/dma/sun6i-dma.c &vchan->vc); vchan 986 drivers/dma/sun6i-dma.c &vchan->vc); vchan 989 drivers/dma/sun6i-dma.c spin_unlock_irqrestore(&vchan->vc.lock, flags); vchan 995 drivers/dma/sun6i-dma.c struct sun6i_vchan *vchan = to_sun6i_vchan(chan); vchan 999 drivers/dma/sun6i-dma.c list_del_init(&vchan->node); vchan 1002 drivers/dma/sun6i-dma.c vchan_free_chan_resources(&vchan->vc); vchan 1009 drivers/dma/sun6i-dma.c struct sun6i_vchan *vchan; vchan 1020 drivers/dma/sun6i-dma.c vchan = to_sun6i_vchan(chan); vchan 1021 drivers/dma/sun6i-dma.c vchan->port = port; vchan 1047 drivers/dma/sun6i-dma.c struct sun6i_vchan *vchan = &sdev->vchans[i]; vchan 1049 drivers/dma/sun6i-dma.c list_del(&vchan->vc.chan.device_node); vchan 1050 drivers/dma/sun6i-dma.c tasklet_kill(&vchan->vc.task); vchan 1356 drivers/dma/sun6i-dma.c struct sun6i_vchan *vchan = &sdc->vchans[i]; vchan 1358 drivers/dma/sun6i-dma.c INIT_LIST_HEAD(&vchan->node); vchan 1359 drivers/dma/sun6i-dma.c vchan->vc.desc_free = sun6i_dma_free_desc; vchan 1360 drivers/dma/sun6i-dma.c vchan_init(&vchan->vc, &sdc->slave); vchan 225 drivers/dma/ti/edma.c struct virt_dma_chan vchan; vchan 748 drivers/dma/ti/edma.c return container_of(c, struct edma_chan, vchan.chan); vchan 767 drivers/dma/ti/edma.c struct device *dev = echan->vchan.chan.device->dev; vchan 772 drivers/dma/ti/edma.c vdesc = vchan_next_desc(&echan->vchan); vchan 861 drivers/dma/ti/edma.c spin_lock_irqsave(&echan->vchan.lock, flags); vchan 878 drivers/dma/ti/edma.c vchan_get_all_descriptors(&echan->vchan, &head); vchan 879 drivers/dma/ti/edma.c spin_unlock_irqrestore(&echan->vchan.lock, flags); vchan 880 drivers/dma/ti/edma.c vchan_dma_desc_free_list(&echan->vchan, &head); vchan 889 drivers/dma/ti/edma.c vchan_synchronize(&echan->vchan); vchan 1146 drivers/dma/ti/edma.c return vchan_tx_prep(&echan->vchan, &edesc->vdesc, tx_flags); vchan 1262 drivers/dma/ti/edma.c return vchan_tx_prep(&echan->vchan, &edesc->vdesc, tx_flags); vchan 1419 drivers/dma/ti/edma.c return vchan_tx_prep(&echan->vchan, &edesc->vdesc, tx_flags); vchan 1424 drivers/dma/ti/edma.c struct device *dev = echan->vchan.chan.device->dev; vchan 1427 drivers/dma/ti/edma.c spin_lock(&echan->vchan.lock); vchan 1432 drivers/dma/ti/edma.c spin_unlock(&echan->vchan.lock); vchan 1456 drivers/dma/ti/edma.c spin_unlock(&echan->vchan.lock); vchan 1508 drivers/dma/ti/edma.c struct device *dev = echan->vchan.chan.device->dev; vchan 1515 drivers/dma/ti/edma.c spin_lock(&echan->vchan.lock); vchan 1545 drivers/dma/ti/edma.c spin_unlock(&echan->vchan.lock); vchan 1688 drivers/dma/ti/edma.c vchan_free_chan_resources(&echan->vchan); vchan 1720 drivers/dma/ti/edma.c spin_lock_irqsave(&echan->vchan.lock, flags); vchan 1721 drivers/dma/ti/edma.c if (vchan_issue_pending(&echan->vchan) && !echan->edesc) vchan 1723 drivers/dma/ti/edma.c spin_unlock_irqrestore(&echan->vchan.lock, flags); vchan 1774 drivers/dma/ti/edma.c dev_dbg_ratelimited(echan->vchan.chan.device->dev, vchan 1843 drivers/dma/ti/edma.c spin_lock_irqsave(&echan->vchan.lock, flags); vchan 1847 drivers/dma/ti/edma.c struct virt_dma_desc *vdesc = vchan_find_desc(&echan->vchan, vchan 1870 drivers/dma/ti/edma.c spin_unlock_irqrestore(&echan->vchan.lock, flags); vchan 1971 drivers/dma/ti/edma.c echan->vchan.desc_free = edma_desc_free; vchan 1974 drivers/dma/ti/edma.c vchan_init(&echan->vchan, m_ddev); vchan 1976 drivers/dma/ti/edma.c vchan_init(&echan->vchan, s_ddev); vchan 2209 drivers/dma/ti/edma.c chan = &echan->vchan.chan; vchan 2486 drivers/dma/ti/edma.c &dmadev->channels, vchan.chan.device_node) { vchan 2487 drivers/dma/ti/edma.c list_del(&echan->vchan.chan.device_node); vchan 2488 drivers/dma/ti/edma.c tasklet_kill(&echan->vchan.task); vchan 112 drivers/dma/zx_dma.c struct zx_dma_chan *vchan; vchan 234 drivers/dma/zx_dma.c p->vchan = NULL; vchan 245 drivers/dma/zx_dma.c if (!p->vchan) { vchan 250 drivers/dma/zx_dma.c p->vchan = c; vchan 261 drivers/dma/zx_dma.c c = p->vchan; vchan 286 drivers/dma/zx_dma.c c = p->vchan; vchan 681 drivers/dma/zx_dma.c p->vchan = NULL; vchan 144 drivers/staging/mt7621-dma/mtk-hsdma.c struct virt_dma_chan vchan; vchan 168 drivers/staging/mt7621-dma/mtk-hsdma.c return container_of(chan->vchan.chan.device, struct mtk_hsdam_engine, vchan 174 drivers/staging/mt7621-dma/mtk-hsdma.c return container_of(c, struct mtk_hsdma_chan, vchan.chan); vchan 291 drivers/staging/mt7621-dma/mtk-hsdma.c spin_lock_bh(&chan->vchan.lock); vchan 294 drivers/staging/mt7621-dma/mtk-hsdma.c vchan_get_all_descriptors(&chan->vchan, &head); vchan 295 drivers/staging/mt7621-dma/mtk-hsdma.c spin_unlock_bh(&chan->vchan.lock); vchan 297 drivers/staging/mt7621-dma/mtk-hsdma.c vchan_dma_desc_free_list(&chan->vchan, &head); vchan 390 drivers/staging/mt7621-dma/mtk-hsdma.c vdesc = vchan_next_desc(&chan->vchan); vchan 408 drivers/staging/mt7621-dma/mtk-hsdma.c spin_lock_bh(&chan->vchan.lock); vchan 422 drivers/staging/mt7621-dma/mtk-hsdma.c spin_unlock_bh(&chan->vchan.lock); vchan 449 drivers/staging/mt7621-dma/mtk-hsdma.c spin_lock_bh(&chan->vchan.lock); vchan 450 drivers/staging/mt7621-dma/mtk-hsdma.c if (vchan_issue_pending(&chan->vchan) && !chan->desc) { vchan 458 drivers/staging/mt7621-dma/mtk-hsdma.c spin_unlock_bh(&chan->vchan.lock); vchan 481 drivers/staging/mt7621-dma/mtk-hsdma.c return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); vchan 704 drivers/staging/mt7621-dma/mtk-hsdma.c chan->vchan.desc_free = mtk_hsdma_desc_free; vchan 705 drivers/staging/mt7621-dma/mtk-hsdma.c vchan_init(&chan->vchan, dd); vchan 108 drivers/staging/ralink-gdma/ralink-gdma.c struct virt_dma_chan vchan; vchan 141 drivers/staging/ralink-gdma/ralink-gdma.c return container_of(chan->vchan.chan.device, struct gdma_dma_dev, vchan 147 drivers/staging/ralink-gdma/ralink-gdma.c return container_of(c, struct gdma_dmaengine_chan, vchan.chan); vchan 228 drivers/staging/ralink-gdma/ralink-gdma.c spin_lock_irqsave(&chan->vchan.lock, flags); vchan 231 drivers/staging/ralink-gdma/ralink-gdma.c vchan_get_all_descriptors(&chan->vchan, &head); vchan 232 drivers/staging/ralink-gdma/ralink-gdma.c spin_unlock_irqrestore(&chan->vchan.lock, flags); vchan 234 drivers/staging/ralink-gdma/ralink-gdma.c vchan_dma_desc_free_list(&chan->vchan, &head); vchan 417 drivers/staging/ralink-gdma/ralink-gdma.c vdesc = vchan_next_desc(&chan->vchan); vchan 436 drivers/staging/ralink-gdma/ralink-gdma.c spin_lock_irqsave(&chan->vchan.lock, flags); vchan 460 drivers/staging/ralink-gdma/ralink-gdma.c spin_unlock_irqrestore(&chan->vchan.lock, flags); vchan 500 drivers/staging/ralink-gdma/ralink-gdma.c spin_lock_irqsave(&chan->vchan.lock, flags); vchan 501 drivers/staging/ralink-gdma/ralink-gdma.c if (vchan_issue_pending(&chan->vchan) && !chan->desc) { vchan 510 drivers/staging/ralink-gdma/ralink-gdma.c spin_unlock_irqrestore(&chan->vchan.lock, flags); vchan 554 drivers/staging/ralink-gdma/ralink-gdma.c return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); vchan 601 drivers/staging/ralink-gdma/ralink-gdma.c return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); vchan 648 drivers/staging/ralink-gdma/ralink-gdma.c return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); vchan 669 drivers/staging/ralink-gdma/ralink-gdma.c spin_lock_irqsave(&chan->vchan.lock, flags); vchan 683 drivers/staging/ralink-gdma/ralink-gdma.c vdesc = vchan_find_desc(&chan->vchan, cookie); vchan 687 drivers/staging/ralink-gdma/ralink-gdma.c spin_unlock_irqrestore(&chan->vchan.lock, flags); vchan 866 drivers/staging/ralink-gdma/ralink-gdma.c chan->vchan.desc_free = gdma_dma_desc_free; vchan 867 drivers/staging/ralink-gdma/ralink-gdma.c vchan_init(&chan->vchan, dd);