vdesc 98 drivers/dma/dma-axi-dmac.c struct virt_dma_desc vdesc; vdesc 152 drivers/dma/dma-axi-dmac.c static struct axi_dmac_desc *to_axi_dmac_desc(struct virt_dma_desc *vdesc) vdesc 154 drivers/dma/dma-axi-dmac.c return container_of(vdesc, struct axi_dmac_desc, vdesc); vdesc 197 drivers/dma/dma-axi-dmac.c struct virt_dma_desc *vdesc; vdesc 210 drivers/dma/dma-axi-dmac.c vdesc = vchan_next_desc(&chan->vchan); vdesc 211 drivers/dma/dma-axi-dmac.c if (!vdesc) vdesc 213 drivers/dma/dma-axi-dmac.c list_move_tail(&vdesc->node, &chan->active_descs); vdesc 214 drivers/dma/dma-axi-dmac.c desc = to_axi_dmac_desc(vdesc); vdesc 253 drivers/dma/dma-axi-dmac.c if (chan->hw_cyclic && desc->cyclic && !desc->vdesc.tx.callback && vdesc 269 drivers/dma/dma-axi-dmac.c struct axi_dmac_desc, vdesc.node); vdesc 294 drivers/dma/dma-axi-dmac.c list_for_each_entry(desc, &chan->active_descs, vdesc.node) { vdesc 330 drivers/dma/dma-axi-dmac.c struct dmaengine_result *rslt = &active->vdesc.tx_result; vdesc 381 drivers/dma/dma-axi-dmac.c vchan_cyclic_callback(&active->vdesc); vdesc 388 drivers/dma/dma-axi-dmac.c list_del(&active->vdesc.node); vdesc 389 drivers/dma/dma-axi-dmac.c vchan_cookie_complete(&active->vdesc); vdesc 564 drivers/dma/dma-axi-dmac.c return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); vdesc 598 drivers/dma/dma-axi-dmac.c return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); vdesc 669 drivers/dma/dma-axi-dmac.c return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); vdesc 677 drivers/dma/dma-axi-dmac.c static void axi_dmac_desc_free(struct virt_dma_desc *vdesc) vdesc 679 drivers/dma/dma-axi-dmac.c kfree(container_of(vdesc, struct axi_dmac_desc, vdesc)); vdesc 119 drivers/dma/dma-jz4780.c struct virt_dma_desc vdesc; vdesc 170 drivers/dma/dma-jz4780.c struct virt_dma_desc *vdesc) vdesc 172 drivers/dma/dma-jz4780.c return container_of(vdesc, struct jz4780_dma_desc, vdesc); vdesc 254 drivers/dma/dma-jz4780.c static void jz4780_dma_desc_free(struct virt_dma_desc *vdesc) vdesc 256 drivers/dma/dma-jz4780.c struct jz4780_dma_desc *desc = to_jz4780_dma_desc(vdesc); vdesc 257 drivers/dma/dma-jz4780.c struct jz4780_dma_chan *jzchan = to_jz4780_dma_chan(vdesc->tx.chan); vdesc 374 drivers/dma/dma-jz4780.c jz4780_dma_desc_free(&jzchan->desc->vdesc); vdesc 395 drivers/dma/dma-jz4780.c return vchan_tx_prep(&jzchan->vchan, &desc->vdesc, flags); vdesc 421 drivers/dma/dma-jz4780.c jz4780_dma_desc_free(&jzchan->desc->vdesc); vdesc 447 drivers/dma/dma-jz4780.c return vchan_tx_prep(&jzchan->vchan, &desc->vdesc, flags); vdesc 475 drivers/dma/dma-jz4780.c return vchan_tx_prep(&jzchan->vchan, &desc->vdesc, flags); vdesc 481 drivers/dma/dma-jz4780.c struct virt_dma_desc *vdesc; vdesc 486 drivers/dma/dma-jz4780.c vdesc = vchan_next_desc(&jzchan->vchan); vdesc 487 drivers/dma/dma-jz4780.c if (!vdesc) vdesc 490 drivers/dma/dma-jz4780.c list_del(&vdesc->node); vdesc 492 drivers/dma/dma-jz4780.c jzchan->desc = to_jz4780_dma_desc(vdesc); vdesc 495 drivers/dma/dma-jz4780.c if (jzchan->desc->type == DMA_CYCLIC && vdesc->tx.callback) { vdesc 578 drivers/dma/dma-jz4780.c vchan_terminate_vdesc(&jzchan->desc->vdesc); vdesc 637 drivers/dma/dma-jz4780.c struct virt_dma_desc *vdesc; vdesc 648 drivers/dma/dma-jz4780.c vdesc = vchan_find_desc(&jzchan->vchan, cookie); vdesc 649 drivers/dma/dma-jz4780.c if (vdesc) { vdesc 652 drivers/dma/dma-jz4780.c to_jz4780_dma_desc(vdesc), 0); vdesc 653 drivers/dma/dma-jz4780.c } else if (cookie == jzchan->desc->vdesc.tx.cookie) { vdesc 659 drivers/dma/dma-jz4780.c if (vdesc && jzchan->desc && vdesc == &jzchan->desc->vdesc vdesc 695 drivers/dma/dma-jz4780.c vchan_cyclic_callback(&jzchan->desc->vdesc); vdesc 701 drivers/dma/dma-jz4780.c vchan_cookie_complete(&desc->vdesc); vdesc 241 drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c static void vchan_desc_put(struct virt_dma_desc *vdesc) vdesc 243 drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c axi_desc_put(vd_to_axi_desc(vdesc)); vdesc 163 drivers/dma/dw-edma/dw-edma-core.c static void vchan_free_desc(struct virt_dma_desc *vdesc) vdesc 165 drivers/dma/dw-edma/dw-edma-core.c dw_edma_free_desc(vd2dw_edma_desc(vdesc)); vdesc 142 drivers/dma/fsl-edma-common.c void fsl_edma_free_desc(struct virt_dma_desc *vdesc) vdesc 147 drivers/dma/fsl-edma-common.c fsl_desc = to_fsl_edma_desc(vdesc); vdesc 266 drivers/dma/fsl-edma-common.c struct virt_dma_desc *vdesc, bool in_progress) vdesc 312 drivers/dma/fsl-edma-common.c struct virt_dma_desc *vdesc; vdesc 324 drivers/dma/fsl-edma-common.c vdesc = vchan_find_desc(&fsl_chan->vchan, cookie); vdesc 325 drivers/dma/fsl-edma-common.c if (fsl_chan->edesc && cookie == fsl_chan->edesc->vdesc.tx.cookie) vdesc 327 drivers/dma/fsl-edma-common.c fsl_edma_desc_residue(fsl_chan, vdesc, true); vdesc 328 drivers/dma/fsl-edma-common.c else if (vdesc) vdesc 330 drivers/dma/fsl-edma-common.c fsl_edma_desc_residue(fsl_chan, vdesc, false); vdesc 507 drivers/dma/fsl-edma-common.c return vchan_tx_prep(&fsl_chan->vchan, &fsl_desc->vdesc, flags); vdesc 579 drivers/dma/fsl-edma-common.c return vchan_tx_prep(&fsl_chan->vchan, &fsl_desc->vdesc, flags); vdesc 585 drivers/dma/fsl-edma-common.c struct virt_dma_desc *vdesc; vdesc 587 drivers/dma/fsl-edma-common.c vdesc = vchan_next_desc(&fsl_chan->vchan); vdesc 588 drivers/dma/fsl-edma-common.c if (!vdesc) vdesc 590 drivers/dma/fsl-edma-common.c fsl_chan->edesc = to_fsl_edma_desc(vdesc); vdesc 132 drivers/dma/fsl-edma-common.h struct virt_dma_desc vdesc; vdesc 220 drivers/dma/fsl-edma-common.h return container_of(vd, struct fsl_edma_desc, vdesc); vdesc 226 drivers/dma/fsl-edma-common.h void fsl_edma_free_desc(struct virt_dma_desc *vdesc); vdesc 49 drivers/dma/fsl-edma.c list_del(&fsl_chan->edesc->vdesc.node); vdesc 50 drivers/dma/fsl-edma.c vchan_cookie_complete(&fsl_chan->edesc->vdesc); vdesc 55 drivers/dma/fsl-edma.c vchan_cyclic_callback(&fsl_chan->edesc->vdesc); vdesc 166 drivers/dma/fsl-qdma.c struct virt_dma_desc vdesc; vdesc 193 drivers/dma/fsl-qdma.c struct virt_dma_desc vdesc; vdesc 289 drivers/dma/fsl-qdma.c return container_of(vd, struct fsl_qdma_comp, vdesc); vdesc 690 drivers/dma/fsl-qdma.c vchan_cookie_complete(&fsl_comp->vdesc); vdesc 934 drivers/dma/fsl-qdma.c return vchan_tx_prep(&fsl_chan->vchan, &fsl_comp->vdesc, flags); vdesc 940 drivers/dma/fsl-qdma.c struct virt_dma_desc *vdesc; vdesc 948 drivers/dma/fsl-qdma.c vdesc = vchan_next_desc(&fsl_chan->vchan); vdesc 949 drivers/dma/fsl-qdma.c if (!vdesc) vdesc 951 drivers/dma/fsl-qdma.c list_del(&vdesc->node); vdesc 952 drivers/dma/fsl-qdma.c fsl_comp = to_fsl_qdma_comp(vdesc); vdesc 967 drivers/dma/fsl-qdma.c static void fsl_qdma_free_desc(struct virt_dma_desc *vdesc) vdesc 973 drivers/dma/fsl-qdma.c fsl_comp = to_fsl_qdma_comp(vdesc); vdesc 110 drivers/dma/hsu/hsu.c struct virt_dma_desc *vdesc; vdesc 113 drivers/dma/hsu/hsu.c vdesc = vchan_next_desc(&hsuc->vchan); vdesc 114 drivers/dma/hsu/hsu.c if (!vdesc) { vdesc 119 drivers/dma/hsu/hsu.c list_del(&vdesc->node); vdesc 120 drivers/dma/hsu/hsu.c hsuc->desc = to_hsu_dma_desc(vdesc); vdesc 222 drivers/dma/hsu/hsu.c vchan_cookie_complete(&desc->vdesc); vdesc 250 drivers/dma/hsu/hsu.c static void hsu_dma_desc_free(struct virt_dma_desc *vdesc) vdesc 252 drivers/dma/hsu/hsu.c struct hsu_dma_desc *desc = to_hsu_dma_desc(vdesc); vdesc 284 drivers/dma/hsu/hsu.c return vchan_tx_prep(&hsuc->vchan, &desc->vdesc, flags); vdesc 319 drivers/dma/hsu/hsu.c struct virt_dma_desc *vdesc; vdesc 329 drivers/dma/hsu/hsu.c vdesc = vchan_find_desc(&hsuc->vchan, cookie); vdesc 330 drivers/dma/hsu/hsu.c if (hsuc->desc && cookie == hsuc->desc->vdesc.tx.cookie) { vdesc 334 drivers/dma/hsu/hsu.c } else if (vdesc) { vdesc 335 drivers/dma/hsu/hsu.c bytes = to_hsu_dma_desc(vdesc)->length; vdesc 393 drivers/dma/hsu/hsu.c hsu_dma_desc_free(&hsuc->desc->vdesc); vdesc 68 drivers/dma/hsu/hsu.h struct virt_dma_desc vdesc; vdesc 77 drivers/dma/hsu/hsu.h static inline struct hsu_dma_desc *to_hsu_dma_desc(struct virt_dma_desc *vdesc) vdesc 79 drivers/dma/hsu/hsu.h return container_of(vdesc, struct hsu_dma_desc, vdesc); vdesc 115 drivers/dma/idma64.c struct virt_dma_desc *vdesc; vdesc 118 drivers/dma/idma64.c vdesc = vchan_next_desc(&idma64c->vchan); vdesc 119 drivers/dma/idma64.c if (!vdesc) { vdesc 124 drivers/dma/idma64.c list_del(&vdesc->node); vdesc 125 drivers/dma/idma64.c idma64c->desc = to_idma64_desc(vdesc); vdesc 151 drivers/dma/idma64.c vchan_cookie_complete(&desc->vdesc); vdesc 222 drivers/dma/idma64.c static void idma64_vdesc_free(struct virt_dma_desc *vdesc) vdesc 224 drivers/dma/idma64.c struct idma64_chan *idma64c = to_idma64_chan(vdesc->tx.chan); vdesc 226 drivers/dma/idma64.c idma64_desc_free(idma64c, to_idma64_desc(vdesc)); vdesc 326 drivers/dma/idma64.c return vchan_tx_prep(&idma64c->vchan, &desc->vdesc, flags); vdesc 369 drivers/dma/idma64.c struct virt_dma_desc *vdesc; vdesc 379 drivers/dma/idma64.c vdesc = vchan_find_desc(&idma64c->vchan, cookie); vdesc 380 drivers/dma/idma64.c if (idma64c->desc && cookie == idma64c->desc->vdesc.tx.cookie) { vdesc 384 drivers/dma/idma64.c } else if (vdesc) { vdesc 385 drivers/dma/idma64.c bytes = to_idma64_desc(vdesc)->length; vdesc 480 drivers/dma/idma64.c idma64_vdesc_free(&idma64c->desc->vdesc); vdesc 114 drivers/dma/idma64.h struct virt_dma_desc vdesc; vdesc 122 drivers/dma/idma64.h static inline struct idma64_desc *to_idma64_desc(struct virt_dma_desc *vdesc) vdesc 124 drivers/dma/idma64.h return container_of(vdesc, struct idma64_desc, vdesc); vdesc 176 drivers/dma/img-mdc-dma.c struct virt_dma_desc *vdesc = container_of(t, struct virt_dma_desc, tx); vdesc 178 drivers/dma/img-mdc-dma.c return container_of(vdesc, struct mdc_tx_desc, vd); vdesc 39 drivers/dma/mcf-edma.c list_del(&mcf_chan->edesc->vdesc.node); vdesc 40 drivers/dma/mcf-edma.c vchan_cookie_complete(&mcf_chan->edesc->vdesc); vdesc 45 drivers/dma/mcf-edma.c vchan_cyclic_callback(&mcf_chan->edesc->vdesc); vdesc 28 drivers/dma/st_fdma.c return container_of(vd, struct st_fdma_desc, vdesc); vdesc 75 drivers/dma/st_fdma.c struct virt_dma_desc *vdesc; vdesc 78 drivers/dma/st_fdma.c vdesc = vchan_next_desc(&fchan->vchan); vdesc 79 drivers/dma/st_fdma.c if (!vdesc) vdesc 82 drivers/dma/st_fdma.c fchan->fdesc = to_st_fdma_desc(vdesc); vdesc 143 drivers/dma/st_fdma.c list_del(&fchan->fdesc->vdesc.node); vdesc 144 drivers/dma/st_fdma.c vchan_cookie_complete(&fchan->fdesc->vdesc); vdesc 148 drivers/dma/st_fdma.c vchan_cyclic_callback(&fchan->fdesc->vdesc); vdesc 224 drivers/dma/st_fdma.c static void st_fdma_free_desc(struct virt_dma_desc *vdesc) vdesc 229 drivers/dma/st_fdma.c fdesc = to_st_fdma_desc(vdesc); vdesc 342 drivers/dma/st_fdma.c return vchan_tx_prep(&fchan->vchan, &fdesc->vdesc, flags); vdesc 502 drivers/dma/st_fdma.c return vchan_tx_prep(&fchan->vchan, &fdesc->vdesc, flags); vdesc 551 drivers/dma/st_fdma.c return vchan_tx_prep(&fchan->vchan, &fdesc->vdesc, flags); vdesc 555 drivers/dma/st_fdma.c struct virt_dma_desc *vdesc, vdesc 594 drivers/dma/st_fdma.c if (fchan->fdesc && cookie == fchan->fdesc->vdesc.tx.cookie) vdesc 96 drivers/dma/st_fdma.h struct virt_dma_desc vdesc; vdesc 185 drivers/dma/stm32-dma.c struct virt_dma_desc vdesc; vdesc 226 drivers/dma/stm32-dma.c static struct stm32_dma_desc *to_stm32_dma_desc(struct virt_dma_desc *vdesc) vdesc 228 drivers/dma/stm32-dma.c return container_of(vdesc, struct stm32_dma_desc, vdesc); vdesc 533 drivers/dma/stm32-dma.c struct virt_dma_desc *vdesc; vdesc 544 drivers/dma/stm32-dma.c vdesc = vchan_next_desc(&chan->vchan); vdesc 545 drivers/dma/stm32-dma.c if (!vdesc) vdesc 548 drivers/dma/stm32-dma.c chan->desc = to_stm32_dma_desc(vdesc); vdesc 619 drivers/dma/stm32-dma.c vchan_cyclic_callback(&chan->desc->vdesc); vdesc 625 drivers/dma/stm32-dma.c list_del(&chan->desc->vdesc.node); vdesc 626 drivers/dma/stm32-dma.c vchan_cookie_complete(&chan->desc->vdesc); vdesc 886 drivers/dma/stm32-dma.c return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); vdesc 971 drivers/dma/stm32-dma.c return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); vdesc 1023 drivers/dma/stm32-dma.c return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); vdesc 1145 drivers/dma/stm32-dma.c struct virt_dma_desc *vdesc; vdesc 1155 drivers/dma/stm32-dma.c vdesc = vchan_find_desc(&chan->vchan, cookie); vdesc 1156 drivers/dma/stm32-dma.c if (chan->desc && cookie == chan->desc->vdesc.tx.cookie) vdesc 1159 drivers/dma/stm32-dma.c else if (vdesc) vdesc 1161 drivers/dma/stm32-dma.c to_stm32_dma_desc(vdesc), 0); vdesc 1208 drivers/dma/stm32-dma.c static void stm32_dma_desc_free(struct virt_dma_desc *vdesc) vdesc 1210 drivers/dma/stm32-dma.c kfree(container_of(vdesc, struct stm32_dma_desc, vdesc)); vdesc 251 drivers/dma/stm32-mdma.c struct virt_dma_desc vdesc; vdesc 296 drivers/dma/stm32-mdma.c static struct stm32_mdma_desc *to_stm32_mdma_desc(struct virt_dma_desc *vdesc) vdesc 298 drivers/dma/stm32-mdma.c return container_of(vdesc, struct stm32_mdma_desc, vdesc); vdesc 368 drivers/dma/stm32-mdma.c static void stm32_mdma_desc_free(struct virt_dma_desc *vdesc) vdesc 370 drivers/dma/stm32-mdma.c struct stm32_mdma_desc *desc = to_stm32_mdma_desc(vdesc); vdesc 371 drivers/dma/stm32-mdma.c struct stm32_mdma_chan *chan = to_stm32_mdma_chan(vdesc->tx.chan); vdesc 810 drivers/dma/stm32-mdma.c return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); vdesc 901 drivers/dma/stm32-mdma.c return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); vdesc 1087 drivers/dma/stm32-mdma.c return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); vdesc 1119 drivers/dma/stm32-mdma.c struct virt_dma_desc *vdesc; vdesc 1124 drivers/dma/stm32-mdma.c vdesc = vchan_next_desc(&chan->vchan); vdesc 1125 drivers/dma/stm32-mdma.c if (!vdesc) { vdesc 1130 drivers/dma/stm32-mdma.c chan->desc = to_stm32_mdma_desc(vdesc); vdesc 1307 drivers/dma/stm32-mdma.c struct virt_dma_desc *vdesc; vdesc 1318 drivers/dma/stm32-mdma.c vdesc = vchan_find_desc(&chan->vchan, cookie); vdesc 1319 drivers/dma/stm32-mdma.c if (chan->desc && cookie == chan->desc->vdesc.tx.cookie) vdesc 1322 drivers/dma/stm32-mdma.c else if (vdesc) vdesc 1324 drivers/dma/stm32-mdma.c to_stm32_mdma_desc(vdesc), 0); vdesc 1334 drivers/dma/stm32-mdma.c list_del(&chan->desc->vdesc.node); vdesc 1335 drivers/dma/stm32-mdma.c vchan_cookie_complete(&chan->desc->vdesc); vdesc 1414 drivers/dma/stm32-mdma.c vchan_cyclic_callback(&chan->desc->vdesc); vdesc 179 drivers/dma/ti/edma.c struct virt_dma_desc vdesc; vdesc 753 drivers/dma/ti/edma.c return container_of(tx, struct edma_desc, vdesc.tx); vdesc 756 drivers/dma/ti/edma.c static void edma_desc_free(struct virt_dma_desc *vdesc) vdesc 758 drivers/dma/ti/edma.c kfree(container_of(vdesc, struct edma_desc, vdesc)); vdesc 765 drivers/dma/ti/edma.c struct virt_dma_desc *vdesc; vdesc 772 drivers/dma/ti/edma.c vdesc = vchan_next_desc(&echan->vchan); vdesc 773 drivers/dma/ti/edma.c if (!vdesc) vdesc 775 drivers/dma/ti/edma.c list_del(&vdesc->node); vdesc 776 drivers/dma/ti/edma.c echan->edesc = to_edma_desc(&vdesc->tx); vdesc 874 drivers/dma/ti/edma.c vchan_terminate_vdesc(&echan->edesc->vdesc); vdesc 1146 drivers/dma/ti/edma.c return vchan_tx_prep(&echan->vchan, &edesc->vdesc, tx_flags); vdesc 1262 drivers/dma/ti/edma.c return vchan_tx_prep(&echan->vchan, &edesc->vdesc, tx_flags); vdesc 1419 drivers/dma/ti/edma.c return vchan_tx_prep(&echan->vchan, &edesc->vdesc, tx_flags); vdesc 1431 drivers/dma/ti/edma.c vchan_cyclic_callback(&edesc->vdesc); vdesc 1437 drivers/dma/ti/edma.c vchan_cookie_complete(&edesc->vdesc); vdesc 1844 drivers/dma/ti/edma.c if (echan->edesc && echan->edesc->vdesc.tx.cookie == cookie) { vdesc 1847 drivers/dma/ti/edma.c struct virt_dma_desc *vdesc = vchan_find_desc(&echan->vchan, vdesc 1850 drivers/dma/ti/edma.c if (vdesc) vdesc 1851 drivers/dma/ti/edma.c txstate->residue = to_edma_desc(&vdesc->tx)->residue; vdesc 1862 drivers/dma/ti/edma.c echan->edesc->vdesc.tx.cookie == cookie) { vdesc 1864 drivers/dma/ti/edma.c vchan_cookie_complete(&echan->edesc->vdesc); vdesc 122 drivers/sh/intc/virq.c struct irq_desc *vdesc = irq_to_desc(entry->irq); vdesc 124 drivers/sh/intc/virq.c if (vdesc) { vdesc 125 drivers/sh/intc/virq.c handle = (unsigned long)irq_desc_get_handler_data(vdesc); vdesc 128 drivers/sh/intc/virq.c generic_handle_irq_desc(vdesc); vdesc 138 drivers/staging/mt7621-dma/mtk-hsdma.c struct virt_dma_desc vdesc; vdesc 178 drivers/staging/mt7621-dma/mtk-hsdma.c struct virt_dma_desc *vdesc) vdesc 180 drivers/staging/mt7621-dma/mtk-hsdma.c return container_of(vdesc, struct mtk_hsdma_desc, vdesc); vdesc 388 drivers/staging/mt7621-dma/mtk-hsdma.c struct virt_dma_desc *vdesc; vdesc 390 drivers/staging/mt7621-dma/mtk-hsdma.c vdesc = vchan_next_desc(&chan->vchan); vdesc 391 drivers/staging/mt7621-dma/mtk-hsdma.c if (!vdesc) { vdesc 395 drivers/staging/mt7621-dma/mtk-hsdma.c chan->desc = to_mtk_hsdma_desc(vdesc); vdesc 412 drivers/staging/mt7621-dma/mtk-hsdma.c list_del(&desc->vdesc.node); vdesc 413 drivers/staging/mt7621-dma/mtk-hsdma.c vchan_cookie_complete(&desc->vdesc); vdesc 481 drivers/staging/mt7621-dma/mtk-hsdma.c return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); vdesc 496 drivers/staging/mt7621-dma/mtk-hsdma.c static void mtk_hsdma_desc_free(struct virt_dma_desc *vdesc) vdesc 498 drivers/staging/mt7621-dma/mtk-hsdma.c kfree(container_of(vdesc, struct mtk_hsdma_desc, vdesc)); vdesc 97 drivers/staging/ralink-gdma/ralink-gdma.c struct virt_dma_desc vdesc; vdesc 150 drivers/staging/ralink-gdma/ralink-gdma.c static struct gdma_dma_desc *to_gdma_dma_desc(struct virt_dma_desc *vdesc) vdesc 152 drivers/staging/ralink-gdma/ralink-gdma.c return container_of(vdesc, struct gdma_dma_desc, vdesc); vdesc 415 drivers/staging/ralink-gdma/ralink-gdma.c struct virt_dma_desc *vdesc; vdesc 417 drivers/staging/ralink-gdma/ralink-gdma.c vdesc = vchan_next_desc(&chan->vchan); vdesc 418 drivers/staging/ralink-gdma/ralink-gdma.c if (!vdesc) { vdesc 422 drivers/staging/ralink-gdma/ralink-gdma.c chan->desc = to_gdma_dma_desc(vdesc); vdesc 440 drivers/staging/ralink-gdma/ralink-gdma.c vchan_cyclic_callback(&desc->vdesc); vdesc 447 drivers/staging/ralink-gdma/ralink-gdma.c list_del(&desc->vdesc.node); vdesc 448 drivers/staging/ralink-gdma/ralink-gdma.c vchan_cookie_complete(&desc->vdesc); vdesc 554 drivers/staging/ralink-gdma/ralink-gdma.c return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); vdesc 601 drivers/staging/ralink-gdma/ralink-gdma.c return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); vdesc 648 drivers/staging/ralink-gdma/ralink-gdma.c return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); vdesc 660 drivers/staging/ralink-gdma/ralink-gdma.c struct virt_dma_desc *vdesc; vdesc 671 drivers/staging/ralink-gdma/ralink-gdma.c if (desc && (cookie == desc->vdesc.tx.cookie)) { vdesc 683 drivers/staging/ralink-gdma/ralink-gdma.c vdesc = vchan_find_desc(&chan->vchan, cookie); vdesc 684 drivers/staging/ralink-gdma/ralink-gdma.c if (vdesc) vdesc 685 drivers/staging/ralink-gdma/ralink-gdma.c state->residue = to_gdma_dma_desc(vdesc)->residue; vdesc 699 drivers/staging/ralink-gdma/ralink-gdma.c static void gdma_dma_desc_free(struct virt_dma_desc *vdesc) vdesc 701 drivers/staging/ralink-gdma/ralink-gdma.c kfree(container_of(vdesc, struct gdma_dma_desc, vdesc)); vdesc 556 fs/nilfs2/ioctl.c struct nilfs_vdesc *vdesc, vdesc 562 fs/nilfs2/ioctl.c if (vdesc->vd_flags == 0) vdesc 564 fs/nilfs2/ioctl.c inode, vdesc->vd_offset, vdesc->vd_blocknr, vdesc 565 fs/nilfs2/ioctl.c vdesc->vd_vblocknr, &bh); vdesc 568 fs/nilfs2/ioctl.c inode, vdesc->vd_blocknr, vdesc->vd_vblocknr, &bh); vdesc 574 fs/nilfs2/ioctl.c __func__, vdesc->vd_flags ? "node" : "data", vdesc 575 fs/nilfs2/ioctl.c (unsigned long long)vdesc->vd_ino, vdesc 576 fs/nilfs2/ioctl.c (unsigned long long)vdesc->vd_cno, vdesc 577 fs/nilfs2/ioctl.c (unsigned long long)vdesc->vd_offset, vdesc 578 fs/nilfs2/ioctl.c (unsigned long long)vdesc->vd_blocknr, vdesc 579 fs/nilfs2/ioctl.c (unsigned long long)vdesc->vd_vblocknr); vdesc 585 fs/nilfs2/ioctl.c __func__, vdesc->vd_flags ? "node" : "data", vdesc 586 fs/nilfs2/ioctl.c (unsigned long long)vdesc->vd_ino, vdesc 587 fs/nilfs2/ioctl.c (unsigned long long)vdesc->vd_cno, vdesc 588 fs/nilfs2/ioctl.c (unsigned long long)vdesc->vd_offset, vdesc 589 fs/nilfs2/ioctl.c (unsigned long long)vdesc->vd_blocknr, vdesc 590 fs/nilfs2/ioctl.c (unsigned long long)vdesc->vd_vblocknr); vdesc 617 fs/nilfs2/ioctl.c struct nilfs_vdesc *vdesc; vdesc 624 fs/nilfs2/ioctl.c for (i = 0, vdesc = buf; i < nmembs; ) { vdesc 625 fs/nilfs2/ioctl.c ino = vdesc->vd_ino; vdesc 626 fs/nilfs2/ioctl.c cno = vdesc->vd_cno; vdesc 644 fs/nilfs2/ioctl.c ret = nilfs_ioctl_move_inode_block(inode, vdesc, vdesc 650 fs/nilfs2/ioctl.c vdesc++; vdesc 652 fs/nilfs2/ioctl.c vdesc->vd_ino == ino && vdesc->vd_cno == cno);