vdesc              98 drivers/dma/dma-axi-dmac.c 	struct virt_dma_desc vdesc;
vdesc             152 drivers/dma/dma-axi-dmac.c static struct axi_dmac_desc *to_axi_dmac_desc(struct virt_dma_desc *vdesc)
vdesc             154 drivers/dma/dma-axi-dmac.c 	return container_of(vdesc, struct axi_dmac_desc, vdesc);
vdesc             197 drivers/dma/dma-axi-dmac.c 	struct virt_dma_desc *vdesc;
vdesc             210 drivers/dma/dma-axi-dmac.c 		vdesc = vchan_next_desc(&chan->vchan);
vdesc             211 drivers/dma/dma-axi-dmac.c 		if (!vdesc)
vdesc             213 drivers/dma/dma-axi-dmac.c 		list_move_tail(&vdesc->node, &chan->active_descs);
vdesc             214 drivers/dma/dma-axi-dmac.c 		desc = to_axi_dmac_desc(vdesc);
vdesc             253 drivers/dma/dma-axi-dmac.c 	if (chan->hw_cyclic && desc->cyclic && !desc->vdesc.tx.callback &&
vdesc             269 drivers/dma/dma-axi-dmac.c 		struct axi_dmac_desc, vdesc.node);
vdesc             294 drivers/dma/dma-axi-dmac.c 		list_for_each_entry(desc, &chan->active_descs, vdesc.node) {
vdesc             330 drivers/dma/dma-axi-dmac.c 	struct dmaengine_result *rslt = &active->vdesc.tx_result;
vdesc             381 drivers/dma/dma-axi-dmac.c 			vchan_cyclic_callback(&active->vdesc);
vdesc             388 drivers/dma/dma-axi-dmac.c 				list_del(&active->vdesc.node);
vdesc             389 drivers/dma/dma-axi-dmac.c 				vchan_cookie_complete(&active->vdesc);
vdesc             564 drivers/dma/dma-axi-dmac.c 	return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags);
vdesc             598 drivers/dma/dma-axi-dmac.c 	return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags);
vdesc             669 drivers/dma/dma-axi-dmac.c 	return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags);
vdesc             677 drivers/dma/dma-axi-dmac.c static void axi_dmac_desc_free(struct virt_dma_desc *vdesc)
vdesc             679 drivers/dma/dma-axi-dmac.c 	kfree(container_of(vdesc, struct axi_dmac_desc, vdesc));
vdesc             119 drivers/dma/dma-jz4780.c 	struct virt_dma_desc vdesc;
vdesc             170 drivers/dma/dma-jz4780.c 	struct virt_dma_desc *vdesc)
vdesc             172 drivers/dma/dma-jz4780.c 	return container_of(vdesc, struct jz4780_dma_desc, vdesc);
vdesc             254 drivers/dma/dma-jz4780.c static void jz4780_dma_desc_free(struct virt_dma_desc *vdesc)
vdesc             256 drivers/dma/dma-jz4780.c 	struct jz4780_dma_desc *desc = to_jz4780_dma_desc(vdesc);
vdesc             257 drivers/dma/dma-jz4780.c 	struct jz4780_dma_chan *jzchan = to_jz4780_dma_chan(vdesc->tx.chan);
vdesc             374 drivers/dma/dma-jz4780.c 			jz4780_dma_desc_free(&jzchan->desc->vdesc);
vdesc             395 drivers/dma/dma-jz4780.c 	return vchan_tx_prep(&jzchan->vchan, &desc->vdesc, flags);
vdesc             421 drivers/dma/dma-jz4780.c 			jz4780_dma_desc_free(&jzchan->desc->vdesc);
vdesc             447 drivers/dma/dma-jz4780.c 	return vchan_tx_prep(&jzchan->vchan, &desc->vdesc, flags);
vdesc             475 drivers/dma/dma-jz4780.c 	return vchan_tx_prep(&jzchan->vchan, &desc->vdesc, flags);
vdesc             481 drivers/dma/dma-jz4780.c 	struct virt_dma_desc *vdesc;
vdesc             486 drivers/dma/dma-jz4780.c 		vdesc = vchan_next_desc(&jzchan->vchan);
vdesc             487 drivers/dma/dma-jz4780.c 		if (!vdesc)
vdesc             490 drivers/dma/dma-jz4780.c 		list_del(&vdesc->node);
vdesc             492 drivers/dma/dma-jz4780.c 		jzchan->desc = to_jz4780_dma_desc(vdesc);
vdesc             495 drivers/dma/dma-jz4780.c 		if (jzchan->desc->type == DMA_CYCLIC && vdesc->tx.callback) {
vdesc             578 drivers/dma/dma-jz4780.c 		vchan_terminate_vdesc(&jzchan->desc->vdesc);
vdesc             637 drivers/dma/dma-jz4780.c 	struct virt_dma_desc *vdesc;
vdesc             648 drivers/dma/dma-jz4780.c 	vdesc = vchan_find_desc(&jzchan->vchan, cookie);
vdesc             649 drivers/dma/dma-jz4780.c 	if (vdesc) {
vdesc             652 drivers/dma/dma-jz4780.c 					to_jz4780_dma_desc(vdesc), 0);
vdesc             653 drivers/dma/dma-jz4780.c 	} else if (cookie == jzchan->desc->vdesc.tx.cookie) {
vdesc             659 drivers/dma/dma-jz4780.c 	if (vdesc && jzchan->desc && vdesc == &jzchan->desc->vdesc
vdesc             695 drivers/dma/dma-jz4780.c 				vchan_cyclic_callback(&jzchan->desc->vdesc);
vdesc             701 drivers/dma/dma-jz4780.c 					vchan_cookie_complete(&desc->vdesc);
vdesc             241 drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c static void vchan_desc_put(struct virt_dma_desc *vdesc)
vdesc             243 drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c 	axi_desc_put(vd_to_axi_desc(vdesc));
vdesc             163 drivers/dma/dw-edma/dw-edma-core.c static void vchan_free_desc(struct virt_dma_desc *vdesc)
vdesc             165 drivers/dma/dw-edma/dw-edma-core.c 	dw_edma_free_desc(vd2dw_edma_desc(vdesc));
vdesc             142 drivers/dma/fsl-edma-common.c void fsl_edma_free_desc(struct virt_dma_desc *vdesc)
vdesc             147 drivers/dma/fsl-edma-common.c 	fsl_desc = to_fsl_edma_desc(vdesc);
vdesc             266 drivers/dma/fsl-edma-common.c 		struct virt_dma_desc *vdesc, bool in_progress)
vdesc             312 drivers/dma/fsl-edma-common.c 	struct virt_dma_desc *vdesc;
vdesc             324 drivers/dma/fsl-edma-common.c 	vdesc = vchan_find_desc(&fsl_chan->vchan, cookie);
vdesc             325 drivers/dma/fsl-edma-common.c 	if (fsl_chan->edesc && cookie == fsl_chan->edesc->vdesc.tx.cookie)
vdesc             327 drivers/dma/fsl-edma-common.c 			fsl_edma_desc_residue(fsl_chan, vdesc, true);
vdesc             328 drivers/dma/fsl-edma-common.c 	else if (vdesc)
vdesc             330 drivers/dma/fsl-edma-common.c 			fsl_edma_desc_residue(fsl_chan, vdesc, false);
vdesc             507 drivers/dma/fsl-edma-common.c 	return vchan_tx_prep(&fsl_chan->vchan, &fsl_desc->vdesc, flags);
vdesc             579 drivers/dma/fsl-edma-common.c 	return vchan_tx_prep(&fsl_chan->vchan, &fsl_desc->vdesc, flags);
vdesc             585 drivers/dma/fsl-edma-common.c 	struct virt_dma_desc *vdesc;
vdesc             587 drivers/dma/fsl-edma-common.c 	vdesc = vchan_next_desc(&fsl_chan->vchan);
vdesc             588 drivers/dma/fsl-edma-common.c 	if (!vdesc)
vdesc             590 drivers/dma/fsl-edma-common.c 	fsl_chan->edesc = to_fsl_edma_desc(vdesc);
vdesc             132 drivers/dma/fsl-edma-common.h 	struct virt_dma_desc		vdesc;
vdesc             220 drivers/dma/fsl-edma-common.h 	return container_of(vd, struct fsl_edma_desc, vdesc);
vdesc             226 drivers/dma/fsl-edma-common.h void fsl_edma_free_desc(struct virt_dma_desc *vdesc);
vdesc              49 drivers/dma/fsl-edma.c 				list_del(&fsl_chan->edesc->vdesc.node);
vdesc              50 drivers/dma/fsl-edma.c 				vchan_cookie_complete(&fsl_chan->edesc->vdesc);
vdesc              55 drivers/dma/fsl-edma.c 				vchan_cyclic_callback(&fsl_chan->edesc->vdesc);
vdesc             166 drivers/dma/fsl-qdma.c 	struct virt_dma_desc		vdesc;
vdesc             193 drivers/dma/fsl-qdma.c 	struct virt_dma_desc    vdesc;
vdesc             289 drivers/dma/fsl-qdma.c 	return container_of(vd, struct fsl_qdma_comp, vdesc);
vdesc             690 drivers/dma/fsl-qdma.c 		vchan_cookie_complete(&fsl_comp->vdesc);
vdesc             934 drivers/dma/fsl-qdma.c 	return vchan_tx_prep(&fsl_chan->vchan, &fsl_comp->vdesc, flags);
vdesc             940 drivers/dma/fsl-qdma.c 	struct virt_dma_desc *vdesc;
vdesc             948 drivers/dma/fsl-qdma.c 	vdesc = vchan_next_desc(&fsl_chan->vchan);
vdesc             949 drivers/dma/fsl-qdma.c 	if (!vdesc)
vdesc             951 drivers/dma/fsl-qdma.c 	list_del(&vdesc->node);
vdesc             952 drivers/dma/fsl-qdma.c 	fsl_comp = to_fsl_qdma_comp(vdesc);
vdesc             967 drivers/dma/fsl-qdma.c static void fsl_qdma_free_desc(struct virt_dma_desc *vdesc)
vdesc             973 drivers/dma/fsl-qdma.c 	fsl_comp = to_fsl_qdma_comp(vdesc);
vdesc             110 drivers/dma/hsu/hsu.c 	struct virt_dma_desc *vdesc;
vdesc             113 drivers/dma/hsu/hsu.c 	vdesc = vchan_next_desc(&hsuc->vchan);
vdesc             114 drivers/dma/hsu/hsu.c 	if (!vdesc) {
vdesc             119 drivers/dma/hsu/hsu.c 	list_del(&vdesc->node);
vdesc             120 drivers/dma/hsu/hsu.c 	hsuc->desc = to_hsu_dma_desc(vdesc);
vdesc             222 drivers/dma/hsu/hsu.c 			vchan_cookie_complete(&desc->vdesc);
vdesc             250 drivers/dma/hsu/hsu.c static void hsu_dma_desc_free(struct virt_dma_desc *vdesc)
vdesc             252 drivers/dma/hsu/hsu.c 	struct hsu_dma_desc *desc = to_hsu_dma_desc(vdesc);
vdesc             284 drivers/dma/hsu/hsu.c 	return vchan_tx_prep(&hsuc->vchan, &desc->vdesc, flags);
vdesc             319 drivers/dma/hsu/hsu.c 	struct virt_dma_desc *vdesc;
vdesc             329 drivers/dma/hsu/hsu.c 	vdesc = vchan_find_desc(&hsuc->vchan, cookie);
vdesc             330 drivers/dma/hsu/hsu.c 	if (hsuc->desc && cookie == hsuc->desc->vdesc.tx.cookie) {
vdesc             334 drivers/dma/hsu/hsu.c 	} else if (vdesc) {
vdesc             335 drivers/dma/hsu/hsu.c 		bytes = to_hsu_dma_desc(vdesc)->length;
vdesc             393 drivers/dma/hsu/hsu.c 		hsu_dma_desc_free(&hsuc->desc->vdesc);
vdesc              68 drivers/dma/hsu/hsu.h 	struct virt_dma_desc vdesc;
vdesc              77 drivers/dma/hsu/hsu.h static inline struct hsu_dma_desc *to_hsu_dma_desc(struct virt_dma_desc *vdesc)
vdesc              79 drivers/dma/hsu/hsu.h 	return container_of(vdesc, struct hsu_dma_desc, vdesc);
vdesc             115 drivers/dma/idma64.c 	struct virt_dma_desc *vdesc;
vdesc             118 drivers/dma/idma64.c 	vdesc = vchan_next_desc(&idma64c->vchan);
vdesc             119 drivers/dma/idma64.c 	if (!vdesc) {
vdesc             124 drivers/dma/idma64.c 	list_del(&vdesc->node);
vdesc             125 drivers/dma/idma64.c 	idma64c->desc = to_idma64_desc(vdesc);
vdesc             151 drivers/dma/idma64.c 			vchan_cookie_complete(&desc->vdesc);
vdesc             222 drivers/dma/idma64.c static void idma64_vdesc_free(struct virt_dma_desc *vdesc)
vdesc             224 drivers/dma/idma64.c 	struct idma64_chan *idma64c = to_idma64_chan(vdesc->tx.chan);
vdesc             226 drivers/dma/idma64.c 	idma64_desc_free(idma64c, to_idma64_desc(vdesc));
vdesc             326 drivers/dma/idma64.c 	return vchan_tx_prep(&idma64c->vchan, &desc->vdesc, flags);
vdesc             369 drivers/dma/idma64.c 	struct virt_dma_desc *vdesc;
vdesc             379 drivers/dma/idma64.c 	vdesc = vchan_find_desc(&idma64c->vchan, cookie);
vdesc             380 drivers/dma/idma64.c 	if (idma64c->desc && cookie == idma64c->desc->vdesc.tx.cookie) {
vdesc             384 drivers/dma/idma64.c 	} else if (vdesc) {
vdesc             385 drivers/dma/idma64.c 		bytes = to_idma64_desc(vdesc)->length;
vdesc             480 drivers/dma/idma64.c 		idma64_vdesc_free(&idma64c->desc->vdesc);
vdesc             114 drivers/dma/idma64.h 	struct virt_dma_desc vdesc;
vdesc             122 drivers/dma/idma64.h static inline struct idma64_desc *to_idma64_desc(struct virt_dma_desc *vdesc)
vdesc             124 drivers/dma/idma64.h 	return container_of(vdesc, struct idma64_desc, vdesc);
vdesc             176 drivers/dma/img-mdc-dma.c 	struct virt_dma_desc *vdesc = container_of(t, struct virt_dma_desc, tx);
vdesc             178 drivers/dma/img-mdc-dma.c 	return container_of(vdesc, struct mdc_tx_desc, vd);
vdesc              39 drivers/dma/mcf-edma.c 				list_del(&mcf_chan->edesc->vdesc.node);
vdesc              40 drivers/dma/mcf-edma.c 				vchan_cookie_complete(&mcf_chan->edesc->vdesc);
vdesc              45 drivers/dma/mcf-edma.c 				vchan_cyclic_callback(&mcf_chan->edesc->vdesc);
vdesc              28 drivers/dma/st_fdma.c 	return container_of(vd, struct st_fdma_desc, vdesc);
vdesc              75 drivers/dma/st_fdma.c 	struct virt_dma_desc *vdesc;
vdesc              78 drivers/dma/st_fdma.c 	vdesc = vchan_next_desc(&fchan->vchan);
vdesc              79 drivers/dma/st_fdma.c 	if (!vdesc)
vdesc              82 drivers/dma/st_fdma.c 	fchan->fdesc = to_st_fdma_desc(vdesc);
vdesc             143 drivers/dma/st_fdma.c 				list_del(&fchan->fdesc->vdesc.node);
vdesc             144 drivers/dma/st_fdma.c 				vchan_cookie_complete(&fchan->fdesc->vdesc);
vdesc             148 drivers/dma/st_fdma.c 				vchan_cyclic_callback(&fchan->fdesc->vdesc);
vdesc             224 drivers/dma/st_fdma.c static void st_fdma_free_desc(struct virt_dma_desc *vdesc)
vdesc             229 drivers/dma/st_fdma.c 	fdesc = to_st_fdma_desc(vdesc);
vdesc             342 drivers/dma/st_fdma.c 	return vchan_tx_prep(&fchan->vchan, &fdesc->vdesc, flags);
vdesc             502 drivers/dma/st_fdma.c 	return vchan_tx_prep(&fchan->vchan, &fdesc->vdesc, flags);
vdesc             551 drivers/dma/st_fdma.c 	return vchan_tx_prep(&fchan->vchan, &fdesc->vdesc, flags);
vdesc             555 drivers/dma/st_fdma.c 				   struct virt_dma_desc *vdesc,
vdesc             594 drivers/dma/st_fdma.c 	if (fchan->fdesc && cookie == fchan->fdesc->vdesc.tx.cookie)
vdesc              96 drivers/dma/st_fdma.h 	struct virt_dma_desc vdesc;
vdesc             185 drivers/dma/stm32-dma.c 	struct virt_dma_desc vdesc;
vdesc             226 drivers/dma/stm32-dma.c static struct stm32_dma_desc *to_stm32_dma_desc(struct virt_dma_desc *vdesc)
vdesc             228 drivers/dma/stm32-dma.c 	return container_of(vdesc, struct stm32_dma_desc, vdesc);
vdesc             533 drivers/dma/stm32-dma.c 	struct virt_dma_desc *vdesc;
vdesc             544 drivers/dma/stm32-dma.c 		vdesc = vchan_next_desc(&chan->vchan);
vdesc             545 drivers/dma/stm32-dma.c 		if (!vdesc)
vdesc             548 drivers/dma/stm32-dma.c 		chan->desc = to_stm32_dma_desc(vdesc);
vdesc             619 drivers/dma/stm32-dma.c 			vchan_cyclic_callback(&chan->desc->vdesc);
vdesc             625 drivers/dma/stm32-dma.c 				list_del(&chan->desc->vdesc.node);
vdesc             626 drivers/dma/stm32-dma.c 				vchan_cookie_complete(&chan->desc->vdesc);
vdesc             886 drivers/dma/stm32-dma.c 	return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags);
vdesc             971 drivers/dma/stm32-dma.c 	return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags);
vdesc            1023 drivers/dma/stm32-dma.c 	return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags);
vdesc            1145 drivers/dma/stm32-dma.c 	struct virt_dma_desc *vdesc;
vdesc            1155 drivers/dma/stm32-dma.c 	vdesc = vchan_find_desc(&chan->vchan, cookie);
vdesc            1156 drivers/dma/stm32-dma.c 	if (chan->desc && cookie == chan->desc->vdesc.tx.cookie)
vdesc            1159 drivers/dma/stm32-dma.c 	else if (vdesc)
vdesc            1161 drivers/dma/stm32-dma.c 						 to_stm32_dma_desc(vdesc), 0);
vdesc            1208 drivers/dma/stm32-dma.c static void stm32_dma_desc_free(struct virt_dma_desc *vdesc)
vdesc            1210 drivers/dma/stm32-dma.c 	kfree(container_of(vdesc, struct stm32_dma_desc, vdesc));
vdesc             251 drivers/dma/stm32-mdma.c 	struct virt_dma_desc vdesc;
vdesc             296 drivers/dma/stm32-mdma.c static struct stm32_mdma_desc *to_stm32_mdma_desc(struct virt_dma_desc *vdesc)
vdesc             298 drivers/dma/stm32-mdma.c 	return container_of(vdesc, struct stm32_mdma_desc, vdesc);
vdesc             368 drivers/dma/stm32-mdma.c static void stm32_mdma_desc_free(struct virt_dma_desc *vdesc)
vdesc             370 drivers/dma/stm32-mdma.c 	struct stm32_mdma_desc *desc = to_stm32_mdma_desc(vdesc);
vdesc             371 drivers/dma/stm32-mdma.c 	struct stm32_mdma_chan *chan = to_stm32_mdma_chan(vdesc->tx.chan);
vdesc             810 drivers/dma/stm32-mdma.c 	return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags);
vdesc             901 drivers/dma/stm32-mdma.c 	return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags);
vdesc            1087 drivers/dma/stm32-mdma.c 	return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags);
vdesc            1119 drivers/dma/stm32-mdma.c 	struct virt_dma_desc *vdesc;
vdesc            1124 drivers/dma/stm32-mdma.c 	vdesc = vchan_next_desc(&chan->vchan);
vdesc            1125 drivers/dma/stm32-mdma.c 	if (!vdesc) {
vdesc            1130 drivers/dma/stm32-mdma.c 	chan->desc = to_stm32_mdma_desc(vdesc);
vdesc            1307 drivers/dma/stm32-mdma.c 	struct virt_dma_desc *vdesc;
vdesc            1318 drivers/dma/stm32-mdma.c 	vdesc = vchan_find_desc(&chan->vchan, cookie);
vdesc            1319 drivers/dma/stm32-mdma.c 	if (chan->desc && cookie == chan->desc->vdesc.tx.cookie)
vdesc            1322 drivers/dma/stm32-mdma.c 	else if (vdesc)
vdesc            1324 drivers/dma/stm32-mdma.c 						  to_stm32_mdma_desc(vdesc), 0);
vdesc            1334 drivers/dma/stm32-mdma.c 	list_del(&chan->desc->vdesc.node);
vdesc            1335 drivers/dma/stm32-mdma.c 	vchan_cookie_complete(&chan->desc->vdesc);
vdesc            1414 drivers/dma/stm32-mdma.c 			vchan_cyclic_callback(&chan->desc->vdesc);
vdesc             179 drivers/dma/ti/edma.c 	struct virt_dma_desc		vdesc;
vdesc             753 drivers/dma/ti/edma.c 	return container_of(tx, struct edma_desc, vdesc.tx);
vdesc             756 drivers/dma/ti/edma.c static void edma_desc_free(struct virt_dma_desc *vdesc)
vdesc             758 drivers/dma/ti/edma.c 	kfree(container_of(vdesc, struct edma_desc, vdesc));
vdesc             765 drivers/dma/ti/edma.c 	struct virt_dma_desc *vdesc;
vdesc             772 drivers/dma/ti/edma.c 		vdesc = vchan_next_desc(&echan->vchan);
vdesc             773 drivers/dma/ti/edma.c 		if (!vdesc)
vdesc             775 drivers/dma/ti/edma.c 		list_del(&vdesc->node);
vdesc             776 drivers/dma/ti/edma.c 		echan->edesc = to_edma_desc(&vdesc->tx);
vdesc             874 drivers/dma/ti/edma.c 		vchan_terminate_vdesc(&echan->edesc->vdesc);
vdesc            1146 drivers/dma/ti/edma.c 	return vchan_tx_prep(&echan->vchan, &edesc->vdesc, tx_flags);
vdesc            1262 drivers/dma/ti/edma.c 	return vchan_tx_prep(&echan->vchan, &edesc->vdesc, tx_flags);
vdesc            1419 drivers/dma/ti/edma.c 	return vchan_tx_prep(&echan->vchan, &edesc->vdesc, tx_flags);
vdesc            1431 drivers/dma/ti/edma.c 			vchan_cyclic_callback(&edesc->vdesc);
vdesc            1437 drivers/dma/ti/edma.c 			vchan_cookie_complete(&edesc->vdesc);
vdesc            1844 drivers/dma/ti/edma.c 	if (echan->edesc && echan->edesc->vdesc.tx.cookie == cookie) {
vdesc            1847 drivers/dma/ti/edma.c 		struct virt_dma_desc *vdesc = vchan_find_desc(&echan->vchan,
vdesc            1850 drivers/dma/ti/edma.c 		if (vdesc)
vdesc            1851 drivers/dma/ti/edma.c 			txstate->residue = to_edma_desc(&vdesc->tx)->residue;
vdesc            1862 drivers/dma/ti/edma.c 	    echan->edesc->vdesc.tx.cookie == cookie) {
vdesc            1864 drivers/dma/ti/edma.c 		vchan_cookie_complete(&echan->edesc->vdesc);
vdesc             122 drivers/sh/intc/virq.c 		struct irq_desc *vdesc = irq_to_desc(entry->irq);
vdesc             124 drivers/sh/intc/virq.c 		if (vdesc) {
vdesc             125 drivers/sh/intc/virq.c 			handle = (unsigned long)irq_desc_get_handler_data(vdesc);
vdesc             128 drivers/sh/intc/virq.c 				generic_handle_irq_desc(vdesc);
vdesc             138 drivers/staging/mt7621-dma/mtk-hsdma.c 	struct virt_dma_desc vdesc;
vdesc             178 drivers/staging/mt7621-dma/mtk-hsdma.c 		struct virt_dma_desc *vdesc)
vdesc             180 drivers/staging/mt7621-dma/mtk-hsdma.c 	return container_of(vdesc, struct mtk_hsdma_desc, vdesc);
vdesc             388 drivers/staging/mt7621-dma/mtk-hsdma.c 	struct virt_dma_desc *vdesc;
vdesc             390 drivers/staging/mt7621-dma/mtk-hsdma.c 	vdesc = vchan_next_desc(&chan->vchan);
vdesc             391 drivers/staging/mt7621-dma/mtk-hsdma.c 	if (!vdesc) {
vdesc             395 drivers/staging/mt7621-dma/mtk-hsdma.c 	chan->desc = to_mtk_hsdma_desc(vdesc);
vdesc             412 drivers/staging/mt7621-dma/mtk-hsdma.c 			list_del(&desc->vdesc.node);
vdesc             413 drivers/staging/mt7621-dma/mtk-hsdma.c 			vchan_cookie_complete(&desc->vdesc);
vdesc             481 drivers/staging/mt7621-dma/mtk-hsdma.c 	return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags);
vdesc             496 drivers/staging/mt7621-dma/mtk-hsdma.c static void mtk_hsdma_desc_free(struct virt_dma_desc *vdesc)
vdesc             498 drivers/staging/mt7621-dma/mtk-hsdma.c 	kfree(container_of(vdesc, struct mtk_hsdma_desc, vdesc));
vdesc              97 drivers/staging/ralink-gdma/ralink-gdma.c 	struct virt_dma_desc vdesc;
vdesc             150 drivers/staging/ralink-gdma/ralink-gdma.c static struct gdma_dma_desc *to_gdma_dma_desc(struct virt_dma_desc *vdesc)
vdesc             152 drivers/staging/ralink-gdma/ralink-gdma.c 	return container_of(vdesc, struct gdma_dma_desc, vdesc);
vdesc             415 drivers/staging/ralink-gdma/ralink-gdma.c 	struct virt_dma_desc *vdesc;
vdesc             417 drivers/staging/ralink-gdma/ralink-gdma.c 	vdesc = vchan_next_desc(&chan->vchan);
vdesc             418 drivers/staging/ralink-gdma/ralink-gdma.c 	if (!vdesc) {
vdesc             422 drivers/staging/ralink-gdma/ralink-gdma.c 	chan->desc = to_gdma_dma_desc(vdesc);
vdesc             440 drivers/staging/ralink-gdma/ralink-gdma.c 			vchan_cyclic_callback(&desc->vdesc);
vdesc             447 drivers/staging/ralink-gdma/ralink-gdma.c 				list_del(&desc->vdesc.node);
vdesc             448 drivers/staging/ralink-gdma/ralink-gdma.c 				vchan_cookie_complete(&desc->vdesc);
vdesc             554 drivers/staging/ralink-gdma/ralink-gdma.c 	return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags);
vdesc             601 drivers/staging/ralink-gdma/ralink-gdma.c 	return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags);
vdesc             648 drivers/staging/ralink-gdma/ralink-gdma.c 	return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags);
vdesc             660 drivers/staging/ralink-gdma/ralink-gdma.c 	struct virt_dma_desc *vdesc;
vdesc             671 drivers/staging/ralink-gdma/ralink-gdma.c 	if (desc && (cookie == desc->vdesc.tx.cookie)) {
vdesc             683 drivers/staging/ralink-gdma/ralink-gdma.c 		vdesc = vchan_find_desc(&chan->vchan, cookie);
vdesc             684 drivers/staging/ralink-gdma/ralink-gdma.c 		if (vdesc)
vdesc             685 drivers/staging/ralink-gdma/ralink-gdma.c 			state->residue = to_gdma_dma_desc(vdesc)->residue;
vdesc             699 drivers/staging/ralink-gdma/ralink-gdma.c static void gdma_dma_desc_free(struct virt_dma_desc *vdesc)
vdesc             701 drivers/staging/ralink-gdma/ralink-gdma.c 	kfree(container_of(vdesc, struct gdma_dma_desc, vdesc));
vdesc             556 fs/nilfs2/ioctl.c 					struct nilfs_vdesc *vdesc,
vdesc             562 fs/nilfs2/ioctl.c 	if (vdesc->vd_flags == 0)
vdesc             564 fs/nilfs2/ioctl.c 			inode, vdesc->vd_offset, vdesc->vd_blocknr,
vdesc             565 fs/nilfs2/ioctl.c 			vdesc->vd_vblocknr, &bh);
vdesc             568 fs/nilfs2/ioctl.c 			inode, vdesc->vd_blocknr, vdesc->vd_vblocknr, &bh);
vdesc             574 fs/nilfs2/ioctl.c 				  __func__, vdesc->vd_flags ? "node" : "data",
vdesc             575 fs/nilfs2/ioctl.c 				  (unsigned long long)vdesc->vd_ino,
vdesc             576 fs/nilfs2/ioctl.c 				  (unsigned long long)vdesc->vd_cno,
vdesc             577 fs/nilfs2/ioctl.c 				  (unsigned long long)vdesc->vd_offset,
vdesc             578 fs/nilfs2/ioctl.c 				  (unsigned long long)vdesc->vd_blocknr,
vdesc             579 fs/nilfs2/ioctl.c 				  (unsigned long long)vdesc->vd_vblocknr);
vdesc             585 fs/nilfs2/ioctl.c 			  __func__, vdesc->vd_flags ? "node" : "data",
vdesc             586 fs/nilfs2/ioctl.c 			  (unsigned long long)vdesc->vd_ino,
vdesc             587 fs/nilfs2/ioctl.c 			  (unsigned long long)vdesc->vd_cno,
vdesc             588 fs/nilfs2/ioctl.c 			  (unsigned long long)vdesc->vd_offset,
vdesc             589 fs/nilfs2/ioctl.c 			  (unsigned long long)vdesc->vd_blocknr,
vdesc             590 fs/nilfs2/ioctl.c 			  (unsigned long long)vdesc->vd_vblocknr);
vdesc             617 fs/nilfs2/ioctl.c 	struct nilfs_vdesc *vdesc;
vdesc             624 fs/nilfs2/ioctl.c 	for (i = 0, vdesc = buf; i < nmembs; ) {
vdesc             625 fs/nilfs2/ioctl.c 		ino = vdesc->vd_ino;
vdesc             626 fs/nilfs2/ioctl.c 		cno = vdesc->vd_cno;
vdesc             644 fs/nilfs2/ioctl.c 			ret = nilfs_ioctl_move_inode_block(inode, vdesc,
vdesc             650 fs/nilfs2/ioctl.c 			vdesc++;
vdesc             652 fs/nilfs2/ioctl.c 			 vdesc->vd_ino == ino && vdesc->vd_cno == cno);