desc_sw 412 drivers/dma/xgene-dma.c struct xgene_dma_desc_sw *desc_sw, desc_sw 421 drivers/dma/xgene-dma.c desc1 = &desc_sw->desc1; desc_sw 422 drivers/dma/xgene-dma.c desc2 = &desc_sw->desc2; desc_sw 451 drivers/dma/xgene-dma.c desc_sw->flags |= XGENE_DMA_FLAG_64B_DESC; desc_sw 592 drivers/dma/xgene-dma.c struct xgene_dma_desc_sw *desc_sw) desc_sw 608 drivers/dma/xgene-dma.c memcpy(desc_hw, &desc_sw->desc1, sizeof(*desc_hw)); desc_sw 614 drivers/dma/xgene-dma.c if (desc_sw->flags & XGENE_DMA_FLAG_64B_DESC) { desc_sw 620 drivers/dma/xgene-dma.c memcpy(desc_hw, &desc_sw->desc2, sizeof(*desc_hw)); desc_sw 624 drivers/dma/xgene-dma.c chan->pending += ((desc_sw->flags & desc_sw 628 drivers/dma/xgene-dma.c iowrite32((desc_sw->flags & XGENE_DMA_FLAG_64B_DESC) ? desc_sw 640 drivers/dma/xgene-dma.c struct xgene_dma_desc_sw *desc_sw, *_desc_sw; desc_sw 655 drivers/dma/xgene-dma.c list_for_each_entry_safe(desc_sw, _desc_sw, &chan->ld_pending, node) { desc_sw 665 drivers/dma/xgene-dma.c xgene_chan_xfer_request(chan, desc_sw); desc_sw 671 drivers/dma/xgene-dma.c list_move_tail(&desc_sw->node, &chan->ld_running); desc_sw 687 drivers/dma/xgene-dma.c struct xgene_dma_desc_sw *desc_sw, *_desc_sw; desc_sw 700 drivers/dma/xgene-dma.c list_for_each_entry_safe(desc_sw, _desc_sw, &chan->ld_running, node) { desc_sw 725 drivers/dma/xgene-dma.c XGENE_DMA_DESC_DUMP(&desc_sw->desc1, desc_sw 728 drivers/dma/xgene-dma.c if (desc_sw->flags & XGENE_DMA_FLAG_64B_DESC) desc_sw 729 drivers/dma/xgene-dma.c XGENE_DMA_DESC_DUMP(&desc_sw->desc2, desc_sw 746 drivers/dma/xgene-dma.c chan->pending -= ((desc_sw->flags & desc_sw 753 drivers/dma/xgene-dma.c list_move_tail(&desc_sw->node, &ld_completed); desc_sw 766 drivers/dma/xgene-dma.c list_for_each_entry_safe(desc_sw, _desc_sw, &ld_completed, node) { desc_sw 767 drivers/dma/xgene-dma.c xgene_dma_run_tx_complete_actions(chan, desc_sw); desc_sw 768 drivers/dma/xgene-dma.c xgene_dma_clean_running_descriptor(chan, desc_sw);