Lines Matching refs:dma_async_tx_descriptor

438 struct dma_async_tx_descriptor {  struct
443 dma_cookie_t (*tx_submit)(struct dma_async_tx_descriptor *tx); argument
448 struct dma_async_tx_descriptor *next; argument
449 struct dma_async_tx_descriptor *parent; argument
455 static inline void dma_set_unmap(struct dma_async_tx_descriptor *tx, in dma_set_unmap() argument
466 static inline void dma_set_unmap(struct dma_async_tx_descriptor *tx, in dma_set_unmap()
480 static inline void dma_descriptor_unmap(struct dma_async_tx_descriptor *tx) in dma_descriptor_unmap()
489 static inline void txd_lock(struct dma_async_tx_descriptor *txd) in txd_lock()
492 static inline void txd_unlock(struct dma_async_tx_descriptor *txd) in txd_unlock()
495 static inline void txd_chain(struct dma_async_tx_descriptor *txd, struct dma_async_tx_descriptor *n… in txd_chain()
499 static inline void txd_clear_parent(struct dma_async_tx_descriptor *txd) in txd_clear_parent()
502 static inline void txd_clear_next(struct dma_async_tx_descriptor *txd) in txd_clear_next()
505 static inline struct dma_async_tx_descriptor *txd_next(struct dma_async_tx_descriptor *txd) in txd_next()
509 static inline struct dma_async_tx_descriptor *txd_parent(struct dma_async_tx_descriptor *txd) in txd_parent()
515 static inline void txd_lock(struct dma_async_tx_descriptor *txd) in txd_lock()
519 static inline void txd_unlock(struct dma_async_tx_descriptor *txd) in txd_unlock()
523 static inline void txd_chain(struct dma_async_tx_descriptor *txd, struct dma_async_tx_descriptor *n… in txd_chain()
528 static inline void txd_clear_parent(struct dma_async_tx_descriptor *txd) in txd_clear_parent()
532 static inline void txd_clear_next(struct dma_async_tx_descriptor *txd) in txd_clear_next()
536 static inline struct dma_async_tx_descriptor *txd_parent(struct dma_async_tx_descriptor *txd) in txd_parent()
540 static inline struct dma_async_tx_descriptor *txd_next(struct dma_async_tx_descriptor *txd) in txd_next()
636 struct dma_async_tx_descriptor *(*device_prep_dma_memcpy)(
639 struct dma_async_tx_descriptor *(*device_prep_dma_xor)(
642 struct dma_async_tx_descriptor *(*device_prep_dma_xor_val)(
645 struct dma_async_tx_descriptor *(*device_prep_dma_pq)(
649 struct dma_async_tx_descriptor *(*device_prep_dma_pq_val)(
653 struct dma_async_tx_descriptor *(*device_prep_dma_interrupt)(
655 struct dma_async_tx_descriptor *(*device_prep_dma_sg)(
661 struct dma_async_tx_descriptor *(*device_prep_slave_sg)(
665 struct dma_async_tx_descriptor *(*device_prep_dma_cyclic)(
669 struct dma_async_tx_descriptor *(*device_prep_interleaved_dma)(
699 static inline struct dma_async_tx_descriptor *dmaengine_prep_slave_single( in dmaengine_prep_slave_single()
712 static inline struct dma_async_tx_descriptor *dmaengine_prep_slave_sg( in dmaengine_prep_slave_sg()
722 static inline struct dma_async_tx_descriptor *dmaengine_prep_rio_sg( in dmaengine_prep_rio_sg()
732 static inline struct dma_async_tx_descriptor *dmaengine_prep_dma_cyclic( in dmaengine_prep_dma_cyclic()
741 static inline struct dma_async_tx_descriptor *dmaengine_prep_interleaved_dma( in dmaengine_prep_interleaved_dma()
748 static inline struct dma_async_tx_descriptor *dmaengine_prep_dma_sg( in dmaengine_prep_dma_sg()
788 static inline dma_cookie_t dmaengine_submit(struct dma_async_tx_descriptor *desc) in dmaengine_submit()
912 void dma_async_tx_descriptor_init(struct dma_async_tx_descriptor *tx,
915 static inline void async_tx_ack(struct dma_async_tx_descriptor *tx) in async_tx_ack()
920 static inline void async_tx_clear_ack(struct dma_async_tx_descriptor *tx) in async_tx_clear_ack()
925 static inline bool async_tx_test_ack(struct dma_async_tx_descriptor *tx) in async_tx_test_ack()
1032 enum dma_status dma_wait_for_async_tx(struct dma_async_tx_descriptor *tx);
1050 static inline enum dma_status dma_wait_for_async_tx(struct dma_async_tx_descriptor *tx) in dma_wait_for_async_tx()
1086 void dma_run_dependencies(struct dma_async_tx_descriptor *tx);