rchan 832 drivers/dma/sh/rcar-dmac.c struct rcar_dmac_chan *rchan = to_rcar_dmac_chan(chan); rchan 834 drivers/dma/sh/rcar-dmac.c spin_lock_irqsave(&rchan->lock, flags); rchan 835 drivers/dma/sh/rcar-dmac.c rcar_dmac_clear_chcr_de(rchan); rchan 836 drivers/dma/sh/rcar-dmac.c spin_unlock_irqrestore(&rchan->lock, flags); rchan 1022 drivers/dma/sh/rcar-dmac.c struct rcar_dmac_chan *rchan = to_rcar_dmac_chan(chan); rchan 1025 drivers/dma/sh/rcar-dmac.c INIT_LIST_HEAD(&rchan->desc.chunks_free); rchan 1026 drivers/dma/sh/rcar-dmac.c INIT_LIST_HEAD(&rchan->desc.pages); rchan 1029 drivers/dma/sh/rcar-dmac.c ret = rcar_dmac_xfer_chunk_alloc(rchan, GFP_KERNEL); rchan 1033 drivers/dma/sh/rcar-dmac.c ret = rcar_dmac_desc_alloc(rchan, GFP_KERNEL); rchan 1042 drivers/dma/sh/rcar-dmac.c struct rcar_dmac_chan *rchan = to_rcar_dmac_chan(chan); rchan 1044 drivers/dma/sh/rcar-dmac.c struct rcar_dmac_chan_map *map = &rchan->map; rchan 1050 drivers/dma/sh/rcar-dmac.c spin_lock_irq(&rchan->lock); rchan 1051 drivers/dma/sh/rcar-dmac.c rcar_dmac_chan_halt(rchan); rchan 1052 drivers/dma/sh/rcar-dmac.c spin_unlock_irq(&rchan->lock); rchan 1058 drivers/dma/sh/rcar-dmac.c synchronize_irq(rchan->irq); rchan 1060 drivers/dma/sh/rcar-dmac.c if (rchan->mid_rid >= 0) { rchan 1062 drivers/dma/sh/rcar-dmac.c clear_bit(rchan->mid_rid, dmac->modules); rchan 1063 drivers/dma/sh/rcar-dmac.c rchan->mid_rid = -EINVAL; rchan 1066 drivers/dma/sh/rcar-dmac.c list_splice_init(&rchan->desc.free, &list); rchan 1067 drivers/dma/sh/rcar-dmac.c list_splice_init(&rchan->desc.pending, &list); rchan 1068 drivers/dma/sh/rcar-dmac.c list_splice_init(&rchan->desc.active, &list); rchan 1069 drivers/dma/sh/rcar-dmac.c list_splice_init(&rchan->desc.done, &list); rchan 1070 drivers/dma/sh/rcar-dmac.c list_splice_init(&rchan->desc.wait, &list); rchan 1072 drivers/dma/sh/rcar-dmac.c rchan->desc.running = NULL; rchan 1075 drivers/dma/sh/rcar-dmac.c rcar_dmac_realloc_hwdesc(rchan, desc, 0); rchan 1077 drivers/dma/sh/rcar-dmac.c list_for_each_entry_safe(page, _page, &rchan->desc.pages, node) { rchan 1096 drivers/dma/sh/rcar-dmac.c struct rcar_dmac_chan *rchan = to_rcar_dmac_chan(chan); rchan 1108 drivers/dma/sh/rcar-dmac.c return rcar_dmac_chan_prep_sg(rchan, &sgl, 1, dma_dest, rchan 1115 drivers/dma/sh/rcar-dmac.c struct rcar_dmac_chan *rchan = to_rcar_dmac_chan(chan); rchan 1116 drivers/dma/sh/rcar-dmac.c struct rcar_dmac_chan_map *map = &rchan->map; rchan 1122 drivers/dma/sh/rcar-dmac.c dev_addr = rchan->src.slave_addr; rchan 1123 drivers/dma/sh/rcar-dmac.c dev_size = rchan->src.xfer_size; rchan 1126 drivers/dma/sh/rcar-dmac.c dev_addr = rchan->dst.slave_addr; rchan 1127 drivers/dma/sh/rcar-dmac.c dev_size = rchan->dst.xfer_size; rchan 1149 drivers/dma/sh/rcar-dmac.c "chan%u: failed to map %zx@%pap", rchan->index, rchan 1155 drivers/dma/sh/rcar-dmac.c rchan->index, dev_size, &dev_addr, &map->addr, rchan 1170 drivers/dma/sh/rcar-dmac.c struct rcar_dmac_chan *rchan = to_rcar_dmac_chan(chan); rchan 1173 drivers/dma/sh/rcar-dmac.c if (rchan->mid_rid < 0 || !sg_len || !sg_dma_len(sgl)) { rchan 1176 drivers/dma/sh/rcar-dmac.c __func__, sg_len, rchan->mid_rid); rchan 1183 drivers/dma/sh/rcar-dmac.c return rcar_dmac_chan_prep_sg(rchan, sgl, sg_len, rchan->map.addr, rchan 1194 drivers/dma/sh/rcar-dmac.c struct rcar_dmac_chan *rchan = to_rcar_dmac_chan(chan); rchan 1201 drivers/dma/sh/rcar-dmac.c if (rchan->mid_rid < 0 || buf_len < period_len) { rchan 1204 drivers/dma/sh/rcar-dmac.c __func__, buf_len, period_len, rchan->mid_rid); rchan 1215 drivers/dma/sh/rcar-dmac.c rchan->index, sg_len, RCAR_DMAC_MAX_SG_LEN); rchan 1238 drivers/dma/sh/rcar-dmac.c desc = rcar_dmac_chan_prep_sg(rchan, sgl, sg_len, rchan->map.addr, rchan 1248 drivers/dma/sh/rcar-dmac.c struct rcar_dmac_chan *rchan = to_rcar_dmac_chan(chan); rchan 1254 drivers/dma/sh/rcar-dmac.c rchan->src.slave_addr = cfg->src_addr; rchan 1255 drivers/dma/sh/rcar-dmac.c rchan->dst.slave_addr = cfg->dst_addr; rchan 1256 drivers/dma/sh/rcar-dmac.c rchan->src.xfer_size = cfg->src_addr_width; rchan 1257 drivers/dma/sh/rcar-dmac.c rchan->dst.xfer_size = cfg->dst_addr_width; rchan 1264 drivers/dma/sh/rcar-dmac.c struct rcar_dmac_chan *rchan = to_rcar_dmac_chan(chan); rchan 1267 drivers/dma/sh/rcar-dmac.c spin_lock_irqsave(&rchan->lock, flags); rchan 1268 drivers/dma/sh/rcar-dmac.c rcar_dmac_chan_halt(rchan); rchan 1269 drivers/dma/sh/rcar-dmac.c spin_unlock_irqrestore(&rchan->lock, flags); rchan 1276 drivers/dma/sh/rcar-dmac.c rcar_dmac_chan_reinit(rchan); rchan 1392 drivers/dma/sh/rcar-dmac.c struct rcar_dmac_chan *rchan = to_rcar_dmac_chan(chan); rchan 1402 drivers/dma/sh/rcar-dmac.c spin_lock_irqsave(&rchan->lock, flags); rchan 1403 drivers/dma/sh/rcar-dmac.c residue = rcar_dmac_chan_get_residue(rchan, cookie); rchan 1404 drivers/dma/sh/rcar-dmac.c cyclic = rchan->desc.running ? rchan->desc.running->cyclic : false; rchan 1405 drivers/dma/sh/rcar-dmac.c spin_unlock_irqrestore(&rchan->lock, flags); rchan 1418 drivers/dma/sh/rcar-dmac.c struct rcar_dmac_chan *rchan = to_rcar_dmac_chan(chan); rchan 1421 drivers/dma/sh/rcar-dmac.c spin_lock_irqsave(&rchan->lock, flags); rchan 1423 drivers/dma/sh/rcar-dmac.c if (list_empty(&rchan->desc.pending)) rchan 1427 drivers/dma/sh/rcar-dmac.c list_splice_tail_init(&rchan->desc.pending, &rchan->desc.active); rchan 1433 drivers/dma/sh/rcar-dmac.c if (!rchan->desc.running) { rchan 1436 drivers/dma/sh/rcar-dmac.c desc = list_first_entry(&rchan->desc.active, rchan 1438 drivers/dma/sh/rcar-dmac.c rchan->desc.running = desc; rchan 1440 drivers/dma/sh/rcar-dmac.c rcar_dmac_chan_start_xfer(rchan); rchan 1444 drivers/dma/sh/rcar-dmac.c spin_unlock_irqrestore(&rchan->lock, flags); rchan 1449 drivers/dma/sh/rcar-dmac.c struct rcar_dmac_chan *rchan = to_rcar_dmac_chan(chan); rchan 1451 drivers/dma/sh/rcar-dmac.c synchronize_irq(rchan->irq); rchan 1671 drivers/dma/sh/rcar-dmac.c struct rcar_dmac_chan *rchan; rchan 1687 drivers/dma/sh/rcar-dmac.c rchan = to_rcar_dmac_chan(chan); rchan 1688 drivers/dma/sh/rcar-dmac.c rchan->mid_rid = dma_spec->args[0]; rchan 1728 drivers/dma/sh/rcar-dmac.c struct rcar_dmac_chan *rchan, rchan 1732 drivers/dma/sh/rcar-dmac.c struct dma_chan *chan = &rchan->chan; rchan 1737 drivers/dma/sh/rcar-dmac.c rchan->index = index; rchan 1738 drivers/dma/sh/rcar-dmac.c rchan->iomem = dmac->iomem + RCAR_DMAC_CHAN_OFFSET(index); rchan 1739 drivers/dma/sh/rcar-dmac.c rchan->mid_rid = -EINVAL; rchan 1741 drivers/dma/sh/rcar-dmac.c spin_lock_init(&rchan->lock); rchan 1743 drivers/dma/sh/rcar-dmac.c INIT_LIST_HEAD(&rchan->desc.free); rchan 1744 drivers/dma/sh/rcar-dmac.c INIT_LIST_HEAD(&rchan->desc.pending); rchan 1745 drivers/dma/sh/rcar-dmac.c INIT_LIST_HEAD(&rchan->desc.active); rchan 1746 drivers/dma/sh/rcar-dmac.c INIT_LIST_HEAD(&rchan->desc.done); rchan 1747 drivers/dma/sh/rcar-dmac.c INIT_LIST_HEAD(&rchan->desc.wait); rchan 1751 drivers/dma/sh/rcar-dmac.c rchan->irq = platform_get_irq_byname(pdev, pdev_irqname); rchan 1752 drivers/dma/sh/rcar-dmac.c if (rchan->irq < 0) rchan 1769 drivers/dma/sh/rcar-dmac.c ret = devm_request_threaded_irq(dmac->dev, rchan->irq, rchan 1772 drivers/dma/sh/rcar-dmac.c irqname, rchan); rchan 1775 drivers/dma/sh/rcar-dmac.c rchan->irq, ret); rchan 370 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c struct rchan *guc_log_relay_chan; rchan 51 drivers/gpu/drm/i915/gt/uc/intel_guc_log.h struct rchan *channel; rchan 1148 drivers/net/wireless/ath/ath10k/core.h struct rchan *rfs_chan_spec_scan; rchan 453 drivers/net/wireless/ath/ath9k/common-spectral.c struct rchan *rc = spec_priv->rfs_chan_spec_scan; rchan 100 drivers/net/wireless/ath/ath9k/common-spectral.h struct rchan *rfs_chan_spec_scan; rchan 17 include/linux/blktrace_api.h struct rchan *rchan; rchan 41 include/linux/relay.h struct rchan *chan; /* associated channel */ rchan 169 include/linux/relay.h struct rchan *relay_open(const char *base_filename, rchan 175 include/linux/relay.h extern int relay_late_setup_files(struct rchan *chan, rchan 178 include/linux/relay.h extern void relay_close(struct rchan *chan); rchan 179 include/linux/relay.h extern void relay_flush(struct rchan *chan); rchan 180 include/linux/relay.h extern void relay_subbufs_consumed(struct rchan *chan, rchan 183 include/linux/relay.h extern void relay_reset(struct rchan *chan); rchan 202 include/linux/relay.h static inline void relay_write(struct rchan *chan, rchan 230 include/linux/relay.h static inline void __relay_write(struct rchan *chan, rchan 255 include/linux/relay.h static inline void *relay_reserve(struct rchan *chan, size_t length) rchan 162 kernel/relay.c static struct rchan_buf *relay_create_buf(struct rchan *chan) rchan 199 kernel/relay.c struct rchan *chan = container_of(kref, struct rchan, kref); rchan 209 kernel/relay.c struct rchan *chan = buf->chan; rchan 387 kernel/relay.c void relay_reset(struct rchan *chan) rchan 415 kernel/relay.c static struct dentry *relay_create_buf_file(struct rchan *chan, rchan 444 kernel/relay.c static struct rchan_buf *relay_open_buf(struct rchan *chan, unsigned int cpu) rchan 501 kernel/relay.c static void setup_callbacks(struct rchan *chan, rchan 524 kernel/relay.c struct rchan *chan; rchan 563 kernel/relay.c struct rchan *relay_open(const char *base_filename, rchan 571 kernel/relay.c struct rchan *chan; rchan 579 kernel/relay.c chan = kzalloc(sizeof(struct rchan), GFP_KERNEL); rchan 654 kernel/relay.c int relay_late_setup_files(struct rchan *chan, rchan 813 kernel/relay.c void relay_subbufs_consumed(struct rchan *chan, rchan 839 kernel/relay.c void relay_close(struct rchan *chan) rchan 872 kernel/relay.c void relay_flush(struct rchan *chan) rchan 90 kernel/trace/blktrace.c if (!bt->rchan) rchan 93 kernel/trace/blktrace.c t = relay_reserve(bt->rchan, sizeof(*t) + len + cgid_len); rchan 272 kernel/trace/blktrace.c t = relay_reserve(bt->rchan, sizeof(*t) + pdu_len + cgid_len); rchan 314 kernel/trace/blktrace.c relay_close(bt->rchan); rchan 527 kernel/trace/blktrace.c bt->rchan = relay_open("trace", dir, buts->buf_size, rchan 529 kernel/trace/blktrace.c if (!bt->rchan) rchan 665 kernel/trace/blktrace.c relay_flush(bt->rchan);