root/sound/soc/uniphier/aio-dma.c

/* [<][>][^][v][top][bottom][index][help] */

DEFINITIONS

This source file includes following definitions.
  1. aiodma_pcm_irq
  2. aiodma_compr_irq
  3. aiodma_irq
  4. uniphier_aiodma_open
  5. uniphier_aiodma_hw_params
  6. uniphier_aiodma_hw_free
  7. uniphier_aiodma_prepare
  8. uniphier_aiodma_trigger
  9. uniphier_aiodma_pointer
  10. uniphier_aiodma_mmap
  11. uniphier_aiodma_new
  12. uniphier_aiodma_free
  13. uniphier_aiodma_soc_register_platform

   1 // SPDX-License-Identifier: GPL-2.0
   2 //
   3 // Socionext UniPhier AIO DMA driver.
   4 //
   5 // Copyright (c) 2016-2018 Socionext Inc.
   6 
   7 #include <linux/dma-mapping.h>
   8 #include <linux/errno.h>
   9 #include <linux/kernel.h>
  10 #include <linux/module.h>
  11 #include <sound/core.h>
  12 #include <sound/pcm.h>
  13 #include <sound/soc.h>
  14 
  15 #include "aio.h"
  16 
  17 static struct snd_pcm_hardware uniphier_aiodma_hw = {
  18         .info = SNDRV_PCM_INFO_MMAP |
  19                 SNDRV_PCM_INFO_MMAP_VALID |
  20                 SNDRV_PCM_INFO_INTERLEAVED,
  21         .period_bytes_min = 256,
  22         .period_bytes_max = 4096,
  23         .periods_min      = 4,
  24         .periods_max      = 1024,
  25         .buffer_bytes_max = 128 * 1024,
  26 };
  27 
  28 static void aiodma_pcm_irq(struct uniphier_aio_sub *sub)
  29 {
  30         struct snd_pcm_runtime *runtime = sub->substream->runtime;
  31         int bytes = runtime->period_size *
  32                 runtime->channels * samples_to_bytes(runtime, 1);
  33         int ret;
  34 
  35         spin_lock(&sub->lock);
  36         ret = aiodma_rb_set_threshold(sub, runtime->dma_bytes,
  37                                       sub->threshold + bytes);
  38         if (!ret)
  39                 sub->threshold += bytes;
  40 
  41         aiodma_rb_sync(sub, runtime->dma_addr, runtime->dma_bytes, bytes);
  42         aiodma_rb_clear_irq(sub);
  43         spin_unlock(&sub->lock);
  44 
  45         snd_pcm_period_elapsed(sub->substream);
  46 }
  47 
  48 static void aiodma_compr_irq(struct uniphier_aio_sub *sub)
  49 {
  50         struct snd_compr_runtime *runtime = sub->cstream->runtime;
  51         int bytes = runtime->fragment_size;
  52         int ret;
  53 
  54         spin_lock(&sub->lock);
  55         ret = aiodma_rb_set_threshold(sub, sub->compr_bytes,
  56                                       sub->threshold + bytes);
  57         if (!ret)
  58                 sub->threshold += bytes;
  59 
  60         aiodma_rb_sync(sub, sub->compr_addr, sub->compr_bytes, bytes);
  61         aiodma_rb_clear_irq(sub);
  62         spin_unlock(&sub->lock);
  63 
  64         snd_compr_fragment_elapsed(sub->cstream);
  65 }
  66 
  67 static irqreturn_t aiodma_irq(int irq, void *p)
  68 {
  69         struct platform_device *pdev = p;
  70         struct uniphier_aio_chip *chip = platform_get_drvdata(pdev);
  71         irqreturn_t ret = IRQ_NONE;
  72         int i, j;
  73 
  74         for (i = 0; i < chip->num_aios; i++) {
  75                 struct uniphier_aio *aio = &chip->aios[i];
  76 
  77                 for (j = 0; j < ARRAY_SIZE(aio->sub); j++) {
  78                         struct uniphier_aio_sub *sub = &aio->sub[j];
  79 
  80                         /* Skip channel that does not trigger */
  81                         if (!sub->running || !aiodma_rb_is_irq(sub))
  82                                 continue;
  83 
  84                         if (sub->substream)
  85                                 aiodma_pcm_irq(sub);
  86                         if (sub->cstream)
  87                                 aiodma_compr_irq(sub);
  88 
  89                         ret = IRQ_HANDLED;
  90                 }
  91         }
  92 
  93         return ret;
  94 }
  95 
  96 static int uniphier_aiodma_open(struct snd_pcm_substream *substream)
  97 {
  98         struct snd_pcm_runtime *runtime = substream->runtime;
  99 
 100         snd_soc_set_runtime_hwparams(substream, &uniphier_aiodma_hw);
 101 
 102         return snd_pcm_hw_constraint_step(runtime, 0,
 103                 SNDRV_PCM_HW_PARAM_BUFFER_BYTES, 256);
 104 }
 105 
 106 static int uniphier_aiodma_hw_params(struct snd_pcm_substream *substream,
 107                                      struct snd_pcm_hw_params *params)
 108 {
 109         snd_pcm_set_runtime_buffer(substream, &substream->dma_buffer);
 110         substream->runtime->dma_bytes = params_buffer_bytes(params);
 111 
 112         return 0;
 113 }
 114 
 115 static int uniphier_aiodma_hw_free(struct snd_pcm_substream *substream)
 116 {
 117         snd_pcm_set_runtime_buffer(substream, NULL);
 118         substream->runtime->dma_bytes = 0;
 119 
 120         return 0;
 121 }
 122 
 123 static int uniphier_aiodma_prepare(struct snd_pcm_substream *substream)
 124 {
 125         struct snd_pcm_runtime *runtime = substream->runtime;
 126         struct snd_soc_pcm_runtime *rtd = snd_pcm_substream_chip(substream);
 127         struct uniphier_aio *aio = uniphier_priv(rtd->cpu_dai);
 128         struct uniphier_aio_sub *sub = &aio->sub[substream->stream];
 129         int bytes = runtime->period_size *
 130                 runtime->channels * samples_to_bytes(runtime, 1);
 131         unsigned long flags;
 132         int ret;
 133 
 134         ret = aiodma_ch_set_param(sub);
 135         if (ret)
 136                 return ret;
 137 
 138         spin_lock_irqsave(&sub->lock, flags);
 139         ret = aiodma_rb_set_buffer(sub, runtime->dma_addr,
 140                                    runtime->dma_addr + runtime->dma_bytes,
 141                                    bytes);
 142         spin_unlock_irqrestore(&sub->lock, flags);
 143         if (ret)
 144                 return ret;
 145 
 146         return 0;
 147 }
 148 
 149 static int uniphier_aiodma_trigger(struct snd_pcm_substream *substream, int cmd)
 150 {
 151         struct snd_pcm_runtime *runtime = substream->runtime;
 152         struct snd_soc_pcm_runtime *rtd = snd_pcm_substream_chip(substream);
 153         struct uniphier_aio *aio = uniphier_priv(rtd->cpu_dai);
 154         struct uniphier_aio_sub *sub = &aio->sub[substream->stream];
 155         struct device *dev = &aio->chip->pdev->dev;
 156         int bytes = runtime->period_size *
 157                 runtime->channels * samples_to_bytes(runtime, 1);
 158         unsigned long flags;
 159 
 160         spin_lock_irqsave(&sub->lock, flags);
 161         switch (cmd) {
 162         case SNDRV_PCM_TRIGGER_START:
 163                 aiodma_rb_sync(sub, runtime->dma_addr, runtime->dma_bytes,
 164                                bytes);
 165                 aiodma_ch_set_enable(sub, 1);
 166                 sub->running = 1;
 167 
 168                 break;
 169         case SNDRV_PCM_TRIGGER_STOP:
 170                 sub->running = 0;
 171                 aiodma_ch_set_enable(sub, 0);
 172 
 173                 break;
 174         default:
 175                 dev_warn(dev, "Unknown trigger(%d) ignored\n", cmd);
 176                 break;
 177         }
 178         spin_unlock_irqrestore(&sub->lock, flags);
 179 
 180         return 0;
 181 }
 182 
 183 static snd_pcm_uframes_t uniphier_aiodma_pointer(
 184                                         struct snd_pcm_substream *substream)
 185 {
 186         struct snd_pcm_runtime *runtime = substream->runtime;
 187         struct snd_soc_pcm_runtime *rtd = snd_pcm_substream_chip(substream);
 188         struct uniphier_aio *aio = uniphier_priv(rtd->cpu_dai);
 189         struct uniphier_aio_sub *sub = &aio->sub[substream->stream];
 190         int bytes = runtime->period_size *
 191                 runtime->channels * samples_to_bytes(runtime, 1);
 192         unsigned long flags;
 193         snd_pcm_uframes_t pos;
 194 
 195         spin_lock_irqsave(&sub->lock, flags);
 196         aiodma_rb_sync(sub, runtime->dma_addr, runtime->dma_bytes, bytes);
 197 
 198         if (sub->swm->dir == PORT_DIR_OUTPUT)
 199                 pos = bytes_to_frames(runtime, sub->rd_offs);
 200         else
 201                 pos = bytes_to_frames(runtime, sub->wr_offs);
 202         spin_unlock_irqrestore(&sub->lock, flags);
 203 
 204         return pos;
 205 }
 206 
 207 static int uniphier_aiodma_mmap(struct snd_pcm_substream *substream,
 208                                 struct vm_area_struct *vma)
 209 {
 210         vma->vm_page_prot = pgprot_writecombine(vma->vm_page_prot);
 211 
 212         return remap_pfn_range(vma, vma->vm_start,
 213                                substream->dma_buffer.addr >> PAGE_SHIFT,
 214                                vma->vm_end - vma->vm_start, vma->vm_page_prot);
 215 }
 216 
 217 static const struct snd_pcm_ops uniphier_aiodma_ops = {
 218         .open      = uniphier_aiodma_open,
 219         .ioctl     = snd_pcm_lib_ioctl,
 220         .hw_params = uniphier_aiodma_hw_params,
 221         .hw_free   = uniphier_aiodma_hw_free,
 222         .prepare   = uniphier_aiodma_prepare,
 223         .trigger   = uniphier_aiodma_trigger,
 224         .pointer   = uniphier_aiodma_pointer,
 225         .mmap      = uniphier_aiodma_mmap,
 226 };
 227 
 228 static int uniphier_aiodma_new(struct snd_soc_pcm_runtime *rtd)
 229 {
 230         struct device *dev = rtd->card->snd_card->dev;
 231         struct snd_pcm *pcm = rtd->pcm;
 232         int ret;
 233 
 234         ret = dma_set_mask_and_coherent(dev, DMA_BIT_MASK(33));
 235         if (ret)
 236                 return ret;
 237 
 238         snd_pcm_lib_preallocate_pages_for_all(pcm,
 239                 SNDRV_DMA_TYPE_DEV, dev,
 240                 uniphier_aiodma_hw.buffer_bytes_max,
 241                 uniphier_aiodma_hw.buffer_bytes_max);
 242         return 0;
 243 }
 244 
 245 static void uniphier_aiodma_free(struct snd_pcm *pcm)
 246 {
 247         snd_pcm_lib_preallocate_free_for_all(pcm);
 248 }
 249 
 250 static const struct snd_soc_component_driver uniphier_soc_platform = {
 251         .pcm_new   = uniphier_aiodma_new,
 252         .pcm_free  = uniphier_aiodma_free,
 253         .ops       = &uniphier_aiodma_ops,
 254         .compr_ops = &uniphier_aio_compr_ops,
 255 };
 256 
 257 static const struct regmap_config aiodma_regmap_config = {
 258         .reg_bits      = 32,
 259         .reg_stride    = 4,
 260         .val_bits      = 32,
 261         .max_register  = 0x7fffc,
 262         .cache_type    = REGCACHE_NONE,
 263 };
 264 
 265 /**
 266  * uniphier_aiodma_soc_register_platform - register the AIO DMA
 267  * @pdev: the platform device
 268  *
 269  * Register and setup the DMA of AIO to transfer the sound data to device.
 270  * This function need to call once at driver startup and need NOT to call
 271  * unregister function.
 272  *
 273  * Return: Zero if successful, otherwise a negative value on error.
 274  */
 275 int uniphier_aiodma_soc_register_platform(struct platform_device *pdev)
 276 {
 277         struct uniphier_aio_chip *chip = platform_get_drvdata(pdev);
 278         struct device *dev = &pdev->dev;
 279         void __iomem *preg;
 280         int irq, ret;
 281 
 282         preg = devm_platform_ioremap_resource(pdev, 0);
 283         if (IS_ERR(preg))
 284                 return PTR_ERR(preg);
 285 
 286         chip->regmap = devm_regmap_init_mmio(dev, preg,
 287                                              &aiodma_regmap_config);
 288         if (IS_ERR(chip->regmap))
 289                 return PTR_ERR(chip->regmap);
 290 
 291         irq = platform_get_irq(pdev, 0);
 292         if (irq < 0)
 293                 return irq;
 294 
 295         ret = devm_request_irq(dev, irq, aiodma_irq,
 296                                IRQF_SHARED, dev_name(dev), pdev);
 297         if (ret)
 298                 return ret;
 299 
 300         return devm_snd_soc_register_component(dev, &uniphier_soc_platform,
 301                                                NULL, 0);
 302 }
 303 EXPORT_SYMBOL_GPL(uniphier_aiodma_soc_register_platform);

/* [<][>][^][v][top][bottom][index][help] */