blob: 6fd1906af3873aa90b351649c2d0b8afe7152e9d [file] [log] [blame]
Lars-Peter Clausen28c44682013-04-15 19:19:50 +02001/*
2 * Copyright (C) 2013, Analog Devices Inc.
3 * Author: Lars-Peter Clausen <lars@metafoo.de>
4 *
5 * This program is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License as published by the
7 * Free Software Foundation; either version 2 of the License, or (at your
8 * option) any later version.
9 *
10 * You should have received a copy of the GNU General Public License along
11 * with this program; if not, write to the Free Software Foundation, Inc.,
12 * 675 Mass Ave, Cambridge, MA 02139, USA.
13 *
14 */
15#include <linux/module.h>
16#include <linux/init.h>
17#include <linux/dmaengine.h>
18#include <linux/slab.h>
19#include <sound/pcm.h>
20#include <sound/pcm_params.h>
21#include <sound/soc.h>
22#include <linux/dma-mapping.h>
23#include <linux/of.h>
Lars-Peter Clausen28c44682013-04-15 19:19:50 +020024
25#include <sound/dmaengine_pcm.h>
26
Lars-Peter Clausenacde50a2015-04-27 12:44:25 +020027/*
28 * The platforms dmaengine driver does not support reporting the amount of
29 * bytes that are still left to transfer.
30 */
31#define SND_DMAENGINE_PCM_FLAG_NO_RESIDUE BIT(31)
32
Lars-Peter Clausen28c44682013-04-15 19:19:50 +020033struct dmaengine_pcm {
Takashi Iwaif82bf8e2013-10-25 18:06:09 +020034 struct dma_chan *chan[SNDRV_PCM_STREAM_LAST + 1];
Lars-Peter Clausen28c44682013-04-15 19:19:50 +020035 const struct snd_dmaengine_pcm_config *config;
36 struct snd_soc_platform platform;
Lars-Peter Clausend1e14062013-04-20 19:29:00 +020037 unsigned int flags;
Lars-Peter Clausen28c44682013-04-15 19:19:50 +020038};
39
40static struct dmaengine_pcm *soc_platform_to_pcm(struct snd_soc_platform *p)
41{
42 return container_of(p, struct dmaengine_pcm, platform);
43}
44
Lars-Peter Clausenc0de42b2013-10-08 15:07:59 +020045static struct device *dmaengine_dma_dev(struct dmaengine_pcm *pcm,
46 struct snd_pcm_substream *substream)
47{
48 if (!pcm->chan[substream->stream])
49 return NULL;
50
51 return pcm->chan[substream->stream]->device->dev;
52}
53
Lars-Peter Clausen28c44682013-04-15 19:19:50 +020054/**
55 * snd_dmaengine_pcm_prepare_slave_config() - Generic prepare_slave_config callback
56 * @substream: PCM substream
57 * @params: hw_params
58 * @slave_config: DMA slave config to prepare
59 *
60 * This function can be used as a generic prepare_slave_config callback for
61 * platforms which make use of the snd_dmaengine_dai_dma_data struct for their
62 * DAI DMA data. Internally the function will first call
63 * snd_hwparams_to_dma_slave_config to fill in the slave config based on the
64 * hw_params, followed by snd_dmaengine_set_config_from_dai_data to fill in the
65 * remaining fields based on the DAI DMA data.
66 */
67int snd_dmaengine_pcm_prepare_slave_config(struct snd_pcm_substream *substream,
68 struct snd_pcm_hw_params *params, struct dma_slave_config *slave_config)
69{
70 struct snd_soc_pcm_runtime *rtd = substream->private_data;
71 struct snd_dmaengine_dai_dma_data *dma_data;
72 int ret;
73
74 dma_data = snd_soc_dai_get_dma_data(rtd->cpu_dai, substream);
75
76 ret = snd_hwparams_to_dma_slave_config(substream, params, slave_config);
77 if (ret)
78 return ret;
79
80 snd_dmaengine_pcm_set_config_from_dai_data(substream, dma_data,
81 slave_config);
82
83 return 0;
84}
85EXPORT_SYMBOL_GPL(snd_dmaengine_pcm_prepare_slave_config);
86
87static int dmaengine_pcm_hw_params(struct snd_pcm_substream *substream,
88 struct snd_pcm_hw_params *params)
89{
90 struct snd_soc_pcm_runtime *rtd = substream->private_data;
91 struct dmaengine_pcm *pcm = soc_platform_to_pcm(rtd->platform);
92 struct dma_chan *chan = snd_dmaengine_pcm_get_chan(substream);
Lars-Peter Clausenfa654e02013-10-08 15:08:00 +020093 int (*prepare_slave_config)(struct snd_pcm_substream *substream,
94 struct snd_pcm_hw_params *params,
95 struct dma_slave_config *slave_config);
Lars-Peter Clausen28c44682013-04-15 19:19:50 +020096 struct dma_slave_config slave_config;
97 int ret;
98
Lee Jonesa894bd72013-11-06 10:16:20 +000099 memset(&slave_config, 0, sizeof(slave_config));
100
Lars-Peter Clausenfa654e02013-10-08 15:08:00 +0200101 if (!pcm->config)
102 prepare_slave_config = snd_dmaengine_pcm_prepare_slave_config;
103 else
104 prepare_slave_config = pcm->config->prepare_slave_config;
105
106 if (prepare_slave_config) {
107 ret = prepare_slave_config(substream, params, &slave_config);
Lars-Peter Clausen28c44682013-04-15 19:19:50 +0200108 if (ret)
109 return ret;
110
111 ret = dmaengine_slave_config(chan, &slave_config);
112 if (ret)
113 return ret;
114 }
115
116 return snd_pcm_lib_malloc_pages(substream, params_buffer_bytes(params));
117}
118
Lars-Peter Clausenc0de42b2013-10-08 15:07:59 +0200119static int dmaengine_pcm_set_runtime_hwparams(struct snd_pcm_substream *substream)
120{
121 struct snd_soc_pcm_runtime *rtd = substream->private_data;
122 struct dmaengine_pcm *pcm = soc_platform_to_pcm(rtd->platform);
123 struct device *dma_dev = dmaengine_dma_dev(pcm, substream);
124 struct dma_chan *chan = pcm->chan[substream->stream];
125 struct snd_dmaengine_dai_dma_data *dma_data;
126 struct dma_slave_caps dma_caps;
127 struct snd_pcm_hardware hw;
Peter Ujfalusi2d38df12014-07-03 07:51:54 +0300128 u32 addr_widths = BIT(DMA_SLAVE_BUSWIDTH_1_BYTE) |
129 BIT(DMA_SLAVE_BUSWIDTH_2_BYTES) |
130 BIT(DMA_SLAVE_BUSWIDTH_4_BYTES);
131 int i, ret;
Lars-Peter Clausenc0de42b2013-10-08 15:07:59 +0200132
Lars-Peter Clausenfa654e02013-10-08 15:08:00 +0200133 if (pcm->config && pcm->config->pcm_hardware)
Lars-Peter Clausenc0de42b2013-10-08 15:07:59 +0200134 return snd_soc_set_runtime_hwparams(substream,
135 pcm->config->pcm_hardware);
136
137 dma_data = snd_soc_dai_get_dma_data(rtd->cpu_dai, substream);
138
139 memset(&hw, 0, sizeof(hw));
140 hw.info = SNDRV_PCM_INFO_MMAP | SNDRV_PCM_INFO_MMAP_VALID |
141 SNDRV_PCM_INFO_INTERLEAVED;
142 hw.periods_min = 2;
143 hw.periods_max = UINT_MAX;
144 hw.period_bytes_min = 256;
145 hw.period_bytes_max = dma_get_max_seg_size(dma_dev);
146 hw.buffer_bytes_max = SIZE_MAX;
147 hw.fifo_size = dma_data->fifo_size;
148
Lars-Peter Clausena22f33b2013-11-30 18:00:45 +0100149 if (pcm->flags & SND_DMAENGINE_PCM_FLAG_NO_RESIDUE)
150 hw.info |= SNDRV_PCM_INFO_BATCH;
151
Lars-Peter Clausenc0de42b2013-10-08 15:07:59 +0200152 ret = dma_get_slave_caps(chan, &dma_caps);
153 if (ret == 0) {
154 if (dma_caps.cmd_pause)
155 hw.info |= SNDRV_PCM_INFO_PAUSE | SNDRV_PCM_INFO_RESUME;
Lars-Peter Clausen478028e2014-01-11 14:02:19 +0100156 if (dma_caps.residue_granularity <= DMA_RESIDUE_GRANULARITY_SEGMENT)
157 hw.info |= SNDRV_PCM_INFO_BATCH;
Peter Ujfalusi2d38df12014-07-03 07:51:54 +0300158
159 if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK)
Maxime Ripardceacbdb2014-11-17 14:41:57 +0100160 addr_widths = dma_caps.dst_addr_widths;
Peter Ujfalusi2d38df12014-07-03 07:51:54 +0300161 else
162 addr_widths = dma_caps.src_addr_widths;
163 }
164
165 /*
166 * Prepare formats mask for valid/allowed sample types. If the dma does
167 * not have support for the given physical word size, it needs to be
168 * masked out so user space can not use the format which produces
169 * corrupted audio.
170 * In case the dma driver does not implement the slave_caps the default
171 * assumption is that it supports 1, 2 and 4 bytes widths.
172 */
173 for (i = 0; i <= SNDRV_PCM_FORMAT_LAST; i++) {
174 int bits = snd_pcm_format_physical_width(i);
175
176 /* Enable only samples with DMA supported physical widths */
177 switch (bits) {
178 case 8:
179 case 16:
180 case 24:
181 case 32:
182 case 64:
183 if (addr_widths & (1 << (bits / 8)))
184 hw.formats |= (1LL << i);
185 break;
186 default:
187 /* Unsupported types */
188 break;
189 }
Lars-Peter Clausenc0de42b2013-10-08 15:07:59 +0200190 }
191
192 return snd_soc_set_runtime_hwparams(substream, &hw);
193}
194
Lars-Peter Clausen28c44682013-04-15 19:19:50 +0200195static int dmaengine_pcm_open(struct snd_pcm_substream *substream)
196{
197 struct snd_soc_pcm_runtime *rtd = substream->private_data;
198 struct dmaengine_pcm *pcm = soc_platform_to_pcm(rtd->platform);
199 struct dma_chan *chan = pcm->chan[substream->stream];
200 int ret;
201
Lars-Peter Clausenc0de42b2013-10-08 15:07:59 +0200202 ret = dmaengine_pcm_set_runtime_hwparams(substream);
Lars-Peter Clausen28c44682013-04-15 19:19:50 +0200203 if (ret)
204 return ret;
205
206 return snd_dmaengine_pcm_open(substream, chan);
207}
208
Lars-Peter Clausenc9998362013-04-15 19:19:51 +0200209static struct dma_chan *dmaengine_pcm_compat_request_channel(
210 struct snd_soc_pcm_runtime *rtd,
211 struct snd_pcm_substream *substream)
212{
213 struct dmaengine_pcm *pcm = soc_platform_to_pcm(rtd->platform);
Mark Brown90130d22013-10-19 21:38:26 +0100214 struct snd_dmaengine_dai_dma_data *dma_data;
Xiubo Liec4f2852014-01-16 16:08:04 +0800215 dma_filter_fn fn = NULL;
Mark Brown90130d22013-10-19 21:38:26 +0100216
217 dma_data = snd_soc_dai_get_dma_data(rtd->cpu_dai, substream);
Lars-Peter Clausenc9998362013-04-15 19:19:51 +0200218
Lars-Peter Clausend1e14062013-04-20 19:29:00 +0200219 if ((pcm->flags & SND_DMAENGINE_PCM_FLAG_HALF_DUPLEX) && pcm->chan[0])
220 return pcm->chan[0];
221
Xiubo Liec4f2852014-01-16 16:08:04 +0800222 if (pcm->config && pcm->config->compat_request_channel)
Lars-Peter Clausenc9998362013-04-15 19:19:51 +0200223 return pcm->config->compat_request_channel(rtd, substream);
224
Xiubo Liec4f2852014-01-16 16:08:04 +0800225 if (pcm->config)
226 fn = pcm->config->compat_filter_fn;
227
228 return snd_dmaengine_pcm_request_channel(fn, dma_data->filter_data);
Lars-Peter Clausenc9998362013-04-15 19:19:51 +0200229}
230
Lars-Peter Clausenacde50a2015-04-27 12:44:25 +0200231static bool dmaengine_pcm_can_report_residue(struct device *dev,
232 struct dma_chan *chan)
Lars-Peter Clausen478028e2014-01-11 14:02:19 +0100233{
234 struct dma_slave_caps dma_caps;
235 int ret;
236
237 ret = dma_get_slave_caps(chan, &dma_caps);
Lars-Peter Clausenacde50a2015-04-27 12:44:25 +0200238 if (ret != 0) {
239 dev_warn(dev, "Failed to get DMA channel capabilities, falling back to period counting: %d\n",
240 ret);
241 return false;
242 }
Lars-Peter Clausen478028e2014-01-11 14:02:19 +0100243
244 if (dma_caps.residue_granularity == DMA_RESIDUE_GRANULARITY_DESCRIPTOR)
245 return false;
246
247 return true;
248}
249
Lars-Peter Clausen28c44682013-04-15 19:19:50 +0200250static int dmaengine_pcm_new(struct snd_soc_pcm_runtime *rtd)
251{
252 struct dmaengine_pcm *pcm = soc_platform_to_pcm(rtd->platform);
253 const struct snd_dmaengine_pcm_config *config = pcm->config;
Mark Brownea73b7d2013-10-19 17:43:51 +0100254 struct device *dev = rtd->platform->dev;
255 struct snd_dmaengine_dai_dma_data *dma_data;
Lars-Peter Clausen28c44682013-04-15 19:19:50 +0200256 struct snd_pcm_substream *substream;
Lars-Peter Clausenfa654e02013-10-08 15:08:00 +0200257 size_t prealloc_buffer_size;
258 size_t max_buffer_size;
Lars-Peter Clausen28c44682013-04-15 19:19:50 +0200259 unsigned int i;
260 int ret;
261
Lars-Peter Clausenfa654e02013-10-08 15:08:00 +0200262 if (config && config->prealloc_buffer_size) {
263 prealloc_buffer_size = config->prealloc_buffer_size;
264 max_buffer_size = config->pcm_hardware->buffer_bytes_max;
265 } else {
266 prealloc_buffer_size = 512 * 1024;
267 max_buffer_size = SIZE_MAX;
268 }
269
270
Lars-Peter Clausen28c44682013-04-15 19:19:50 +0200271 for (i = SNDRV_PCM_STREAM_PLAYBACK; i <= SNDRV_PCM_STREAM_CAPTURE; i++) {
272 substream = rtd->pcm->streams[i].substream;
273 if (!substream)
274 continue;
275
Mark Brownea73b7d2013-10-19 17:43:51 +0100276 dma_data = snd_soc_dai_get_dma_data(rtd->cpu_dai, substream);
277
278 if (!pcm->chan[i] &&
279 (pcm->flags & SND_DMAENGINE_PCM_FLAG_CUSTOM_CHANNEL_NAME))
280 pcm->chan[i] = dma_request_slave_channel(dev,
281 dma_data->chan_name);
282
Lars-Peter Clausend1e14062013-04-20 19:29:00 +0200283 if (!pcm->chan[i] && (pcm->flags & SND_DMAENGINE_PCM_FLAG_COMPAT)) {
Lars-Peter Clausenc9998362013-04-15 19:19:51 +0200284 pcm->chan[i] = dmaengine_pcm_compat_request_channel(rtd,
285 substream);
286 }
287
Lars-Peter Clausen28c44682013-04-15 19:19:50 +0200288 if (!pcm->chan[i]) {
289 dev_err(rtd->platform->dev,
290 "Missing dma channel for stream: %d\n", i);
Lars-Peter Clausende7621e2015-01-02 13:56:07 +0100291 return -EINVAL;
Lars-Peter Clausen28c44682013-04-15 19:19:50 +0200292 }
293
294 ret = snd_pcm_lib_preallocate_pages(substream,
Nicolin Chenca2b0292013-11-07 14:45:16 +0800295 SNDRV_DMA_TYPE_DEV_IRAM,
Lars-Peter Clausen28c44682013-04-15 19:19:50 +0200296 dmaengine_dma_dev(pcm, substream),
Lars-Peter Clausenfa654e02013-10-08 15:08:00 +0200297 prealloc_buffer_size,
298 max_buffer_size);
Lars-Peter Clausen28c44682013-04-15 19:19:50 +0200299 if (ret)
Lars-Peter Clausende7621e2015-01-02 13:56:07 +0100300 return ret;
Lars-Peter Clausen478028e2014-01-11 14:02:19 +0100301
Lars-Peter Clausenacde50a2015-04-27 12:44:25 +0200302 if (!dmaengine_pcm_can_report_residue(dev, pcm->chan[i]))
Lars-Peter Clausen478028e2014-01-11 14:02:19 +0100303 pcm->flags |= SND_DMAENGINE_PCM_FLAG_NO_RESIDUE;
Lars-Peter Clausen28c44682013-04-15 19:19:50 +0200304 }
305
306 return 0;
Lars-Peter Clausen28c44682013-04-15 19:19:50 +0200307}
308
Lars-Peter Clausen93b943e2014-01-11 14:02:18 +0100309static snd_pcm_uframes_t dmaengine_pcm_pointer(
310 struct snd_pcm_substream *substream)
311{
312 struct snd_soc_pcm_runtime *rtd = substream->private_data;
313 struct dmaengine_pcm *pcm = soc_platform_to_pcm(rtd->platform);
314
315 if (pcm->flags & SND_DMAENGINE_PCM_FLAG_NO_RESIDUE)
316 return snd_dmaengine_pcm_pointer_no_residue(substream);
317 else
318 return snd_dmaengine_pcm_pointer(substream);
319}
320
Lars-Peter Clausen28c44682013-04-15 19:19:50 +0200321static const struct snd_pcm_ops dmaengine_pcm_ops = {
322 .open = dmaengine_pcm_open,
323 .close = snd_dmaengine_pcm_close,
324 .ioctl = snd_pcm_lib_ioctl,
325 .hw_params = dmaengine_pcm_hw_params,
326 .hw_free = snd_pcm_lib_free_pages,
327 .trigger = snd_dmaengine_pcm_trigger,
Lars-Peter Clausen93b943e2014-01-11 14:02:18 +0100328 .pointer = dmaengine_pcm_pointer,
Lars-Peter Clausen28c44682013-04-15 19:19:50 +0200329};
330
331static const struct snd_soc_platform_driver dmaengine_pcm_platform = {
Lars-Peter Clausenf1d45cc32014-08-19 15:51:19 +0200332 .component_driver = {
333 .probe_order = SND_SOC_COMP_ORDER_LATE,
334 },
Lars-Peter Clausen28c44682013-04-15 19:19:50 +0200335 .ops = &dmaengine_pcm_ops,
336 .pcm_new = dmaengine_pcm_new,
Lars-Peter Clausen28c44682013-04-15 19:19:50 +0200337};
338
339static const char * const dmaengine_pcm_dma_channel_names[] = {
340 [SNDRV_PCM_STREAM_PLAYBACK] = "tx",
341 [SNDRV_PCM_STREAM_CAPTURE] = "rx",
342};
343
Stephen Warren5eda87b2013-12-10 11:11:02 -0700344static int dmaengine_pcm_request_chan_of(struct dmaengine_pcm *pcm,
Stephen Warren194c7de2013-12-03 14:26:34 -0700345 struct device *dev, const struct snd_dmaengine_pcm_config *config)
Lars-Peter Clausend1e14062013-04-20 19:29:00 +0200346{
347 unsigned int i;
Stephen Warren11b3a7a2013-12-03 14:26:32 -0700348 const char *name;
Stephen Warren5eda87b2013-12-10 11:11:02 -0700349 struct dma_chan *chan;
Lars-Peter Clausend1e14062013-04-20 19:29:00 +0200350
Mark Brownea73b7d2013-10-19 17:43:51 +0100351 if ((pcm->flags & (SND_DMAENGINE_PCM_FLAG_NO_DT |
352 SND_DMAENGINE_PCM_FLAG_CUSTOM_CHANNEL_NAME)) ||
353 !dev->of_node)
Stephen Warren5eda87b2013-12-10 11:11:02 -0700354 return 0;
Lars-Peter Clausend1e14062013-04-20 19:29:00 +0200355
Xiubo Li2b67f8b2013-12-17 15:16:40 +0800356 if (config && config->dma_dev) {
Stephen Warren194c7de2013-12-03 14:26:34 -0700357 /*
358 * If this warning is seen, it probably means that your Linux
359 * device structure does not match your HW device structure.
360 * It would be best to refactor the Linux device structure to
361 * correctly match the HW structure.
362 */
363 dev_warn(dev, "DMA channels sourced from device %s",
364 dev_name(config->dma_dev));
365 dev = config->dma_dev;
366 }
367
Stephen Warren11b3a7a2013-12-03 14:26:32 -0700368 for (i = SNDRV_PCM_STREAM_PLAYBACK; i <= SNDRV_PCM_STREAM_CAPTURE;
369 i++) {
370 if (pcm->flags & SND_DMAENGINE_PCM_FLAG_HALF_DUPLEX)
371 name = "rx-tx";
372 else
373 name = dmaengine_pcm_dma_channel_names[i];
Xiubo Li2b67f8b2013-12-17 15:16:40 +0800374 if (config && config->chan_names[i])
Stephen Warren194c7de2013-12-03 14:26:34 -0700375 name = config->chan_names[i];
Stephen Warren5eda87b2013-12-10 11:11:02 -0700376 chan = dma_request_slave_channel_reason(dev, name);
377 if (IS_ERR(chan)) {
Stephen Warrene9036c22013-12-11 11:20:50 -0700378 if (PTR_ERR(chan) == -EPROBE_DEFER)
Stephen Warren5eda87b2013-12-10 11:11:02 -0700379 return -EPROBE_DEFER;
380 pcm->chan[i] = NULL;
381 } else {
382 pcm->chan[i] = chan;
383 }
Stephen Warren11b3a7a2013-12-03 14:26:32 -0700384 if (pcm->flags & SND_DMAENGINE_PCM_FLAG_HALF_DUPLEX)
385 break;
Lars-Peter Clausend1e14062013-04-20 19:29:00 +0200386 }
Stephen Warren11b3a7a2013-12-03 14:26:32 -0700387
388 if (pcm->flags & SND_DMAENGINE_PCM_FLAG_HALF_DUPLEX)
389 pcm->chan[1] = pcm->chan[0];
Stephen Warren5eda87b2013-12-10 11:11:02 -0700390
391 return 0;
Lars-Peter Clausend1e14062013-04-20 19:29:00 +0200392}
393
Stephen Warren6b9f3e62013-12-03 14:26:33 -0700394static void dmaengine_pcm_release_chan(struct dmaengine_pcm *pcm)
395{
396 unsigned int i;
397
398 for (i = SNDRV_PCM_STREAM_PLAYBACK; i <= SNDRV_PCM_STREAM_CAPTURE;
399 i++) {
400 if (!pcm->chan[i])
401 continue;
402 dma_release_channel(pcm->chan[i]);
403 if (pcm->flags & SND_DMAENGINE_PCM_FLAG_HALF_DUPLEX)
404 break;
405 }
406}
407
Lars-Peter Clausen28c44682013-04-15 19:19:50 +0200408/**
409 * snd_dmaengine_pcm_register - Register a dmaengine based PCM device
410 * @dev: The parent device for the PCM device
411 * @config: Platform specific PCM configuration
412 * @flags: Platform specific quirks
413 */
414int snd_dmaengine_pcm_register(struct device *dev,
415 const struct snd_dmaengine_pcm_config *config, unsigned int flags)
416{
417 struct dmaengine_pcm *pcm;
Stephen Warren6b9f3e62013-12-03 14:26:33 -0700418 int ret;
Lars-Peter Clausen28c44682013-04-15 19:19:50 +0200419
Lars-Peter Clausen28c44682013-04-15 19:19:50 +0200420 pcm = kzalloc(sizeof(*pcm), GFP_KERNEL);
421 if (!pcm)
422 return -ENOMEM;
423
424 pcm->config = config;
Lars-Peter Clausend1e14062013-04-20 19:29:00 +0200425 pcm->flags = flags;
Lars-Peter Clausen28c44682013-04-15 19:19:50 +0200426
Stephen Warren5eda87b2013-12-10 11:11:02 -0700427 ret = dmaengine_pcm_request_chan_of(pcm, dev, config);
428 if (ret)
429 goto err_free_dma;
Lars-Peter Clausen28c44682013-04-15 19:19:50 +0200430
Lars-Peter Clausen93b943e2014-01-11 14:02:18 +0100431 ret = snd_soc_add_platform(dev, &pcm->platform,
432 &dmaengine_pcm_platform);
Stephen Warren6b9f3e62013-12-03 14:26:33 -0700433 if (ret)
434 goto err_free_dma;
435
436 return 0;
437
438err_free_dma:
439 dmaengine_pcm_release_chan(pcm);
440 kfree(pcm);
441 return ret;
Lars-Peter Clausen28c44682013-04-15 19:19:50 +0200442}
443EXPORT_SYMBOL_GPL(snd_dmaengine_pcm_register);
444
445/**
446 * snd_dmaengine_pcm_unregister - Removes a dmaengine based PCM device
447 * @dev: Parent device the PCM was register with
448 *
449 * Removes a dmaengine based PCM device previously registered with
450 * snd_dmaengine_pcm_register.
451 */
452void snd_dmaengine_pcm_unregister(struct device *dev)
453{
454 struct snd_soc_platform *platform;
455 struct dmaengine_pcm *pcm;
Lars-Peter Clausen28c44682013-04-15 19:19:50 +0200456
457 platform = snd_soc_lookup_platform(dev);
458 if (!platform)
459 return;
460
461 pcm = soc_platform_to_pcm(platform);
462
Lars-Peter Clausen28c44682013-04-15 19:19:50 +0200463 snd_soc_remove_platform(platform);
Stephen Warren6b9f3e62013-12-03 14:26:33 -0700464 dmaengine_pcm_release_chan(pcm);
Lars-Peter Clausen28c44682013-04-15 19:19:50 +0200465 kfree(pcm);
466}
467EXPORT_SYMBOL_GPL(snd_dmaengine_pcm_unregister);
468
469MODULE_LICENSE("GPL");