]> asedeno.scripts.mit.edu Git - linux.git/blob - sound/soc/sprd/sprd-pcm-compress.c
Merge tag 'mips_fixes_5.2_2' of git://git.kernel.org/pub/scm/linux/kernel/git/mips/linux
[linux.git] / sound / soc / sprd / sprd-pcm-compress.c
1 // SPDX-License-Identifier: GPL-2.0
2 // Copyright (C) 2019 Spreadtrum Communications Inc.
3
4 #include <linux/dma-mapping.h>
5 #include <linux/dmaengine.h>
6 #include <linux/dma/sprd-dma.h>
7 #include <linux/kernel.h>
8 #include <linux/module.h>
9 #include <sound/pcm.h>
10 #include <sound/pcm_params.h>
11 #include <sound/soc.h>
12 #include <sound/compress_driver.h>
13
14 #include "sprd-pcm-dma.h"
15
16 #define SPRD_COMPR_DMA_CHANS            2
17
18 /* Default values if userspace does not set */
19 #define SPRD_COMPR_MIN_FRAGMENT_SIZE    SZ_8K
20 #define SPRD_COMPR_MAX_FRAGMENT_SIZE    SZ_128K
21 #define SPRD_COMPR_MIN_NUM_FRAGMENTS    4
22 #define SPRD_COMPR_MAX_NUM_FRAGMENTS    64
23
24 /* DSP FIFO size */
25 #define SPRD_COMPR_MCDT_EMPTY_WMK       0
26 #define SPRD_COMPR_MCDT_FIFO_SIZE       512
27
28 /* Stage 0 IRAM buffer size definition */
29 #define SPRD_COMPR_IRAM_BUF_SIZE        SZ_32K
30 #define SPRD_COMPR_IRAM_INFO_SIZE       (sizeof(struct sprd_compr_playinfo))
31 #define SPRD_COMPR_IRAM_LINKLIST_SIZE   (1024 - SPRD_COMPR_IRAM_INFO_SIZE)
32 #define SPRD_COMPR_IRAM_SIZE            (SPRD_COMPR_IRAM_BUF_SIZE + \
33                                          SPRD_COMPR_IRAM_INFO_SIZE + \
34                                          SPRD_COMPR_IRAM_LINKLIST_SIZE)
35
36 /* Stage 1 DDR buffer size definition */
37 #define SPRD_COMPR_AREA_BUF_SIZE        SZ_2M
38 #define SPRD_COMPR_AREA_LINKLIST_SIZE   1024
39 #define SPRD_COMPR_AREA_SIZE            (SPRD_COMPR_AREA_BUF_SIZE + \
40                                          SPRD_COMPR_AREA_LINKLIST_SIZE)
41
42 struct sprd_compr_dma {
43         struct dma_chan *chan;
44         struct dma_async_tx_descriptor *desc;
45         dma_cookie_t cookie;
46         dma_addr_t phys;
47         void *virt;
48         int trans_len;
49 };
50
51 /*
52  * The Spreadtrum Audio compress offload mode will use 2-stage DMA transfer to
53  * save power. That means we can request 2 dma channels, one for source channel,
54  * and another one for destination channel. Once the source channel's transaction
55  * is done, it will trigger the destination channel's transaction automatically
56  * by hardware signal.
57  *
58  * For 2-stage DMA transfer, we can allocate 2 buffers: IRAM buffer (always
59  * power-on) and DDR buffer. The source channel will transfer data from IRAM
60  * buffer to the DSP fifo to decoding/encoding, once IRAM buffer is empty by
61  * transferring done, the destination channel will start to transfer data from
62  * DDR buffer to IRAM buffer.
63  *
64  * Since the DSP fifo is only 512B, IRAM buffer is allocated by 32K, and DDR
65  * buffer is larger to 2M. That means only the IRAM 32k data is transferred
66  * done, we can wake up the AP system to transfer data from DDR to IRAM, and
67  * other time the AP system can be suspended to save power.
68  */
69 struct sprd_compr_stream {
70         struct snd_compr_stream *cstream;
71         struct sprd_compr_ops *compr_ops;
72         struct sprd_compr_dma dma[SPRD_COMPR_DMA_CHANS];
73
74         /* DMA engine channel number */
75         int num_channels;
76
77         /* Stage 0 IRAM buffer */
78         struct snd_dma_buffer iram_buffer;
79         /* Stage 1 DDR buffer */
80         struct snd_dma_buffer compr_buffer;
81
82         /* DSP play information IRAM buffer */
83         dma_addr_t info_phys;
84         void *info_area;
85         int info_size;
86
87         /* Data size copied to IRAM buffer */
88         int copied_total;
89         /* Total received data size from userspace */
90         int received_total;
91         /* Stage 0 IRAM buffer received data size */
92         int received_stage0;
93         /* Stage 1 DDR buffer received data size */
94         int received_stage1;
95         /* Stage 1 DDR buffer pointer */
96         int stage1_pointer;
97 };
98
99 static int sprd_platform_compr_trigger(struct snd_compr_stream *cstream,
100                                        int cmd);
101
102 static void sprd_platform_compr_drain_notify(void *arg)
103 {
104         struct snd_compr_stream *cstream = arg;
105         struct snd_compr_runtime *runtime = cstream->runtime;
106         struct sprd_compr_stream *stream = runtime->private_data;
107
108         memset(stream->info_area, 0, sizeof(struct sprd_compr_playinfo));
109
110         snd_compr_drain_notify(cstream);
111 }
112
113 static void sprd_platform_compr_dma_complete(void *data)
114 {
115         struct snd_compr_stream *cstream = data;
116         struct snd_compr_runtime *runtime = cstream->runtime;
117         struct sprd_compr_stream *stream = runtime->private_data;
118         struct sprd_compr_dma *dma = &stream->dma[1];
119
120         /* Update data size copied to IRAM buffer */
121         stream->copied_total += dma->trans_len;
122         if (stream->copied_total > stream->received_total)
123                 stream->copied_total = stream->received_total;
124
125         snd_compr_fragment_elapsed(cstream);
126 }
127
128 static int sprd_platform_compr_dma_config(struct snd_compr_stream *cstream,
129                                           struct snd_compr_params *params,
130                                           int channel)
131 {
132         struct snd_compr_runtime *runtime = cstream->runtime;
133         struct sprd_compr_stream *stream = runtime->private_data;
134         struct snd_soc_pcm_runtime *rtd = cstream->private_data;
135         struct snd_soc_component *component =
136                 snd_soc_rtdcom_lookup(rtd, DRV_NAME);
137         struct device *dev = component->dev;
138         struct sprd_compr_data *data = snd_soc_dai_get_drvdata(rtd->cpu_dai);
139         struct sprd_pcm_dma_params *dma_params = data->dma_params;
140         struct sprd_compr_dma *dma = &stream->dma[channel];
141         struct dma_slave_config config = { };
142         struct sprd_dma_linklist link = { };
143         enum dma_transfer_direction dir;
144         struct scatterlist *sg, *sgt;
145         enum dma_slave_buswidth bus_width;
146         int period, period_cnt, sg_num = 2;
147         dma_addr_t src_addr, dst_addr;
148         unsigned long flags;
149         int ret, j;
150
151         if (!dma_params) {
152                 dev_err(dev, "no dma parameters setting\n");
153                 return -EINVAL;
154         }
155
156         dma->chan = dma_request_slave_channel(dev,
157                                               dma_params->chan_name[channel]);
158         if (!dma->chan) {
159                 dev_err(dev, "failed to request dma channel\n");
160                 return -ENODEV;
161         }
162
163         sgt = sg = devm_kcalloc(dev, sg_num, sizeof(*sg), GFP_KERNEL);
164         if (!sg) {
165                 ret = -ENOMEM;
166                 goto sg_err;
167         }
168
169         switch (channel) {
170         case 0:
171                 bus_width = DMA_SLAVE_BUSWIDTH_4_BYTES;
172                 period = (SPRD_COMPR_MCDT_FIFO_SIZE - SPRD_COMPR_MCDT_EMPTY_WMK) * 4;
173                 period_cnt = params->buffer.fragment_size / period;
174                 src_addr = stream->iram_buffer.addr;
175                 dst_addr = dma_params->dev_phys[channel];
176                 flags = SPRD_DMA_FLAGS(SPRD_DMA_SRC_CHN1,
177                                        SPRD_DMA_TRANS_DONE_TRG,
178                                        SPRD_DMA_FRAG_REQ,
179                                        SPRD_DMA_TRANS_INT);
180                 break;
181
182         case 1:
183                 bus_width = DMA_SLAVE_BUSWIDTH_2_BYTES;
184                 period = params->buffer.fragment_size;
185                 period_cnt = params->buffer.fragments;
186                 src_addr = stream->compr_buffer.addr;
187                 dst_addr = stream->iram_buffer.addr;
188                 flags = SPRD_DMA_FLAGS(SPRD_DMA_DST_CHN1,
189                                        SPRD_DMA_TRANS_DONE_TRG,
190                                        SPRD_DMA_FRAG_REQ,
191                                        SPRD_DMA_TRANS_INT);
192                 break;
193
194         default:
195                 ret = -EINVAL;
196                 goto config_err;
197         }
198
199         dma->trans_len = period * period_cnt;
200
201         config.src_maxburst = period;
202         config.src_addr_width = bus_width;
203         config.dst_addr_width = bus_width;
204         if (cstream->direction == SND_COMPRESS_PLAYBACK) {
205                 config.src_addr = src_addr;
206                 config.dst_addr = dst_addr;
207                 dir = DMA_MEM_TO_DEV;
208         } else {
209                 config.src_addr = dst_addr;
210                 config.dst_addr = src_addr;
211                 dir = DMA_DEV_TO_MEM;
212         }
213
214         sg_init_table(sgt, sg_num);
215         for (j = 0; j < sg_num; j++, sgt++) {
216                 sg_dma_len(sgt) = dma->trans_len;
217                 sg_dma_address(sgt) = dst_addr;
218         }
219
220         /*
221          * Configure the link-list address for the DMA engine link-list
222          * mode.
223          */
224         link.virt_addr = (unsigned long)dma->virt;
225         link.phy_addr = dma->phys;
226
227         ret = dmaengine_slave_config(dma->chan, &config);
228         if (ret) {
229                 dev_err(dev,
230                         "failed to set slave configuration: %d\n", ret);
231                 goto config_err;
232         }
233
234         /*
235          * We configure the DMA request mode, interrupt mode, channel
236          * mode and channel trigger mode by the flags.
237          */
238         dma->desc = dma->chan->device->device_prep_slave_sg(dma->chan, sg,
239                                                             sg_num, dir,
240                                                             flags, &link);
241         if (!dma->desc) {
242                 dev_err(dev, "failed to prepare slave sg\n");
243                 ret = -ENOMEM;
244                 goto config_err;
245         }
246
247         /* Only channel 1 transfer can wake up the AP system. */
248         if (!params->no_wake_mode && channel == 1) {
249                 dma->desc->callback = sprd_platform_compr_dma_complete;
250                 dma->desc->callback_param = cstream;
251         }
252
253         devm_kfree(dev, sg);
254
255         return 0;
256
257 config_err:
258         devm_kfree(dev, sg);
259 sg_err:
260         dma_release_channel(dma->chan);
261         return ret;
262 }
263
264 static int sprd_platform_compr_set_params(struct snd_compr_stream *cstream,
265                                           struct snd_compr_params *params)
266 {
267         struct snd_compr_runtime *runtime = cstream->runtime;
268         struct sprd_compr_stream *stream = runtime->private_data;
269         struct snd_soc_pcm_runtime *rtd = cstream->private_data;
270         struct snd_soc_component *component =
271                 snd_soc_rtdcom_lookup(rtd, DRV_NAME);
272         struct device *dev = component->dev;
273         struct sprd_compr_params compr_params = { };
274         int ret;
275
276         /*
277          * Configure the DMA engine 2-stage transfer mode. Channel 1 set as the
278          * destination channel, and channel 0 set as the source channel, that
279          * means once the source channel's transaction is done, it will trigger
280          * the destination channel's transaction automatically.
281          */
282         ret = sprd_platform_compr_dma_config(cstream, params, 1);
283         if (ret) {
284                 dev_err(dev, "failed to config stage 1 DMA: %d\n", ret);
285                 return ret;
286         }
287
288         ret = sprd_platform_compr_dma_config(cstream, params, 0);
289         if (ret) {
290                 dev_err(dev, "failed to config stage 0 DMA: %d\n", ret);
291                 goto config_err;
292         }
293
294         compr_params.direction = cstream->direction;
295         compr_params.sample_rate = params->codec.sample_rate;
296         compr_params.channels = stream->num_channels;
297         compr_params.info_phys = stream->info_phys;
298         compr_params.info_size = stream->info_size;
299         compr_params.rate = params->codec.bit_rate;
300         compr_params.format = params->codec.id;
301
302         ret = stream->compr_ops->set_params(cstream->direction, &compr_params);
303         if (ret) {
304                 dev_err(dev, "failed to set parameters: %d\n", ret);
305                 goto params_err;
306         }
307
308         return 0;
309
310 params_err:
311         dma_release_channel(stream->dma[0].chan);
312 config_err:
313         dma_release_channel(stream->dma[1].chan);
314         return ret;
315 }
316
317 static int sprd_platform_compr_open(struct snd_compr_stream *cstream)
318 {
319         struct snd_compr_runtime *runtime = cstream->runtime;
320         struct snd_soc_pcm_runtime *rtd = cstream->private_data;
321         struct snd_soc_component *component =
322                 snd_soc_rtdcom_lookup(rtd, DRV_NAME);
323         struct device *dev = component->dev;
324         struct sprd_compr_data *data = snd_soc_dai_get_drvdata(rtd->cpu_dai);
325         struct sprd_compr_stream *stream;
326         struct sprd_compr_callback cb;
327         int stream_id = cstream->direction, ret;
328
329         ret = dma_coerce_mask_and_coherent(dev, DMA_BIT_MASK(32));
330         if (ret)
331                 return ret;
332
333         stream = devm_kzalloc(dev, sizeof(*stream), GFP_KERNEL);
334         if (!stream)
335                 return -ENOMEM;
336
337         stream->cstream = cstream;
338         stream->num_channels = 2;
339         stream->compr_ops = data->ops;
340
341         /*
342          * Allocate the stage 0 IRAM buffer size, including the DMA 0
343          * link-list size and play information of DSP address size.
344          */
345         ret = snd_dma_alloc_pages(SNDRV_DMA_TYPE_DEV_IRAM, dev,
346                                   SPRD_COMPR_IRAM_SIZE, &stream->iram_buffer);
347         if (ret < 0)
348                 goto err_iram;
349
350         /* Use to save link-list configuration for DMA 0. */
351         stream->dma[0].virt = stream->iram_buffer.area + SPRD_COMPR_IRAM_SIZE;
352         stream->dma[0].phys = stream->iram_buffer.addr + SPRD_COMPR_IRAM_SIZE;
353
354         /* Use to update the current data offset of DSP. */
355         stream->info_phys = stream->iram_buffer.addr + SPRD_COMPR_IRAM_SIZE +
356                 SPRD_COMPR_IRAM_LINKLIST_SIZE;
357         stream->info_area = stream->iram_buffer.area + SPRD_COMPR_IRAM_SIZE +
358                 SPRD_COMPR_IRAM_LINKLIST_SIZE;
359         stream->info_size = SPRD_COMPR_IRAM_INFO_SIZE;
360
361         /*
362          * Allocate the stage 1 DDR buffer size, including the DMA 1 link-list
363          * size.
364          */
365         ret = snd_dma_alloc_pages(SNDRV_DMA_TYPE_DEV, dev,
366                                   SPRD_COMPR_AREA_SIZE, &stream->compr_buffer);
367         if (ret < 0)
368                 goto err_compr;
369
370         /* Use to save link-list configuration for DMA 1. */
371         stream->dma[1].virt = stream->compr_buffer.area + SPRD_COMPR_AREA_SIZE;
372         stream->dma[1].phys = stream->compr_buffer.addr + SPRD_COMPR_AREA_SIZE;
373
374         cb.drain_notify = sprd_platform_compr_drain_notify;
375         cb.drain_data = cstream;
376         ret = stream->compr_ops->open(stream_id, &cb);
377         if (ret) {
378                 dev_err(dev, "failed to open compress platform: %d\n", ret);
379                 goto err_open;
380         }
381
382         runtime->private_data = stream;
383         return 0;
384
385 err_open:
386         snd_dma_free_pages(&stream->compr_buffer);
387 err_compr:
388         snd_dma_free_pages(&stream->iram_buffer);
389 err_iram:
390         devm_kfree(dev, stream);
391
392         return ret;
393 }
394
395 static int sprd_platform_compr_free(struct snd_compr_stream *cstream)
396 {
397         struct snd_compr_runtime *runtime = cstream->runtime;
398         struct sprd_compr_stream *stream = runtime->private_data;
399         struct snd_soc_pcm_runtime *rtd = cstream->private_data;
400         struct snd_soc_component *component =
401                 snd_soc_rtdcom_lookup(rtd, DRV_NAME);
402         struct device *dev = component->dev;
403         int stream_id = cstream->direction, i;
404
405         for (i = 0; i < stream->num_channels; i++) {
406                 struct sprd_compr_dma *dma = &stream->dma[i];
407
408                 if (dma->chan) {
409                         dma_release_channel(dma->chan);
410                         dma->chan = NULL;
411                 }
412         }
413
414         snd_dma_free_pages(&stream->compr_buffer);
415         snd_dma_free_pages(&stream->iram_buffer);
416
417         stream->compr_ops->close(stream_id);
418
419         devm_kfree(dev, stream);
420         return 0;
421 }
422
423 static int sprd_platform_compr_trigger(struct snd_compr_stream *cstream,
424                                        int cmd)
425 {
426         struct snd_compr_runtime *runtime = cstream->runtime;
427         struct sprd_compr_stream *stream = runtime->private_data;
428         struct snd_soc_pcm_runtime *rtd = cstream->private_data;
429         struct snd_soc_component *component =
430                 snd_soc_rtdcom_lookup(rtd, DRV_NAME);
431         struct device *dev = component->dev;
432         int channels = stream->num_channels, ret = 0, i;
433         int stream_id = cstream->direction;
434
435         if (cstream->direction != SND_COMPRESS_PLAYBACK) {
436                 dev_err(dev, "unsupported compress direction\n");
437                 return -EINVAL;
438         }
439
440         switch (cmd) {
441         case SNDRV_PCM_TRIGGER_START:
442                 for (i = channels - 1; i >= 0; i--) {
443                         struct sprd_compr_dma *dma = &stream->dma[i];
444
445                         if (!dma->desc)
446                                 continue;
447
448                         dma->cookie = dmaengine_submit(dma->desc);
449                         ret = dma_submit_error(dma->cookie);
450                         if (ret) {
451                                 dev_err(dev, "failed to submit request: %d\n",
452                                         ret);
453                                 return ret;
454                         }
455                 }
456
457                 for (i = channels - 1; i >= 0; i--) {
458                         struct sprd_compr_dma *dma = &stream->dma[i];
459
460                         if (dma->chan)
461                                 dma_async_issue_pending(dma->chan);
462                 }
463
464                 ret = stream->compr_ops->start(stream_id);
465                 break;
466
467         case SNDRV_PCM_TRIGGER_STOP:
468                 for (i = channels - 1; i >= 0; i--) {
469                         struct sprd_compr_dma *dma = &stream->dma[i];
470
471                         if (dma->chan)
472                                 dmaengine_terminate_async(dma->chan);
473                 }
474
475                 stream->copied_total = 0;
476                 stream->stage1_pointer  = 0;
477                 stream->received_total = 0;
478                 stream->received_stage0 = 0;
479                 stream->received_stage1 = 0;
480
481                 ret = stream->compr_ops->stop(stream_id);
482                 break;
483
484         case SNDRV_PCM_TRIGGER_SUSPEND:
485         case SNDRV_PCM_TRIGGER_PAUSE_PUSH:
486                 for (i = channels - 1; i >= 0; i--) {
487                         struct sprd_compr_dma *dma = &stream->dma[i];
488
489                         if (dma->chan)
490                                 dmaengine_pause(dma->chan);
491                 }
492
493                 ret = stream->compr_ops->pause(stream_id);
494                 break;
495
496         case SNDRV_PCM_TRIGGER_RESUME:
497         case SNDRV_PCM_TRIGGER_PAUSE_RELEASE:
498                 for (i = channels - 1; i >= 0; i--) {
499                         struct sprd_compr_dma *dma = &stream->dma[i];
500
501                         if (dma->chan)
502                                 dmaengine_resume(dma->chan);
503                 }
504
505                 ret = stream->compr_ops->pause_release(stream_id);
506                 break;
507
508         case SND_COMPR_TRIGGER_PARTIAL_DRAIN:
509         case SND_COMPR_TRIGGER_DRAIN:
510                 ret = stream->compr_ops->drain(stream->received_total);
511                 break;
512
513         default:
514                 ret = -EINVAL;
515                 break;
516         }
517
518         return ret;
519 }
520
521 static int sprd_platform_compr_pointer(struct snd_compr_stream *cstream,
522                                        struct snd_compr_tstamp *tstamp)
523 {
524         struct snd_compr_runtime *runtime = cstream->runtime;
525         struct sprd_compr_stream *stream = runtime->private_data;
526         struct sprd_compr_playinfo *info =
527                 (struct sprd_compr_playinfo *)stream->info_area;
528
529         tstamp->copied_total = stream->copied_total;
530         tstamp->pcm_io_frames = info->current_data_offset;
531
532         return 0;
533 }
534
535 static int sprd_platform_compr_copy(struct snd_compr_stream *cstream,
536                                     char __user *buf, size_t count)
537 {
538         struct snd_compr_runtime *runtime = cstream->runtime;
539         struct sprd_compr_stream *stream = runtime->private_data;
540         int avail_bytes, data_count = count;
541         void *dst;
542
543         /*
544          * We usually set fragment size as 32K, and the stage 0 IRAM buffer
545          * size is 32K too. So if now the received data size of the stage 0
546          * IRAM buffer is less than 32K, that means we have some available
547          * spaces for the stage 0 IRAM buffer.
548          */
549         if (stream->received_stage0 < runtime->fragment_size) {
550                 avail_bytes = runtime->fragment_size - stream->received_stage0;
551                 dst = stream->iram_buffer.area + stream->received_stage0;
552
553                 if (avail_bytes >= data_count) {
554                         /*
555                          * Copy data to the stage 0 IRAM buffer directly if
556                          * spaces are enough.
557                          */
558                         if (copy_from_user(dst, buf, data_count))
559                                 return -EFAULT;
560
561                         stream->received_stage0 += data_count;
562                         stream->copied_total += data_count;
563                         goto copy_done;
564                 } else {
565                         /*
566                          * If the data count is larger than the available spaces
567                          * of the the stage 0 IRAM buffer, we should copy one
568                          * partial data to the stage 0 IRAM buffer, and copy
569                          * the left to the stage 1 DDR buffer.
570                          */
571                         if (copy_from_user(dst, buf, avail_bytes))
572                                 return -EFAULT;
573
574                         data_count -= avail_bytes;
575                         stream->received_stage0 += avail_bytes;
576                         stream->copied_total += avail_bytes;
577                         buf += avail_bytes;
578                 }
579         }
580
581         /*
582          * Copy data to the stage 1 DDR buffer if no spaces for the stage 0 IRAM
583          * buffer.
584          */
585         dst = stream->compr_buffer.area + stream->stage1_pointer;
586         if (data_count < stream->compr_buffer.bytes - stream->stage1_pointer) {
587                 if (copy_from_user(dst, buf, data_count))
588                         return -EFAULT;
589
590                 stream->stage1_pointer += data_count;
591         } else {
592                 avail_bytes = stream->compr_buffer.bytes - stream->stage1_pointer;
593
594                 if (copy_from_user(dst, buf, avail_bytes))
595                         return -EFAULT;
596
597                 if (copy_from_user(stream->compr_buffer.area, buf + avail_bytes,
598                                    data_count - avail_bytes))
599                         return -EFAULT;
600
601                 stream->stage1_pointer = data_count - avail_bytes;
602         }
603
604         stream->received_stage1 += data_count;
605
606 copy_done:
607         /* Update the copied data size. */
608         stream->received_total += count;
609         return count;
610 }
611
612 static int sprd_platform_compr_get_caps(struct snd_compr_stream *cstream,
613                                         struct snd_compr_caps *caps)
614 {
615         caps->direction = cstream->direction;
616         caps->min_fragment_size = SPRD_COMPR_MIN_FRAGMENT_SIZE;
617         caps->max_fragment_size = SPRD_COMPR_MAX_FRAGMENT_SIZE;
618         caps->min_fragments = SPRD_COMPR_MIN_NUM_FRAGMENTS;
619         caps->max_fragments = SPRD_COMPR_MAX_NUM_FRAGMENTS;
620         caps->num_codecs = 2;
621         caps->codecs[0] = SND_AUDIOCODEC_MP3;
622         caps->codecs[1] = SND_AUDIOCODEC_AAC;
623
624         return 0;
625 }
626
627 static int
628 sprd_platform_compr_get_codec_caps(struct snd_compr_stream *cstream,
629                                    struct snd_compr_codec_caps *codec)
630 {
631         switch (codec->codec) {
632         case SND_AUDIOCODEC_MP3:
633                 codec->num_descriptors = 2;
634                 codec->descriptor[0].max_ch = 2;
635                 codec->descriptor[0].bit_rate[0] = 320;
636                 codec->descriptor[0].bit_rate[1] = 128;
637                 codec->descriptor[0].num_bitrates = 2;
638                 codec->descriptor[0].profiles = 0;
639                 codec->descriptor[0].modes = SND_AUDIOCHANMODE_MP3_STEREO;
640                 codec->descriptor[0].formats = 0;
641                 break;
642
643         case SND_AUDIOCODEC_AAC:
644                 codec->num_descriptors = 2;
645                 codec->descriptor[1].max_ch = 2;
646                 codec->descriptor[1].bit_rate[0] = 320;
647                 codec->descriptor[1].bit_rate[1] = 128;
648                 codec->descriptor[1].num_bitrates = 2;
649                 codec->descriptor[1].profiles = 0;
650                 codec->descriptor[1].modes = 0;
651                 codec->descriptor[1].formats = 0;
652                 break;
653
654         default:
655                 return -EINVAL;
656         }
657
658         return 0;
659 }
660
661 const struct snd_compr_ops sprd_platform_compr_ops = {
662         .open = sprd_platform_compr_open,
663         .free = sprd_platform_compr_free,
664         .set_params = sprd_platform_compr_set_params,
665         .trigger = sprd_platform_compr_trigger,
666         .pointer = sprd_platform_compr_pointer,
667         .copy = sprd_platform_compr_copy,
668         .get_caps = sprd_platform_compr_get_caps,
669         .get_codec_caps = sprd_platform_compr_get_codec_caps,
670 };
671
672 MODULE_DESCRIPTION("Spreadtrum ASoC Compress Platform Driver");
673 MODULE_LICENSE("GPL v2");
674 MODULE_ALIAS("platform:compress-platform");