| // SPDX-License-Identifier: GPL-2.0 |
| /* |
| * mtk-afe-fe-dais.c -- Mediatek afe fe dai operator |
| * |
| * Copyright (c) 2016 MediaTek Inc. |
| * Author: Garlic Tseng <garlic.tseng@mediatek.com> |
| */ |
| |
| #include <linux/io.h> |
| #include <linux/module.h> |
| #include <linux/pm_runtime.h> |
| #include <linux/regmap.h> |
| #include <sound/soc.h> |
| #include "mtk-afe-platform-driver.h" |
| #include <sound/pcm_params.h> |
| #include "mtk-afe-fe-dai.h" |
| #include "mtk-base-afe.h" |
| |
| #define AFE_BASE_END_OFFSET 8 |
| |
| static int mtk_regmap_update_bits(struct regmap *map, int reg, |
| unsigned int mask, |
| unsigned int val, int shift) |
| { |
| if (reg < 0 || WARN_ON_ONCE(shift < 0)) |
| return 0; |
| return regmap_update_bits(map, reg, mask << shift, val << shift); |
| } |
| |
| static int mtk_regmap_write(struct regmap *map, int reg, unsigned int val) |
| { |
| if (reg < 0) |
| return 0; |
| return regmap_write(map, reg, val); |
| } |
| |
| int mtk_afe_fe_startup(struct snd_pcm_substream *substream, |
| struct snd_soc_dai *dai) |
| { |
| struct snd_soc_pcm_runtime *rtd = asoc_substream_to_rtd(substream); |
| struct mtk_base_afe *afe = snd_soc_dai_get_drvdata(dai); |
| struct snd_pcm_runtime *runtime = substream->runtime; |
| int memif_num = asoc_rtd_to_cpu(rtd, 0)->id; |
| struct mtk_base_afe_memif *memif = &afe->memif[memif_num]; |
| const struct snd_pcm_hardware *mtk_afe_hardware = afe->mtk_afe_hardware; |
| int ret; |
| |
| memif->substream = substream; |
| |
| snd_pcm_hw_constraint_step(substream->runtime, 0, |
| SNDRV_PCM_HW_PARAM_BUFFER_BYTES, 16); |
| /* enable agent */ |
| mtk_regmap_update_bits(afe->regmap, memif->data->agent_disable_reg, |
| 1, 0, memif->data->agent_disable_shift); |
| |
| snd_soc_set_runtime_hwparams(substream, mtk_afe_hardware); |
| |
| /* |
| * Capture cannot use ping-pong buffer since hw_ptr at IRQ may be |
| * smaller than period_size due to AFE's internal buffer. |
| * This easily leads to overrun when avail_min is period_size. |
| * One more period can hold the possible unread buffer. |
| */ |
| if (substream->stream == SNDRV_PCM_STREAM_CAPTURE) { |
| int periods_max = mtk_afe_hardware->periods_max; |
| |
| ret = snd_pcm_hw_constraint_minmax(runtime, |
| SNDRV_PCM_HW_PARAM_PERIODS, |
| 3, periods_max); |
| if (ret < 0) { |
| dev_err(afe->dev, "hw_constraint_minmax failed\n"); |
| return ret; |
| } |
| } |
| |
| ret = snd_pcm_hw_constraint_integer(runtime, |
| SNDRV_PCM_HW_PARAM_PERIODS); |
| if (ret < 0) |
| dev_err(afe->dev, "snd_pcm_hw_constraint_integer failed\n"); |
| |
| /* dynamic allocate irq to memif */ |
| if (memif->irq_usage < 0) { |
| int irq_id = mtk_dynamic_irq_acquire(afe); |
| |
| if (irq_id != afe->irqs_size) { |
| /* link */ |
| memif->irq_usage = irq_id; |
| } else { |
| dev_err(afe->dev, "%s() error: no more asys irq\n", |
| __func__); |
| ret = -EBUSY; |
| } |
| } |
| return ret; |
| } |
| EXPORT_SYMBOL_GPL(mtk_afe_fe_startup); |
| |
| void mtk_afe_fe_shutdown(struct snd_pcm_substream *substream, |
| struct snd_soc_dai *dai) |
| { |
| struct snd_soc_pcm_runtime *rtd = asoc_substream_to_rtd(substream); |
| struct mtk_base_afe *afe = snd_soc_dai_get_drvdata(dai); |
| struct mtk_base_afe_memif *memif = &afe->memif[asoc_rtd_to_cpu(rtd, 0)->id]; |
| int irq_id; |
| |
| irq_id = memif->irq_usage; |
| |
| mtk_regmap_update_bits(afe->regmap, memif->data->agent_disable_reg, |
| 1, 1, memif->data->agent_disable_shift); |
| |
| if (!memif->const_irq) { |
| mtk_dynamic_irq_release(afe, irq_id); |
| memif->irq_usage = -1; |
| memif->substream = NULL; |
| } |
| } |
| EXPORT_SYMBOL_GPL(mtk_afe_fe_shutdown); |
| |
| int mtk_afe_fe_hw_params(struct snd_pcm_substream *substream, |
| struct snd_pcm_hw_params *params, |
| struct snd_soc_dai *dai) |
| { |
| struct snd_soc_pcm_runtime *rtd = asoc_substream_to_rtd(substream); |
| struct mtk_base_afe *afe = snd_soc_dai_get_drvdata(dai); |
| int id = asoc_rtd_to_cpu(rtd, 0)->id; |
| struct mtk_base_afe_memif *memif = &afe->memif[id]; |
| int ret; |
| unsigned int channels = params_channels(params); |
| unsigned int rate = params_rate(params); |
| snd_pcm_format_t format = params_format(params); |
| |
| if (afe->request_dram_resource) |
| afe->request_dram_resource(afe->dev); |
| |
| dev_dbg(afe->dev, "%s(), %s, ch %d, rate %d, fmt %d, dma_addr %pad, dma_area %p, dma_bytes 0x%zx\n", |
| __func__, memif->data->name, |
| channels, rate, format, |
| &substream->runtime->dma_addr, |
| substream->runtime->dma_area, |
| substream->runtime->dma_bytes); |
| |
| memset_io((void __force __iomem *)substream->runtime->dma_area, 0, |
| substream->runtime->dma_bytes); |
| |
| /* set addr */ |
| ret = mtk_memif_set_addr(afe, id, |
| substream->runtime->dma_area, |
| substream->runtime->dma_addr, |
| substream->runtime->dma_bytes); |
| if (ret) { |
| dev_err(afe->dev, "%s(), error, id %d, set addr, ret %d\n", |
| __func__, id, ret); |
| return ret; |
| } |
| |
| /* set channel */ |
| ret = mtk_memif_set_channel(afe, id, channels); |
| if (ret) { |
| dev_err(afe->dev, "%s(), error, id %d, set channel %d, ret %d\n", |
| __func__, id, channels, ret); |
| return ret; |
| } |
| |
| /* set rate */ |
| ret = mtk_memif_set_rate_substream(substream, id, rate); |
| if (ret) { |
| dev_err(afe->dev, "%s(), error, id %d, set rate %d, ret %d\n", |
| __func__, id, rate, ret); |
| return ret; |
| } |
| |
| /* set format */ |
| ret = mtk_memif_set_format(afe, id, format); |
| if (ret) { |
| dev_err(afe->dev, "%s(), error, id %d, set format %d, ret %d\n", |
| __func__, id, format, ret); |
| return ret; |
| } |
| |
| return 0; |
| } |
| EXPORT_SYMBOL_GPL(mtk_afe_fe_hw_params); |
| |
| int mtk_afe_fe_hw_free(struct snd_pcm_substream *substream, |
| struct snd_soc_dai *dai) |
| { |
| struct mtk_base_afe *afe = snd_soc_dai_get_drvdata(dai); |
| |
| if (afe->release_dram_resource) |
| afe->release_dram_resource(afe->dev); |
| |
| return 0; |
| } |
| EXPORT_SYMBOL_GPL(mtk_afe_fe_hw_free); |
| |
| int mtk_afe_fe_trigger(struct snd_pcm_substream *substream, int cmd, |
| struct snd_soc_dai *dai) |
| { |
| struct snd_soc_pcm_runtime *rtd = asoc_substream_to_rtd(substream); |
| struct snd_pcm_runtime * const runtime = substream->runtime; |
| struct mtk_base_afe *afe = snd_soc_dai_get_drvdata(dai); |
| int id = asoc_rtd_to_cpu(rtd, 0)->id; |
| struct mtk_base_afe_memif *memif = &afe->memif[id]; |
| struct mtk_base_afe_irq *irqs = &afe->irqs[memif->irq_usage]; |
| const struct mtk_base_irq_data *irq_data = irqs->irq_data; |
| unsigned int counter = runtime->period_size; |
| int fs; |
| int ret; |
| |
| dev_dbg(afe->dev, "%s %s cmd=%d\n", __func__, memif->data->name, cmd); |
| |
| switch (cmd) { |
| case SNDRV_PCM_TRIGGER_START: |
| case SNDRV_PCM_TRIGGER_RESUME: |
| ret = mtk_memif_set_enable(afe, id); |
| if (ret) { |
| dev_err(afe->dev, "%s(), error, id %d, memif enable, ret %d\n", |
| __func__, id, ret); |
| return ret; |
| } |
| |
| /* set irq counter */ |
| mtk_regmap_update_bits(afe->regmap, irq_data->irq_cnt_reg, |
| irq_data->irq_cnt_maskbit, counter, |
| irq_data->irq_cnt_shift); |
| |
| /* set irq fs */ |
| fs = afe->irq_fs(substream, runtime->rate); |
| |
| if (fs < 0) |
| return -EINVAL; |
| |
| mtk_regmap_update_bits(afe->regmap, irq_data->irq_fs_reg, |
| irq_data->irq_fs_maskbit, fs, |
| irq_data->irq_fs_shift); |
| |
| /* enable interrupt */ |
| mtk_regmap_update_bits(afe->regmap, irq_data->irq_en_reg, |
| 1, 1, irq_data->irq_en_shift); |
| |
| return 0; |
| case SNDRV_PCM_TRIGGER_STOP: |
| case SNDRV_PCM_TRIGGER_SUSPEND: |
| ret = mtk_memif_set_disable(afe, id); |
| if (ret) { |
| dev_err(afe->dev, "%s(), error, id %d, memif enable, ret %d\n", |
| __func__, id, ret); |
| } |
| |
| /* disable interrupt */ |
| mtk_regmap_update_bits(afe->regmap, irq_data->irq_en_reg, |
| 1, 0, irq_data->irq_en_shift); |
| /* and clear pending IRQ */ |
| mtk_regmap_write(afe->regmap, irq_data->irq_clr_reg, |
| 1 << irq_data->irq_clr_shift); |
| return ret; |
| default: |
| return -EINVAL; |
| } |
| } |
| EXPORT_SYMBOL_GPL(mtk_afe_fe_trigger); |
| |
| int mtk_afe_fe_prepare(struct snd_pcm_substream *substream, |
| struct snd_soc_dai *dai) |
| { |
| struct snd_soc_pcm_runtime *rtd = asoc_substream_to_rtd(substream); |
| struct mtk_base_afe *afe = snd_soc_dai_get_drvdata(dai); |
| int id = asoc_rtd_to_cpu(rtd, 0)->id; |
| int pbuf_size; |
| |
| if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK) { |
| if (afe->get_memif_pbuf_size) { |
| pbuf_size = afe->get_memif_pbuf_size(substream); |
| mtk_memif_set_pbuf_size(afe, id, pbuf_size); |
| } |
| } |
| return 0; |
| } |
| EXPORT_SYMBOL_GPL(mtk_afe_fe_prepare); |
| |
| const struct snd_soc_dai_ops mtk_afe_fe_ops = { |
| .startup = mtk_afe_fe_startup, |
| .shutdown = mtk_afe_fe_shutdown, |
| .hw_params = mtk_afe_fe_hw_params, |
| .hw_free = mtk_afe_fe_hw_free, |
| .prepare = mtk_afe_fe_prepare, |
| .trigger = mtk_afe_fe_trigger, |
| }; |
| EXPORT_SYMBOL_GPL(mtk_afe_fe_ops); |
| |
| int mtk_dynamic_irq_acquire(struct mtk_base_afe *afe) |
| { |
| int i; |
| |
| mutex_lock(&afe->irq_alloc_lock); |
| for (i = 0; i < afe->irqs_size; ++i) { |
| if (afe->irqs[i].irq_occupyed == 0) { |
| afe->irqs[i].irq_occupyed = 1; |
| mutex_unlock(&afe->irq_alloc_lock); |
| return i; |
| } |
| } |
| mutex_unlock(&afe->irq_alloc_lock); |
| return afe->irqs_size; |
| } |
| EXPORT_SYMBOL_GPL(mtk_dynamic_irq_acquire); |
| |
| int mtk_dynamic_irq_release(struct mtk_base_afe *afe, int irq_id) |
| { |
| mutex_lock(&afe->irq_alloc_lock); |
| if (irq_id >= 0 && irq_id < afe->irqs_size) { |
| afe->irqs[irq_id].irq_occupyed = 0; |
| mutex_unlock(&afe->irq_alloc_lock); |
| return 0; |
| } |
| mutex_unlock(&afe->irq_alloc_lock); |
| return -EINVAL; |
| } |
| EXPORT_SYMBOL_GPL(mtk_dynamic_irq_release); |
| |
| int mtk_afe_suspend(struct snd_soc_component *component) |
| { |
| struct mtk_base_afe *afe = snd_soc_component_get_drvdata(component); |
| struct device *dev = afe->dev; |
| struct regmap *regmap = afe->regmap; |
| int i; |
| |
| if (pm_runtime_status_suspended(dev) || afe->suspended) |
| return 0; |
| |
| if (!afe->reg_back_up) |
| afe->reg_back_up = |
| devm_kcalloc(dev, afe->reg_back_up_list_num, |
| sizeof(unsigned int), GFP_KERNEL); |
| |
| if (afe->reg_back_up) { |
| for (i = 0; i < afe->reg_back_up_list_num; i++) |
| regmap_read(regmap, afe->reg_back_up_list[i], |
| &afe->reg_back_up[i]); |
| } |
| |
| afe->suspended = true; |
| afe->runtime_suspend(dev); |
| return 0; |
| } |
| EXPORT_SYMBOL_GPL(mtk_afe_suspend); |
| |
| int mtk_afe_resume(struct snd_soc_component *component) |
| { |
| struct mtk_base_afe *afe = snd_soc_component_get_drvdata(component); |
| struct device *dev = afe->dev; |
| struct regmap *regmap = afe->regmap; |
| int i; |
| |
| if (pm_runtime_status_suspended(dev) || !afe->suspended) |
| return 0; |
| |
| afe->runtime_resume(dev); |
| |
| if (!afe->reg_back_up) { |
| dev_dbg(dev, "%s no reg_backup\n", __func__); |
| } else { |
| for (i = 0; i < afe->reg_back_up_list_num; i++) |
| mtk_regmap_write(regmap, afe->reg_back_up_list[i], |
| afe->reg_back_up[i]); |
| } |
| |
| afe->suspended = false; |
| return 0; |
| } |
| EXPORT_SYMBOL_GPL(mtk_afe_resume); |
| |
| int mtk_memif_set_enable(struct mtk_base_afe *afe, int id) |
| { |
| struct mtk_base_afe_memif *memif = &afe->memif[id]; |
| |
| if (memif->data->enable_shift < 0) { |
| dev_warn(afe->dev, "%s(), error, id %d, enable_shift < 0\n", |
| __func__, id); |
| return 0; |
| } |
| return mtk_regmap_update_bits(afe->regmap, memif->data->enable_reg, |
| 1, 1, memif->data->enable_shift); |
| } |
| EXPORT_SYMBOL_GPL(mtk_memif_set_enable); |
| |
| int mtk_memif_set_disable(struct mtk_base_afe *afe, int id) |
| { |
| struct mtk_base_afe_memif *memif = &afe->memif[id]; |
| |
| if (memif->data->enable_shift < 0) { |
| dev_warn(afe->dev, "%s(), error, id %d, enable_shift < 0\n", |
| __func__, id); |
| return 0; |
| } |
| return mtk_regmap_update_bits(afe->regmap, memif->data->enable_reg, |
| 1, 0, memif->data->enable_shift); |
| } |
| EXPORT_SYMBOL_GPL(mtk_memif_set_disable); |
| |
| int mtk_memif_set_addr(struct mtk_base_afe *afe, int id, |
| unsigned char *dma_area, |
| dma_addr_t dma_addr, |
| size_t dma_bytes) |
| { |
| struct mtk_base_afe_memif *memif = &afe->memif[id]; |
| int msb_at_bit33 = upper_32_bits(dma_addr) ? 1 : 0; |
| unsigned int phys_buf_addr = lower_32_bits(dma_addr); |
| unsigned int phys_buf_addr_upper_32 = upper_32_bits(dma_addr); |
| |
| memif->dma_area = dma_area; |
| memif->dma_addr = dma_addr; |
| memif->dma_bytes = dma_bytes; |
| |
| /* start */ |
| mtk_regmap_write(afe->regmap, memif->data->reg_ofs_base, |
| phys_buf_addr); |
| /* end */ |
| if (memif->data->reg_ofs_end) |
| mtk_regmap_write(afe->regmap, |
| memif->data->reg_ofs_end, |
| phys_buf_addr + dma_bytes - 1); |
| else |
| mtk_regmap_write(afe->regmap, |
| memif->data->reg_ofs_base + |
| AFE_BASE_END_OFFSET, |
| phys_buf_addr + dma_bytes - 1); |
| |
| /* set start, end, upper 32 bits */ |
| if (memif->data->reg_ofs_base_msb) { |
| mtk_regmap_write(afe->regmap, memif->data->reg_ofs_base_msb, |
| phys_buf_addr_upper_32); |
| mtk_regmap_write(afe->regmap, |
| memif->data->reg_ofs_end_msb, |
| phys_buf_addr_upper_32); |
| } |
| |
| /* |
| * set MSB to 33-bit, for memif address |
| * only for memif base address, if msb_end_reg exists |
| */ |
| if (memif->data->msb_reg) |
| mtk_regmap_update_bits(afe->regmap, memif->data->msb_reg, |
| 1, msb_at_bit33, memif->data->msb_shift); |
| |
| /* set MSB to 33-bit, for memif end address */ |
| if (memif->data->msb_end_reg) |
| mtk_regmap_update_bits(afe->regmap, memif->data->msb_end_reg, |
| 1, msb_at_bit33, |
| memif->data->msb_end_shift); |
| |
| return 0; |
| } |
| EXPORT_SYMBOL_GPL(mtk_memif_set_addr); |
| |
| int mtk_memif_set_channel(struct mtk_base_afe *afe, |
| int id, unsigned int channel) |
| { |
| struct mtk_base_afe_memif *memif = &afe->memif[id]; |
| unsigned int mono; |
| |
| if (memif->data->mono_shift < 0) |
| return 0; |
| |
| if (memif->data->quad_ch_mask) { |
| unsigned int quad_ch = (channel == 4) ? 1 : 0; |
| |
| mtk_regmap_update_bits(afe->regmap, memif->data->quad_ch_reg, |
| memif->data->quad_ch_mask, |
| quad_ch, memif->data->quad_ch_shift); |
| } |
| |
| if (memif->data->mono_invert) |
| mono = (channel == 1) ? 0 : 1; |
| else |
| mono = (channel == 1) ? 1 : 0; |
| |
| /* for specific configuration of memif mono mode */ |
| if (memif->data->int_odd_flag_reg) |
| mtk_regmap_update_bits(afe->regmap, |
| memif->data->int_odd_flag_reg, |
| 1, mono, |
| memif->data->int_odd_flag_shift); |
| |
| return mtk_regmap_update_bits(afe->regmap, memif->data->mono_reg, |
| 1, mono, memif->data->mono_shift); |
| } |
| EXPORT_SYMBOL_GPL(mtk_memif_set_channel); |
| |
| static int mtk_memif_set_rate_fs(struct mtk_base_afe *afe, |
| int id, int fs) |
| { |
| struct mtk_base_afe_memif *memif = &afe->memif[id]; |
| |
| if (memif->data->fs_shift >= 0) |
| mtk_regmap_update_bits(afe->regmap, memif->data->fs_reg, |
| memif->data->fs_maskbit, |
| fs, memif->data->fs_shift); |
| |
| return 0; |
| } |
| |
| int mtk_memif_set_rate(struct mtk_base_afe *afe, |
| int id, unsigned int rate) |
| { |
| int fs = 0; |
| |
| if (!afe->get_dai_fs) { |
| dev_err(afe->dev, "%s(), error, afe->get_dai_fs == NULL\n", |
| __func__); |
| return -EINVAL; |
| } |
| |
| fs = afe->get_dai_fs(afe, id, rate); |
| |
| if (fs < 0) |
| return -EINVAL; |
| |
| return mtk_memif_set_rate_fs(afe, id, fs); |
| } |
| EXPORT_SYMBOL_GPL(mtk_memif_set_rate); |
| |
| int mtk_memif_set_rate_substream(struct snd_pcm_substream *substream, |
| int id, unsigned int rate) |
| { |
| struct snd_soc_pcm_runtime *rtd = asoc_substream_to_rtd(substream); |
| struct snd_soc_component *component = |
| snd_soc_rtdcom_lookup(rtd, AFE_PCM_NAME); |
| struct mtk_base_afe *afe = snd_soc_component_get_drvdata(component); |
| |
| int fs = 0; |
| |
| if (!afe->memif_fs) { |
| dev_err(afe->dev, "%s(), error, afe->memif_fs == NULL\n", |
| __func__); |
| return -EINVAL; |
| } |
| |
| fs = afe->memif_fs(substream, rate); |
| |
| if (fs < 0) |
| return -EINVAL; |
| |
| return mtk_memif_set_rate_fs(afe, id, fs); |
| } |
| EXPORT_SYMBOL_GPL(mtk_memif_set_rate_substream); |
| |
| int mtk_memif_set_format(struct mtk_base_afe *afe, |
| int id, snd_pcm_format_t format) |
| { |
| struct mtk_base_afe_memif *memif = &afe->memif[id]; |
| int hd_audio = 0; |
| int hd_align = 0; |
| |
| /* set hd mode */ |
| switch (format) { |
| case SNDRV_PCM_FORMAT_S16_LE: |
| case SNDRV_PCM_FORMAT_U16_LE: |
| hd_audio = 0; |
| break; |
| case SNDRV_PCM_FORMAT_S32_LE: |
| case SNDRV_PCM_FORMAT_U32_LE: |
| if (afe->memif_32bit_supported) { |
| hd_audio = 2; |
| hd_align = 0; |
| } else { |
| hd_audio = 1; |
| hd_align = 1; |
| } |
| break; |
| case SNDRV_PCM_FORMAT_S24_LE: |
| case SNDRV_PCM_FORMAT_U24_LE: |
| hd_audio = 1; |
| break; |
| default: |
| dev_err(afe->dev, "%s() error: unsupported format %d\n", |
| __func__, format); |
| break; |
| } |
| |
| mtk_regmap_update_bits(afe->regmap, memif->data->hd_reg, |
| 0x3, hd_audio, memif->data->hd_shift); |
| |
| mtk_regmap_update_bits(afe->regmap, memif->data->hd_align_reg, |
| 0x1, hd_align, memif->data->hd_align_mshift); |
| |
| return 0; |
| } |
| EXPORT_SYMBOL_GPL(mtk_memif_set_format); |
| |
| int mtk_memif_set_pbuf_size(struct mtk_base_afe *afe, |
| int id, int pbuf_size) |
| { |
| const struct mtk_base_memif_data *memif_data = afe->memif[id].data; |
| |
| if (memif_data->pbuf_mask == 0 || memif_data->minlen_mask == 0) |
| return 0; |
| |
| mtk_regmap_update_bits(afe->regmap, memif_data->pbuf_reg, |
| memif_data->pbuf_mask, |
| pbuf_size, memif_data->pbuf_shift); |
| |
| mtk_regmap_update_bits(afe->regmap, memif_data->minlen_reg, |
| memif_data->minlen_mask, |
| pbuf_size, memif_data->minlen_shift); |
| return 0; |
| } |
| EXPORT_SYMBOL_GPL(mtk_memif_set_pbuf_size); |
| |
| MODULE_DESCRIPTION("Mediatek simple fe dai operator"); |
| MODULE_AUTHOR("Garlic Tseng <garlic.tseng@mediatek.com>"); |
| MODULE_LICENSE("GPL v2"); |