| // SPDX-License-Identifier: GPL-2.0-only |
| /* |
| * Copyright (c) 2010-2011,2013-2015 The Linux Foundation. All rights reserved. |
| * |
| * lpass-cpu.c -- ALSA SoC CPU DAI driver for QTi LPASS |
| */ |
| |
| #include <linux/clk.h> |
| #include <linux/kernel.h> |
| #include <linux/module.h> |
| #include <linux/of.h> |
| #include <linux/of_device.h> |
| #include <linux/platform_device.h> |
| #include <sound/pcm.h> |
| #include <sound/pcm_params.h> |
| #include <linux/regmap.h> |
| #include <sound/soc.h> |
| #include <sound/soc-dai.h> |
| #include "lpass-lpaif-reg.h" |
| #include "lpass.h" |
| |
| #define LPASS_CPU_MAX_MI2S_LINES 4 |
| #define LPASS_CPU_I2S_SD0_MASK BIT(0) |
| #define LPASS_CPU_I2S_SD1_MASK BIT(1) |
| #define LPASS_CPU_I2S_SD2_MASK BIT(2) |
| #define LPASS_CPU_I2S_SD3_MASK BIT(3) |
| #define LPASS_CPU_I2S_SD0_1_MASK GENMASK(1, 0) |
| #define LPASS_CPU_I2S_SD2_3_MASK GENMASK(3, 2) |
| #define LPASS_CPU_I2S_SD0_1_2_MASK GENMASK(2, 0) |
| #define LPASS_CPU_I2S_SD0_1_2_3_MASK GENMASK(3, 0) |
| #define LPASS_REG_READ 1 |
| #define LPASS_REG_WRITE 0 |
| |
| /* |
| * Channel maps for Quad channel playbacks on MI2S Secondary |
| */ |
| static struct snd_pcm_chmap_elem lpass_quad_chmaps[] = { |
| { .channels = 4, |
| .map = { SNDRV_CHMAP_FL, SNDRV_CHMAP_RL, |
| SNDRV_CHMAP_FR, SNDRV_CHMAP_RR } }, |
| { } |
| }; |
| static int lpass_cpu_init_i2sctl_bitfields(struct device *dev, |
| struct lpaif_i2sctl *i2sctl, struct regmap *map) |
| { |
| struct lpass_data *drvdata = dev_get_drvdata(dev); |
| struct lpass_variant *v = drvdata->variant; |
| |
| i2sctl->loopback = devm_regmap_field_alloc(dev, map, v->loopback); |
| i2sctl->spken = devm_regmap_field_alloc(dev, map, v->spken); |
| i2sctl->spkmode = devm_regmap_field_alloc(dev, map, v->spkmode); |
| i2sctl->spkmono = devm_regmap_field_alloc(dev, map, v->spkmono); |
| i2sctl->micen = devm_regmap_field_alloc(dev, map, v->micen); |
| i2sctl->micmode = devm_regmap_field_alloc(dev, map, v->micmode); |
| i2sctl->micmono = devm_regmap_field_alloc(dev, map, v->micmono); |
| i2sctl->wssrc = devm_regmap_field_alloc(dev, map, v->wssrc); |
| i2sctl->bitwidth = devm_regmap_field_alloc(dev, map, v->bitwidth); |
| |
| if (IS_ERR(i2sctl->loopback) || IS_ERR(i2sctl->spken) || |
| IS_ERR(i2sctl->spkmode) || IS_ERR(i2sctl->spkmono) || |
| IS_ERR(i2sctl->micen) || IS_ERR(i2sctl->micmode) || |
| IS_ERR(i2sctl->micmono) || IS_ERR(i2sctl->wssrc) || |
| IS_ERR(i2sctl->bitwidth)) |
| return -EINVAL; |
| |
| return 0; |
| } |
| |
| static int lpass_cpu_daiops_set_sysclk(struct snd_soc_dai *dai, int clk_id, |
| unsigned int freq, int dir) |
| { |
| struct lpass_data *drvdata = snd_soc_dai_get_drvdata(dai); |
| int ret; |
| |
| ret = clk_set_rate(drvdata->mi2s_osr_clk[dai->driver->id], freq); |
| if (ret) |
| dev_err(dai->dev, "error setting mi2s osrclk to %u: %d\n", |
| freq, ret); |
| |
| return ret; |
| } |
| |
| static int lpass_cpu_daiops_startup(struct snd_pcm_substream *substream, |
| struct snd_soc_dai *dai) |
| { |
| struct lpass_data *drvdata = snd_soc_dai_get_drvdata(dai); |
| int ret; |
| |
| ret = clk_prepare_enable(drvdata->mi2s_osr_clk[dai->driver->id]); |
| if (ret) { |
| dev_err(dai->dev, "error in enabling mi2s osr clk: %d\n", ret); |
| return ret; |
| } |
| ret = clk_prepare(drvdata->mi2s_bit_clk[dai->driver->id]); |
| if (ret) { |
| dev_err(dai->dev, "error in enabling mi2s bit clk: %d\n", ret); |
| clk_disable_unprepare(drvdata->mi2s_osr_clk[dai->driver->id]); |
| return ret; |
| } |
| return 0; |
| } |
| |
| static void lpass_cpu_daiops_shutdown(struct snd_pcm_substream *substream, |
| struct snd_soc_dai *dai) |
| { |
| struct lpass_data *drvdata = snd_soc_dai_get_drvdata(dai); |
| struct lpaif_i2sctl *i2sctl = drvdata->i2sctl; |
| unsigned int id = dai->driver->id; |
| |
| clk_disable_unprepare(drvdata->mi2s_osr_clk[dai->driver->id]); |
| /* |
| * Ensure LRCLK is disabled even in device node validation. |
| * Will not impact if disabled in lpass_cpu_daiops_trigger() |
| * suspend. |
| */ |
| if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK) |
| regmap_fields_write(i2sctl->spken, id, LPAIF_I2SCTL_SPKEN_DISABLE); |
| else |
| regmap_fields_write(i2sctl->micen, id, LPAIF_I2SCTL_MICEN_DISABLE); |
| |
| /* |
| * BCLK may not be enabled if lpass_cpu_daiops_prepare is called before |
| * lpass_cpu_daiops_shutdown. It's paired with the clk_enable in |
| * lpass_cpu_daiops_prepare. |
| */ |
| if (drvdata->mi2s_was_prepared[dai->driver->id]) { |
| drvdata->mi2s_was_prepared[dai->driver->id] = false; |
| clk_disable(drvdata->mi2s_bit_clk[dai->driver->id]); |
| } |
| |
| clk_unprepare(drvdata->mi2s_bit_clk[dai->driver->id]); |
| } |
| |
| static int lpass_cpu_daiops_hw_params(struct snd_pcm_substream *substream, |
| struct snd_pcm_hw_params *params, struct snd_soc_dai *dai) |
| { |
| struct lpass_data *drvdata = snd_soc_dai_get_drvdata(dai); |
| struct lpaif_i2sctl *i2sctl = drvdata->i2sctl; |
| unsigned int id = dai->driver->id; |
| snd_pcm_format_t format = params_format(params); |
| unsigned int channels = params_channels(params); |
| unsigned int rate = params_rate(params); |
| unsigned int mode; |
| unsigned int regval; |
| int bitwidth, ret; |
| |
| bitwidth = snd_pcm_format_width(format); |
| if (bitwidth < 0) { |
| dev_err(dai->dev, "invalid bit width given: %d\n", bitwidth); |
| return bitwidth; |
| } |
| |
| ret = regmap_fields_write(i2sctl->loopback, id, |
| LPAIF_I2SCTL_LOOPBACK_DISABLE); |
| if (ret) { |
| dev_err(dai->dev, "error updating loopback field: %d\n", ret); |
| return ret; |
| } |
| |
| ret = regmap_fields_write(i2sctl->wssrc, id, |
| LPAIF_I2SCTL_WSSRC_INTERNAL); |
| if (ret) { |
| dev_err(dai->dev, "error updating wssrc field: %d\n", ret); |
| return ret; |
| } |
| |
| switch (bitwidth) { |
| case 16: |
| regval = LPAIF_I2SCTL_BITWIDTH_16; |
| break; |
| case 24: |
| regval = LPAIF_I2SCTL_BITWIDTH_24; |
| break; |
| case 32: |
| regval = LPAIF_I2SCTL_BITWIDTH_32; |
| break; |
| default: |
| dev_err(dai->dev, "invalid bitwidth given: %d\n", bitwidth); |
| return -EINVAL; |
| } |
| |
| ret = regmap_fields_write(i2sctl->bitwidth, id, regval); |
| if (ret) { |
| dev_err(dai->dev, "error updating bitwidth field: %d\n", ret); |
| return ret; |
| } |
| |
| if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK) |
| mode = drvdata->mi2s_playback_sd_mode[id]; |
| else |
| mode = drvdata->mi2s_capture_sd_mode[id]; |
| |
| if (!mode) { |
| dev_err(dai->dev, "no line is assigned\n"); |
| return -EINVAL; |
| } |
| |
| switch (channels) { |
| case 1: |
| case 2: |
| switch (mode) { |
| case LPAIF_I2SCTL_MODE_QUAD01: |
| case LPAIF_I2SCTL_MODE_6CH: |
| case LPAIF_I2SCTL_MODE_8CH: |
| mode = LPAIF_I2SCTL_MODE_SD0; |
| break; |
| case LPAIF_I2SCTL_MODE_QUAD23: |
| mode = LPAIF_I2SCTL_MODE_SD2; |
| break; |
| } |
| |
| break; |
| case 4: |
| if (mode < LPAIF_I2SCTL_MODE_QUAD01) { |
| dev_err(dai->dev, "cannot configure 4 channels with mode %d\n", |
| mode); |
| return -EINVAL; |
| } |
| |
| switch (mode) { |
| case LPAIF_I2SCTL_MODE_6CH: |
| case LPAIF_I2SCTL_MODE_8CH: |
| mode = LPAIF_I2SCTL_MODE_QUAD01; |
| break; |
| } |
| break; |
| case 6: |
| if (mode < LPAIF_I2SCTL_MODE_6CH) { |
| dev_err(dai->dev, "cannot configure 6 channels with mode %d\n", |
| mode); |
| return -EINVAL; |
| } |
| |
| switch (mode) { |
| case LPAIF_I2SCTL_MODE_8CH: |
| mode = LPAIF_I2SCTL_MODE_6CH; |
| break; |
| } |
| break; |
| case 8: |
| if (mode < LPAIF_I2SCTL_MODE_8CH) { |
| dev_err(dai->dev, "cannot configure 8 channels with mode %d\n", |
| mode); |
| return -EINVAL; |
| } |
| break; |
| default: |
| dev_err(dai->dev, "invalid channels given: %u\n", channels); |
| return -EINVAL; |
| } |
| |
| if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK) { |
| ret = regmap_fields_write(i2sctl->spkmode, id, |
| LPAIF_I2SCTL_SPKMODE(mode)); |
| if (ret) { |
| dev_err(dai->dev, "error writing to i2sctl spkr mode: %d\n", |
| ret); |
| return ret; |
| } |
| if (channels >= 2) |
| ret = regmap_fields_write(i2sctl->spkmono, id, |
| LPAIF_I2SCTL_SPKMONO_STEREO); |
| else |
| ret = regmap_fields_write(i2sctl->spkmono, id, |
| LPAIF_I2SCTL_SPKMONO_MONO); |
| } else { |
| ret = regmap_fields_write(i2sctl->micmode, id, |
| LPAIF_I2SCTL_MICMODE(mode)); |
| if (ret) { |
| dev_err(dai->dev, "error writing to i2sctl mic mode: %d\n", |
| ret); |
| return ret; |
| } |
| if (channels >= 2) |
| ret = regmap_fields_write(i2sctl->micmono, id, |
| LPAIF_I2SCTL_MICMONO_STEREO); |
| else |
| ret = regmap_fields_write(i2sctl->micmono, id, |
| LPAIF_I2SCTL_MICMONO_MONO); |
| } |
| |
| if (ret) { |
| dev_err(dai->dev, "error writing to i2sctl channels mode: %d\n", |
| ret); |
| return ret; |
| } |
| |
| ret = clk_set_rate(drvdata->mi2s_bit_clk[id], |
| rate * bitwidth * 2); |
| if (ret) { |
| dev_err(dai->dev, "error setting mi2s bitclk to %u: %d\n", |
| rate * bitwidth * 2, ret); |
| return ret; |
| } |
| |
| return 0; |
| } |
| |
| static int lpass_cpu_daiops_trigger(struct snd_pcm_substream *substream, |
| int cmd, struct snd_soc_dai *dai) |
| { |
| struct lpass_data *drvdata = snd_soc_dai_get_drvdata(dai); |
| struct lpaif_i2sctl *i2sctl = drvdata->i2sctl; |
| unsigned int id = dai->driver->id; |
| int ret = -EINVAL; |
| |
| switch (cmd) { |
| case SNDRV_PCM_TRIGGER_START: |
| case SNDRV_PCM_TRIGGER_RESUME: |
| case SNDRV_PCM_TRIGGER_PAUSE_RELEASE: |
| /* |
| * Ensure lpass BCLK/LRCLK is enabled during |
| * device resume as lpass_cpu_daiops_prepare() is not called |
| * after the device resumes. We don't check mi2s_was_prepared before |
| * enable/disable BCLK in trigger events because: |
| * 1. These trigger events are paired, so the BCLK |
| * enable_count is balanced. |
| * 2. the BCLK can be shared (ex: headset and headset mic), |
| * we need to increase the enable_count so that we don't |
| * turn off the shared BCLK while other devices are using |
| * it. |
| */ |
| if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK) { |
| ret = regmap_fields_write(i2sctl->spken, id, |
| LPAIF_I2SCTL_SPKEN_ENABLE); |
| } else { |
| ret = regmap_fields_write(i2sctl->micen, id, |
| LPAIF_I2SCTL_MICEN_ENABLE); |
| } |
| if (ret) |
| dev_err(dai->dev, "error writing to i2sctl reg: %d\n", |
| ret); |
| |
| ret = clk_enable(drvdata->mi2s_bit_clk[id]); |
| if (ret) { |
| dev_err(dai->dev, "error in enabling mi2s bit clk: %d\n", ret); |
| clk_disable(drvdata->mi2s_osr_clk[id]); |
| return ret; |
| } |
| break; |
| case SNDRV_PCM_TRIGGER_STOP: |
| case SNDRV_PCM_TRIGGER_SUSPEND: |
| case SNDRV_PCM_TRIGGER_PAUSE_PUSH: |
| /* |
| * To ensure lpass BCLK/LRCLK is disabled during |
| * device suspend. |
| */ |
| if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK) { |
| ret = regmap_fields_write(i2sctl->spken, id, |
| LPAIF_I2SCTL_SPKEN_DISABLE); |
| } else { |
| ret = regmap_fields_write(i2sctl->micen, id, |
| LPAIF_I2SCTL_MICEN_DISABLE); |
| } |
| if (ret) |
| dev_err(dai->dev, "error writing to i2sctl reg: %d\n", |
| ret); |
| |
| clk_disable(drvdata->mi2s_bit_clk[dai->driver->id]); |
| |
| break; |
| } |
| |
| return ret; |
| } |
| |
| static int lpass_cpu_daiops_prepare(struct snd_pcm_substream *substream, |
| struct snd_soc_dai *dai) |
| { |
| struct lpass_data *drvdata = snd_soc_dai_get_drvdata(dai); |
| struct lpaif_i2sctl *i2sctl = drvdata->i2sctl; |
| unsigned int id = dai->driver->id; |
| int ret; |
| |
| /* |
| * Ensure lpass BCLK/LRCLK is enabled bit before playback/capture |
| * data flow starts. This allows other codec to have some delay before |
| * the data flow. |
| * (ex: to drop start up pop noise before capture starts). |
| */ |
| if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK) |
| ret = regmap_fields_write(i2sctl->spken, id, LPAIF_I2SCTL_SPKEN_ENABLE); |
| else |
| ret = regmap_fields_write(i2sctl->micen, id, LPAIF_I2SCTL_MICEN_ENABLE); |
| |
| if (ret) { |
| dev_err(dai->dev, "error writing to i2sctl reg: %d\n", ret); |
| return ret; |
| } |
| |
| /* |
| * Check mi2s_was_prepared before enabling BCLK as lpass_cpu_daiops_prepare can |
| * be called multiple times. It's paired with the clk_disable in |
| * lpass_cpu_daiops_shutdown. |
| */ |
| if (!drvdata->mi2s_was_prepared[dai->driver->id]) { |
| ret = clk_enable(drvdata->mi2s_bit_clk[id]); |
| if (ret) { |
| dev_err(dai->dev, "error in enabling mi2s bit clk: %d\n", ret); |
| return ret; |
| } |
| drvdata->mi2s_was_prepared[dai->driver->id] = true; |
| } |
| return 0; |
| } |
| |
| const struct snd_soc_dai_ops asoc_qcom_lpass_cpu_dai_ops = { |
| .set_sysclk = lpass_cpu_daiops_set_sysclk, |
| .startup = lpass_cpu_daiops_startup, |
| .shutdown = lpass_cpu_daiops_shutdown, |
| .hw_params = lpass_cpu_daiops_hw_params, |
| .trigger = lpass_cpu_daiops_trigger, |
| .prepare = lpass_cpu_daiops_prepare, |
| }; |
| EXPORT_SYMBOL_GPL(asoc_qcom_lpass_cpu_dai_ops); |
| |
| int lpass_cpu_pcm_new(struct snd_soc_pcm_runtime *rtd, |
| struct snd_soc_dai *dai) |
| { |
| int ret; |
| struct snd_soc_dai_driver *drv = dai->driver; |
| struct lpass_data *drvdata = snd_soc_dai_get_drvdata(dai); |
| |
| if (drvdata->mi2s_playback_sd_mode[dai->id] == LPAIF_I2SCTL_MODE_QUAD01) { |
| ret = snd_pcm_add_chmap_ctls(rtd->pcm, SNDRV_PCM_STREAM_PLAYBACK, |
| lpass_quad_chmaps, drv->playback.channels_max, 0, |
| NULL); |
| if (ret < 0) |
| return ret; |
| } |
| |
| return 0; |
| } |
| EXPORT_SYMBOL_GPL(lpass_cpu_pcm_new); |
| |
| int asoc_qcom_lpass_cpu_dai_probe(struct snd_soc_dai *dai) |
| { |
| struct lpass_data *drvdata = snd_soc_dai_get_drvdata(dai); |
| int ret; |
| |
| /* ensure audio hardware is disabled */ |
| ret = regmap_write(drvdata->lpaif_map, |
| LPAIF_I2SCTL_REG(drvdata->variant, dai->driver->id), 0); |
| if (ret) |
| dev_err(dai->dev, "error writing to i2sctl reg: %d\n", ret); |
| |
| return ret; |
| } |
| EXPORT_SYMBOL_GPL(asoc_qcom_lpass_cpu_dai_probe); |
| |
| static int asoc_qcom_of_xlate_dai_name(struct snd_soc_component *component, |
| const struct of_phandle_args *args, |
| const char **dai_name) |
| { |
| struct lpass_data *drvdata = snd_soc_component_get_drvdata(component); |
| struct lpass_variant *variant = drvdata->variant; |
| int id = args->args[0]; |
| int ret = -EINVAL; |
| int i; |
| |
| for (i = 0; i < variant->num_dai; i++) { |
| if (variant->dai_driver[i].id == id) { |
| *dai_name = variant->dai_driver[i].name; |
| ret = 0; |
| break; |
| } |
| } |
| |
| return ret; |
| } |
| |
| static const struct snd_soc_component_driver lpass_cpu_comp_driver = { |
| .name = "lpass-cpu", |
| .of_xlate_dai_name = asoc_qcom_of_xlate_dai_name, |
| .legacy_dai_naming = 1, |
| }; |
| |
| static bool lpass_cpu_regmap_writeable(struct device *dev, unsigned int reg) |
| { |
| struct lpass_data *drvdata = dev_get_drvdata(dev); |
| struct lpass_variant *v = drvdata->variant; |
| int i; |
| |
| for (i = 0; i < v->i2s_ports; ++i) |
| if (reg == LPAIF_I2SCTL_REG(v, i)) |
| return true; |
| |
| for (i = 0; i < v->irq_ports; ++i) { |
| if (reg == LPAIF_IRQEN_REG(v, i)) |
| return true; |
| if (reg == LPAIF_IRQCLEAR_REG(v, i)) |
| return true; |
| } |
| |
| for (i = 0; i < v->rdma_channels; ++i) { |
| if (reg == LPAIF_RDMACTL_REG(v, i)) |
| return true; |
| if (reg == LPAIF_RDMABASE_REG(v, i)) |
| return true; |
| if (reg == LPAIF_RDMABUFF_REG(v, i)) |
| return true; |
| if (reg == LPAIF_RDMAPER_REG(v, i)) |
| return true; |
| } |
| |
| for (i = 0; i < v->wrdma_channels; ++i) { |
| if (reg == LPAIF_WRDMACTL_REG(v, i + v->wrdma_channel_start)) |
| return true; |
| if (reg == LPAIF_WRDMABASE_REG(v, i + v->wrdma_channel_start)) |
| return true; |
| if (reg == LPAIF_WRDMABUFF_REG(v, i + v->wrdma_channel_start)) |
| return true; |
| if (reg == LPAIF_WRDMAPER_REG(v, i + v->wrdma_channel_start)) |
| return true; |
| } |
| |
| return false; |
| } |
| |
| static bool lpass_cpu_regmap_readable(struct device *dev, unsigned int reg) |
| { |
| struct lpass_data *drvdata = dev_get_drvdata(dev); |
| struct lpass_variant *v = drvdata->variant; |
| int i; |
| |
| for (i = 0; i < v->i2s_ports; ++i) |
| if (reg == LPAIF_I2SCTL_REG(v, i)) |
| return true; |
| |
| for (i = 0; i < v->irq_ports; ++i) { |
| if (reg == LPAIF_IRQCLEAR_REG(v, i)) |
| return true; |
| if (reg == LPAIF_IRQEN_REG(v, i)) |
| return true; |
| if (reg == LPAIF_IRQSTAT_REG(v, i)) |
| return true; |
| } |
| |
| for (i = 0; i < v->rdma_channels; ++i) { |
| if (reg == LPAIF_RDMACTL_REG(v, i)) |
| return true; |
| if (reg == LPAIF_RDMABASE_REG(v, i)) |
| return true; |
| if (reg == LPAIF_RDMABUFF_REG(v, i)) |
| return true; |
| if (reg == LPAIF_RDMACURR_REG(v, i)) |
| return true; |
| if (reg == LPAIF_RDMAPER_REG(v, i)) |
| return true; |
| } |
| |
| for (i = 0; i < v->wrdma_channels; ++i) { |
| if (reg == LPAIF_WRDMACTL_REG(v, i + v->wrdma_channel_start)) |
| return true; |
| if (reg == LPAIF_WRDMABASE_REG(v, i + v->wrdma_channel_start)) |
| return true; |
| if (reg == LPAIF_WRDMABUFF_REG(v, i + v->wrdma_channel_start)) |
| return true; |
| if (reg == LPAIF_WRDMACURR_REG(v, i + v->wrdma_channel_start)) |
| return true; |
| if (reg == LPAIF_WRDMAPER_REG(v, i + v->wrdma_channel_start)) |
| return true; |
| } |
| |
| return false; |
| } |
| |
| static bool lpass_cpu_regmap_volatile(struct device *dev, unsigned int reg) |
| { |
| struct lpass_data *drvdata = dev_get_drvdata(dev); |
| struct lpass_variant *v = drvdata->variant; |
| int i; |
| |
| for (i = 0; i < v->irq_ports; ++i) { |
| if (reg == LPAIF_IRQCLEAR_REG(v, i)) |
| return true; |
| if (reg == LPAIF_IRQSTAT_REG(v, i)) |
| return true; |
| } |
| |
| for (i = 0; i < v->rdma_channels; ++i) |
| if (reg == LPAIF_RDMACURR_REG(v, i)) |
| return true; |
| |
| for (i = 0; i < v->wrdma_channels; ++i) |
| if (reg == LPAIF_WRDMACURR_REG(v, i + v->wrdma_channel_start)) |
| return true; |
| |
| return false; |
| } |
| |
| static struct regmap_config lpass_cpu_regmap_config = { |
| .name = "lpass_cpu", |
| .reg_bits = 32, |
| .reg_stride = 4, |
| .val_bits = 32, |
| .writeable_reg = lpass_cpu_regmap_writeable, |
| .readable_reg = lpass_cpu_regmap_readable, |
| .volatile_reg = lpass_cpu_regmap_volatile, |
| .cache_type = REGCACHE_FLAT, |
| }; |
| |
| static int lpass_hdmi_init_bitfields(struct device *dev, struct regmap *map) |
| { |
| struct lpass_data *drvdata = dev_get_drvdata(dev); |
| struct lpass_variant *v = drvdata->variant; |
| unsigned int i; |
| struct lpass_hdmi_tx_ctl *tx_ctl; |
| struct regmap_field *legacy_en; |
| struct lpass_vbit_ctrl *vbit_ctl; |
| struct regmap_field *tx_parity; |
| struct lpass_dp_metadata_ctl *meta_ctl; |
| struct lpass_sstream_ctl *sstream_ctl; |
| struct regmap_field *ch_msb; |
| struct regmap_field *ch_lsb; |
| struct lpass_hdmitx_dmactl *tx_dmactl; |
| int rval; |
| |
| tx_ctl = devm_kzalloc(dev, sizeof(*tx_ctl), GFP_KERNEL); |
| if (!tx_ctl) |
| return -ENOMEM; |
| |
| QCOM_REGMAP_FIELD_ALLOC(dev, map, v->soft_reset, tx_ctl->soft_reset); |
| QCOM_REGMAP_FIELD_ALLOC(dev, map, v->force_reset, tx_ctl->force_reset); |
| drvdata->tx_ctl = tx_ctl; |
| |
| QCOM_REGMAP_FIELD_ALLOC(dev, map, v->legacy_en, legacy_en); |
| drvdata->hdmitx_legacy_en = legacy_en; |
| |
| vbit_ctl = devm_kzalloc(dev, sizeof(*vbit_ctl), GFP_KERNEL); |
| if (!vbit_ctl) |
| return -ENOMEM; |
| |
| QCOM_REGMAP_FIELD_ALLOC(dev, map, v->replace_vbit, vbit_ctl->replace_vbit); |
| QCOM_REGMAP_FIELD_ALLOC(dev, map, v->vbit_stream, vbit_ctl->vbit_stream); |
| drvdata->vbit_ctl = vbit_ctl; |
| |
| |
| QCOM_REGMAP_FIELD_ALLOC(dev, map, v->calc_en, tx_parity); |
| drvdata->hdmitx_parity_calc_en = tx_parity; |
| |
| meta_ctl = devm_kzalloc(dev, sizeof(*meta_ctl), GFP_KERNEL); |
| if (!meta_ctl) |
| return -ENOMEM; |
| |
| rval = devm_regmap_field_bulk_alloc(dev, map, &meta_ctl->mute, &v->mute, 7); |
| if (rval) |
| return rval; |
| drvdata->meta_ctl = meta_ctl; |
| |
| sstream_ctl = devm_kzalloc(dev, sizeof(*sstream_ctl), GFP_KERNEL); |
| if (!sstream_ctl) |
| return -ENOMEM; |
| |
| rval = devm_regmap_field_bulk_alloc(dev, map, &sstream_ctl->sstream_en, &v->sstream_en, 9); |
| if (rval) |
| return rval; |
| |
| drvdata->sstream_ctl = sstream_ctl; |
| |
| for (i = 0; i < LPASS_MAX_HDMI_DMA_CHANNELS; i++) { |
| QCOM_REGMAP_FIELD_ALLOC(dev, map, v->msb_bits, ch_msb); |
| drvdata->hdmitx_ch_msb[i] = ch_msb; |
| |
| QCOM_REGMAP_FIELD_ALLOC(dev, map, v->lsb_bits, ch_lsb); |
| drvdata->hdmitx_ch_lsb[i] = ch_lsb; |
| |
| tx_dmactl = devm_kzalloc(dev, sizeof(*tx_dmactl), GFP_KERNEL); |
| if (!tx_dmactl) |
| return -ENOMEM; |
| |
| QCOM_REGMAP_FIELD_ALLOC(dev, map, v->use_hw_chs, tx_dmactl->use_hw_chs); |
| QCOM_REGMAP_FIELD_ALLOC(dev, map, v->use_hw_usr, tx_dmactl->use_hw_usr); |
| QCOM_REGMAP_FIELD_ALLOC(dev, map, v->hw_chs_sel, tx_dmactl->hw_chs_sel); |
| QCOM_REGMAP_FIELD_ALLOC(dev, map, v->hw_usr_sel, tx_dmactl->hw_usr_sel); |
| drvdata->hdmi_tx_dmactl[i] = tx_dmactl; |
| } |
| return 0; |
| } |
| |
| static bool lpass_hdmi_regmap_writeable(struct device *dev, unsigned int reg) |
| { |
| struct lpass_data *drvdata = dev_get_drvdata(dev); |
| struct lpass_variant *v = drvdata->variant; |
| int i; |
| |
| if (reg == LPASS_HDMI_TX_CTL_ADDR(v)) |
| return true; |
| if (reg == LPASS_HDMI_TX_LEGACY_ADDR(v)) |
| return true; |
| if (reg == LPASS_HDMI_TX_VBIT_CTL_ADDR(v)) |
| return true; |
| if (reg == LPASS_HDMI_TX_PARITY_ADDR(v)) |
| return true; |
| if (reg == LPASS_HDMI_TX_DP_ADDR(v)) |
| return true; |
| if (reg == LPASS_HDMI_TX_SSTREAM_ADDR(v)) |
| return true; |
| if (reg == LPASS_HDMITX_APP_IRQEN_REG(v)) |
| return true; |
| if (reg == LPASS_HDMITX_APP_IRQCLEAR_REG(v)) |
| return true; |
| |
| for (i = 0; i < v->hdmi_rdma_channels; i++) { |
| if (reg == LPASS_HDMI_TX_CH_LSB_ADDR(v, i)) |
| return true; |
| if (reg == LPASS_HDMI_TX_CH_MSB_ADDR(v, i)) |
| return true; |
| if (reg == LPASS_HDMI_TX_DMA_ADDR(v, i)) |
| return true; |
| } |
| |
| for (i = 0; i < v->hdmi_rdma_channels; ++i) { |
| if (reg == LPAIF_HDMI_RDMACTL_REG(v, i)) |
| return true; |
| if (reg == LPAIF_HDMI_RDMABASE_REG(v, i)) |
| return true; |
| if (reg == LPAIF_HDMI_RDMABUFF_REG(v, i)) |
| return true; |
| if (reg == LPAIF_HDMI_RDMAPER_REG(v, i)) |
| return true; |
| } |
| return false; |
| } |
| |
| static bool lpass_hdmi_regmap_readable(struct device *dev, unsigned int reg) |
| { |
| struct lpass_data *drvdata = dev_get_drvdata(dev); |
| struct lpass_variant *v = drvdata->variant; |
| int i; |
| |
| if (reg == LPASS_HDMI_TX_CTL_ADDR(v)) |
| return true; |
| if (reg == LPASS_HDMI_TX_LEGACY_ADDR(v)) |
| return true; |
| if (reg == LPASS_HDMI_TX_VBIT_CTL_ADDR(v)) |
| return true; |
| |
| for (i = 0; i < v->hdmi_rdma_channels; i++) { |
| if (reg == LPASS_HDMI_TX_CH_LSB_ADDR(v, i)) |
| return true; |
| if (reg == LPASS_HDMI_TX_CH_MSB_ADDR(v, i)) |
| return true; |
| if (reg == LPASS_HDMI_TX_DMA_ADDR(v, i)) |
| return true; |
| } |
| |
| if (reg == LPASS_HDMI_TX_PARITY_ADDR(v)) |
| return true; |
| if (reg == LPASS_HDMI_TX_DP_ADDR(v)) |
| return true; |
| if (reg == LPASS_HDMI_TX_SSTREAM_ADDR(v)) |
| return true; |
| if (reg == LPASS_HDMITX_APP_IRQEN_REG(v)) |
| return true; |
| if (reg == LPASS_HDMITX_APP_IRQSTAT_REG(v)) |
| return true; |
| |
| for (i = 0; i < v->hdmi_rdma_channels; ++i) { |
| if (reg == LPAIF_HDMI_RDMACTL_REG(v, i)) |
| return true; |
| if (reg == LPAIF_HDMI_RDMABASE_REG(v, i)) |
| return true; |
| if (reg == LPAIF_HDMI_RDMABUFF_REG(v, i)) |
| return true; |
| if (reg == LPAIF_HDMI_RDMAPER_REG(v, i)) |
| return true; |
| if (reg == LPAIF_HDMI_RDMACURR_REG(v, i)) |
| return true; |
| } |
| |
| return false; |
| } |
| |
| static bool lpass_hdmi_regmap_volatile(struct device *dev, unsigned int reg) |
| { |
| struct lpass_data *drvdata = dev_get_drvdata(dev); |
| struct lpass_variant *v = drvdata->variant; |
| int i; |
| |
| if (reg == LPASS_HDMITX_APP_IRQSTAT_REG(v)) |
| return true; |
| if (reg == LPASS_HDMI_TX_LEGACY_ADDR(v)) |
| return true; |
| if (reg == LPASS_HDMI_TX_VBIT_CTL_ADDR(v)) |
| return true; |
| if (reg == LPASS_HDMI_TX_PARITY_ADDR(v)) |
| return true; |
| |
| for (i = 0; i < v->hdmi_rdma_channels; ++i) { |
| if (reg == LPAIF_HDMI_RDMACURR_REG(v, i)) |
| return true; |
| if (reg == LPASS_HDMI_TX_DMA_ADDR(v, i)) |
| return true; |
| if (reg == LPASS_HDMI_TX_CH_LSB_ADDR(v, i)) |
| return true; |
| if (reg == LPASS_HDMI_TX_CH_MSB_ADDR(v, i)) |
| return true; |
| } |
| return false; |
| } |
| |
| static struct regmap_config lpass_hdmi_regmap_config = { |
| .name = "lpass_hdmi", |
| .reg_bits = 32, |
| .reg_stride = 4, |
| .val_bits = 32, |
| .writeable_reg = lpass_hdmi_regmap_writeable, |
| .readable_reg = lpass_hdmi_regmap_readable, |
| .volatile_reg = lpass_hdmi_regmap_volatile, |
| .cache_type = REGCACHE_FLAT, |
| }; |
| |
| static bool __lpass_rxtx_regmap_accessible(struct device *dev, unsigned int reg, bool rw) |
| { |
| struct lpass_data *drvdata = dev_get_drvdata(dev); |
| struct lpass_variant *v = drvdata->variant; |
| int i; |
| |
| for (i = 0; i < v->rxtx_irq_ports; ++i) { |
| if (reg == LPAIF_RXTX_IRQCLEAR_REG(v, i)) |
| return true; |
| if (reg == LPAIF_RXTX_IRQEN_REG(v, i)) |
| return true; |
| if (reg == LPAIF_RXTX_IRQSTAT_REG(v, i)) |
| return true; |
| } |
| |
| for (i = 0; i < v->rxtx_rdma_channels; ++i) { |
| if (reg == LPAIF_CDC_RXTX_RDMACTL_REG(v, i, LPASS_CDC_DMA_RX0)) |
| return true; |
| if (reg == LPAIF_CDC_RXTX_RDMABASE_REG(v, i, LPASS_CDC_DMA_RX0)) |
| return true; |
| if (reg == LPAIF_CDC_RXTX_RDMABUFF_REG(v, i, LPASS_CDC_DMA_RX0)) |
| return true; |
| if (rw == LPASS_REG_READ) { |
| if (reg == LPAIF_CDC_RXTX_RDMACURR_REG(v, i, LPASS_CDC_DMA_RX0)) |
| return true; |
| } |
| if (reg == LPAIF_CDC_RXTX_RDMAPER_REG(v, i, LPASS_CDC_DMA_RX0)) |
| return true; |
| if (reg == LPAIF_CDC_RXTX_RDMA_INTF_REG(v, i, LPASS_CDC_DMA_RX0)) |
| return true; |
| } |
| |
| for (i = 0; i < v->rxtx_wrdma_channels; ++i) { |
| if (reg == LPAIF_CDC_RXTX_WRDMACTL_REG(v, i + v->rxtx_wrdma_channel_start, |
| LPASS_CDC_DMA_TX3)) |
| return true; |
| if (reg == LPAIF_CDC_RXTX_WRDMABASE_REG(v, i + v->rxtx_wrdma_channel_start, |
| LPASS_CDC_DMA_TX3)) |
| return true; |
| if (reg == LPAIF_CDC_RXTX_WRDMABUFF_REG(v, i + v->rxtx_wrdma_channel_start, |
| LPASS_CDC_DMA_TX3)) |
| return true; |
| if (rw == LPASS_REG_READ) { |
| if (reg == LPAIF_CDC_RXTX_WRDMACURR_REG(v, i, LPASS_CDC_DMA_RX0)) |
| return true; |
| } |
| if (reg == LPAIF_CDC_RXTX_WRDMAPER_REG(v, i + v->rxtx_wrdma_channel_start, |
| LPASS_CDC_DMA_TX3)) |
| return true; |
| if (reg == LPAIF_CDC_RXTX_WRDMA_INTF_REG(v, i + v->rxtx_wrdma_channel_start, |
| LPASS_CDC_DMA_TX3)) |
| return true; |
| } |
| return false; |
| } |
| |
| static bool lpass_rxtx_regmap_writeable(struct device *dev, unsigned int reg) |
| { |
| return __lpass_rxtx_regmap_accessible(dev, reg, LPASS_REG_WRITE); |
| } |
| |
| static bool lpass_rxtx_regmap_readable(struct device *dev, unsigned int reg) |
| { |
| return __lpass_rxtx_regmap_accessible(dev, reg, LPASS_REG_READ); |
| } |
| |
| static bool lpass_rxtx_regmap_volatile(struct device *dev, unsigned int reg) |
| { |
| struct lpass_data *drvdata = dev_get_drvdata(dev); |
| struct lpass_variant *v = drvdata->variant; |
| int i; |
| |
| for (i = 0; i < v->rxtx_irq_ports; ++i) { |
| if (reg == LPAIF_RXTX_IRQCLEAR_REG(v, i)) |
| return true; |
| if (reg == LPAIF_RXTX_IRQSTAT_REG(v, i)) |
| return true; |
| } |
| |
| for (i = 0; i < v->rxtx_rdma_channels; ++i) |
| if (reg == LPAIF_CDC_RXTX_RDMACURR_REG(v, i, LPASS_CDC_DMA_RX0)) |
| return true; |
| |
| for (i = 0; i < v->rxtx_wrdma_channels; ++i) |
| if (reg == LPAIF_CDC_RXTX_WRDMACURR_REG(v, i + v->rxtx_wrdma_channel_start, |
| LPASS_CDC_DMA_TX3)) |
| return true; |
| |
| return false; |
| } |
| |
| static bool __lpass_va_regmap_accessible(struct device *dev, unsigned int reg, bool rw) |
| { |
| struct lpass_data *drvdata = dev_get_drvdata(dev); |
| struct lpass_variant *v = drvdata->variant; |
| int i; |
| |
| for (i = 0; i < v->va_irq_ports; ++i) { |
| if (reg == LPAIF_VA_IRQCLEAR_REG(v, i)) |
| return true; |
| if (reg == LPAIF_VA_IRQEN_REG(v, i)) |
| return true; |
| if (reg == LPAIF_VA_IRQSTAT_REG(v, i)) |
| return true; |
| } |
| |
| for (i = 0; i < v->va_wrdma_channels; ++i) { |
| if (reg == LPAIF_CDC_VA_WRDMACTL_REG(v, i + v->va_wrdma_channel_start, |
| LPASS_CDC_DMA_VA_TX0)) |
| return true; |
| if (reg == LPAIF_CDC_VA_WRDMABASE_REG(v, i + v->va_wrdma_channel_start, |
| LPASS_CDC_DMA_VA_TX0)) |
| return true; |
| if (reg == LPAIF_CDC_VA_WRDMABUFF_REG(v, i + v->va_wrdma_channel_start, |
| LPASS_CDC_DMA_VA_TX0)) |
| return true; |
| if (rw == LPASS_REG_READ) { |
| if (reg == LPAIF_CDC_VA_WRDMACURR_REG(v, i + v->va_wrdma_channel_start, |
| LPASS_CDC_DMA_VA_TX0)) |
| return true; |
| } |
| if (reg == LPAIF_CDC_VA_WRDMAPER_REG(v, i + v->va_wrdma_channel_start, |
| LPASS_CDC_DMA_VA_TX0)) |
| return true; |
| if (reg == LPAIF_CDC_VA_WRDMA_INTF_REG(v, i + v->va_wrdma_channel_start, |
| LPASS_CDC_DMA_VA_TX0)) |
| return true; |
| } |
| return false; |
| } |
| |
| static bool lpass_va_regmap_writeable(struct device *dev, unsigned int reg) |
| { |
| return __lpass_va_regmap_accessible(dev, reg, LPASS_REG_WRITE); |
| } |
| |
| static bool lpass_va_regmap_readable(struct device *dev, unsigned int reg) |
| { |
| return __lpass_va_regmap_accessible(dev, reg, LPASS_REG_READ); |
| } |
| |
| static bool lpass_va_regmap_volatile(struct device *dev, unsigned int reg) |
| { |
| struct lpass_data *drvdata = dev_get_drvdata(dev); |
| struct lpass_variant *v = drvdata->variant; |
| int i; |
| |
| for (i = 0; i < v->va_irq_ports; ++i) { |
| if (reg == LPAIF_VA_IRQCLEAR_REG(v, i)) |
| return true; |
| if (reg == LPAIF_VA_IRQSTAT_REG(v, i)) |
| return true; |
| } |
| |
| for (i = 0; i < v->va_wrdma_channels; ++i) { |
| if (reg == LPAIF_CDC_VA_WRDMACURR_REG(v, i + v->va_wrdma_channel_start, |
| LPASS_CDC_DMA_VA_TX0)) |
| return true; |
| } |
| |
| return false; |
| } |
| |
| static struct regmap_config lpass_rxtx_regmap_config = { |
| .reg_bits = 32, |
| .reg_stride = 4, |
| .val_bits = 32, |
| .writeable_reg = lpass_rxtx_regmap_writeable, |
| .readable_reg = lpass_rxtx_regmap_readable, |
| .volatile_reg = lpass_rxtx_regmap_volatile, |
| .cache_type = REGCACHE_FLAT, |
| }; |
| |
| static struct regmap_config lpass_va_regmap_config = { |
| .reg_bits = 32, |
| .reg_stride = 4, |
| .val_bits = 32, |
| .writeable_reg = lpass_va_regmap_writeable, |
| .readable_reg = lpass_va_regmap_readable, |
| .volatile_reg = lpass_va_regmap_volatile, |
| .cache_type = REGCACHE_FLAT, |
| }; |
| |
| static unsigned int of_lpass_cpu_parse_sd_lines(struct device *dev, |
| struct device_node *node, |
| const char *name) |
| { |
| unsigned int lines[LPASS_CPU_MAX_MI2S_LINES]; |
| unsigned int sd_line_mask = 0; |
| int num_lines, i; |
| |
| num_lines = of_property_read_variable_u32_array(node, name, lines, 0, |
| LPASS_CPU_MAX_MI2S_LINES); |
| if (num_lines < 0) |
| return LPAIF_I2SCTL_MODE_NONE; |
| |
| for (i = 0; i < num_lines; i++) |
| sd_line_mask |= BIT(lines[i]); |
| |
| switch (sd_line_mask) { |
| case LPASS_CPU_I2S_SD0_MASK: |
| return LPAIF_I2SCTL_MODE_SD0; |
| case LPASS_CPU_I2S_SD1_MASK: |
| return LPAIF_I2SCTL_MODE_SD1; |
| case LPASS_CPU_I2S_SD2_MASK: |
| return LPAIF_I2SCTL_MODE_SD2; |
| case LPASS_CPU_I2S_SD3_MASK: |
| return LPAIF_I2SCTL_MODE_SD3; |
| case LPASS_CPU_I2S_SD0_1_MASK: |
| return LPAIF_I2SCTL_MODE_QUAD01; |
| case LPASS_CPU_I2S_SD2_3_MASK: |
| return LPAIF_I2SCTL_MODE_QUAD23; |
| case LPASS_CPU_I2S_SD0_1_2_MASK: |
| return LPAIF_I2SCTL_MODE_6CH; |
| case LPASS_CPU_I2S_SD0_1_2_3_MASK: |
| return LPAIF_I2SCTL_MODE_8CH; |
| default: |
| dev_err(dev, "Unsupported SD line mask: %#x\n", sd_line_mask); |
| return LPAIF_I2SCTL_MODE_NONE; |
| } |
| } |
| |
| static void of_lpass_cpu_parse_dai_data(struct device *dev, |
| struct lpass_data *data) |
| { |
| struct device_node *node; |
| int ret, i, id; |
| |
| /* Allow all channels by default for backwards compatibility */ |
| for (i = 0; i < data->variant->num_dai; i++) { |
| id = data->variant->dai_driver[i].id; |
| data->mi2s_playback_sd_mode[id] = LPAIF_I2SCTL_MODE_8CH; |
| data->mi2s_capture_sd_mode[id] = LPAIF_I2SCTL_MODE_8CH; |
| } |
| |
| for_each_child_of_node(dev->of_node, node) { |
| ret = of_property_read_u32(node, "reg", &id); |
| if (ret || id < 0) { |
| dev_err(dev, "valid dai id not found: %d\n", ret); |
| continue; |
| } |
| if (id == LPASS_DP_RX) { |
| data->hdmi_port_enable = 1; |
| } else if (is_cdc_dma_port(id)) { |
| data->codec_dma_enable = 1; |
| } else { |
| data->mi2s_playback_sd_mode[id] = |
| of_lpass_cpu_parse_sd_lines(dev, node, |
| "qcom,playback-sd-lines"); |
| data->mi2s_capture_sd_mode[id] = |
| of_lpass_cpu_parse_sd_lines(dev, node, |
| "qcom,capture-sd-lines"); |
| } |
| } |
| } |
| |
| static int of_lpass_cdc_dma_clks_parse(struct device *dev, |
| struct lpass_data *data) |
| { |
| data->codec_mem0 = devm_clk_get(dev, "audio_cc_codec_mem0"); |
| if (IS_ERR(data->codec_mem0)) |
| return PTR_ERR(data->codec_mem0); |
| |
| data->codec_mem1 = devm_clk_get(dev, "audio_cc_codec_mem1"); |
| if (IS_ERR(data->codec_mem1)) |
| return PTR_ERR(data->codec_mem1); |
| |
| data->codec_mem2 = devm_clk_get(dev, "audio_cc_codec_mem2"); |
| if (IS_ERR(data->codec_mem2)) |
| return PTR_ERR(data->codec_mem2); |
| |
| data->va_mem0 = devm_clk_get(dev, "aon_cc_va_mem0"); |
| if (IS_ERR(data->va_mem0)) |
| return PTR_ERR(data->va_mem0); |
| |
| return 0; |
| } |
| |
| int asoc_qcom_lpass_cpu_platform_probe(struct platform_device *pdev) |
| { |
| struct lpass_data *drvdata; |
| struct device_node *dsp_of_node; |
| struct resource *res; |
| struct lpass_variant *variant; |
| struct device *dev = &pdev->dev; |
| const struct of_device_id *match; |
| int ret, i, dai_id; |
| |
| dsp_of_node = of_parse_phandle(pdev->dev.of_node, "qcom,adsp", 0); |
| if (dsp_of_node) { |
| dev_err(dev, "DSP exists and holds audio resources\n"); |
| of_node_put(dsp_of_node); |
| return -EBUSY; |
| } |
| |
| drvdata = devm_kzalloc(dev, sizeof(struct lpass_data), GFP_KERNEL); |
| if (!drvdata) |
| return -ENOMEM; |
| platform_set_drvdata(pdev, drvdata); |
| |
| match = of_match_device(dev->driver->of_match_table, dev); |
| if (!match || !match->data) |
| return -EINVAL; |
| |
| if (of_device_is_compatible(dev->of_node, "qcom,lpass-cpu-apq8016")) { |
| dev_warn(dev, "%s compatible is deprecated\n", |
| match->compatible); |
| } |
| |
| drvdata->variant = (struct lpass_variant *)match->data; |
| variant = drvdata->variant; |
| |
| of_lpass_cpu_parse_dai_data(dev, drvdata); |
| |
| if (drvdata->codec_dma_enable) { |
| drvdata->rxtx_lpaif = |
| devm_platform_ioremap_resource_byname(pdev, "lpass-rxtx-lpaif"); |
| if (IS_ERR(drvdata->rxtx_lpaif)) |
| return PTR_ERR(drvdata->rxtx_lpaif); |
| |
| drvdata->va_lpaif = devm_platform_ioremap_resource_byname(pdev, "lpass-va-lpaif"); |
| if (IS_ERR(drvdata->va_lpaif)) |
| return PTR_ERR(drvdata->va_lpaif); |
| |
| lpass_rxtx_regmap_config.max_register = LPAIF_CDC_RXTX_WRDMAPER_REG(variant, |
| variant->rxtx_wrdma_channels + |
| variant->rxtx_wrdma_channel_start, LPASS_CDC_DMA_TX3); |
| |
| drvdata->rxtx_lpaif_map = devm_regmap_init_mmio(dev, drvdata->rxtx_lpaif, |
| &lpass_rxtx_regmap_config); |
| if (IS_ERR(drvdata->rxtx_lpaif_map)) |
| return PTR_ERR(drvdata->rxtx_lpaif_map); |
| |
| lpass_va_regmap_config.max_register = LPAIF_CDC_VA_WRDMAPER_REG(variant, |
| variant->va_wrdma_channels + |
| variant->va_wrdma_channel_start, LPASS_CDC_DMA_VA_TX0); |
| |
| drvdata->va_lpaif_map = devm_regmap_init_mmio(dev, drvdata->va_lpaif, |
| &lpass_va_regmap_config); |
| if (IS_ERR(drvdata->va_lpaif_map)) |
| return PTR_ERR(drvdata->va_lpaif_map); |
| |
| ret = of_lpass_cdc_dma_clks_parse(dev, drvdata); |
| if (ret) { |
| dev_err(dev, "failed to get cdc dma clocks %d\n", ret); |
| return ret; |
| } |
| |
| res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "lpass-rxtx-cdc-dma-lpm"); |
| drvdata->rxtx_cdc_dma_lpm_buf = res->start; |
| |
| res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "lpass-va-cdc-dma-lpm"); |
| drvdata->va_cdc_dma_lpm_buf = res->start; |
| } |
| |
| drvdata->lpaif = devm_platform_ioremap_resource_byname(pdev, "lpass-lpaif"); |
| if (IS_ERR(drvdata->lpaif)) |
| return PTR_ERR(drvdata->lpaif); |
| |
| lpass_cpu_regmap_config.max_register = LPAIF_WRDMAPER_REG(variant, |
| variant->wrdma_channels + |
| variant->wrdma_channel_start); |
| |
| drvdata->lpaif_map = devm_regmap_init_mmio(dev, drvdata->lpaif, |
| &lpass_cpu_regmap_config); |
| if (IS_ERR(drvdata->lpaif_map)) { |
| dev_err(dev, "error initializing regmap: %ld\n", |
| PTR_ERR(drvdata->lpaif_map)); |
| return PTR_ERR(drvdata->lpaif_map); |
| } |
| |
| if (drvdata->hdmi_port_enable) { |
| drvdata->hdmiif = devm_platform_ioremap_resource_byname(pdev, "lpass-hdmiif"); |
| if (IS_ERR(drvdata->hdmiif)) |
| return PTR_ERR(drvdata->hdmiif); |
| |
| lpass_hdmi_regmap_config.max_register = LPAIF_HDMI_RDMAPER_REG(variant, |
| variant->hdmi_rdma_channels - 1); |
| drvdata->hdmiif_map = devm_regmap_init_mmio(dev, drvdata->hdmiif, |
| &lpass_hdmi_regmap_config); |
| if (IS_ERR(drvdata->hdmiif_map)) { |
| dev_err(dev, "error initializing regmap: %ld\n", |
| PTR_ERR(drvdata->hdmiif_map)); |
| return PTR_ERR(drvdata->hdmiif_map); |
| } |
| } |
| |
| if (variant->init) { |
| ret = variant->init(pdev); |
| if (ret) { |
| dev_err(dev, "error initializing variant: %d\n", ret); |
| return ret; |
| } |
| } |
| |
| for (i = 0; i < variant->num_dai; i++) { |
| dai_id = variant->dai_driver[i].id; |
| if (dai_id == LPASS_DP_RX || is_cdc_dma_port(dai_id)) |
| continue; |
| |
| drvdata->mi2s_osr_clk[dai_id] = devm_clk_get_optional(dev, |
| variant->dai_osr_clk_names[i]); |
| drvdata->mi2s_bit_clk[dai_id] = devm_clk_get(dev, |
| variant->dai_bit_clk_names[i]); |
| if (IS_ERR(drvdata->mi2s_bit_clk[dai_id])) { |
| dev_err(dev, |
| "error getting %s: %ld\n", |
| variant->dai_bit_clk_names[i], |
| PTR_ERR(drvdata->mi2s_bit_clk[dai_id])); |
| return PTR_ERR(drvdata->mi2s_bit_clk[dai_id]); |
| } |
| if (drvdata->mi2s_playback_sd_mode[dai_id] == |
| LPAIF_I2SCTL_MODE_QUAD01) { |
| variant->dai_driver[dai_id].playback.channels_min = 4; |
| variant->dai_driver[dai_id].playback.channels_max = 4; |
| } |
| } |
| |
| /* Allocation for i2sctl regmap fields */ |
| drvdata->i2sctl = devm_kzalloc(&pdev->dev, sizeof(struct lpaif_i2sctl), |
| GFP_KERNEL); |
| |
| /* Initialize bitfields for dai I2SCTL register */ |
| ret = lpass_cpu_init_i2sctl_bitfields(dev, drvdata->i2sctl, |
| drvdata->lpaif_map); |
| if (ret) { |
| dev_err(dev, "error init i2sctl field: %d\n", ret); |
| return ret; |
| } |
| |
| if (drvdata->hdmi_port_enable) { |
| ret = lpass_hdmi_init_bitfields(dev, drvdata->hdmiif_map); |
| if (ret) { |
| dev_err(dev, "%s error hdmi init failed\n", __func__); |
| return ret; |
| } |
| } |
| ret = devm_snd_soc_register_component(dev, |
| &lpass_cpu_comp_driver, |
| variant->dai_driver, |
| variant->num_dai); |
| if (ret) { |
| dev_err(dev, "error registering cpu driver: %d\n", ret); |
| goto err; |
| } |
| |
| ret = asoc_qcom_lpass_platform_register(pdev); |
| if (ret) { |
| dev_err(dev, "error registering platform driver: %d\n", ret); |
| goto err; |
| } |
| |
| err: |
| return ret; |
| } |
| EXPORT_SYMBOL_GPL(asoc_qcom_lpass_cpu_platform_probe); |
| |
| int asoc_qcom_lpass_cpu_platform_remove(struct platform_device *pdev) |
| { |
| struct lpass_data *drvdata = platform_get_drvdata(pdev); |
| |
| if (drvdata->variant->exit) |
| drvdata->variant->exit(pdev); |
| |
| |
| return 0; |
| } |
| EXPORT_SYMBOL_GPL(asoc_qcom_lpass_cpu_platform_remove); |
| |
| void asoc_qcom_lpass_cpu_platform_shutdown(struct platform_device *pdev) |
| { |
| struct lpass_data *drvdata = platform_get_drvdata(pdev); |
| |
| if (drvdata->variant->exit) |
| drvdata->variant->exit(pdev); |
| |
| } |
| EXPORT_SYMBOL_GPL(asoc_qcom_lpass_cpu_platform_shutdown); |
| |
| MODULE_DESCRIPTION("QTi LPASS CPU Driver"); |
| MODULE_LICENSE("GPL v2"); |