| // SPDX-License-Identifier: GPL-2.0 |
| /* Copyright (c) 2012, The Linux Foundation. All rights reserved. |
| * |
| * Description: CoreSight Trace Memory Controller driver |
| */ |
| |
| #include <linux/kernel.h> |
| #include <linux/init.h> |
| #include <linux/types.h> |
| #include <linux/device.h> |
| #include <linux/idr.h> |
| #include <linux/io.h> |
| #include <linux/err.h> |
| #include <linux/fs.h> |
| #include <linux/miscdevice.h> |
| #include <linux/mutex.h> |
| #include <linux/property.h> |
| #include <linux/uaccess.h> |
| #include <linux/slab.h> |
| #include <linux/dma-mapping.h> |
| #include <linux/spinlock.h> |
| #include <linux/pm_runtime.h> |
| #include <linux/of.h> |
| #include <linux/coresight.h> |
| #include <linux/amba/bus.h> |
| |
| #include "coresight-priv.h" |
| #include "coresight-tmc.h" |
| |
| DEFINE_CORESIGHT_DEVLIST(etb_devs, "tmc_etb"); |
| DEFINE_CORESIGHT_DEVLIST(etf_devs, "tmc_etf"); |
| DEFINE_CORESIGHT_DEVLIST(etr_devs, "tmc_etr"); |
| |
| void tmc_wait_for_tmcready(struct tmc_drvdata *drvdata) |
| { |
| /* Ensure formatter, unformatter and hardware fifo are empty */ |
| if (coresight_timeout(drvdata->base, |
| TMC_STS, TMC_STS_TMCREADY_BIT, 1)) { |
| dev_err(&drvdata->csdev->dev, |
| "timeout while waiting for TMC to be Ready\n"); |
| } |
| } |
| |
| void tmc_flush_and_stop(struct tmc_drvdata *drvdata) |
| { |
| u32 ffcr; |
| |
| ffcr = readl_relaxed(drvdata->base + TMC_FFCR); |
| ffcr |= TMC_FFCR_STOP_ON_FLUSH; |
| writel_relaxed(ffcr, drvdata->base + TMC_FFCR); |
| ffcr |= BIT(TMC_FFCR_FLUSHMAN_BIT); |
| writel_relaxed(ffcr, drvdata->base + TMC_FFCR); |
| /* Ensure flush completes */ |
| if (coresight_timeout(drvdata->base, |
| TMC_FFCR, TMC_FFCR_FLUSHMAN_BIT, 0)) { |
| dev_err(&drvdata->csdev->dev, |
| "timeout while waiting for completion of Manual Flush\n"); |
| } |
| |
| tmc_wait_for_tmcready(drvdata); |
| } |
| |
| void tmc_enable_hw(struct tmc_drvdata *drvdata) |
| { |
| writel_relaxed(TMC_CTL_CAPT_EN, drvdata->base + TMC_CTL); |
| } |
| |
| void tmc_disable_hw(struct tmc_drvdata *drvdata) |
| { |
| writel_relaxed(0x0, drvdata->base + TMC_CTL); |
| } |
| |
| u32 tmc_get_memwidth_mask(struct tmc_drvdata *drvdata) |
| { |
| u32 mask = 0; |
| |
| /* |
| * When moving RRP or an offset address forward, the new values must |
| * be byte-address aligned to the width of the trace memory databus |
| * _and_ to a frame boundary (16 byte), whichever is the biggest. For |
| * example, for 32-bit, 64-bit and 128-bit wide trace memory, the four |
| * LSBs must be 0s. For 256-bit wide trace memory, the five LSBs must |
| * be 0s. |
| */ |
| switch (drvdata->memwidth) { |
| case TMC_MEM_INTF_WIDTH_32BITS: |
| case TMC_MEM_INTF_WIDTH_64BITS: |
| case TMC_MEM_INTF_WIDTH_128BITS: |
| mask = GENMASK(31, 4); |
| break; |
| case TMC_MEM_INTF_WIDTH_256BITS: |
| mask = GENMASK(31, 5); |
| break; |
| } |
| |
| return mask; |
| } |
| |
| static int tmc_read_prepare(struct tmc_drvdata *drvdata) |
| { |
| int ret = 0; |
| |
| switch (drvdata->config_type) { |
| case TMC_CONFIG_TYPE_ETB: |
| case TMC_CONFIG_TYPE_ETF: |
| ret = tmc_read_prepare_etb(drvdata); |
| break; |
| case TMC_CONFIG_TYPE_ETR: |
| ret = tmc_read_prepare_etr(drvdata); |
| break; |
| default: |
| ret = -EINVAL; |
| } |
| |
| if (!ret) |
| dev_dbg(&drvdata->csdev->dev, "TMC read start\n"); |
| |
| return ret; |
| } |
| |
| static int tmc_read_unprepare(struct tmc_drvdata *drvdata) |
| { |
| int ret = 0; |
| |
| switch (drvdata->config_type) { |
| case TMC_CONFIG_TYPE_ETB: |
| case TMC_CONFIG_TYPE_ETF: |
| ret = tmc_read_unprepare_etb(drvdata); |
| break; |
| case TMC_CONFIG_TYPE_ETR: |
| ret = tmc_read_unprepare_etr(drvdata); |
| break; |
| default: |
| ret = -EINVAL; |
| } |
| |
| if (!ret) |
| dev_dbg(&drvdata->csdev->dev, "TMC read end\n"); |
| |
| return ret; |
| } |
| |
| static int tmc_open(struct inode *inode, struct file *file) |
| { |
| int ret; |
| struct tmc_drvdata *drvdata = container_of(file->private_data, |
| struct tmc_drvdata, miscdev); |
| |
| ret = tmc_read_prepare(drvdata); |
| if (ret) |
| return ret; |
| |
| nonseekable_open(inode, file); |
| |
| dev_dbg(&drvdata->csdev->dev, "%s: successfully opened\n", __func__); |
| return 0; |
| } |
| |
| static inline ssize_t tmc_get_sysfs_trace(struct tmc_drvdata *drvdata, |
| loff_t pos, size_t len, char **bufpp) |
| { |
| switch (drvdata->config_type) { |
| case TMC_CONFIG_TYPE_ETB: |
| case TMC_CONFIG_TYPE_ETF: |
| return tmc_etb_get_sysfs_trace(drvdata, pos, len, bufpp); |
| case TMC_CONFIG_TYPE_ETR: |
| return tmc_etr_get_sysfs_trace(drvdata, pos, len, bufpp); |
| } |
| |
| return -EINVAL; |
| } |
| |
| static ssize_t tmc_read(struct file *file, char __user *data, size_t len, |
| loff_t *ppos) |
| { |
| char *bufp; |
| ssize_t actual; |
| struct tmc_drvdata *drvdata = container_of(file->private_data, |
| struct tmc_drvdata, miscdev); |
| actual = tmc_get_sysfs_trace(drvdata, *ppos, len, &bufp); |
| if (actual <= 0) |
| return 0; |
| |
| if (copy_to_user(data, bufp, actual)) { |
| dev_dbg(&drvdata->csdev->dev, |
| "%s: copy_to_user failed\n", __func__); |
| return -EFAULT; |
| } |
| |
| *ppos += actual; |
| dev_dbg(&drvdata->csdev->dev, "%zu bytes copied\n", actual); |
| |
| return actual; |
| } |
| |
| static int tmc_release(struct inode *inode, struct file *file) |
| { |
| int ret; |
| struct tmc_drvdata *drvdata = container_of(file->private_data, |
| struct tmc_drvdata, miscdev); |
| |
| ret = tmc_read_unprepare(drvdata); |
| if (ret) |
| return ret; |
| |
| dev_dbg(&drvdata->csdev->dev, "%s: released\n", __func__); |
| return 0; |
| } |
| |
| static const struct file_operations tmc_fops = { |
| .owner = THIS_MODULE, |
| .open = tmc_open, |
| .read = tmc_read, |
| .release = tmc_release, |
| .llseek = no_llseek, |
| }; |
| |
| static enum tmc_mem_intf_width tmc_get_memwidth(u32 devid) |
| { |
| enum tmc_mem_intf_width memwidth; |
| |
| /* |
| * Excerpt from the TRM: |
| * |
| * DEVID::MEMWIDTH[10:8] |
| * 0x2 Memory interface databus is 32 bits wide. |
| * 0x3 Memory interface databus is 64 bits wide. |
| * 0x4 Memory interface databus is 128 bits wide. |
| * 0x5 Memory interface databus is 256 bits wide. |
| */ |
| switch (BMVAL(devid, 8, 10)) { |
| case 0x2: |
| memwidth = TMC_MEM_INTF_WIDTH_32BITS; |
| break; |
| case 0x3: |
| memwidth = TMC_MEM_INTF_WIDTH_64BITS; |
| break; |
| case 0x4: |
| memwidth = TMC_MEM_INTF_WIDTH_128BITS; |
| break; |
| case 0x5: |
| memwidth = TMC_MEM_INTF_WIDTH_256BITS; |
| break; |
| default: |
| memwidth = 0; |
| } |
| |
| return memwidth; |
| } |
| |
| #define coresight_tmc_reg(name, offset) \ |
| coresight_simple_reg32(struct tmc_drvdata, name, offset) |
| #define coresight_tmc_reg64(name, lo_off, hi_off) \ |
| coresight_simple_reg64(struct tmc_drvdata, name, lo_off, hi_off) |
| |
| coresight_tmc_reg(rsz, TMC_RSZ); |
| coresight_tmc_reg(sts, TMC_STS); |
| coresight_tmc_reg(trg, TMC_TRG); |
| coresight_tmc_reg(ctl, TMC_CTL); |
| coresight_tmc_reg(ffsr, TMC_FFSR); |
| coresight_tmc_reg(ffcr, TMC_FFCR); |
| coresight_tmc_reg(mode, TMC_MODE); |
| coresight_tmc_reg(pscr, TMC_PSCR); |
| coresight_tmc_reg(axictl, TMC_AXICTL); |
| coresight_tmc_reg(authstatus, TMC_AUTHSTATUS); |
| coresight_tmc_reg(devid, CORESIGHT_DEVID); |
| coresight_tmc_reg64(rrp, TMC_RRP, TMC_RRPHI); |
| coresight_tmc_reg64(rwp, TMC_RWP, TMC_RWPHI); |
| coresight_tmc_reg64(dba, TMC_DBALO, TMC_DBAHI); |
| |
| static struct attribute *coresight_tmc_mgmt_attrs[] = { |
| &dev_attr_rsz.attr, |
| &dev_attr_sts.attr, |
| &dev_attr_rrp.attr, |
| &dev_attr_rwp.attr, |
| &dev_attr_trg.attr, |
| &dev_attr_ctl.attr, |
| &dev_attr_ffsr.attr, |
| &dev_attr_ffcr.attr, |
| &dev_attr_mode.attr, |
| &dev_attr_pscr.attr, |
| &dev_attr_devid.attr, |
| &dev_attr_dba.attr, |
| &dev_attr_axictl.attr, |
| &dev_attr_authstatus.attr, |
| NULL, |
| }; |
| |
| static ssize_t trigger_cntr_show(struct device *dev, |
| struct device_attribute *attr, char *buf) |
| { |
| struct tmc_drvdata *drvdata = dev_get_drvdata(dev->parent); |
| unsigned long val = drvdata->trigger_cntr; |
| |
| return sprintf(buf, "%#lx\n", val); |
| } |
| |
| static ssize_t trigger_cntr_store(struct device *dev, |
| struct device_attribute *attr, |
| const char *buf, size_t size) |
| { |
| int ret; |
| unsigned long val; |
| struct tmc_drvdata *drvdata = dev_get_drvdata(dev->parent); |
| |
| ret = kstrtoul(buf, 16, &val); |
| if (ret) |
| return ret; |
| |
| drvdata->trigger_cntr = val; |
| return size; |
| } |
| static DEVICE_ATTR_RW(trigger_cntr); |
| |
| static ssize_t buffer_size_show(struct device *dev, |
| struct device_attribute *attr, char *buf) |
| { |
| struct tmc_drvdata *drvdata = dev_get_drvdata(dev->parent); |
| |
| return sprintf(buf, "%#x\n", drvdata->size); |
| } |
| |
| static ssize_t buffer_size_store(struct device *dev, |
| struct device_attribute *attr, |
| const char *buf, size_t size) |
| { |
| int ret; |
| unsigned long val; |
| struct tmc_drvdata *drvdata = dev_get_drvdata(dev->parent); |
| |
| /* Only permitted for TMC-ETRs */ |
| if (drvdata->config_type != TMC_CONFIG_TYPE_ETR) |
| return -EPERM; |
| |
| ret = kstrtoul(buf, 0, &val); |
| if (ret) |
| return ret; |
| /* The buffer size should be page aligned */ |
| if (val & (PAGE_SIZE - 1)) |
| return -EINVAL; |
| drvdata->size = val; |
| return size; |
| } |
| |
| static DEVICE_ATTR_RW(buffer_size); |
| |
| static struct attribute *coresight_tmc_attrs[] = { |
| &dev_attr_trigger_cntr.attr, |
| &dev_attr_buffer_size.attr, |
| NULL, |
| }; |
| |
| static const struct attribute_group coresight_tmc_group = { |
| .attrs = coresight_tmc_attrs, |
| }; |
| |
| static const struct attribute_group coresight_tmc_mgmt_group = { |
| .attrs = coresight_tmc_mgmt_attrs, |
| .name = "mgmt", |
| }; |
| |
| static const struct attribute_group *coresight_tmc_groups[] = { |
| &coresight_tmc_group, |
| &coresight_tmc_mgmt_group, |
| NULL, |
| }; |
| |
| static inline bool tmc_etr_can_use_sg(struct device *dev) |
| { |
| return fwnode_property_present(dev->fwnode, "arm,scatter-gather"); |
| } |
| |
| static inline bool tmc_etr_has_non_secure_access(struct tmc_drvdata *drvdata) |
| { |
| u32 auth = readl_relaxed(drvdata->base + TMC_AUTHSTATUS); |
| |
| return (auth & TMC_AUTH_NSID_MASK) == 0x3; |
| } |
| |
| /* Detect and initialise the capabilities of a TMC ETR */ |
| static int tmc_etr_setup_caps(struct device *parent, u32 devid, void *dev_caps) |
| { |
| int rc; |
| u32 dma_mask = 0; |
| struct tmc_drvdata *drvdata = dev_get_drvdata(parent); |
| |
| if (!tmc_etr_has_non_secure_access(drvdata)) |
| return -EACCES; |
| |
| /* Set the unadvertised capabilities */ |
| tmc_etr_init_caps(drvdata, (u32)(unsigned long)dev_caps); |
| |
| if (!(devid & TMC_DEVID_NOSCAT) && tmc_etr_can_use_sg(parent)) |
| tmc_etr_set_cap(drvdata, TMC_ETR_SG); |
| |
| /* Check if the AXI address width is available */ |
| if (devid & TMC_DEVID_AXIAW_VALID) |
| dma_mask = ((devid >> TMC_DEVID_AXIAW_SHIFT) & |
| TMC_DEVID_AXIAW_MASK); |
| |
| /* |
| * Unless specified in the device configuration, ETR uses a 40-bit |
| * AXI master in place of the embedded SRAM of ETB/ETF. |
| */ |
| switch (dma_mask) { |
| case 32: |
| case 40: |
| case 44: |
| case 48: |
| case 52: |
| dev_info(parent, "Detected dma mask %dbits\n", dma_mask); |
| break; |
| default: |
| dma_mask = 40; |
| } |
| |
| rc = dma_set_mask_and_coherent(parent, DMA_BIT_MASK(dma_mask)); |
| if (rc) |
| dev_err(parent, "Failed to setup DMA mask: %d\n", rc); |
| return rc; |
| } |
| |
| static u32 tmc_etr_get_default_buffer_size(struct device *dev) |
| { |
| u32 size; |
| |
| if (fwnode_property_read_u32(dev->fwnode, "arm,buffer-size", &size)) |
| size = SZ_1M; |
| return size; |
| } |
| |
| static int tmc_probe(struct amba_device *adev, const struct amba_id *id) |
| { |
| int ret = 0; |
| u32 devid; |
| void __iomem *base; |
| struct device *dev = &adev->dev; |
| struct coresight_platform_data *pdata = NULL; |
| struct tmc_drvdata *drvdata; |
| struct resource *res = &adev->res; |
| struct coresight_desc desc = { 0 }; |
| struct coresight_dev_list *dev_list = NULL; |
| |
| ret = -ENOMEM; |
| drvdata = devm_kzalloc(dev, sizeof(*drvdata), GFP_KERNEL); |
| if (!drvdata) |
| goto out; |
| |
| dev_set_drvdata(dev, drvdata); |
| |
| /* Validity for the resource is already checked by the AMBA core */ |
| base = devm_ioremap_resource(dev, res); |
| if (IS_ERR(base)) { |
| ret = PTR_ERR(base); |
| goto out; |
| } |
| |
| drvdata->base = base; |
| |
| spin_lock_init(&drvdata->spinlock); |
| |
| devid = readl_relaxed(drvdata->base + CORESIGHT_DEVID); |
| drvdata->config_type = BMVAL(devid, 6, 7); |
| drvdata->memwidth = tmc_get_memwidth(devid); |
| /* This device is not associated with a session */ |
| drvdata->pid = -1; |
| |
| if (drvdata->config_type == TMC_CONFIG_TYPE_ETR) |
| drvdata->size = tmc_etr_get_default_buffer_size(dev); |
| else |
| drvdata->size = readl_relaxed(drvdata->base + TMC_RSZ) * 4; |
| |
| desc.dev = dev; |
| desc.groups = coresight_tmc_groups; |
| |
| switch (drvdata->config_type) { |
| case TMC_CONFIG_TYPE_ETB: |
| desc.type = CORESIGHT_DEV_TYPE_SINK; |
| desc.subtype.sink_subtype = CORESIGHT_DEV_SUBTYPE_SINK_BUFFER; |
| desc.ops = &tmc_etb_cs_ops; |
| dev_list = &etb_devs; |
| break; |
| case TMC_CONFIG_TYPE_ETR: |
| desc.type = CORESIGHT_DEV_TYPE_SINK; |
| desc.subtype.sink_subtype = CORESIGHT_DEV_SUBTYPE_SINK_SYSMEM; |
| desc.ops = &tmc_etr_cs_ops; |
| ret = tmc_etr_setup_caps(dev, devid, |
| coresight_get_uci_data(id)); |
| if (ret) |
| goto out; |
| idr_init(&drvdata->idr); |
| mutex_init(&drvdata->idr_mutex); |
| dev_list = &etr_devs; |
| break; |
| case TMC_CONFIG_TYPE_ETF: |
| desc.type = CORESIGHT_DEV_TYPE_LINKSINK; |
| desc.subtype.sink_subtype = CORESIGHT_DEV_SUBTYPE_SINK_BUFFER; |
| desc.subtype.link_subtype = CORESIGHT_DEV_SUBTYPE_LINK_FIFO; |
| desc.ops = &tmc_etf_cs_ops; |
| dev_list = &etf_devs; |
| break; |
| default: |
| pr_err("%s: Unsupported TMC config\n", desc.name); |
| ret = -EINVAL; |
| goto out; |
| } |
| |
| desc.name = coresight_alloc_device_name(dev_list, dev); |
| if (!desc.name) { |
| ret = -ENOMEM; |
| goto out; |
| } |
| |
| pdata = coresight_get_platform_data(dev); |
| if (IS_ERR(pdata)) { |
| ret = PTR_ERR(pdata); |
| goto out; |
| } |
| adev->dev.platform_data = pdata; |
| desc.pdata = pdata; |
| |
| drvdata->csdev = coresight_register(&desc); |
| if (IS_ERR(drvdata->csdev)) { |
| ret = PTR_ERR(drvdata->csdev); |
| goto out; |
| } |
| |
| drvdata->miscdev.name = desc.name; |
| drvdata->miscdev.minor = MISC_DYNAMIC_MINOR; |
| drvdata->miscdev.fops = &tmc_fops; |
| ret = misc_register(&drvdata->miscdev); |
| if (ret) |
| coresight_unregister(drvdata->csdev); |
| else |
| pm_runtime_put(&adev->dev); |
| out: |
| return ret; |
| } |
| |
| static void tmc_shutdown(struct amba_device *adev) |
| { |
| unsigned long flags; |
| struct tmc_drvdata *drvdata = amba_get_drvdata(adev); |
| |
| spin_lock_irqsave(&drvdata->spinlock, flags); |
| |
| if (drvdata->mode == CS_MODE_DISABLED) |
| goto out; |
| |
| if (drvdata->config_type == TMC_CONFIG_TYPE_ETR) |
| tmc_etr_disable_hw(drvdata); |
| |
| /* |
| * We do not care about coresight unregister here unlike remove |
| * callback which is required for making coresight modular since |
| * the system is going down after this. |
| */ |
| out: |
| spin_unlock_irqrestore(&drvdata->spinlock, flags); |
| } |
| |
| static const struct amba_id tmc_ids[] = { |
| CS_AMBA_ID(0x000bb961), |
| /* Coresight SoC 600 TMC-ETR/ETS */ |
| CS_AMBA_ID_DATA(0x000bb9e8, (unsigned long)CORESIGHT_SOC_600_ETR_CAPS), |
| /* Coresight SoC 600 TMC-ETB */ |
| CS_AMBA_ID(0x000bb9e9), |
| /* Coresight SoC 600 TMC-ETF */ |
| CS_AMBA_ID(0x000bb9ea), |
| { 0, 0}, |
| }; |
| |
| static struct amba_driver tmc_driver = { |
| .drv = { |
| .name = "coresight-tmc", |
| .owner = THIS_MODULE, |
| .suppress_bind_attrs = true, |
| }, |
| .probe = tmc_probe, |
| .shutdown = tmc_shutdown, |
| .id_table = tmc_ids, |
| }; |
| builtin_amba_driver(tmc_driver); |