| // SPDX-License-Identifier: GPL-2.0 |
| // PCI1xxxx SPI driver |
| // Copyright (C) 2022 Microchip Technology Inc. |
| // Authors: Tharun Kumar P <tharunkumar.pasumarthi@microchip.com> |
| // Kumaravel Thiagarajan <Kumaravel.Thiagarajan@microchip.com> |
| |
| |
| #include <linux/module.h> |
| #include <linux/pci.h> |
| #include <linux/spi/spi.h> |
| #include <linux/delay.h> |
| |
| #define DRV_NAME "spi-pci1xxxx" |
| |
| #define SYS_FREQ_DEFAULT (62500000) |
| |
| #define PCI1XXXX_SPI_MAX_CLOCK_HZ (30000000) |
| #define PCI1XXXX_SPI_CLK_20MHZ (20000000) |
| #define PCI1XXXX_SPI_CLK_15MHZ (15000000) |
| #define PCI1XXXX_SPI_CLK_12MHZ (12000000) |
| #define PCI1XXXX_SPI_CLK_10MHZ (10000000) |
| #define PCI1XXXX_SPI_MIN_CLOCK_HZ (2000000) |
| |
| #define PCI1XXXX_SPI_BUFFER_SIZE (320) |
| |
| #define SPI_MST_CTL_DEVSEL_MASK (GENMASK(27, 25)) |
| #define SPI_MST_CTL_CMD_LEN_MASK (GENMASK(16, 8)) |
| #define SPI_MST_CTL_SPEED_MASK (GENMASK(7, 5)) |
| #define SPI_MSI_VECTOR_SEL_MASK (GENMASK(4, 4)) |
| |
| #define SPI_MST_CTL_FORCE_CE (BIT(4)) |
| #define SPI_MST_CTL_MODE_SEL (BIT(2)) |
| #define SPI_MST_CTL_GO (BIT(0)) |
| |
| #define SPI_MST1_ADDR_BASE (0x800) |
| |
| /* x refers to SPI Host Controller HW instance id in the below macros - 0 or 1 */ |
| |
| #define SPI_MST_CMD_BUF_OFFSET(x) (((x) * SPI_MST1_ADDR_BASE) + 0x00) |
| #define SPI_MST_RSP_BUF_OFFSET(x) (((x) * SPI_MST1_ADDR_BASE) + 0x200) |
| #define SPI_MST_CTL_REG_OFFSET(x) (((x) * SPI_MST1_ADDR_BASE) + 0x400) |
| #define SPI_MST_EVENT_REG_OFFSET(x) (((x) * SPI_MST1_ADDR_BASE) + 0x420) |
| #define SPI_MST_EVENT_MASK_REG_OFFSET(x) (((x) * SPI_MST1_ADDR_BASE) + 0x424) |
| #define SPI_MST_PAD_CTL_REG_OFFSET(x) (((x) * SPI_MST1_ADDR_BASE) + 0x460) |
| #define SPIALERT_MST_DB_REG_OFFSET(x) (((x) * SPI_MST1_ADDR_BASE) + 0x464) |
| #define SPIALERT_MST_VAL_REG_OFFSET(x) (((x) * SPI_MST1_ADDR_BASE) + 0x468) |
| #define SPI_PCI_CTRL_REG_OFFSET(x) (((x) * SPI_MST1_ADDR_BASE) + 0x480) |
| |
| #define PCI1XXXX_IRQ_FLAGS (IRQF_NO_SUSPEND | IRQF_TRIGGER_NONE) |
| #define SPI_MAX_DATA_LEN 320 |
| |
| #define PCI1XXXX_SPI_TIMEOUT (msecs_to_jiffies(100)) |
| |
| #define SPI_INTR BIT(8) |
| #define SPI_FORCE_CE BIT(4) |
| |
| #define SPI_CHIP_SEL_COUNT 7 |
| #define VENDOR_ID_MCHP 0x1055 |
| |
| #define SPI_SUSPEND_CONFIG 0x101 |
| #define SPI_RESUME_CONFIG 0x303 |
| |
| struct pci1xxxx_spi_internal { |
| u8 hw_inst; |
| bool spi_xfer_in_progress; |
| int irq; |
| struct completion spi_xfer_done; |
| struct spi_master *spi_host; |
| struct pci1xxxx_spi *parent; |
| struct { |
| unsigned int dev_sel : 3; |
| unsigned int msi_vector_sel : 1; |
| } prev_val; |
| }; |
| |
| struct pci1xxxx_spi { |
| struct pci_dev *dev; |
| u8 total_hw_instances; |
| void __iomem *reg_base; |
| struct pci1xxxx_spi_internal *spi_int[]; |
| }; |
| |
| static const struct pci_device_id pci1xxxx_spi_pci_id_table[] = { |
| { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa004, PCI_ANY_ID, 0x0001), 0, 0, 0x02}, |
| { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa004, PCI_ANY_ID, 0x0002), 0, 0, 0x01}, |
| { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa004, PCI_ANY_ID, 0x0003), 0, 0, 0x11}, |
| { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa004, PCI_ANY_ID, PCI_ANY_ID), 0, 0, 0x01}, |
| { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa014, PCI_ANY_ID, 0x0001), 0, 0, 0x02}, |
| { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa014, PCI_ANY_ID, 0x0002), 0, 0, 0x01}, |
| { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa014, PCI_ANY_ID, 0x0003), 0, 0, 0x11}, |
| { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa014, PCI_ANY_ID, PCI_ANY_ID), 0, 0, 0x01}, |
| { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa024, PCI_ANY_ID, 0x0001), 0, 0, 0x02}, |
| { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa024, PCI_ANY_ID, 0x0002), 0, 0, 0x01}, |
| { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa024, PCI_ANY_ID, 0x0003), 0, 0, 0x11}, |
| { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa024, PCI_ANY_ID, PCI_ANY_ID), 0, 0, 0x01}, |
| { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa034, PCI_ANY_ID, 0x0001), 0, 0, 0x02}, |
| { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa034, PCI_ANY_ID, 0x0002), 0, 0, 0x01}, |
| { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa034, PCI_ANY_ID, 0x0003), 0, 0, 0x11}, |
| { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa034, PCI_ANY_ID, PCI_ANY_ID), 0, 0, 0x01}, |
| { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa044, PCI_ANY_ID, 0x0001), 0, 0, 0x02}, |
| { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa044, PCI_ANY_ID, 0x0002), 0, 0, 0x01}, |
| { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa044, PCI_ANY_ID, 0x0003), 0, 0, 0x11}, |
| { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa044, PCI_ANY_ID, PCI_ANY_ID), 0, 0, 0x01}, |
| { 0, } |
| }; |
| |
| MODULE_DEVICE_TABLE(pci, pci1xxxx_spi_pci_id_table); |
| |
| static void pci1xxxx_spi_set_cs(struct spi_device *spi, bool enable) |
| { |
| struct pci1xxxx_spi_internal *p = spi_controller_get_devdata(spi->controller); |
| struct pci1xxxx_spi *par = p->parent; |
| u32 regval; |
| |
| /* Set the DEV_SEL bits of the SPI_MST_CTL_REG */ |
| regval = readl(par->reg_base + SPI_MST_CTL_REG_OFFSET(p->hw_inst)); |
| if (enable) { |
| regval &= ~SPI_MST_CTL_DEVSEL_MASK; |
| regval |= (spi->chip_select << 25); |
| writel(regval, |
| par->reg_base + SPI_MST_CTL_REG_OFFSET(p->hw_inst)); |
| } else { |
| regval &= ~(spi->chip_select << 25); |
| writel(regval, |
| par->reg_base + SPI_MST_CTL_REG_OFFSET(p->hw_inst)); |
| |
| } |
| } |
| |
| static u8 pci1xxxx_get_clock_div(u32 hz) |
| { |
| u8 val = 0; |
| |
| if (hz >= PCI1XXXX_SPI_MAX_CLOCK_HZ) |
| val = 2; |
| else if ((hz < PCI1XXXX_SPI_MAX_CLOCK_HZ) && (hz >= PCI1XXXX_SPI_CLK_20MHZ)) |
| val = 3; |
| else if ((hz < PCI1XXXX_SPI_CLK_20MHZ) && (hz >= PCI1XXXX_SPI_CLK_15MHZ)) |
| val = 4; |
| else if ((hz < PCI1XXXX_SPI_CLK_15MHZ) && (hz >= PCI1XXXX_SPI_CLK_12MHZ)) |
| val = 5; |
| else if ((hz < PCI1XXXX_SPI_CLK_12MHZ) && (hz >= PCI1XXXX_SPI_CLK_10MHZ)) |
| val = 6; |
| else if ((hz < PCI1XXXX_SPI_CLK_10MHZ) && (hz >= PCI1XXXX_SPI_MIN_CLOCK_HZ)) |
| val = 7; |
| else |
| val = 2; |
| |
| return val; |
| } |
| |
| static int pci1xxxx_spi_transfer_one(struct spi_controller *spi_ctlr, |
| struct spi_device *spi, struct spi_transfer *xfer) |
| { |
| struct pci1xxxx_spi_internal *p = spi_controller_get_devdata(spi_ctlr); |
| int mode, len, loop_iter, transfer_len; |
| struct pci1xxxx_spi *par = p->parent; |
| unsigned long bytes_transfered; |
| unsigned long bytes_recvd; |
| unsigned long loop_count; |
| u8 *rx_buf, result; |
| const u8 *tx_buf; |
| u32 regval; |
| u8 clkdiv; |
| |
| p->spi_xfer_in_progress = true; |
| mode = spi->mode; |
| clkdiv = pci1xxxx_get_clock_div(xfer->speed_hz); |
| tx_buf = xfer->tx_buf; |
| rx_buf = xfer->rx_buf; |
| transfer_len = xfer->len; |
| regval = readl(par->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst)); |
| writel(regval, par->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst)); |
| |
| if (tx_buf) { |
| bytes_transfered = 0; |
| bytes_recvd = 0; |
| loop_count = transfer_len / SPI_MAX_DATA_LEN; |
| if (transfer_len % SPI_MAX_DATA_LEN != 0) |
| loop_count += 1; |
| |
| for (loop_iter = 0; loop_iter < loop_count; loop_iter++) { |
| len = SPI_MAX_DATA_LEN; |
| if ((transfer_len % SPI_MAX_DATA_LEN != 0) && |
| (loop_iter == loop_count - 1)) |
| len = transfer_len % SPI_MAX_DATA_LEN; |
| |
| reinit_completion(&p->spi_xfer_done); |
| memcpy_toio(par->reg_base + SPI_MST_CMD_BUF_OFFSET(p->hw_inst), |
| &tx_buf[bytes_transfered], len); |
| bytes_transfered += len; |
| regval = readl(par->reg_base + |
| SPI_MST_CTL_REG_OFFSET(p->hw_inst)); |
| regval &= ~(SPI_MST_CTL_MODE_SEL | SPI_MST_CTL_CMD_LEN_MASK | |
| SPI_MST_CTL_SPEED_MASK); |
| |
| if (mode == SPI_MODE_3) |
| regval |= SPI_MST_CTL_MODE_SEL; |
| else |
| regval &= ~SPI_MST_CTL_MODE_SEL; |
| |
| regval |= ((clkdiv << 5) | SPI_FORCE_CE | (len << 8)); |
| regval &= ~SPI_MST_CTL_CMD_LEN_MASK; |
| writel(regval, par->reg_base + |
| SPI_MST_CTL_REG_OFFSET(p->hw_inst)); |
| regval = readl(par->reg_base + |
| SPI_MST_CTL_REG_OFFSET(p->hw_inst)); |
| regval |= SPI_MST_CTL_GO; |
| writel(regval, par->reg_base + |
| SPI_MST_CTL_REG_OFFSET(p->hw_inst)); |
| |
| /* Wait for DMA_TERM interrupt */ |
| result = wait_for_completion_timeout(&p->spi_xfer_done, |
| PCI1XXXX_SPI_TIMEOUT); |
| if (!result) |
| return -ETIMEDOUT; |
| |
| if (rx_buf) { |
| memcpy_fromio(&rx_buf[bytes_recvd], par->reg_base + |
| SPI_MST_RSP_BUF_OFFSET(p->hw_inst), len); |
| bytes_recvd += len; |
| } |
| } |
| } |
| |
| regval = readl(par->reg_base + SPI_MST_CTL_REG_OFFSET(p->hw_inst)); |
| regval &= ~SPI_FORCE_CE; |
| writel(regval, par->reg_base + SPI_MST_CTL_REG_OFFSET(p->hw_inst)); |
| p->spi_xfer_in_progress = false; |
| |
| return 0; |
| } |
| |
| static irqreturn_t pci1xxxx_spi_isr(int irq, void *dev) |
| { |
| struct pci1xxxx_spi_internal *p = dev; |
| irqreturn_t spi_int_fired = IRQ_NONE; |
| u32 regval; |
| |
| /* Clear the SPI GO_BIT Interrupt */ |
| regval = readl(p->parent->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst)); |
| if (regval & SPI_INTR) { |
| /* Clear xfer_done */ |
| complete(&p->spi_xfer_done); |
| spi_int_fired = IRQ_HANDLED; |
| } |
| |
| writel(regval, p->parent->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst)); |
| |
| return spi_int_fired; |
| } |
| |
| static int pci1xxxx_spi_probe(struct pci_dev *pdev, const struct pci_device_id *ent) |
| { |
| u8 hw_inst_cnt, iter, start, only_sec_inst; |
| struct pci1xxxx_spi_internal *spi_sub_ptr; |
| struct device *dev = &pdev->dev; |
| struct pci1xxxx_spi *spi_bus; |
| struct spi_master *spi_host; |
| u32 regval; |
| int ret; |
| |
| hw_inst_cnt = ent->driver_data & 0x0f; |
| start = (ent->driver_data & 0xf0) >> 4; |
| if (start == 1) |
| only_sec_inst = 1; |
| else |
| only_sec_inst = 0; |
| |
| spi_bus = devm_kzalloc(&pdev->dev, |
| struct_size(spi_bus, spi_int, hw_inst_cnt), |
| GFP_KERNEL); |
| if (!spi_bus) |
| return -ENOMEM; |
| |
| spi_bus->dev = pdev; |
| spi_bus->total_hw_instances = hw_inst_cnt; |
| pci_set_master(pdev); |
| |
| for (iter = 0; iter < hw_inst_cnt; iter++) { |
| spi_bus->spi_int[iter] = devm_kzalloc(&pdev->dev, |
| sizeof(struct pci1xxxx_spi_internal), |
| GFP_KERNEL); |
| spi_sub_ptr = spi_bus->spi_int[iter]; |
| spi_sub_ptr->spi_host = devm_spi_alloc_master(dev, sizeof(struct spi_master)); |
| if (!spi_sub_ptr->spi_host) |
| return -ENOMEM; |
| |
| spi_sub_ptr->parent = spi_bus; |
| spi_sub_ptr->spi_xfer_in_progress = false; |
| |
| if (!iter) { |
| ret = pcim_enable_device(pdev); |
| if (ret) |
| return -ENOMEM; |
| |
| ret = pci_request_regions(pdev, DRV_NAME); |
| if (ret) |
| return -ENOMEM; |
| |
| spi_bus->reg_base = pcim_iomap(pdev, 0, pci_resource_len(pdev, 0)); |
| if (!spi_bus->reg_base) { |
| ret = -EINVAL; |
| goto error; |
| } |
| |
| ret = pci_alloc_irq_vectors(pdev, hw_inst_cnt, hw_inst_cnt, |
| PCI_IRQ_ALL_TYPES); |
| if (ret < 0) { |
| dev_err(&pdev->dev, "Error allocating MSI vectors\n"); |
| goto error; |
| } |
| |
| init_completion(&spi_sub_ptr->spi_xfer_done); |
| /* Initialize Interrupts - SPI_INT */ |
| regval = readl(spi_bus->reg_base + |
| SPI_MST_EVENT_MASK_REG_OFFSET(spi_sub_ptr->hw_inst)); |
| regval &= ~SPI_INTR; |
| writel(regval, spi_bus->reg_base + |
| SPI_MST_EVENT_MASK_REG_OFFSET(spi_sub_ptr->hw_inst)); |
| spi_sub_ptr->irq = pci_irq_vector(pdev, 0); |
| |
| ret = devm_request_irq(&pdev->dev, spi_sub_ptr->irq, |
| pci1xxxx_spi_isr, PCI1XXXX_IRQ_FLAGS, |
| pci_name(pdev), spi_sub_ptr); |
| if (ret < 0) { |
| dev_err(&pdev->dev, "Unable to request irq : %d", |
| spi_sub_ptr->irq); |
| ret = -ENODEV; |
| goto error; |
| } |
| |
| /* This register is only applicable for 1st instance */ |
| regval = readl(spi_bus->reg_base + SPI_PCI_CTRL_REG_OFFSET(0)); |
| if (!only_sec_inst) |
| regval |= (BIT(4)); |
| else |
| regval &= ~(BIT(4)); |
| |
| writel(regval, spi_bus->reg_base + SPI_PCI_CTRL_REG_OFFSET(0)); |
| } |
| |
| spi_sub_ptr->hw_inst = start++; |
| |
| if (iter == 1) { |
| init_completion(&spi_sub_ptr->spi_xfer_done); |
| /* Initialize Interrupts - SPI_INT */ |
| regval = readl(spi_bus->reg_base + |
| SPI_MST_EVENT_MASK_REG_OFFSET(spi_sub_ptr->hw_inst)); |
| regval &= ~SPI_INTR; |
| writel(regval, spi_bus->reg_base + |
| SPI_MST_EVENT_MASK_REG_OFFSET(spi_sub_ptr->hw_inst)); |
| spi_sub_ptr->irq = pci_irq_vector(pdev, iter); |
| ret = devm_request_irq(&pdev->dev, spi_sub_ptr->irq, |
| pci1xxxx_spi_isr, PCI1XXXX_IRQ_FLAGS, |
| pci_name(pdev), spi_sub_ptr); |
| if (ret < 0) { |
| dev_err(&pdev->dev, "Unable to request irq : %d", |
| spi_sub_ptr->irq); |
| ret = -ENODEV; |
| goto error; |
| } |
| } |
| |
| spi_host = spi_sub_ptr->spi_host; |
| spi_host->num_chipselect = SPI_CHIP_SEL_COUNT; |
| spi_host->mode_bits = SPI_MODE_0 | SPI_MODE_3 | SPI_RX_DUAL | |
| SPI_TX_DUAL | SPI_LOOP; |
| spi_host->transfer_one = pci1xxxx_spi_transfer_one; |
| spi_host->set_cs = pci1xxxx_spi_set_cs; |
| spi_host->bits_per_word_mask = SPI_BPW_MASK(8); |
| spi_host->max_speed_hz = PCI1XXXX_SPI_MAX_CLOCK_HZ; |
| spi_host->min_speed_hz = PCI1XXXX_SPI_MIN_CLOCK_HZ; |
| spi_host->flags = SPI_MASTER_MUST_TX; |
| spi_master_set_devdata(spi_host, spi_sub_ptr); |
| ret = devm_spi_register_master(dev, spi_host); |
| if (ret) |
| goto error; |
| } |
| pci_set_drvdata(pdev, spi_bus); |
| |
| return 0; |
| |
| error: |
| pci_release_regions(pdev); |
| return ret; |
| } |
| |
| static void store_restore_config(struct pci1xxxx_spi *spi_ptr, |
| struct pci1xxxx_spi_internal *spi_sub_ptr, |
| u8 inst, bool store) |
| { |
| u32 regval; |
| |
| if (store) { |
| regval = readl(spi_ptr->reg_base + |
| SPI_MST_CTL_REG_OFFSET(spi_sub_ptr->hw_inst)); |
| regval &= SPI_MST_CTL_DEVSEL_MASK; |
| spi_sub_ptr->prev_val.dev_sel = (regval >> 25) & 7; |
| regval = readl(spi_ptr->reg_base + |
| SPI_PCI_CTRL_REG_OFFSET(spi_sub_ptr->hw_inst)); |
| regval &= SPI_MSI_VECTOR_SEL_MASK; |
| spi_sub_ptr->prev_val.msi_vector_sel = (regval >> 4) & 1; |
| } else { |
| regval = readl(spi_ptr->reg_base + SPI_MST_CTL_REG_OFFSET(inst)); |
| regval &= ~SPI_MST_CTL_DEVSEL_MASK; |
| regval |= (spi_sub_ptr->prev_val.dev_sel << 25); |
| writel(regval, |
| spi_ptr->reg_base + SPI_MST_CTL_REG_OFFSET(inst)); |
| writel((spi_sub_ptr->prev_val.msi_vector_sel << 4), |
| spi_ptr->reg_base + SPI_PCI_CTRL_REG_OFFSET(inst)); |
| } |
| } |
| |
| static int pci1xxxx_spi_resume(struct device *dev) |
| { |
| struct pci1xxxx_spi *spi_ptr = dev_get_drvdata(dev); |
| struct pci1xxxx_spi_internal *spi_sub_ptr; |
| u32 regval = SPI_RESUME_CONFIG; |
| u8 iter; |
| |
| for (iter = 0; iter < spi_ptr->total_hw_instances; iter++) { |
| spi_sub_ptr = spi_ptr->spi_int[iter]; |
| spi_master_resume(spi_sub_ptr->spi_host); |
| writel(regval, spi_ptr->reg_base + |
| SPI_MST_EVENT_MASK_REG_OFFSET(iter)); |
| |
| /* Restore config at resume */ |
| store_restore_config(spi_ptr, spi_sub_ptr, iter, 0); |
| } |
| |
| return 0; |
| } |
| |
| static int pci1xxxx_spi_suspend(struct device *dev) |
| { |
| struct pci1xxxx_spi *spi_ptr = dev_get_drvdata(dev); |
| struct pci1xxxx_spi_internal *spi_sub_ptr; |
| u32 reg1 = SPI_SUSPEND_CONFIG; |
| u8 iter; |
| |
| for (iter = 0; iter < spi_ptr->total_hw_instances; iter++) { |
| spi_sub_ptr = spi_ptr->spi_int[iter]; |
| |
| while (spi_sub_ptr->spi_xfer_in_progress) |
| msleep(20); |
| |
| /* Store existing config before suspend */ |
| store_restore_config(spi_ptr, spi_sub_ptr, iter, 1); |
| spi_master_suspend(spi_sub_ptr->spi_host); |
| writel(reg1, spi_ptr->reg_base + |
| SPI_MST_EVENT_MASK_REG_OFFSET(iter)); |
| } |
| |
| return 0; |
| } |
| |
| static DEFINE_SIMPLE_DEV_PM_OPS(spi_pm_ops, pci1xxxx_spi_suspend, |
| pci1xxxx_spi_resume); |
| |
| static struct pci_driver pci1xxxx_spi_driver = { |
| .name = DRV_NAME, |
| .id_table = pci1xxxx_spi_pci_id_table, |
| .probe = pci1xxxx_spi_probe, |
| .driver = { |
| .pm = pm_sleep_ptr(&spi_pm_ops), |
| }, |
| }; |
| |
| module_pci_driver(pci1xxxx_spi_driver); |
| |
| MODULE_DESCRIPTION("Microchip Technology Inc. pci1xxxx SPI bus driver"); |
| MODULE_AUTHOR("Tharun Kumar P<tharunkumar.pasumarthi@microchip.com>"); |
| MODULE_AUTHOR("Kumaravel Thiagarajan<kumaravel.thiagarajan@microchip.com>"); |
| MODULE_LICENSE("GPL v2"); |