| // SPDX-License-Identifier: GPL-2.0 |
| // |
| // Copyright (C) 2018 Macronix International Co., Ltd. |
| // |
| // Authors: |
| // Mason Yang <masonccyang@mxic.com.tw> |
| // zhengxunli <zhengxunli@mxic.com.tw> |
| // Boris Brezillon <boris.brezillon@bootlin.com> |
| // |
| |
| #include <linux/clk.h> |
| #include <linux/io.h> |
| #include <linux/iopoll.h> |
| #include <linux/module.h> |
| #include <linux/mtd/nand.h> |
| #include <linux/mtd/nand-ecc-mxic.h> |
| #include <linux/platform_device.h> |
| #include <linux/pm_runtime.h> |
| #include <linux/spi/spi.h> |
| #include <linux/spi/spi-mem.h> |
| |
| #define HC_CFG 0x0 |
| #define HC_CFG_IF_CFG(x) ((x) << 27) |
| #define HC_CFG_DUAL_SLAVE BIT(31) |
| #define HC_CFG_INDIVIDUAL BIT(30) |
| #define HC_CFG_NIO(x) (((x) / 4) << 27) |
| #define HC_CFG_TYPE(s, t) ((t) << (23 + ((s) * 2))) |
| #define HC_CFG_TYPE_SPI_NOR 0 |
| #define HC_CFG_TYPE_SPI_NAND 1 |
| #define HC_CFG_TYPE_SPI_RAM 2 |
| #define HC_CFG_TYPE_RAW_NAND 3 |
| #define HC_CFG_SLV_ACT(x) ((x) << 21) |
| #define HC_CFG_CLK_PH_EN BIT(20) |
| #define HC_CFG_CLK_POL_INV BIT(19) |
| #define HC_CFG_BIG_ENDIAN BIT(18) |
| #define HC_CFG_DATA_PASS BIT(17) |
| #define HC_CFG_IDLE_SIO_LVL(x) ((x) << 16) |
| #define HC_CFG_MAN_START_EN BIT(3) |
| #define HC_CFG_MAN_START BIT(2) |
| #define HC_CFG_MAN_CS_EN BIT(1) |
| #define HC_CFG_MAN_CS_ASSERT BIT(0) |
| |
| #define INT_STS 0x4 |
| #define INT_STS_EN 0x8 |
| #define INT_SIG_EN 0xc |
| #define INT_STS_ALL GENMASK(31, 0) |
| #define INT_RDY_PIN BIT(26) |
| #define INT_RDY_SR BIT(25) |
| #define INT_LNR_SUSP BIT(24) |
| #define INT_ECC_ERR BIT(17) |
| #define INT_CRC_ERR BIT(16) |
| #define INT_LWR_DIS BIT(12) |
| #define INT_LRD_DIS BIT(11) |
| #define INT_SDMA_INT BIT(10) |
| #define INT_DMA_FINISH BIT(9) |
| #define INT_RX_NOT_FULL BIT(3) |
| #define INT_RX_NOT_EMPTY BIT(2) |
| #define INT_TX_NOT_FULL BIT(1) |
| #define INT_TX_EMPTY BIT(0) |
| |
| #define HC_EN 0x10 |
| #define HC_EN_BIT BIT(0) |
| |
| #define TXD(x) (0x14 + ((x) * 4)) |
| #define RXD 0x24 |
| |
| #define SS_CTRL(s) (0x30 + ((s) * 4)) |
| #define LRD_CFG 0x44 |
| #define LWR_CFG 0x80 |
| #define RWW_CFG 0x70 |
| #define OP_READ BIT(23) |
| #define OP_DUMMY_CYC(x) ((x) << 17) |
| #define OP_ADDR_BYTES(x) ((x) << 14) |
| #define OP_CMD_BYTES(x) (((x) - 1) << 13) |
| #define OP_OCTA_CRC_EN BIT(12) |
| #define OP_DQS_EN BIT(11) |
| #define OP_ENHC_EN BIT(10) |
| #define OP_PREAMBLE_EN BIT(9) |
| #define OP_DATA_DDR BIT(8) |
| #define OP_DATA_BUSW(x) ((x) << 6) |
| #define OP_ADDR_DDR BIT(5) |
| #define OP_ADDR_BUSW(x) ((x) << 3) |
| #define OP_CMD_DDR BIT(2) |
| #define OP_CMD_BUSW(x) (x) |
| #define OP_BUSW_1 0 |
| #define OP_BUSW_2 1 |
| #define OP_BUSW_4 2 |
| #define OP_BUSW_8 3 |
| |
| #define OCTA_CRC 0x38 |
| #define OCTA_CRC_IN_EN(s) BIT(3 + ((s) * 16)) |
| #define OCTA_CRC_CHUNK(s, x) ((fls((x) / 32)) << (1 + ((s) * 16))) |
| #define OCTA_CRC_OUT_EN(s) BIT(0 + ((s) * 16)) |
| |
| #define ONFI_DIN_CNT(s) (0x3c + (s)) |
| |
| #define LRD_CTRL 0x48 |
| #define RWW_CTRL 0x74 |
| #define LWR_CTRL 0x84 |
| #define LMODE_EN BIT(31) |
| #define LMODE_SLV_ACT(x) ((x) << 21) |
| #define LMODE_CMD1(x) ((x) << 8) |
| #define LMODE_CMD0(x) (x) |
| |
| #define LRD_ADDR 0x4c |
| #define LWR_ADDR 0x88 |
| #define LRD_RANGE 0x50 |
| #define LWR_RANGE 0x8c |
| |
| #define AXI_SLV_ADDR 0x54 |
| |
| #define DMAC_RD_CFG 0x58 |
| #define DMAC_WR_CFG 0x94 |
| #define DMAC_CFG_PERIPH_EN BIT(31) |
| #define DMAC_CFG_ALLFLUSH_EN BIT(30) |
| #define DMAC_CFG_LASTFLUSH_EN BIT(29) |
| #define DMAC_CFG_QE(x) (((x) + 1) << 16) |
| #define DMAC_CFG_BURST_LEN(x) (((x) + 1) << 12) |
| #define DMAC_CFG_BURST_SZ(x) ((x) << 8) |
| #define DMAC_CFG_DIR_READ BIT(1) |
| #define DMAC_CFG_START BIT(0) |
| |
| #define DMAC_RD_CNT 0x5c |
| #define DMAC_WR_CNT 0x98 |
| |
| #define SDMA_ADDR 0x60 |
| |
| #define DMAM_CFG 0x64 |
| #define DMAM_CFG_START BIT(31) |
| #define DMAM_CFG_CONT BIT(30) |
| #define DMAM_CFG_SDMA_GAP(x) (fls((x) / 8192) << 2) |
| #define DMAM_CFG_DIR_READ BIT(1) |
| #define DMAM_CFG_EN BIT(0) |
| |
| #define DMAM_CNT 0x68 |
| |
| #define LNR_TIMER_TH 0x6c |
| |
| #define RDM_CFG0 0x78 |
| #define RDM_CFG0_POLY(x) (x) |
| |
| #define RDM_CFG1 0x7c |
| #define RDM_CFG1_RDM_EN BIT(31) |
| #define RDM_CFG1_SEED(x) (x) |
| |
| #define LWR_SUSP_CTRL 0x90 |
| #define LWR_SUSP_CTRL_EN BIT(31) |
| |
| #define DMAS_CTRL 0x9c |
| #define DMAS_CTRL_EN BIT(31) |
| #define DMAS_CTRL_DIR_READ BIT(30) |
| |
| #define DATA_STROB 0xa0 |
| #define DATA_STROB_EDO_EN BIT(2) |
| #define DATA_STROB_INV_POL BIT(1) |
| #define DATA_STROB_DELAY_2CYC BIT(0) |
| |
| #define IDLY_CODE(x) (0xa4 + ((x) * 4)) |
| #define IDLY_CODE_VAL(x, v) ((v) << (((x) % 4) * 8)) |
| |
| #define GPIO 0xc4 |
| #define GPIO_PT(x) BIT(3 + ((x) * 16)) |
| #define GPIO_RESET(x) BIT(2 + ((x) * 16)) |
| #define GPIO_HOLDB(x) BIT(1 + ((x) * 16)) |
| #define GPIO_WPB(x) BIT((x) * 16) |
| |
| #define HC_VER 0xd0 |
| |
| #define HW_TEST(x) (0xe0 + ((x) * 4)) |
| |
| struct mxic_spi { |
| struct device *dev; |
| struct clk *ps_clk; |
| struct clk *send_clk; |
| struct clk *send_dly_clk; |
| void __iomem *regs; |
| u32 cur_speed_hz; |
| struct { |
| void __iomem *map; |
| dma_addr_t dma; |
| size_t size; |
| } linear; |
| |
| struct { |
| bool use_pipelined_conf; |
| struct nand_ecc_engine *pipelined_engine; |
| void *ctx; |
| } ecc; |
| }; |
| |
| static int mxic_spi_clk_enable(struct mxic_spi *mxic) |
| { |
| int ret; |
| |
| ret = clk_prepare_enable(mxic->send_clk); |
| if (ret) |
| return ret; |
| |
| ret = clk_prepare_enable(mxic->send_dly_clk); |
| if (ret) |
| goto err_send_dly_clk; |
| |
| return ret; |
| |
| err_send_dly_clk: |
| clk_disable_unprepare(mxic->send_clk); |
| |
| return ret; |
| } |
| |
| static void mxic_spi_clk_disable(struct mxic_spi *mxic) |
| { |
| clk_disable_unprepare(mxic->send_clk); |
| clk_disable_unprepare(mxic->send_dly_clk); |
| } |
| |
| static void mxic_spi_set_input_delay_dqs(struct mxic_spi *mxic, u8 idly_code) |
| { |
| writel(IDLY_CODE_VAL(0, idly_code) | |
| IDLY_CODE_VAL(1, idly_code) | |
| IDLY_CODE_VAL(2, idly_code) | |
| IDLY_CODE_VAL(3, idly_code), |
| mxic->regs + IDLY_CODE(0)); |
| writel(IDLY_CODE_VAL(4, idly_code) | |
| IDLY_CODE_VAL(5, idly_code) | |
| IDLY_CODE_VAL(6, idly_code) | |
| IDLY_CODE_VAL(7, idly_code), |
| mxic->regs + IDLY_CODE(1)); |
| } |
| |
| static int mxic_spi_clk_setup(struct mxic_spi *mxic, unsigned long freq) |
| { |
| int ret; |
| |
| ret = clk_set_rate(mxic->send_clk, freq); |
| if (ret) |
| return ret; |
| |
| ret = clk_set_rate(mxic->send_dly_clk, freq); |
| if (ret) |
| return ret; |
| |
| /* |
| * A constant delay range from 0x0 ~ 0x1F for input delay, |
| * the unit is 78 ps, the max input delay is 2.418 ns. |
| */ |
| mxic_spi_set_input_delay_dqs(mxic, 0xf); |
| |
| /* |
| * Phase degree = 360 * freq * output-delay |
| * where output-delay is a constant value 1 ns in FPGA. |
| * |
| * Get Phase degree = 360 * freq * 1 ns |
| * = 360 * freq * 1 sec / 1000000000 |
| * = 9 * freq / 25000000 |
| */ |
| ret = clk_set_phase(mxic->send_dly_clk, 9 * freq / 25000000); |
| if (ret) |
| return ret; |
| |
| return 0; |
| } |
| |
| static int mxic_spi_set_freq(struct mxic_spi *mxic, unsigned long freq) |
| { |
| int ret; |
| |
| if (mxic->cur_speed_hz == freq) |
| return 0; |
| |
| mxic_spi_clk_disable(mxic); |
| ret = mxic_spi_clk_setup(mxic, freq); |
| if (ret) |
| return ret; |
| |
| ret = mxic_spi_clk_enable(mxic); |
| if (ret) |
| return ret; |
| |
| mxic->cur_speed_hz = freq; |
| |
| return 0; |
| } |
| |
| static void mxic_spi_hw_init(struct mxic_spi *mxic) |
| { |
| writel(0, mxic->regs + DATA_STROB); |
| writel(INT_STS_ALL, mxic->regs + INT_STS_EN); |
| writel(0, mxic->regs + HC_EN); |
| writel(0, mxic->regs + LRD_CFG); |
| writel(0, mxic->regs + LRD_CTRL); |
| writel(HC_CFG_NIO(1) | HC_CFG_TYPE(0, HC_CFG_TYPE_SPI_NOR) | |
| HC_CFG_SLV_ACT(0) | HC_CFG_MAN_CS_EN | HC_CFG_IDLE_SIO_LVL(1), |
| mxic->regs + HC_CFG); |
| } |
| |
| static u32 mxic_spi_prep_hc_cfg(struct spi_device *spi, u32 flags) |
| { |
| int nio = 1; |
| |
| if (spi->mode & (SPI_TX_OCTAL | SPI_RX_OCTAL)) |
| nio = 8; |
| else if (spi->mode & (SPI_TX_QUAD | SPI_RX_QUAD)) |
| nio = 4; |
| else if (spi->mode & (SPI_TX_DUAL | SPI_RX_DUAL)) |
| nio = 2; |
| |
| return flags | HC_CFG_NIO(nio) | |
| HC_CFG_TYPE(spi_get_chipselect(spi, 0), HC_CFG_TYPE_SPI_NOR) | |
| HC_CFG_SLV_ACT(spi_get_chipselect(spi, 0)) | HC_CFG_IDLE_SIO_LVL(1); |
| } |
| |
| static u32 mxic_spi_mem_prep_op_cfg(const struct spi_mem_op *op, |
| unsigned int data_len) |
| { |
| u32 cfg = OP_CMD_BYTES(op->cmd.nbytes) | |
| OP_CMD_BUSW(fls(op->cmd.buswidth) - 1) | |
| (op->cmd.dtr ? OP_CMD_DDR : 0); |
| |
| if (op->addr.nbytes) |
| cfg |= OP_ADDR_BYTES(op->addr.nbytes) | |
| OP_ADDR_BUSW(fls(op->addr.buswidth) - 1) | |
| (op->addr.dtr ? OP_ADDR_DDR : 0); |
| |
| if (op->dummy.nbytes) |
| cfg |= OP_DUMMY_CYC(op->dummy.nbytes); |
| |
| /* Direct mapping data.nbytes field is not populated */ |
| if (data_len) { |
| cfg |= OP_DATA_BUSW(fls(op->data.buswidth) - 1) | |
| (op->data.dtr ? OP_DATA_DDR : 0); |
| if (op->data.dir == SPI_MEM_DATA_IN) { |
| cfg |= OP_READ; |
| if (op->data.dtr) |
| cfg |= OP_DQS_EN; |
| } |
| } |
| |
| return cfg; |
| } |
| |
| static int mxic_spi_data_xfer(struct mxic_spi *mxic, const void *txbuf, |
| void *rxbuf, unsigned int len) |
| { |
| unsigned int pos = 0; |
| |
| while (pos < len) { |
| unsigned int nbytes = len - pos; |
| u32 data = 0xffffffff; |
| u32 sts; |
| int ret; |
| |
| if (nbytes > 4) |
| nbytes = 4; |
| |
| if (txbuf) |
| memcpy(&data, txbuf + pos, nbytes); |
| |
| ret = readl_poll_timeout(mxic->regs + INT_STS, sts, |
| sts & INT_TX_EMPTY, 0, USEC_PER_SEC); |
| if (ret) |
| return ret; |
| |
| writel(data, mxic->regs + TXD(nbytes % 4)); |
| |
| ret = readl_poll_timeout(mxic->regs + INT_STS, sts, |
| sts & INT_TX_EMPTY, 0, USEC_PER_SEC); |
| if (ret) |
| return ret; |
| |
| ret = readl_poll_timeout(mxic->regs + INT_STS, sts, |
| sts & INT_RX_NOT_EMPTY, 0, |
| USEC_PER_SEC); |
| if (ret) |
| return ret; |
| |
| data = readl(mxic->regs + RXD); |
| if (rxbuf) { |
| data >>= (8 * (4 - nbytes)); |
| memcpy(rxbuf + pos, &data, nbytes); |
| } |
| WARN_ON(readl(mxic->regs + INT_STS) & INT_RX_NOT_EMPTY); |
| |
| pos += nbytes; |
| } |
| |
| return 0; |
| } |
| |
| static ssize_t mxic_spi_mem_dirmap_read(struct spi_mem_dirmap_desc *desc, |
| u64 offs, size_t len, void *buf) |
| { |
| struct mxic_spi *mxic = spi_controller_get_devdata(desc->mem->spi->controller); |
| int ret; |
| u32 sts; |
| |
| if (WARN_ON(offs + desc->info.offset + len > U32_MAX)) |
| return -EINVAL; |
| |
| writel(mxic_spi_prep_hc_cfg(desc->mem->spi, 0), mxic->regs + HC_CFG); |
| |
| writel(mxic_spi_mem_prep_op_cfg(&desc->info.op_tmpl, len), |
| mxic->regs + LRD_CFG); |
| writel(desc->info.offset + offs, mxic->regs + LRD_ADDR); |
| len = min_t(size_t, len, mxic->linear.size); |
| writel(len, mxic->regs + LRD_RANGE); |
| writel(LMODE_CMD0(desc->info.op_tmpl.cmd.opcode) | |
| LMODE_SLV_ACT(spi_get_chipselect(desc->mem->spi, 0)) | |
| LMODE_EN, |
| mxic->regs + LRD_CTRL); |
| |
| if (mxic->ecc.use_pipelined_conf && desc->info.op_tmpl.data.ecc) { |
| ret = mxic_ecc_process_data_pipelined(mxic->ecc.pipelined_engine, |
| NAND_PAGE_READ, |
| mxic->linear.dma + offs); |
| if (ret) |
| return ret; |
| } else { |
| memcpy_fromio(buf, mxic->linear.map, len); |
| } |
| |
| writel(INT_LRD_DIS, mxic->regs + INT_STS); |
| writel(0, mxic->regs + LRD_CTRL); |
| |
| ret = readl_poll_timeout(mxic->regs + INT_STS, sts, |
| sts & INT_LRD_DIS, 0, USEC_PER_SEC); |
| if (ret) |
| return ret; |
| |
| return len; |
| } |
| |
| static ssize_t mxic_spi_mem_dirmap_write(struct spi_mem_dirmap_desc *desc, |
| u64 offs, size_t len, |
| const void *buf) |
| { |
| struct mxic_spi *mxic = spi_controller_get_devdata(desc->mem->spi->controller); |
| u32 sts; |
| int ret; |
| |
| if (WARN_ON(offs + desc->info.offset + len > U32_MAX)) |
| return -EINVAL; |
| |
| writel(mxic_spi_prep_hc_cfg(desc->mem->spi, 0), mxic->regs + HC_CFG); |
| |
| writel(mxic_spi_mem_prep_op_cfg(&desc->info.op_tmpl, len), |
| mxic->regs + LWR_CFG); |
| writel(desc->info.offset + offs, mxic->regs + LWR_ADDR); |
| len = min_t(size_t, len, mxic->linear.size); |
| writel(len, mxic->regs + LWR_RANGE); |
| writel(LMODE_CMD0(desc->info.op_tmpl.cmd.opcode) | |
| LMODE_SLV_ACT(spi_get_chipselect(desc->mem->spi, 0)) | |
| LMODE_EN, |
| mxic->regs + LWR_CTRL); |
| |
| if (mxic->ecc.use_pipelined_conf && desc->info.op_tmpl.data.ecc) { |
| ret = mxic_ecc_process_data_pipelined(mxic->ecc.pipelined_engine, |
| NAND_PAGE_WRITE, |
| mxic->linear.dma + offs); |
| if (ret) |
| return ret; |
| } else { |
| memcpy_toio(mxic->linear.map, buf, len); |
| } |
| |
| writel(INT_LWR_DIS, mxic->regs + INT_STS); |
| writel(0, mxic->regs + LWR_CTRL); |
| |
| ret = readl_poll_timeout(mxic->regs + INT_STS, sts, |
| sts & INT_LWR_DIS, 0, USEC_PER_SEC); |
| if (ret) |
| return ret; |
| |
| return len; |
| } |
| |
| static bool mxic_spi_mem_supports_op(struct spi_mem *mem, |
| const struct spi_mem_op *op) |
| { |
| if (op->data.buswidth > 8 || op->addr.buswidth > 8 || |
| op->dummy.buswidth > 8 || op->cmd.buswidth > 8) |
| return false; |
| |
| if (op->data.nbytes && op->dummy.nbytes && |
| op->data.buswidth != op->dummy.buswidth) |
| return false; |
| |
| if (op->addr.nbytes > 7) |
| return false; |
| |
| return spi_mem_default_supports_op(mem, op); |
| } |
| |
| static int mxic_spi_mem_dirmap_create(struct spi_mem_dirmap_desc *desc) |
| { |
| struct mxic_spi *mxic = spi_controller_get_devdata(desc->mem->spi->controller); |
| |
| if (!mxic->linear.map) |
| return -EOPNOTSUPP; |
| |
| if (desc->info.offset + desc->info.length > U32_MAX) |
| return -EINVAL; |
| |
| if (!mxic_spi_mem_supports_op(desc->mem, &desc->info.op_tmpl)) |
| return -EOPNOTSUPP; |
| |
| return 0; |
| } |
| |
| static int mxic_spi_mem_exec_op(struct spi_mem *mem, |
| const struct spi_mem_op *op) |
| { |
| struct mxic_spi *mxic = spi_controller_get_devdata(mem->spi->controller); |
| int i, ret; |
| u8 addr[8], cmd[2]; |
| |
| ret = mxic_spi_set_freq(mxic, mem->spi->max_speed_hz); |
| if (ret) |
| return ret; |
| |
| writel(mxic_spi_prep_hc_cfg(mem->spi, HC_CFG_MAN_CS_EN), |
| mxic->regs + HC_CFG); |
| |
| writel(HC_EN_BIT, mxic->regs + HC_EN); |
| |
| writel(mxic_spi_mem_prep_op_cfg(op, op->data.nbytes), |
| mxic->regs + SS_CTRL(spi_get_chipselect(mem->spi, 0))); |
| |
| writel(readl(mxic->regs + HC_CFG) | HC_CFG_MAN_CS_ASSERT, |
| mxic->regs + HC_CFG); |
| |
| for (i = 0; i < op->cmd.nbytes; i++) |
| cmd[i] = op->cmd.opcode >> (8 * (op->cmd.nbytes - i - 1)); |
| |
| ret = mxic_spi_data_xfer(mxic, cmd, NULL, op->cmd.nbytes); |
| if (ret) |
| goto out; |
| |
| for (i = 0; i < op->addr.nbytes; i++) |
| addr[i] = op->addr.val >> (8 * (op->addr.nbytes - i - 1)); |
| |
| ret = mxic_spi_data_xfer(mxic, addr, NULL, op->addr.nbytes); |
| if (ret) |
| goto out; |
| |
| ret = mxic_spi_data_xfer(mxic, NULL, NULL, op->dummy.nbytes); |
| if (ret) |
| goto out; |
| |
| ret = mxic_spi_data_xfer(mxic, |
| op->data.dir == SPI_MEM_DATA_OUT ? |
| op->data.buf.out : NULL, |
| op->data.dir == SPI_MEM_DATA_IN ? |
| op->data.buf.in : NULL, |
| op->data.nbytes); |
| |
| out: |
| writel(readl(mxic->regs + HC_CFG) & ~HC_CFG_MAN_CS_ASSERT, |
| mxic->regs + HC_CFG); |
| writel(0, mxic->regs + HC_EN); |
| |
| return ret; |
| } |
| |
| static const struct spi_controller_mem_ops mxic_spi_mem_ops = { |
| .supports_op = mxic_spi_mem_supports_op, |
| .exec_op = mxic_spi_mem_exec_op, |
| .dirmap_create = mxic_spi_mem_dirmap_create, |
| .dirmap_read = mxic_spi_mem_dirmap_read, |
| .dirmap_write = mxic_spi_mem_dirmap_write, |
| }; |
| |
| static const struct spi_controller_mem_caps mxic_spi_mem_caps = { |
| .dtr = true, |
| .ecc = true, |
| }; |
| |
| static void mxic_spi_set_cs(struct spi_device *spi, bool lvl) |
| { |
| struct mxic_spi *mxic = spi_controller_get_devdata(spi->controller); |
| |
| if (!lvl) { |
| writel(readl(mxic->regs + HC_CFG) | HC_CFG_MAN_CS_EN, |
| mxic->regs + HC_CFG); |
| writel(HC_EN_BIT, mxic->regs + HC_EN); |
| writel(readl(mxic->regs + HC_CFG) | HC_CFG_MAN_CS_ASSERT, |
| mxic->regs + HC_CFG); |
| } else { |
| writel(readl(mxic->regs + HC_CFG) & ~HC_CFG_MAN_CS_ASSERT, |
| mxic->regs + HC_CFG); |
| writel(0, mxic->regs + HC_EN); |
| } |
| } |
| |
| static int mxic_spi_transfer_one(struct spi_controller *host, |
| struct spi_device *spi, |
| struct spi_transfer *t) |
| { |
| struct mxic_spi *mxic = spi_controller_get_devdata(host); |
| unsigned int busw = OP_BUSW_1; |
| int ret; |
| |
| if (t->rx_buf && t->tx_buf) { |
| if (((spi->mode & SPI_TX_QUAD) && |
| !(spi->mode & SPI_RX_QUAD)) || |
| ((spi->mode & SPI_TX_DUAL) && |
| !(spi->mode & SPI_RX_DUAL))) |
| return -ENOTSUPP; |
| } |
| |
| ret = mxic_spi_set_freq(mxic, t->speed_hz); |
| if (ret) |
| return ret; |
| |
| if (t->tx_buf) { |
| if (spi->mode & SPI_TX_QUAD) |
| busw = OP_BUSW_4; |
| else if (spi->mode & SPI_TX_DUAL) |
| busw = OP_BUSW_2; |
| } else if (t->rx_buf) { |
| if (spi->mode & SPI_RX_QUAD) |
| busw = OP_BUSW_4; |
| else if (spi->mode & SPI_RX_DUAL) |
| busw = OP_BUSW_2; |
| } |
| |
| writel(OP_CMD_BYTES(1) | OP_CMD_BUSW(busw) | |
| OP_DATA_BUSW(busw) | (t->rx_buf ? OP_READ : 0), |
| mxic->regs + SS_CTRL(0)); |
| |
| ret = mxic_spi_data_xfer(mxic, t->tx_buf, t->rx_buf, t->len); |
| if (ret) |
| return ret; |
| |
| spi_finalize_current_transfer(host); |
| |
| return 0; |
| } |
| |
| /* ECC wrapper */ |
| static int mxic_spi_mem_ecc_init_ctx(struct nand_device *nand) |
| { |
| struct nand_ecc_engine_ops *ops = mxic_ecc_get_pipelined_ops(); |
| struct mxic_spi *mxic = nand->ecc.engine->priv; |
| |
| mxic->ecc.use_pipelined_conf = true; |
| |
| return ops->init_ctx(nand); |
| } |
| |
| static void mxic_spi_mem_ecc_cleanup_ctx(struct nand_device *nand) |
| { |
| struct nand_ecc_engine_ops *ops = mxic_ecc_get_pipelined_ops(); |
| struct mxic_spi *mxic = nand->ecc.engine->priv; |
| |
| mxic->ecc.use_pipelined_conf = false; |
| |
| ops->cleanup_ctx(nand); |
| } |
| |
| static int mxic_spi_mem_ecc_prepare_io_req(struct nand_device *nand, |
| struct nand_page_io_req *req) |
| { |
| struct nand_ecc_engine_ops *ops = mxic_ecc_get_pipelined_ops(); |
| |
| return ops->prepare_io_req(nand, req); |
| } |
| |
| static int mxic_spi_mem_ecc_finish_io_req(struct nand_device *nand, |
| struct nand_page_io_req *req) |
| { |
| struct nand_ecc_engine_ops *ops = mxic_ecc_get_pipelined_ops(); |
| |
| return ops->finish_io_req(nand, req); |
| } |
| |
| static struct nand_ecc_engine_ops mxic_spi_mem_ecc_engine_pipelined_ops = { |
| .init_ctx = mxic_spi_mem_ecc_init_ctx, |
| .cleanup_ctx = mxic_spi_mem_ecc_cleanup_ctx, |
| .prepare_io_req = mxic_spi_mem_ecc_prepare_io_req, |
| .finish_io_req = mxic_spi_mem_ecc_finish_io_req, |
| }; |
| |
| static void mxic_spi_mem_ecc_remove(struct mxic_spi *mxic) |
| { |
| if (mxic->ecc.pipelined_engine) { |
| mxic_ecc_put_pipelined_engine(mxic->ecc.pipelined_engine); |
| nand_ecc_unregister_on_host_hw_engine(mxic->ecc.pipelined_engine); |
| } |
| } |
| |
| static int mxic_spi_mem_ecc_probe(struct platform_device *pdev, |
| struct mxic_spi *mxic) |
| { |
| struct nand_ecc_engine *eng; |
| |
| if (!mxic_ecc_get_pipelined_ops()) |
| return -EOPNOTSUPP; |
| |
| eng = mxic_ecc_get_pipelined_engine(pdev); |
| if (IS_ERR(eng)) |
| return PTR_ERR(eng); |
| |
| eng->dev = &pdev->dev; |
| eng->integration = NAND_ECC_ENGINE_INTEGRATION_PIPELINED; |
| eng->ops = &mxic_spi_mem_ecc_engine_pipelined_ops; |
| eng->priv = mxic; |
| mxic->ecc.pipelined_engine = eng; |
| nand_ecc_register_on_host_hw_engine(eng); |
| |
| return 0; |
| } |
| |
| static int __maybe_unused mxic_spi_runtime_suspend(struct device *dev) |
| { |
| struct spi_controller *host = dev_get_drvdata(dev); |
| struct mxic_spi *mxic = spi_controller_get_devdata(host); |
| |
| mxic_spi_clk_disable(mxic); |
| clk_disable_unprepare(mxic->ps_clk); |
| |
| return 0; |
| } |
| |
| static int __maybe_unused mxic_spi_runtime_resume(struct device *dev) |
| { |
| struct spi_controller *host = dev_get_drvdata(dev); |
| struct mxic_spi *mxic = spi_controller_get_devdata(host); |
| int ret; |
| |
| ret = clk_prepare_enable(mxic->ps_clk); |
| if (ret) { |
| dev_err(dev, "Cannot enable ps_clock.\n"); |
| return ret; |
| } |
| |
| return mxic_spi_clk_enable(mxic); |
| } |
| |
| static const struct dev_pm_ops mxic_spi_dev_pm_ops = { |
| SET_RUNTIME_PM_OPS(mxic_spi_runtime_suspend, |
| mxic_spi_runtime_resume, NULL) |
| }; |
| |
| static int mxic_spi_probe(struct platform_device *pdev) |
| { |
| struct spi_controller *host; |
| struct resource *res; |
| struct mxic_spi *mxic; |
| int ret; |
| |
| host = devm_spi_alloc_host(&pdev->dev, sizeof(struct mxic_spi)); |
| if (!host) |
| return -ENOMEM; |
| |
| platform_set_drvdata(pdev, host); |
| |
| mxic = spi_controller_get_devdata(host); |
| mxic->dev = &pdev->dev; |
| |
| host->dev.of_node = pdev->dev.of_node; |
| |
| mxic->ps_clk = devm_clk_get(&pdev->dev, "ps_clk"); |
| if (IS_ERR(mxic->ps_clk)) |
| return PTR_ERR(mxic->ps_clk); |
| |
| mxic->send_clk = devm_clk_get(&pdev->dev, "send_clk"); |
| if (IS_ERR(mxic->send_clk)) |
| return PTR_ERR(mxic->send_clk); |
| |
| mxic->send_dly_clk = devm_clk_get(&pdev->dev, "send_dly_clk"); |
| if (IS_ERR(mxic->send_dly_clk)) |
| return PTR_ERR(mxic->send_dly_clk); |
| |
| mxic->regs = devm_platform_ioremap_resource_byname(pdev, "regs"); |
| if (IS_ERR(mxic->regs)) |
| return PTR_ERR(mxic->regs); |
| |
| res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "dirmap"); |
| mxic->linear.map = devm_ioremap_resource(&pdev->dev, res); |
| if (!IS_ERR(mxic->linear.map)) { |
| mxic->linear.dma = res->start; |
| mxic->linear.size = resource_size(res); |
| } else { |
| mxic->linear.map = NULL; |
| } |
| |
| pm_runtime_enable(&pdev->dev); |
| host->auto_runtime_pm = true; |
| |
| host->num_chipselect = 1; |
| host->mem_ops = &mxic_spi_mem_ops; |
| host->mem_caps = &mxic_spi_mem_caps; |
| |
| host->set_cs = mxic_spi_set_cs; |
| host->transfer_one = mxic_spi_transfer_one; |
| host->bits_per_word_mask = SPI_BPW_MASK(8); |
| host->mode_bits = SPI_CPOL | SPI_CPHA | |
| SPI_RX_DUAL | SPI_TX_DUAL | |
| SPI_RX_QUAD | SPI_TX_QUAD | |
| SPI_RX_OCTAL | SPI_TX_OCTAL; |
| |
| mxic_spi_hw_init(mxic); |
| |
| ret = mxic_spi_mem_ecc_probe(pdev, mxic); |
| if (ret == -EPROBE_DEFER) { |
| pm_runtime_disable(&pdev->dev); |
| return ret; |
| } |
| |
| ret = spi_register_controller(host); |
| if (ret) { |
| dev_err(&pdev->dev, "spi_register_controller failed\n"); |
| pm_runtime_disable(&pdev->dev); |
| mxic_spi_mem_ecc_remove(mxic); |
| } |
| |
| return ret; |
| } |
| |
| static void mxic_spi_remove(struct platform_device *pdev) |
| { |
| struct spi_controller *host = platform_get_drvdata(pdev); |
| struct mxic_spi *mxic = spi_controller_get_devdata(host); |
| |
| pm_runtime_disable(&pdev->dev); |
| mxic_spi_mem_ecc_remove(mxic); |
| spi_unregister_controller(host); |
| } |
| |
| static const struct of_device_id mxic_spi_of_ids[] = { |
| { .compatible = "mxicy,mx25f0a-spi", }, |
| { /* sentinel */ } |
| }; |
| MODULE_DEVICE_TABLE(of, mxic_spi_of_ids); |
| |
| static struct platform_driver mxic_spi_driver = { |
| .probe = mxic_spi_probe, |
| .remove_new = mxic_spi_remove, |
| .driver = { |
| .name = "mxic-spi", |
| .of_match_table = mxic_spi_of_ids, |
| .pm = &mxic_spi_dev_pm_ops, |
| }, |
| }; |
| module_platform_driver(mxic_spi_driver); |
| |
| MODULE_AUTHOR("Mason Yang <masonccyang@mxic.com.tw>"); |
| MODULE_DESCRIPTION("MX25F0A SPI controller driver"); |
| MODULE_LICENSE("GPL v2"); |