| // SPDX-License-Identifier: GPL-2.0+ |
| /* Copyright (c) 2015-2016 Quantenna Communications. All rights reserved. */ |
| |
| #include <linux/types.h> |
| #include <linux/io.h> |
| |
| #include "shm_ipc.h" |
| |
| #undef pr_fmt |
| #define pr_fmt(fmt) "qtnfmac shm_ipc: %s: " fmt, __func__ |
| |
| static bool qtnf_shm_ipc_has_new_data(struct qtnf_shm_ipc *ipc) |
| { |
| const u32 flags = readl(&ipc->shm_region->headroom.hdr.flags); |
| |
| return (flags & QTNF_SHM_IPC_NEW_DATA); |
| } |
| |
| static void qtnf_shm_handle_new_data(struct qtnf_shm_ipc *ipc) |
| { |
| size_t size; |
| bool rx_buff_ok = true; |
| struct qtnf_shm_ipc_region_header __iomem *shm_reg_hdr; |
| |
| shm_reg_hdr = &ipc->shm_region->headroom.hdr; |
| |
| size = readw(&shm_reg_hdr->data_len); |
| |
| if (unlikely(size == 0 || size > QTN_IPC_MAX_DATA_SZ)) { |
| pr_err("wrong rx packet size: %zu\n", size); |
| rx_buff_ok = false; |
| } |
| |
| if (likely(rx_buff_ok)) { |
| ipc->rx_packet_count++; |
| ipc->rx_callback.fn(ipc->rx_callback.arg, |
| ipc->shm_region->data, size); |
| } |
| |
| writel(QTNF_SHM_IPC_ACK, &shm_reg_hdr->flags); |
| readl(&shm_reg_hdr->flags); /* flush PCIe write */ |
| |
| ipc->interrupt.fn(ipc->interrupt.arg); |
| } |
| |
| static void qtnf_shm_ipc_irq_work(struct work_struct *work) |
| { |
| struct qtnf_shm_ipc *ipc = container_of(work, struct qtnf_shm_ipc, |
| irq_work); |
| |
| while (qtnf_shm_ipc_has_new_data(ipc)) |
| qtnf_shm_handle_new_data(ipc); |
| } |
| |
| static void qtnf_shm_ipc_irq_inbound_handler(struct qtnf_shm_ipc *ipc) |
| { |
| u32 flags; |
| |
| flags = readl(&ipc->shm_region->headroom.hdr.flags); |
| |
| if (flags & QTNF_SHM_IPC_NEW_DATA) |
| queue_work(ipc->workqueue, &ipc->irq_work); |
| } |
| |
| static void qtnf_shm_ipc_irq_outbound_handler(struct qtnf_shm_ipc *ipc) |
| { |
| u32 flags; |
| |
| if (!READ_ONCE(ipc->waiting_for_ack)) |
| return; |
| |
| flags = readl(&ipc->shm_region->headroom.hdr.flags); |
| |
| if (flags & QTNF_SHM_IPC_ACK) { |
| WRITE_ONCE(ipc->waiting_for_ack, 0); |
| complete(&ipc->tx_completion); |
| } |
| } |
| |
| int qtnf_shm_ipc_init(struct qtnf_shm_ipc *ipc, |
| enum qtnf_shm_ipc_direction direction, |
| struct qtnf_shm_ipc_region __iomem *shm_region, |
| struct workqueue_struct *workqueue, |
| const struct qtnf_shm_ipc_int *interrupt, |
| const struct qtnf_shm_ipc_rx_callback *rx_callback) |
| { |
| BUILD_BUG_ON(offsetof(struct qtnf_shm_ipc_region, data) != |
| QTN_IPC_REG_HDR_SZ); |
| BUILD_BUG_ON(sizeof(struct qtnf_shm_ipc_region) > QTN_IPC_REG_SZ); |
| |
| ipc->shm_region = shm_region; |
| ipc->direction = direction; |
| ipc->interrupt = *interrupt; |
| ipc->rx_callback = *rx_callback; |
| ipc->tx_packet_count = 0; |
| ipc->rx_packet_count = 0; |
| ipc->workqueue = workqueue; |
| ipc->waiting_for_ack = 0; |
| ipc->tx_timeout_count = 0; |
| |
| switch (direction) { |
| case QTNF_SHM_IPC_OUTBOUND: |
| ipc->irq_handler = qtnf_shm_ipc_irq_outbound_handler; |
| break; |
| case QTNF_SHM_IPC_INBOUND: |
| ipc->irq_handler = qtnf_shm_ipc_irq_inbound_handler; |
| break; |
| default: |
| return -EINVAL; |
| } |
| |
| INIT_WORK(&ipc->irq_work, qtnf_shm_ipc_irq_work); |
| init_completion(&ipc->tx_completion); |
| |
| return 0; |
| } |
| |
| void qtnf_shm_ipc_free(struct qtnf_shm_ipc *ipc) |
| { |
| complete_all(&ipc->tx_completion); |
| } |
| |
| int qtnf_shm_ipc_send(struct qtnf_shm_ipc *ipc, const u8 *buf, size_t size) |
| { |
| int ret = 0; |
| struct qtnf_shm_ipc_region_header __iomem *shm_reg_hdr; |
| |
| shm_reg_hdr = &ipc->shm_region->headroom.hdr; |
| |
| if (unlikely(size > QTN_IPC_MAX_DATA_SZ)) |
| return -E2BIG; |
| |
| ipc->tx_packet_count++; |
| |
| writew(size, &shm_reg_hdr->data_len); |
| memcpy_toio(ipc->shm_region->data, buf, size); |
| |
| /* sync previous writes before proceeding */ |
| dma_wmb(); |
| |
| WRITE_ONCE(ipc->waiting_for_ack, 1); |
| |
| /* sync previous memory write before announcing new data ready */ |
| wmb(); |
| |
| writel(QTNF_SHM_IPC_NEW_DATA, &shm_reg_hdr->flags); |
| readl(&shm_reg_hdr->flags); /* flush PCIe write */ |
| |
| ipc->interrupt.fn(ipc->interrupt.arg); |
| |
| if (!wait_for_completion_timeout(&ipc->tx_completion, |
| QTN_SHM_IPC_ACK_TIMEOUT)) { |
| ret = -ETIMEDOUT; |
| ipc->tx_timeout_count++; |
| pr_err("TX ACK timeout\n"); |
| } |
| |
| /* now we're not waiting for ACK even in case of timeout */ |
| WRITE_ONCE(ipc->waiting_for_ack, 0); |
| |
| return ret; |
| } |