| // SPDX-License-Identifier: GPL-2.0+ |
| /* |
| * Copyright (c) 2021 Aspeed Technology Inc. |
| */ |
| |
| #include <linux/clk.h> |
| #include <linux/delay.h> |
| #include <linux/dma-mapping.h> |
| #include <linux/interrupt.h> |
| #include <linux/kernel.h> |
| #include <linux/module.h> |
| #include <linux/of.h> |
| #include <linux/platform_device.h> |
| #include <linux/prefetch.h> |
| #include <linux/usb/ch9.h> |
| #include <linux/usb/gadget.h> |
| #include <linux/slab.h> |
| |
| #define AST_UDC_NUM_ENDPOINTS (1 + 4) |
| #define AST_UDC_EP0_MAX_PACKET 64 /* EP0's max packet size */ |
| #define AST_UDC_EPn_MAX_PACKET 1024 /* Generic EPs max packet size */ |
| #define AST_UDC_DESCS_COUNT 256 /* Use 256 stages descriptor mode (32/256) */ |
| #define AST_UDC_DESC_MODE 1 /* Single/Multiple Stage(s) Descriptor Mode */ |
| |
| #define AST_UDC_EP_DMA_SIZE (AST_UDC_EPn_MAX_PACKET + 8 * AST_UDC_DESCS_COUNT) |
| |
| /***************************** |
| * * |
| * UDC register definitions * |
| * * |
| *****************************/ |
| |
| #define AST_UDC_FUNC_CTRL 0x00 /* Root Function Control & Status Register */ |
| #define AST_UDC_CONFIG 0x04 /* Root Configuration Setting Register */ |
| #define AST_UDC_IER 0x08 /* Interrupt Control Register */ |
| #define AST_UDC_ISR 0x0C /* Interrupt Status Register */ |
| #define AST_UDC_EP_ACK_IER 0x10 /* Programmable ep Pool ACK Interrupt Enable Reg */ |
| #define AST_UDC_EP_NAK_IER 0x14 /* Programmable ep Pool NAK Interrupt Enable Reg */ |
| #define AST_UDC_EP_ACK_ISR 0x18 /* Programmable ep Pool ACK Interrupt Status Reg */ |
| #define AST_UDC_EP_NAK_ISR 0x1C /* Programmable ep Pool NAK Interrupt Status Reg */ |
| #define AST_UDC_DEV_RESET 0x20 /* Device Controller Soft Reset Enable Register */ |
| #define AST_UDC_STS 0x24 /* USB Status Register */ |
| #define AST_VHUB_EP_DATA 0x28 /* Programmable ep Pool Data Toggle Value Set */ |
| #define AST_VHUB_ISO_TX_FAIL 0x2C /* Isochronous Transaction Fail Accumulator */ |
| #define AST_UDC_EP0_CTRL 0x30 /* Endpoint 0 Control/Status Register */ |
| #define AST_UDC_EP0_DATA_BUFF 0x34 /* Base Address of ep0 IN/OUT Data Buffer Reg */ |
| #define AST_UDC_SETUP0 0x80 /* Root Device Setup Data Buffer0 */ |
| #define AST_UDC_SETUP1 0x84 /* Root Device Setup Data Buffer1 */ |
| |
| |
| /* Main control reg */ |
| #define USB_PHY_CLK_EN BIT(31) |
| #define USB_FIFO_DYN_PWRD_EN BIT(19) |
| #define USB_EP_LONG_DESC BIT(18) |
| #define USB_BIST_TEST_PASS BIT(13) |
| #define USB_BIST_TURN_ON BIT(12) |
| #define USB_PHY_RESET_DIS BIT(11) |
| #define USB_TEST_MODE(x) ((x) << 8) |
| #define USB_FORCE_TIMER_HS BIT(7) |
| #define USB_FORCE_HS BIT(6) |
| #define USB_REMOTE_WAKEUP_12MS BIT(5) |
| #define USB_REMOTE_WAKEUP_EN BIT(4) |
| #define USB_AUTO_REMOTE_WAKEUP_EN BIT(3) |
| #define USB_STOP_CLK_IN_SUPEND BIT(2) |
| #define USB_UPSTREAM_FS BIT(1) |
| #define USB_UPSTREAM_EN BIT(0) |
| |
| /* Main config reg */ |
| #define UDC_CFG_SET_ADDR(x) ((x) & UDC_CFG_ADDR_MASK) |
| #define UDC_CFG_ADDR_MASK GENMASK(6, 0) |
| |
| /* Interrupt ctrl & status reg */ |
| #define UDC_IRQ_EP_POOL_NAK BIT(17) |
| #define UDC_IRQ_EP_POOL_ACK_STALL BIT(16) |
| #define UDC_IRQ_BUS_RESUME BIT(8) |
| #define UDC_IRQ_BUS_SUSPEND BIT(7) |
| #define UDC_IRQ_BUS_RESET BIT(6) |
| #define UDC_IRQ_EP0_IN_DATA_NAK BIT(4) |
| #define UDC_IRQ_EP0_IN_ACK_STALL BIT(3) |
| #define UDC_IRQ_EP0_OUT_NAK BIT(2) |
| #define UDC_IRQ_EP0_OUT_ACK_STALL BIT(1) |
| #define UDC_IRQ_EP0_SETUP BIT(0) |
| #define UDC_IRQ_ACK_ALL (0x1ff) |
| |
| /* EP isr reg */ |
| #define USB_EP3_ISR BIT(3) |
| #define USB_EP2_ISR BIT(2) |
| #define USB_EP1_ISR BIT(1) |
| #define USB_EP0_ISR BIT(0) |
| #define UDC_IRQ_EP_ACK_ALL (0xf) |
| |
| /*Soft reset reg */ |
| #define ROOT_UDC_SOFT_RESET BIT(0) |
| |
| /* USB status reg */ |
| #define UDC_STS_HIGHSPEED BIT(27) |
| |
| /* Programmable EP data toggle */ |
| #define EP_TOGGLE_SET_EPNUM(x) ((x) & 0x3) |
| |
| /* EP0 ctrl reg */ |
| #define EP0_GET_RX_LEN(x) ((x >> 16) & 0x7f) |
| #define EP0_TX_LEN(x) ((x & 0x7f) << 8) |
| #define EP0_RX_BUFF_RDY BIT(2) |
| #define EP0_TX_BUFF_RDY BIT(1) |
| #define EP0_STALL BIT(0) |
| |
| /************************************* |
| * * |
| * per-endpoint register definitions * |
| * * |
| *************************************/ |
| |
| #define AST_UDC_EP_CONFIG 0x00 /* Endpoint Configuration Register */ |
| #define AST_UDC_EP_DMA_CTRL 0x04 /* DMA Descriptor List Control/Status Register */ |
| #define AST_UDC_EP_DMA_BUFF 0x08 /* DMA Descriptor/Buffer Base Address */ |
| #define AST_UDC_EP_DMA_STS 0x0C /* DMA Descriptor List R/W Pointer and Status */ |
| |
| #define AST_UDC_EP_BASE 0x200 |
| #define AST_UDC_EP_OFFSET 0x10 |
| |
| /* EP config reg */ |
| #define EP_SET_MAX_PKT(x) ((x & 0x3ff) << 16) |
| #define EP_DATA_FETCH_CTRL(x) ((x & 0x3) << 14) |
| #define EP_AUTO_DATA_DISABLE (0x1 << 13) |
| #define EP_SET_EP_STALL (0x1 << 12) |
| #define EP_SET_EP_NUM(x) ((x & 0xf) << 8) |
| #define EP_SET_TYPE_MASK(x) ((x) << 5) |
| #define EP_TYPE_BULK (0x1) |
| #define EP_TYPE_INT (0x2) |
| #define EP_TYPE_ISO (0x3) |
| #define EP_DIR_OUT (0x1 << 4) |
| #define EP_ALLOCATED_MASK (0x7 << 1) |
| #define EP_ENABLE BIT(0) |
| |
| /* EP DMA ctrl reg */ |
| #define EP_DMA_CTRL_GET_PROC_STS(x) ((x >> 4) & 0xf) |
| #define EP_DMA_CTRL_STS_RX_IDLE 0x0 |
| #define EP_DMA_CTRL_STS_TX_IDLE 0x8 |
| #define EP_DMA_CTRL_IN_LONG_MODE (0x1 << 3) |
| #define EP_DMA_CTRL_RESET (0x1 << 2) |
| #define EP_DMA_SINGLE_STAGE (0x1 << 1) |
| #define EP_DMA_DESC_MODE (0x1 << 0) |
| |
| /* EP DMA status reg */ |
| #define EP_DMA_SET_TX_SIZE(x) ((x & 0x7ff) << 16) |
| #define EP_DMA_GET_TX_SIZE(x) (((x) >> 16) & 0x7ff) |
| #define EP_DMA_GET_RPTR(x) (((x) >> 8) & 0xff) |
| #define EP_DMA_GET_WPTR(x) ((x) & 0xff) |
| #define EP_DMA_SINGLE_KICK (1 << 0) /* WPTR = 1 for single mode */ |
| |
| /* EP desc reg */ |
| #define AST_EP_DMA_DESC_INTR_ENABLE BIT(31) |
| #define AST_EP_DMA_DESC_PID_DATA0 (0 << 14) |
| #define AST_EP_DMA_DESC_PID_DATA2 BIT(14) |
| #define AST_EP_DMA_DESC_PID_DATA1 (2 << 14) |
| #define AST_EP_DMA_DESC_PID_MDATA (3 << 14) |
| #define EP_DESC1_IN_LEN(x) ((x) & 0x1fff) |
| #define AST_EP_DMA_DESC_MAX_LEN (7680) /* Max packet length for trasmit in 1 desc */ |
| |
| struct ast_udc_request { |
| struct usb_request req; |
| struct list_head queue; |
| unsigned mapped:1; |
| unsigned int actual_dma_length; |
| u32 saved_dma_wptr; |
| }; |
| |
| #define to_ast_req(__req) container_of(__req, struct ast_udc_request, req) |
| |
| struct ast_dma_desc { |
| u32 des_0; |
| u32 des_1; |
| }; |
| |
| struct ast_udc_ep { |
| struct usb_ep ep; |
| |
| /* Request queue */ |
| struct list_head queue; |
| |
| struct ast_udc_dev *udc; |
| void __iomem *ep_reg; |
| void *epn_buf; |
| dma_addr_t epn_buf_dma; |
| const struct usb_endpoint_descriptor *desc; |
| |
| /* DMA Descriptors */ |
| struct ast_dma_desc *descs; |
| dma_addr_t descs_dma; |
| u32 descs_wptr; |
| u32 chunk_max; |
| |
| bool dir_in:1; |
| unsigned stopped:1; |
| bool desc_mode:1; |
| }; |
| |
| #define to_ast_ep(__ep) container_of(__ep, struct ast_udc_ep, ep) |
| |
| struct ast_udc_dev { |
| struct platform_device *pdev; |
| void __iomem *reg; |
| int irq; |
| spinlock_t lock; |
| struct clk *clk; |
| struct work_struct wake_work; |
| |
| /* EP0 DMA buffers allocated in one chunk */ |
| void *ep0_buf; |
| dma_addr_t ep0_buf_dma; |
| struct ast_udc_ep ep[AST_UDC_NUM_ENDPOINTS]; |
| |
| struct usb_gadget gadget; |
| struct usb_gadget_driver *driver; |
| void __iomem *creq; |
| enum usb_device_state suspended_from; |
| int desc_mode; |
| |
| /* Force full speed only */ |
| bool force_usb1:1; |
| unsigned is_control_tx:1; |
| bool wakeup_en:1; |
| }; |
| |
| #define to_ast_dev(__g) container_of(__g, struct ast_udc_dev, gadget) |
| |
| static const char * const ast_ep_name[] = { |
| "ep0", "ep1", "ep2", "ep3", "ep4" |
| }; |
| |
| #ifdef AST_UDC_DEBUG_ALL |
| #define AST_UDC_DEBUG |
| #define AST_SETUP_DEBUG |
| #define AST_EP_DEBUG |
| #define AST_ISR_DEBUG |
| #endif |
| |
| #ifdef AST_SETUP_DEBUG |
| #define SETUP_DBG(u, fmt, ...) \ |
| dev_dbg(&(u)->pdev->dev, "%s() " fmt, __func__, ##__VA_ARGS__) |
| #else |
| #define SETUP_DBG(u, fmt, ...) |
| #endif |
| |
| #ifdef AST_EP_DEBUG |
| #define EP_DBG(e, fmt, ...) \ |
| dev_dbg(&(e)->udc->pdev->dev, "%s():%s " fmt, __func__, \ |
| (e)->ep.name, ##__VA_ARGS__) |
| #else |
| #define EP_DBG(ep, fmt, ...) ((void)(ep)) |
| #endif |
| |
| #ifdef AST_UDC_DEBUG |
| #define UDC_DBG(u, fmt, ...) \ |
| dev_dbg(&(u)->pdev->dev, "%s() " fmt, __func__, ##__VA_ARGS__) |
| #else |
| #define UDC_DBG(u, fmt, ...) |
| #endif |
| |
| #ifdef AST_ISR_DEBUG |
| #define ISR_DBG(u, fmt, ...) \ |
| dev_dbg(&(u)->pdev->dev, "%s() " fmt, __func__, ##__VA_ARGS__) |
| #else |
| #define ISR_DBG(u, fmt, ...) |
| #endif |
| |
| /*-------------------------------------------------------------------------*/ |
| #define ast_udc_read(udc, offset) \ |
| readl((udc)->reg + (offset)) |
| #define ast_udc_write(udc, val, offset) \ |
| writel((val), (udc)->reg + (offset)) |
| |
| #define ast_ep_read(ep, reg) \ |
| readl((ep)->ep_reg + (reg)) |
| #define ast_ep_write(ep, val, reg) \ |
| writel((val), (ep)->ep_reg + (reg)) |
| |
| /*-------------------------------------------------------------------------*/ |
| |
| static void ast_udc_done(struct ast_udc_ep *ep, struct ast_udc_request *req, |
| int status) |
| { |
| struct ast_udc_dev *udc = ep->udc; |
| |
| EP_DBG(ep, "req @%p, len (%d/%d), buf:0x%x, dir:0x%x\n", |
| req, req->req.actual, req->req.length, |
| (u32)req->req.buf, ep->dir_in); |
| |
| list_del(&req->queue); |
| |
| if (req->req.status == -EINPROGRESS) |
| req->req.status = status; |
| else |
| status = req->req.status; |
| |
| if (status && status != -ESHUTDOWN) |
| EP_DBG(ep, "done req:%p, status:%d\n", req, status); |
| |
| spin_unlock(&udc->lock); |
| usb_gadget_giveback_request(&ep->ep, &req->req); |
| spin_lock(&udc->lock); |
| } |
| |
| static void ast_udc_nuke(struct ast_udc_ep *ep, int status) |
| { |
| int count = 0; |
| |
| while (!list_empty(&ep->queue)) { |
| struct ast_udc_request *req; |
| |
| req = list_entry(ep->queue.next, struct ast_udc_request, |
| queue); |
| ast_udc_done(ep, req, status); |
| count++; |
| } |
| |
| if (count) |
| EP_DBG(ep, "Nuked %d request(s)\n", count); |
| } |
| |
| /* |
| * Stop activity on all endpoints. |
| * Device controller for which EP activity is to be stopped. |
| * |
| * All the endpoints are stopped and any pending transfer requests if any on |
| * the endpoint are terminated. |
| */ |
| static void ast_udc_stop_activity(struct ast_udc_dev *udc) |
| { |
| struct ast_udc_ep *ep; |
| int i; |
| |
| for (i = 0; i < AST_UDC_NUM_ENDPOINTS; i++) { |
| ep = &udc->ep[i]; |
| ep->stopped = 1; |
| ast_udc_nuke(ep, -ESHUTDOWN); |
| } |
| } |
| |
| static int ast_udc_ep_enable(struct usb_ep *_ep, |
| const struct usb_endpoint_descriptor *desc) |
| { |
| u16 maxpacket = usb_endpoint_maxp(desc); |
| struct ast_udc_ep *ep = to_ast_ep(_ep); |
| struct ast_udc_dev *udc = ep->udc; |
| u8 epnum = usb_endpoint_num(desc); |
| unsigned long flags; |
| u32 ep_conf = 0; |
| u8 dir_in; |
| u8 type; |
| |
| if (!_ep || !ep || !desc || desc->bDescriptorType != USB_DT_ENDPOINT || |
| maxpacket == 0 || maxpacket > ep->ep.maxpacket) { |
| EP_DBG(ep, "Failed, invalid EP enable param\n"); |
| return -EINVAL; |
| } |
| |
| if (!udc->driver) { |
| EP_DBG(ep, "bogus device state\n"); |
| return -ESHUTDOWN; |
| } |
| |
| EP_DBG(ep, "maxpacket:0x%x\n", maxpacket); |
| |
| spin_lock_irqsave(&udc->lock, flags); |
| |
| ep->desc = desc; |
| ep->stopped = 0; |
| ep->ep.maxpacket = maxpacket; |
| ep->chunk_max = AST_EP_DMA_DESC_MAX_LEN; |
| |
| if (maxpacket < AST_UDC_EPn_MAX_PACKET) |
| ep_conf = EP_SET_MAX_PKT(maxpacket); |
| |
| ep_conf |= EP_SET_EP_NUM(epnum); |
| |
| type = usb_endpoint_type(desc); |
| dir_in = usb_endpoint_dir_in(desc); |
| ep->dir_in = dir_in; |
| if (!ep->dir_in) |
| ep_conf |= EP_DIR_OUT; |
| |
| EP_DBG(ep, "type %d, dir_in %d\n", type, dir_in); |
| switch (type) { |
| case USB_ENDPOINT_XFER_ISOC: |
| ep_conf |= EP_SET_TYPE_MASK(EP_TYPE_ISO); |
| break; |
| |
| case USB_ENDPOINT_XFER_BULK: |
| ep_conf |= EP_SET_TYPE_MASK(EP_TYPE_BULK); |
| break; |
| |
| case USB_ENDPOINT_XFER_INT: |
| ep_conf |= EP_SET_TYPE_MASK(EP_TYPE_INT); |
| break; |
| } |
| |
| ep->desc_mode = udc->desc_mode && ep->descs_dma && ep->dir_in; |
| if (ep->desc_mode) { |
| ast_ep_write(ep, EP_DMA_CTRL_RESET, AST_UDC_EP_DMA_CTRL); |
| ast_ep_write(ep, 0, AST_UDC_EP_DMA_STS); |
| ast_ep_write(ep, ep->descs_dma, AST_UDC_EP_DMA_BUFF); |
| |
| /* Enable Long Descriptor Mode */ |
| ast_ep_write(ep, EP_DMA_CTRL_IN_LONG_MODE | EP_DMA_DESC_MODE, |
| AST_UDC_EP_DMA_CTRL); |
| |
| ep->descs_wptr = 0; |
| |
| } else { |
| ast_ep_write(ep, EP_DMA_CTRL_RESET, AST_UDC_EP_DMA_CTRL); |
| ast_ep_write(ep, EP_DMA_SINGLE_STAGE, AST_UDC_EP_DMA_CTRL); |
| ast_ep_write(ep, 0, AST_UDC_EP_DMA_STS); |
| } |
| |
| /* Cleanup data toggle just in case */ |
| ast_udc_write(udc, EP_TOGGLE_SET_EPNUM(epnum), AST_VHUB_EP_DATA); |
| |
| /* Enable EP */ |
| ast_ep_write(ep, ep_conf | EP_ENABLE, AST_UDC_EP_CONFIG); |
| |
| EP_DBG(ep, "ep_config: 0x%x\n", ast_ep_read(ep, AST_UDC_EP_CONFIG)); |
| |
| spin_unlock_irqrestore(&udc->lock, flags); |
| |
| return 0; |
| } |
| |
| static int ast_udc_ep_disable(struct usb_ep *_ep) |
| { |
| struct ast_udc_ep *ep = to_ast_ep(_ep); |
| struct ast_udc_dev *udc = ep->udc; |
| unsigned long flags; |
| |
| spin_lock_irqsave(&udc->lock, flags); |
| |
| ep->ep.desc = NULL; |
| ep->stopped = 1; |
| |
| ast_udc_nuke(ep, -ESHUTDOWN); |
| ast_ep_write(ep, 0, AST_UDC_EP_CONFIG); |
| |
| spin_unlock_irqrestore(&udc->lock, flags); |
| |
| return 0; |
| } |
| |
| static struct usb_request *ast_udc_ep_alloc_request(struct usb_ep *_ep, |
| gfp_t gfp_flags) |
| { |
| struct ast_udc_ep *ep = to_ast_ep(_ep); |
| struct ast_udc_request *req; |
| |
| req = kzalloc(sizeof(struct ast_udc_request), gfp_flags); |
| if (!req) { |
| EP_DBG(ep, "request allocation failed\n"); |
| return NULL; |
| } |
| |
| INIT_LIST_HEAD(&req->queue); |
| |
| return &req->req; |
| } |
| |
| static void ast_udc_ep_free_request(struct usb_ep *_ep, |
| struct usb_request *_req) |
| { |
| struct ast_udc_request *req = to_ast_req(_req); |
| |
| kfree(req); |
| } |
| |
| static int ast_dma_descriptor_setup(struct ast_udc_ep *ep, u32 dma_buf, |
| u16 tx_len, struct ast_udc_request *req) |
| { |
| struct ast_udc_dev *udc = ep->udc; |
| struct device *dev = &udc->pdev->dev; |
| bool last = false; |
| int chunk, count; |
| u32 offset; |
| |
| if (!ep->descs) { |
| dev_warn(dev, "%s: Empty DMA descs list failure\n", |
| ep->ep.name); |
| return -EINVAL; |
| } |
| |
| chunk = tx_len; |
| offset = count = 0; |
| |
| EP_DBG(ep, "req @%p, %s:%d, %s:0x%x, %s:0x%x\n", req, |
| "wptr", ep->descs_wptr, "dma_buf", dma_buf, |
| "tx_len", tx_len); |
| |
| /* Create Descriptor Lists */ |
| while (chunk >= 0 && !last && count < AST_UDC_DESCS_COUNT) { |
| |
| ep->descs[ep->descs_wptr].des_0 = dma_buf + offset; |
| |
| if (chunk > ep->chunk_max) { |
| ep->descs[ep->descs_wptr].des_1 = ep->chunk_max; |
| } else { |
| ep->descs[ep->descs_wptr].des_1 = chunk; |
| last = true; |
| } |
| |
| chunk -= ep->chunk_max; |
| |
| EP_DBG(ep, "descs[%d]: 0x%x 0x%x\n", |
| ep->descs_wptr, |
| ep->descs[ep->descs_wptr].des_0, |
| ep->descs[ep->descs_wptr].des_1); |
| |
| if (count == 0) |
| req->saved_dma_wptr = ep->descs_wptr; |
| |
| ep->descs_wptr++; |
| count++; |
| |
| if (ep->descs_wptr >= AST_UDC_DESCS_COUNT) |
| ep->descs_wptr = 0; |
| |
| offset = ep->chunk_max * count; |
| } |
| |
| return 0; |
| } |
| |
| static void ast_udc_epn_kick(struct ast_udc_ep *ep, struct ast_udc_request *req) |
| { |
| u32 tx_len; |
| u32 last; |
| |
| last = req->req.length - req->req.actual; |
| tx_len = last > ep->ep.maxpacket ? ep->ep.maxpacket : last; |
| |
| EP_DBG(ep, "kick req @%p, len:%d, dir:%d\n", |
| req, tx_len, ep->dir_in); |
| |
| ast_ep_write(ep, req->req.dma + req->req.actual, AST_UDC_EP_DMA_BUFF); |
| |
| /* Start DMA */ |
| ast_ep_write(ep, EP_DMA_SET_TX_SIZE(tx_len), AST_UDC_EP_DMA_STS); |
| ast_ep_write(ep, EP_DMA_SET_TX_SIZE(tx_len) | EP_DMA_SINGLE_KICK, |
| AST_UDC_EP_DMA_STS); |
| } |
| |
| static void ast_udc_epn_kick_desc(struct ast_udc_ep *ep, |
| struct ast_udc_request *req) |
| { |
| u32 descs_max_size; |
| u32 tx_len; |
| u32 last; |
| |
| descs_max_size = AST_EP_DMA_DESC_MAX_LEN * AST_UDC_DESCS_COUNT; |
| |
| last = req->req.length - req->req.actual; |
| tx_len = last > descs_max_size ? descs_max_size : last; |
| |
| EP_DBG(ep, "kick req @%p, %s:%d, %s:0x%x, %s:0x%x (%d/%d), %s:0x%x\n", |
| req, "tx_len", tx_len, "dir_in", ep->dir_in, |
| "dma", req->req.dma + req->req.actual, |
| req->req.actual, req->req.length, |
| "descs_max_size", descs_max_size); |
| |
| if (!ast_dma_descriptor_setup(ep, req->req.dma + req->req.actual, |
| tx_len, req)) |
| req->actual_dma_length += tx_len; |
| |
| /* make sure CPU done everything before triggering DMA */ |
| mb(); |
| |
| ast_ep_write(ep, ep->descs_wptr, AST_UDC_EP_DMA_STS); |
| |
| EP_DBG(ep, "descs_wptr:%d, dstat:0x%x, dctrl:0x%x\n", |
| ep->descs_wptr, |
| ast_ep_read(ep, AST_UDC_EP_DMA_STS), |
| ast_ep_read(ep, AST_UDC_EP_DMA_CTRL)); |
| } |
| |
| static void ast_udc_ep0_queue(struct ast_udc_ep *ep, |
| struct ast_udc_request *req) |
| { |
| struct ast_udc_dev *udc = ep->udc; |
| u32 tx_len; |
| u32 last; |
| |
| last = req->req.length - req->req.actual; |
| tx_len = last > ep->ep.maxpacket ? ep->ep.maxpacket : last; |
| |
| ast_udc_write(udc, req->req.dma + req->req.actual, |
| AST_UDC_EP0_DATA_BUFF); |
| |
| if (ep->dir_in) { |
| /* IN requests, send data */ |
| SETUP_DBG(udc, "IN: %s:0x%x, %s:0x%x, %s:%d (%d/%d), %s:%d\n", |
| "buf", (u32)req->req.buf, |
| "dma", req->req.dma + req->req.actual, |
| "tx_len", tx_len, |
| req->req.actual, req->req.length, |
| "dir_in", ep->dir_in); |
| |
| req->req.actual += tx_len; |
| ast_udc_write(udc, EP0_TX_LEN(tx_len), AST_UDC_EP0_CTRL); |
| ast_udc_write(udc, EP0_TX_LEN(tx_len) | EP0_TX_BUFF_RDY, |
| AST_UDC_EP0_CTRL); |
| |
| } else { |
| /* OUT requests, receive data */ |
| SETUP_DBG(udc, "OUT: %s:%x, %s:%x, %s:(%d/%d), %s:%d\n", |
| "buf", (u32)req->req.buf, |
| "dma", req->req.dma + req->req.actual, |
| "len", req->req.actual, req->req.length, |
| "dir_in", ep->dir_in); |
| |
| if (!req->req.length) { |
| /* 0 len request, send tx as completion */ |
| ast_udc_write(udc, EP0_TX_BUFF_RDY, AST_UDC_EP0_CTRL); |
| ep->dir_in = 0x1; |
| } else |
| ast_udc_write(udc, EP0_RX_BUFF_RDY, AST_UDC_EP0_CTRL); |
| } |
| } |
| |
| static int ast_udc_ep_queue(struct usb_ep *_ep, struct usb_request *_req, |
| gfp_t gfp_flags) |
| { |
| struct ast_udc_request *req = to_ast_req(_req); |
| struct ast_udc_ep *ep = to_ast_ep(_ep); |
| struct ast_udc_dev *udc = ep->udc; |
| struct device *dev = &udc->pdev->dev; |
| unsigned long flags; |
| int rc; |
| |
| if (unlikely(!_req || !_req->complete || !_req->buf || !_ep)) { |
| dev_warn(dev, "Invalid EP request !\n"); |
| return -EINVAL; |
| } |
| |
| if (ep->stopped) { |
| dev_warn(dev, "%s is already stopped !\n", _ep->name); |
| return -ESHUTDOWN; |
| } |
| |
| spin_lock_irqsave(&udc->lock, flags); |
| |
| list_add_tail(&req->queue, &ep->queue); |
| |
| req->req.actual = 0; |
| req->req.status = -EINPROGRESS; |
| req->actual_dma_length = 0; |
| |
| rc = usb_gadget_map_request(&udc->gadget, &req->req, ep->dir_in); |
| if (rc) { |
| EP_DBG(ep, "Request mapping failure %d\n", rc); |
| dev_warn(dev, "Request mapping failure %d\n", rc); |
| goto end; |
| } |
| |
| EP_DBG(ep, "enqueue req @%p\n", req); |
| EP_DBG(ep, "l=%d, dma:0x%x, zero:%d, is_in:%d\n", |
| _req->length, _req->dma, _req->zero, ep->dir_in); |
| |
| /* EP0 request enqueue */ |
| if (ep->ep.desc == NULL) { |
| if ((req->req.dma % 4) != 0) { |
| dev_warn(dev, "EP0 req dma alignment error\n"); |
| rc = -ESHUTDOWN; |
| goto end; |
| } |
| |
| ast_udc_ep0_queue(ep, req); |
| goto end; |
| } |
| |
| /* EPn request enqueue */ |
| if (list_is_singular(&ep->queue)) { |
| if (ep->desc_mode) |
| ast_udc_epn_kick_desc(ep, req); |
| else |
| ast_udc_epn_kick(ep, req); |
| } |
| |
| end: |
| spin_unlock_irqrestore(&udc->lock, flags); |
| |
| return rc; |
| } |
| |
| static int ast_udc_ep_dequeue(struct usb_ep *_ep, struct usb_request *_req) |
| { |
| struct ast_udc_ep *ep = to_ast_ep(_ep); |
| struct ast_udc_dev *udc = ep->udc; |
| struct ast_udc_request *req; |
| unsigned long flags; |
| int rc = 0; |
| |
| spin_lock_irqsave(&udc->lock, flags); |
| |
| /* make sure it's actually queued on this endpoint */ |
| list_for_each_entry(req, &ep->queue, queue) { |
| if (&req->req == _req) { |
| list_del_init(&req->queue); |
| ast_udc_done(ep, req, -ESHUTDOWN); |
| _req->status = -ECONNRESET; |
| break; |
| } |
| } |
| |
| /* dequeue request not found */ |
| if (&req->req != _req) |
| rc = -EINVAL; |
| |
| spin_unlock_irqrestore(&udc->lock, flags); |
| |
| return rc; |
| } |
| |
| static int ast_udc_ep_set_halt(struct usb_ep *_ep, int value) |
| { |
| struct ast_udc_ep *ep = to_ast_ep(_ep); |
| struct ast_udc_dev *udc = ep->udc; |
| unsigned long flags; |
| int epnum; |
| u32 ctrl; |
| |
| EP_DBG(ep, "val:%d\n", value); |
| |
| spin_lock_irqsave(&udc->lock, flags); |
| |
| epnum = usb_endpoint_num(ep->desc); |
| |
| /* EP0 */ |
| if (epnum == 0) { |
| ctrl = ast_udc_read(udc, AST_UDC_EP0_CTRL); |
| if (value) |
| ctrl |= EP0_STALL; |
| else |
| ctrl &= ~EP0_STALL; |
| |
| ast_udc_write(udc, ctrl, AST_UDC_EP0_CTRL); |
| |
| } else { |
| /* EPn */ |
| ctrl = ast_udc_read(udc, AST_UDC_EP_CONFIG); |
| if (value) |
| ctrl |= EP_SET_EP_STALL; |
| else |
| ctrl &= ~EP_SET_EP_STALL; |
| |
| ast_ep_write(ep, ctrl, AST_UDC_EP_CONFIG); |
| |
| /* only epn is stopped and waits for clear */ |
| ep->stopped = value ? 1 : 0; |
| } |
| |
| spin_unlock_irqrestore(&udc->lock, flags); |
| |
| return 0; |
| } |
| |
| static const struct usb_ep_ops ast_udc_ep_ops = { |
| .enable = ast_udc_ep_enable, |
| .disable = ast_udc_ep_disable, |
| .alloc_request = ast_udc_ep_alloc_request, |
| .free_request = ast_udc_ep_free_request, |
| .queue = ast_udc_ep_queue, |
| .dequeue = ast_udc_ep_dequeue, |
| .set_halt = ast_udc_ep_set_halt, |
| /* there's only imprecise fifo status reporting */ |
| }; |
| |
| static void ast_udc_ep0_rx(struct ast_udc_dev *udc) |
| { |
| ast_udc_write(udc, udc->ep0_buf_dma, AST_UDC_EP0_DATA_BUFF); |
| ast_udc_write(udc, EP0_RX_BUFF_RDY, AST_UDC_EP0_CTRL); |
| } |
| |
| static void ast_udc_ep0_tx(struct ast_udc_dev *udc) |
| { |
| ast_udc_write(udc, udc->ep0_buf_dma, AST_UDC_EP0_DATA_BUFF); |
| ast_udc_write(udc, EP0_TX_BUFF_RDY, AST_UDC_EP0_CTRL); |
| } |
| |
| static void ast_udc_ep0_out(struct ast_udc_dev *udc) |
| { |
| struct device *dev = &udc->pdev->dev; |
| struct ast_udc_ep *ep = &udc->ep[0]; |
| struct ast_udc_request *req; |
| u16 rx_len; |
| |
| if (list_empty(&ep->queue)) |
| return; |
| |
| req = list_entry(ep->queue.next, struct ast_udc_request, queue); |
| |
| rx_len = EP0_GET_RX_LEN(ast_udc_read(udc, AST_UDC_EP0_CTRL)); |
| req->req.actual += rx_len; |
| |
| SETUP_DBG(udc, "req %p (%d/%d)\n", req, |
| req->req.actual, req->req.length); |
| |
| if ((rx_len < ep->ep.maxpacket) || |
| (req->req.actual == req->req.length)) { |
| ast_udc_ep0_tx(udc); |
| if (!ep->dir_in) |
| ast_udc_done(ep, req, 0); |
| |
| } else { |
| if (rx_len > req->req.length) { |
| // Issue Fix |
| dev_warn(dev, "Something wrong (%d/%d)\n", |
| req->req.actual, req->req.length); |
| ast_udc_ep0_tx(udc); |
| ast_udc_done(ep, req, 0); |
| return; |
| } |
| |
| ep->dir_in = 0; |
| |
| /* More works */ |
| ast_udc_ep0_queue(ep, req); |
| } |
| } |
| |
| static void ast_udc_ep0_in(struct ast_udc_dev *udc) |
| { |
| struct ast_udc_ep *ep = &udc->ep[0]; |
| struct ast_udc_request *req; |
| |
| if (list_empty(&ep->queue)) { |
| if (udc->is_control_tx) { |
| ast_udc_ep0_rx(udc); |
| udc->is_control_tx = 0; |
| } |
| |
| return; |
| } |
| |
| req = list_entry(ep->queue.next, struct ast_udc_request, queue); |
| |
| SETUP_DBG(udc, "req %p (%d/%d)\n", req, |
| req->req.actual, req->req.length); |
| |
| if (req->req.length == req->req.actual) { |
| if (req->req.length) |
| ast_udc_ep0_rx(udc); |
| |
| if (ep->dir_in) |
| ast_udc_done(ep, req, 0); |
| |
| } else { |
| /* More works */ |
| ast_udc_ep0_queue(ep, req); |
| } |
| } |
| |
| static void ast_udc_epn_handle(struct ast_udc_dev *udc, u16 ep_num) |
| { |
| struct ast_udc_ep *ep = &udc->ep[ep_num]; |
| struct ast_udc_request *req; |
| u16 len = 0; |
| |
| if (list_empty(&ep->queue)) |
| return; |
| |
| req = list_first_entry(&ep->queue, struct ast_udc_request, queue); |
| |
| len = EP_DMA_GET_TX_SIZE(ast_ep_read(ep, AST_UDC_EP_DMA_STS)); |
| req->req.actual += len; |
| |
| EP_DBG(ep, "req @%p, length:(%d/%d), %s:0x%x\n", req, |
| req->req.actual, req->req.length, "len", len); |
| |
| /* Done this request */ |
| if (req->req.length == req->req.actual) { |
| ast_udc_done(ep, req, 0); |
| req = list_first_entry_or_null(&ep->queue, |
| struct ast_udc_request, |
| queue); |
| |
| } else { |
| /* Check for short packet */ |
| if (len < ep->ep.maxpacket) { |
| ast_udc_done(ep, req, 0); |
| req = list_first_entry_or_null(&ep->queue, |
| struct ast_udc_request, |
| queue); |
| } |
| } |
| |
| /* More requests */ |
| if (req) |
| ast_udc_epn_kick(ep, req); |
| } |
| |
| static void ast_udc_epn_handle_desc(struct ast_udc_dev *udc, u16 ep_num) |
| { |
| struct ast_udc_ep *ep = &udc->ep[ep_num]; |
| struct device *dev = &udc->pdev->dev; |
| struct ast_udc_request *req; |
| u32 proc_sts, wr_ptr, rd_ptr; |
| u32 len_in_desc, ctrl; |
| u16 total_len = 0; |
| int i; |
| |
| if (list_empty(&ep->queue)) { |
| dev_warn(dev, "%s request queue empty!\n", ep->ep.name); |
| return; |
| } |
| |
| req = list_first_entry(&ep->queue, struct ast_udc_request, queue); |
| |
| ctrl = ast_ep_read(ep, AST_UDC_EP_DMA_CTRL); |
| proc_sts = EP_DMA_CTRL_GET_PROC_STS(ctrl); |
| |
| /* Check processing status is idle */ |
| if (proc_sts != EP_DMA_CTRL_STS_RX_IDLE && |
| proc_sts != EP_DMA_CTRL_STS_TX_IDLE) { |
| dev_warn(dev, "EP DMA CTRL: 0x%x, PS:0x%x\n", |
| ast_ep_read(ep, AST_UDC_EP_DMA_CTRL), |
| proc_sts); |
| return; |
| } |
| |
| ctrl = ast_ep_read(ep, AST_UDC_EP_DMA_STS); |
| rd_ptr = EP_DMA_GET_RPTR(ctrl); |
| wr_ptr = EP_DMA_GET_WPTR(ctrl); |
| |
| if (rd_ptr != wr_ptr) { |
| dev_warn(dev, "desc list is not empty ! %s:%d, %s:%d\n", |
| "rptr", rd_ptr, "wptr", wr_ptr); |
| return; |
| } |
| |
| EP_DBG(ep, "rd_ptr:%d, wr_ptr:%d\n", rd_ptr, wr_ptr); |
| i = req->saved_dma_wptr; |
| |
| do { |
| len_in_desc = EP_DESC1_IN_LEN(ep->descs[i].des_1); |
| EP_DBG(ep, "desc[%d] len: %d\n", i, len_in_desc); |
| total_len += len_in_desc; |
| i++; |
| if (i >= AST_UDC_DESCS_COUNT) |
| i = 0; |
| |
| } while (i != wr_ptr); |
| |
| req->req.actual += total_len; |
| |
| EP_DBG(ep, "req @%p, length:(%d/%d), %s:0x%x\n", req, |
| req->req.actual, req->req.length, "len", total_len); |
| |
| /* Done this request */ |
| if (req->req.length == req->req.actual) { |
| ast_udc_done(ep, req, 0); |
| req = list_first_entry_or_null(&ep->queue, |
| struct ast_udc_request, |
| queue); |
| |
| } else { |
| /* Check for short packet */ |
| if (total_len < ep->ep.maxpacket) { |
| ast_udc_done(ep, req, 0); |
| req = list_first_entry_or_null(&ep->queue, |
| struct ast_udc_request, |
| queue); |
| } |
| } |
| |
| /* More requests & dma descs not setup yet */ |
| if (req && (req->actual_dma_length == req->req.actual)) { |
| EP_DBG(ep, "More requests\n"); |
| ast_udc_epn_kick_desc(ep, req); |
| } |
| } |
| |
| static void ast_udc_ep0_data_tx(struct ast_udc_dev *udc, u8 *tx_data, u32 len) |
| { |
| if (len) { |
| memcpy(udc->ep0_buf, tx_data, len); |
| |
| ast_udc_write(udc, udc->ep0_buf_dma, AST_UDC_EP0_DATA_BUFF); |
| ast_udc_write(udc, EP0_TX_LEN(len), AST_UDC_EP0_CTRL); |
| ast_udc_write(udc, EP0_TX_LEN(len) | EP0_TX_BUFF_RDY, |
| AST_UDC_EP0_CTRL); |
| udc->is_control_tx = 1; |
| |
| } else |
| ast_udc_write(udc, EP0_TX_BUFF_RDY, AST_UDC_EP0_CTRL); |
| } |
| |
| static void ast_udc_getstatus(struct ast_udc_dev *udc) |
| { |
| struct usb_ctrlrequest crq; |
| struct ast_udc_ep *ep; |
| u16 status = 0; |
| u16 epnum = 0; |
| |
| memcpy_fromio(&crq, udc->creq, sizeof(crq)); |
| |
| switch (crq.bRequestType & USB_RECIP_MASK) { |
| case USB_RECIP_DEVICE: |
| /* Get device status */ |
| status = 1 << USB_DEVICE_SELF_POWERED; |
| break; |
| case USB_RECIP_INTERFACE: |
| break; |
| case USB_RECIP_ENDPOINT: |
| epnum = crq.wIndex & USB_ENDPOINT_NUMBER_MASK; |
| if (epnum >= AST_UDC_NUM_ENDPOINTS) |
| goto stall; |
| status = udc->ep[epnum].stopped; |
| break; |
| default: |
| goto stall; |
| } |
| |
| ep = &udc->ep[epnum]; |
| EP_DBG(ep, "status: 0x%x\n", status); |
| ast_udc_ep0_data_tx(udc, (u8 *)&status, sizeof(status)); |
| |
| return; |
| |
| stall: |
| EP_DBG(ep, "Can't respond request\n"); |
| ast_udc_write(udc, ast_udc_read(udc, AST_UDC_EP0_CTRL) | EP0_STALL, |
| AST_UDC_EP0_CTRL); |
| } |
| |
| static void ast_udc_ep0_handle_setup(struct ast_udc_dev *udc) |
| { |
| struct ast_udc_ep *ep = &udc->ep[0]; |
| struct ast_udc_request *req; |
| struct usb_ctrlrequest crq; |
| int req_num = 0; |
| int rc = 0; |
| u32 reg; |
| |
| memcpy_fromio(&crq, udc->creq, sizeof(crq)); |
| |
| SETUP_DBG(udc, "SETUP packet: %02x/%02x/%04x/%04x/%04x\n", |
| crq.bRequestType, crq.bRequest, le16_to_cpu(crq.wValue), |
| le16_to_cpu(crq.wIndex), le16_to_cpu(crq.wLength)); |
| |
| /* |
| * Cleanup ep0 request(s) in queue because |
| * there is a new control setup comes. |
| */ |
| list_for_each_entry(req, &udc->ep[0].queue, queue) { |
| req_num++; |
| EP_DBG(ep, "there is req %p in ep0 queue !\n", req); |
| } |
| |
| if (req_num) |
| ast_udc_nuke(&udc->ep[0], -ETIMEDOUT); |
| |
| udc->ep[0].dir_in = crq.bRequestType & USB_DIR_IN; |
| |
| if ((crq.bRequestType & USB_TYPE_MASK) == USB_TYPE_STANDARD) { |
| switch (crq.bRequest) { |
| case USB_REQ_SET_ADDRESS: |
| if (ast_udc_read(udc, AST_UDC_STS) & UDC_STS_HIGHSPEED) |
| udc->gadget.speed = USB_SPEED_HIGH; |
| else |
| udc->gadget.speed = USB_SPEED_FULL; |
| |
| SETUP_DBG(udc, "set addr: 0x%x\n", crq.wValue); |
| reg = ast_udc_read(udc, AST_UDC_CONFIG); |
| reg &= ~UDC_CFG_ADDR_MASK; |
| reg |= UDC_CFG_SET_ADDR(crq.wValue); |
| ast_udc_write(udc, reg, AST_UDC_CONFIG); |
| goto req_complete; |
| |
| case USB_REQ_CLEAR_FEATURE: |
| SETUP_DBG(udc, "ep0: CLEAR FEATURE\n"); |
| goto req_driver; |
| |
| case USB_REQ_SET_FEATURE: |
| SETUP_DBG(udc, "ep0: SET FEATURE\n"); |
| goto req_driver; |
| |
| case USB_REQ_GET_STATUS: |
| ast_udc_getstatus(udc); |
| return; |
| |
| default: |
| goto req_driver; |
| } |
| |
| } |
| |
| req_driver: |
| if (udc->driver) { |
| SETUP_DBG(udc, "Forwarding %s to gadget...\n", |
| udc->gadget.name); |
| |
| spin_unlock(&udc->lock); |
| rc = udc->driver->setup(&udc->gadget, &crq); |
| spin_lock(&udc->lock); |
| |
| } else { |
| SETUP_DBG(udc, "No gadget for request !\n"); |
| } |
| |
| if (rc >= 0) |
| return; |
| |
| /* Stall if gadget failed */ |
| SETUP_DBG(udc, "Stalling, rc:0x%x\n", rc); |
| ast_udc_write(udc, ast_udc_read(udc, AST_UDC_EP0_CTRL) | EP0_STALL, |
| AST_UDC_EP0_CTRL); |
| return; |
| |
| req_complete: |
| SETUP_DBG(udc, "ep0: Sending IN status without data\n"); |
| ast_udc_write(udc, EP0_TX_BUFF_RDY, AST_UDC_EP0_CTRL); |
| } |
| |
| static irqreturn_t ast_udc_isr(int irq, void *data) |
| { |
| struct ast_udc_dev *udc = (struct ast_udc_dev *)data; |
| struct ast_udc_ep *ep; |
| u32 isr, ep_isr; |
| int i; |
| |
| spin_lock(&udc->lock); |
| |
| isr = ast_udc_read(udc, AST_UDC_ISR); |
| if (!isr) |
| goto done; |
| |
| /* Ack interrupts */ |
| ast_udc_write(udc, isr, AST_UDC_ISR); |
| |
| if (isr & UDC_IRQ_BUS_RESET) { |
| ISR_DBG(udc, "UDC_IRQ_BUS_RESET\n"); |
| udc->gadget.speed = USB_SPEED_UNKNOWN; |
| |
| ep = &udc->ep[1]; |
| EP_DBG(ep, "dctrl:0x%x\n", |
| ast_ep_read(ep, AST_UDC_EP_DMA_CTRL)); |
| |
| if (udc->driver && udc->driver->reset) { |
| spin_unlock(&udc->lock); |
| udc->driver->reset(&udc->gadget); |
| spin_lock(&udc->lock); |
| } |
| } |
| |
| if (isr & UDC_IRQ_BUS_SUSPEND) { |
| ISR_DBG(udc, "UDC_IRQ_BUS_SUSPEND\n"); |
| udc->suspended_from = udc->gadget.state; |
| usb_gadget_set_state(&udc->gadget, USB_STATE_SUSPENDED); |
| |
| if (udc->driver && udc->driver->suspend) { |
| spin_unlock(&udc->lock); |
| udc->driver->suspend(&udc->gadget); |
| spin_lock(&udc->lock); |
| } |
| } |
| |
| if (isr & UDC_IRQ_BUS_RESUME) { |
| ISR_DBG(udc, "UDC_IRQ_BUS_RESUME\n"); |
| usb_gadget_set_state(&udc->gadget, udc->suspended_from); |
| |
| if (udc->driver && udc->driver->resume) { |
| spin_unlock(&udc->lock); |
| udc->driver->resume(&udc->gadget); |
| spin_lock(&udc->lock); |
| } |
| } |
| |
| if (isr & UDC_IRQ_EP0_IN_ACK_STALL) { |
| ISR_DBG(udc, "UDC_IRQ_EP0_IN_ACK_STALL\n"); |
| ast_udc_ep0_in(udc); |
| } |
| |
| if (isr & UDC_IRQ_EP0_OUT_ACK_STALL) { |
| ISR_DBG(udc, "UDC_IRQ_EP0_OUT_ACK_STALL\n"); |
| ast_udc_ep0_out(udc); |
| } |
| |
| if (isr & UDC_IRQ_EP0_SETUP) { |
| ISR_DBG(udc, "UDC_IRQ_EP0_SETUP\n"); |
| ast_udc_ep0_handle_setup(udc); |
| } |
| |
| if (isr & UDC_IRQ_EP_POOL_ACK_STALL) { |
| ISR_DBG(udc, "UDC_IRQ_EP_POOL_ACK_STALL\n"); |
| ep_isr = ast_udc_read(udc, AST_UDC_EP_ACK_ISR); |
| |
| /* Ack EP interrupts */ |
| ast_udc_write(udc, ep_isr, AST_UDC_EP_ACK_ISR); |
| |
| /* Handle each EP */ |
| for (i = 0; i < AST_UDC_NUM_ENDPOINTS - 1; i++) { |
| if (ep_isr & (0x1 << i)) { |
| ep = &udc->ep[i + 1]; |
| if (ep->desc_mode) |
| ast_udc_epn_handle_desc(udc, i + 1); |
| else |
| ast_udc_epn_handle(udc, i + 1); |
| } |
| } |
| } |
| |
| done: |
| spin_unlock(&udc->lock); |
| return IRQ_HANDLED; |
| } |
| |
| static int ast_udc_gadget_getframe(struct usb_gadget *gadget) |
| { |
| struct ast_udc_dev *udc = to_ast_dev(gadget); |
| |
| return (ast_udc_read(udc, AST_UDC_STS) >> 16) & 0x7ff; |
| } |
| |
| static void ast_udc_wake_work(struct work_struct *work) |
| { |
| struct ast_udc_dev *udc = container_of(work, struct ast_udc_dev, |
| wake_work); |
| unsigned long flags; |
| u32 ctrl; |
| |
| spin_lock_irqsave(&udc->lock, flags); |
| |
| UDC_DBG(udc, "Wakeup Host !\n"); |
| ctrl = ast_udc_read(udc, AST_UDC_FUNC_CTRL); |
| ast_udc_write(udc, ctrl | USB_REMOTE_WAKEUP_EN, AST_UDC_FUNC_CTRL); |
| |
| spin_unlock_irqrestore(&udc->lock, flags); |
| } |
| |
| static void ast_udc_wakeup_all(struct ast_udc_dev *udc) |
| { |
| /* |
| * A device is trying to wake the world, because this |
| * can recurse into the device, we break the call chain |
| * using a work queue |
| */ |
| schedule_work(&udc->wake_work); |
| } |
| |
| static int ast_udc_wakeup(struct usb_gadget *gadget) |
| { |
| struct ast_udc_dev *udc = to_ast_dev(gadget); |
| unsigned long flags; |
| int rc = 0; |
| |
| spin_lock_irqsave(&udc->lock, flags); |
| |
| if (!udc->wakeup_en) { |
| UDC_DBG(udc, "Remote Wakeup is disabled\n"); |
| rc = -EINVAL; |
| goto err; |
| } |
| |
| UDC_DBG(udc, "Device initiated wakeup\n"); |
| ast_udc_wakeup_all(udc); |
| |
| err: |
| spin_unlock_irqrestore(&udc->lock, flags); |
| return rc; |
| } |
| |
| /* |
| * Activate/Deactivate link with host |
| */ |
| static int ast_udc_pullup(struct usb_gadget *gadget, int is_on) |
| { |
| struct ast_udc_dev *udc = to_ast_dev(gadget); |
| unsigned long flags; |
| u32 ctrl; |
| |
| spin_lock_irqsave(&udc->lock, flags); |
| |
| UDC_DBG(udc, "is_on: %d\n", is_on); |
| if (is_on) |
| ctrl = ast_udc_read(udc, AST_UDC_FUNC_CTRL) | USB_UPSTREAM_EN; |
| else |
| ctrl = ast_udc_read(udc, AST_UDC_FUNC_CTRL) & ~USB_UPSTREAM_EN; |
| |
| ast_udc_write(udc, ctrl, AST_UDC_FUNC_CTRL); |
| |
| spin_unlock_irqrestore(&udc->lock, flags); |
| |
| return 0; |
| } |
| |
| static int ast_udc_start(struct usb_gadget *gadget, |
| struct usb_gadget_driver *driver) |
| { |
| struct ast_udc_dev *udc = to_ast_dev(gadget); |
| struct ast_udc_ep *ep; |
| unsigned long flags; |
| int i; |
| |
| spin_lock_irqsave(&udc->lock, flags); |
| |
| UDC_DBG(udc, "\n"); |
| udc->driver = driver; |
| udc->gadget.dev.of_node = udc->pdev->dev.of_node; |
| |
| for (i = 0; i < AST_UDC_NUM_ENDPOINTS; i++) { |
| ep = &udc->ep[i]; |
| ep->stopped = 0; |
| } |
| |
| spin_unlock_irqrestore(&udc->lock, flags); |
| |
| return 0; |
| } |
| |
| static int ast_udc_stop(struct usb_gadget *gadget) |
| { |
| struct ast_udc_dev *udc = to_ast_dev(gadget); |
| unsigned long flags; |
| u32 ctrl; |
| |
| spin_lock_irqsave(&udc->lock, flags); |
| |
| UDC_DBG(udc, "\n"); |
| ctrl = ast_udc_read(udc, AST_UDC_FUNC_CTRL) & ~USB_UPSTREAM_EN; |
| ast_udc_write(udc, ctrl, AST_UDC_FUNC_CTRL); |
| |
| udc->gadget.speed = USB_SPEED_UNKNOWN; |
| udc->driver = NULL; |
| |
| ast_udc_stop_activity(udc); |
| usb_gadget_set_state(&udc->gadget, USB_STATE_NOTATTACHED); |
| |
| spin_unlock_irqrestore(&udc->lock, flags); |
| |
| return 0; |
| } |
| |
| static const struct usb_gadget_ops ast_udc_ops = { |
| .get_frame = ast_udc_gadget_getframe, |
| .wakeup = ast_udc_wakeup, |
| .pullup = ast_udc_pullup, |
| .udc_start = ast_udc_start, |
| .udc_stop = ast_udc_stop, |
| }; |
| |
| /* |
| * Support 1 Control Endpoint. |
| * Support multiple programmable endpoints that can be configured to |
| * Bulk IN/OUT, Interrupt IN/OUT, and Isochronous IN/OUT type endpoint. |
| */ |
| static void ast_udc_init_ep(struct ast_udc_dev *udc) |
| { |
| struct ast_udc_ep *ep; |
| int i; |
| |
| for (i = 0; i < AST_UDC_NUM_ENDPOINTS; i++) { |
| ep = &udc->ep[i]; |
| ep->ep.name = ast_ep_name[i]; |
| if (i == 0) { |
| ep->ep.caps.type_control = true; |
| } else { |
| ep->ep.caps.type_iso = true; |
| ep->ep.caps.type_bulk = true; |
| ep->ep.caps.type_int = true; |
| } |
| ep->ep.caps.dir_in = true; |
| ep->ep.caps.dir_out = true; |
| |
| ep->ep.ops = &ast_udc_ep_ops; |
| ep->udc = udc; |
| |
| INIT_LIST_HEAD(&ep->queue); |
| |
| if (i == 0) { |
| usb_ep_set_maxpacket_limit(&ep->ep, |
| AST_UDC_EP0_MAX_PACKET); |
| continue; |
| } |
| |
| ep->ep_reg = udc->reg + AST_UDC_EP_BASE + |
| (AST_UDC_EP_OFFSET * (i - 1)); |
| |
| ep->epn_buf = udc->ep0_buf + (i * AST_UDC_EP_DMA_SIZE); |
| ep->epn_buf_dma = udc->ep0_buf_dma + (i * AST_UDC_EP_DMA_SIZE); |
| usb_ep_set_maxpacket_limit(&ep->ep, AST_UDC_EPn_MAX_PACKET); |
| |
| ep->descs = ep->epn_buf + AST_UDC_EPn_MAX_PACKET; |
| ep->descs_dma = ep->epn_buf_dma + AST_UDC_EPn_MAX_PACKET; |
| ep->descs_wptr = 0; |
| |
| list_add_tail(&ep->ep.ep_list, &udc->gadget.ep_list); |
| } |
| } |
| |
| static void ast_udc_init_dev(struct ast_udc_dev *udc) |
| { |
| INIT_WORK(&udc->wake_work, ast_udc_wake_work); |
| } |
| |
| static void ast_udc_init_hw(struct ast_udc_dev *udc) |
| { |
| u32 ctrl; |
| |
| /* Enable PHY */ |
| ctrl = USB_PHY_CLK_EN | USB_PHY_RESET_DIS; |
| ast_udc_write(udc, ctrl, AST_UDC_FUNC_CTRL); |
| |
| udelay(1); |
| ast_udc_write(udc, 0, AST_UDC_DEV_RESET); |
| |
| /* Set descriptor ring size */ |
| if (AST_UDC_DESCS_COUNT == 256) { |
| ctrl |= USB_EP_LONG_DESC; |
| ast_udc_write(udc, ctrl, AST_UDC_FUNC_CTRL); |
| } |
| |
| /* Mask & ack all interrupts before installing the handler */ |
| ast_udc_write(udc, 0, AST_UDC_IER); |
| ast_udc_write(udc, UDC_IRQ_ACK_ALL, AST_UDC_ISR); |
| |
| /* Enable some interrupts */ |
| ctrl = UDC_IRQ_EP_POOL_ACK_STALL | UDC_IRQ_BUS_RESUME | |
| UDC_IRQ_BUS_SUSPEND | UDC_IRQ_BUS_RESET | |
| UDC_IRQ_EP0_IN_ACK_STALL | UDC_IRQ_EP0_OUT_ACK_STALL | |
| UDC_IRQ_EP0_SETUP; |
| ast_udc_write(udc, ctrl, AST_UDC_IER); |
| |
| /* Cleanup and enable ep ACK interrupts */ |
| ast_udc_write(udc, UDC_IRQ_EP_ACK_ALL, AST_UDC_EP_ACK_IER); |
| ast_udc_write(udc, UDC_IRQ_EP_ACK_ALL, AST_UDC_EP_ACK_ISR); |
| |
| ast_udc_write(udc, 0, AST_UDC_EP0_CTRL); |
| } |
| |
| static void ast_udc_remove(struct platform_device *pdev) |
| { |
| struct ast_udc_dev *udc = platform_get_drvdata(pdev); |
| unsigned long flags; |
| u32 ctrl; |
| |
| usb_del_gadget_udc(&udc->gadget); |
| if (udc->driver) { |
| /* |
| * This is broken as only some cleanup is skipped, *udev is |
| * freed and the register mapping goes away. Any further usage |
| * probably crashes. Also the device is unbound, so the skipped |
| * cleanup is never catched up later. |
| */ |
| dev_alert(&pdev->dev, |
| "Driver is busy and still going away. Fasten your seat belts!\n"); |
| return; |
| } |
| |
| spin_lock_irqsave(&udc->lock, flags); |
| |
| /* Disable upstream port connection */ |
| ctrl = ast_udc_read(udc, AST_UDC_FUNC_CTRL) & ~USB_UPSTREAM_EN; |
| ast_udc_write(udc, ctrl, AST_UDC_FUNC_CTRL); |
| |
| clk_disable_unprepare(udc->clk); |
| |
| spin_unlock_irqrestore(&udc->lock, flags); |
| |
| if (udc->ep0_buf) |
| dma_free_coherent(&pdev->dev, |
| AST_UDC_EP_DMA_SIZE * AST_UDC_NUM_ENDPOINTS, |
| udc->ep0_buf, |
| udc->ep0_buf_dma); |
| |
| udc->ep0_buf = NULL; |
| } |
| |
| static int ast_udc_probe(struct platform_device *pdev) |
| { |
| enum usb_device_speed max_speed; |
| struct device *dev = &pdev->dev; |
| struct ast_udc_dev *udc; |
| int rc; |
| |
| udc = devm_kzalloc(&pdev->dev, sizeof(struct ast_udc_dev), GFP_KERNEL); |
| if (!udc) |
| return -ENOMEM; |
| |
| udc->gadget.dev.parent = dev; |
| udc->pdev = pdev; |
| spin_lock_init(&udc->lock); |
| |
| udc->gadget.ops = &ast_udc_ops; |
| udc->gadget.ep0 = &udc->ep[0].ep; |
| udc->gadget.name = "aspeed-udc"; |
| udc->gadget.dev.init_name = "gadget"; |
| |
| udc->reg = devm_platform_ioremap_resource(pdev, 0); |
| if (IS_ERR(udc->reg)) { |
| dev_err(&pdev->dev, "Failed to map resources\n"); |
| return PTR_ERR(udc->reg); |
| } |
| |
| platform_set_drvdata(pdev, udc); |
| |
| udc->clk = devm_clk_get(&pdev->dev, NULL); |
| if (IS_ERR(udc->clk)) { |
| rc = PTR_ERR(udc->clk); |
| goto err; |
| } |
| rc = clk_prepare_enable(udc->clk); |
| if (rc) { |
| dev_err(&pdev->dev, "Failed to enable clock (0x%x)\n", rc); |
| goto err; |
| } |
| |
| /* Check if we need to limit the HW to USB1 */ |
| max_speed = usb_get_maximum_speed(&pdev->dev); |
| if (max_speed != USB_SPEED_UNKNOWN && max_speed < USB_SPEED_HIGH) |
| udc->force_usb1 = true; |
| |
| /* |
| * Allocate DMA buffers for all EPs in one chunk |
| */ |
| udc->ep0_buf = dma_alloc_coherent(&pdev->dev, |
| AST_UDC_EP_DMA_SIZE * |
| AST_UDC_NUM_ENDPOINTS, |
| &udc->ep0_buf_dma, GFP_KERNEL); |
| |
| udc->gadget.speed = USB_SPEED_UNKNOWN; |
| udc->gadget.max_speed = USB_SPEED_HIGH; |
| udc->creq = udc->reg + AST_UDC_SETUP0; |
| |
| /* |
| * Support single stage mode or 32/256 stages descriptor mode. |
| * Set default as Descriptor Mode. |
| */ |
| udc->desc_mode = AST_UDC_DESC_MODE; |
| |
| dev_info(&pdev->dev, "DMA %s\n", udc->desc_mode ? |
| "descriptor mode" : "single mode"); |
| |
| INIT_LIST_HEAD(&udc->gadget.ep_list); |
| INIT_LIST_HEAD(&udc->gadget.ep0->ep_list); |
| |
| /* Initialized udc ep */ |
| ast_udc_init_ep(udc); |
| |
| /* Initialized udc device */ |
| ast_udc_init_dev(udc); |
| |
| /* Initialized udc hardware */ |
| ast_udc_init_hw(udc); |
| |
| /* Find interrupt and install handler */ |
| udc->irq = platform_get_irq(pdev, 0); |
| if (udc->irq < 0) { |
| rc = udc->irq; |
| goto err; |
| } |
| |
| rc = devm_request_irq(&pdev->dev, udc->irq, ast_udc_isr, 0, |
| KBUILD_MODNAME, udc); |
| if (rc) { |
| dev_err(&pdev->dev, "Failed to request interrupt\n"); |
| goto err; |
| } |
| |
| rc = usb_add_gadget_udc(&pdev->dev, &udc->gadget); |
| if (rc) { |
| dev_err(&pdev->dev, "Failed to add gadget udc\n"); |
| goto err; |
| } |
| |
| dev_info(&pdev->dev, "Initialized udc in USB%s mode\n", |
| udc->force_usb1 ? "1" : "2"); |
| |
| return 0; |
| |
| err: |
| dev_err(&pdev->dev, "Failed to udc probe, rc:0x%x\n", rc); |
| ast_udc_remove(pdev); |
| |
| return rc; |
| } |
| |
| static const struct of_device_id ast_udc_of_dt_ids[] = { |
| { .compatible = "aspeed,ast2600-udc", }, |
| {} |
| }; |
| |
| MODULE_DEVICE_TABLE(of, ast_udc_of_dt_ids); |
| |
| static struct platform_driver ast_udc_driver = { |
| .probe = ast_udc_probe, |
| .remove_new = ast_udc_remove, |
| .driver = { |
| .name = KBUILD_MODNAME, |
| .of_match_table = ast_udc_of_dt_ids, |
| }, |
| }; |
| |
| module_platform_driver(ast_udc_driver); |
| |
| MODULE_DESCRIPTION("ASPEED UDC driver"); |
| MODULE_AUTHOR("Neal Liu <neal_liu@aspeedtech.com>"); |
| MODULE_LICENSE("GPL"); |