| // SPDX-License-Identifier: GPL-2.0+ |
| /* |
| * bcm63xx_udc.c -- BCM63xx UDC high/full speed USB device controller |
| * |
| * Copyright (C) 2012 Kevin Cernekee <cernekee@gmail.com> |
| * Copyright (C) 2012 Broadcom Corporation |
| */ |
| |
| #include <linux/bitops.h> |
| #include <linux/bug.h> |
| #include <linux/clk.h> |
| #include <linux/compiler.h> |
| #include <linux/debugfs.h> |
| #include <linux/delay.h> |
| #include <linux/device.h> |
| #include <linux/dma-mapping.h> |
| #include <linux/errno.h> |
| #include <linux/interrupt.h> |
| #include <linux/ioport.h> |
| #include <linux/kernel.h> |
| #include <linux/list.h> |
| #include <linux/module.h> |
| #include <linux/moduleparam.h> |
| #include <linux/platform_device.h> |
| #include <linux/sched.h> |
| #include <linux/seq_file.h> |
| #include <linux/slab.h> |
| #include <linux/timer.h> |
| #include <linux/usb/ch9.h> |
| #include <linux/usb/gadget.h> |
| #include <linux/workqueue.h> |
| |
| #include <bcm63xx_cpu.h> |
| #include <bcm63xx_iudma.h> |
| #include <bcm63xx_dev_usb_usbd.h> |
| #include <bcm63xx_io.h> |
| #include <bcm63xx_regs.h> |
| |
| #define DRV_MODULE_NAME "bcm63xx_udc" |
| |
| static const char bcm63xx_ep0name[] = "ep0"; |
| |
| static const struct { |
| const char *name; |
| const struct usb_ep_caps caps; |
| } bcm63xx_ep_info[] = { |
| #define EP_INFO(_name, _caps) \ |
| { \ |
| .name = _name, \ |
| .caps = _caps, \ |
| } |
| |
| EP_INFO(bcm63xx_ep0name, |
| USB_EP_CAPS(USB_EP_CAPS_TYPE_CONTROL, USB_EP_CAPS_DIR_ALL)), |
| EP_INFO("ep1in-bulk", |
| USB_EP_CAPS(USB_EP_CAPS_TYPE_BULK, USB_EP_CAPS_DIR_IN)), |
| EP_INFO("ep2out-bulk", |
| USB_EP_CAPS(USB_EP_CAPS_TYPE_BULK, USB_EP_CAPS_DIR_OUT)), |
| EP_INFO("ep3in-int", |
| USB_EP_CAPS(USB_EP_CAPS_TYPE_INT, USB_EP_CAPS_DIR_IN)), |
| EP_INFO("ep4out-int", |
| USB_EP_CAPS(USB_EP_CAPS_TYPE_INT, USB_EP_CAPS_DIR_OUT)), |
| |
| #undef EP_INFO |
| }; |
| |
| static bool use_fullspeed; |
| module_param(use_fullspeed, bool, S_IRUGO); |
| MODULE_PARM_DESC(use_fullspeed, "true for fullspeed only"); |
| |
| /* |
| * RX IRQ coalescing options: |
| * |
| * false (default) - one IRQ per DATAx packet. Slow but reliable. The |
| * driver is able to pass the "testusb" suite and recover from conditions like: |
| * |
| * 1) Device queues up a 2048-byte RX IUDMA transaction on an OUT bulk ep |
| * 2) Host sends 512 bytes of data |
| * 3) Host decides to reconfigure the device and sends SET_INTERFACE |
| * 4) Device shuts down the endpoint and cancels the RX transaction |
| * |
| * true - one IRQ per transfer, for transfers <= 2048B. Generates |
| * considerably fewer IRQs, but error recovery is less robust. Does not |
| * reliably pass "testusb". |
| * |
| * TX always uses coalescing, because we can cancel partially complete TX |
| * transfers by repeatedly flushing the FIFO. The hardware doesn't allow |
| * this on RX. |
| */ |
| static bool irq_coalesce; |
| module_param(irq_coalesce, bool, S_IRUGO); |
| MODULE_PARM_DESC(irq_coalesce, "take one IRQ per RX transfer"); |
| |
| #define BCM63XX_NUM_EP 5 |
| #define BCM63XX_NUM_IUDMA 6 |
| #define BCM63XX_NUM_FIFO_PAIRS 3 |
| |
| #define IUDMA_RESET_TIMEOUT_US 10000 |
| |
| #define IUDMA_EP0_RXCHAN 0 |
| #define IUDMA_EP0_TXCHAN 1 |
| |
| #define IUDMA_MAX_FRAGMENT 2048 |
| #define BCM63XX_MAX_CTRL_PKT 64 |
| |
| #define BCMEP_CTRL 0x00 |
| #define BCMEP_ISOC 0x01 |
| #define BCMEP_BULK 0x02 |
| #define BCMEP_INTR 0x03 |
| |
| #define BCMEP_OUT 0x00 |
| #define BCMEP_IN 0x01 |
| |
| #define BCM63XX_SPD_FULL 1 |
| #define BCM63XX_SPD_HIGH 0 |
| |
| #define IUDMA_DMAC_OFFSET 0x200 |
| #define IUDMA_DMAS_OFFSET 0x400 |
| |
| enum bcm63xx_ep0_state { |
| EP0_REQUEUE, |
| EP0_IDLE, |
| EP0_IN_DATA_PHASE_SETUP, |
| EP0_IN_DATA_PHASE_COMPLETE, |
| EP0_OUT_DATA_PHASE_SETUP, |
| EP0_OUT_DATA_PHASE_COMPLETE, |
| EP0_OUT_STATUS_PHASE, |
| EP0_IN_FAKE_STATUS_PHASE, |
| EP0_SHUTDOWN, |
| }; |
| |
| static const char __maybe_unused bcm63xx_ep0_state_names[][32] = { |
| "REQUEUE", |
| "IDLE", |
| "IN_DATA_PHASE_SETUP", |
| "IN_DATA_PHASE_COMPLETE", |
| "OUT_DATA_PHASE_SETUP", |
| "OUT_DATA_PHASE_COMPLETE", |
| "OUT_STATUS_PHASE", |
| "IN_FAKE_STATUS_PHASE", |
| "SHUTDOWN", |
| }; |
| |
| /** |
| * struct iudma_ch_cfg - Static configuration for an IUDMA channel. |
| * @ep_num: USB endpoint number. |
| * @n_bds: Number of buffer descriptors in the ring. |
| * @ep_type: Endpoint type (control, bulk, interrupt). |
| * @dir: Direction (in, out). |
| * @n_fifo_slots: Number of FIFO entries to allocate for this channel. |
| * @max_pkt_hs: Maximum packet size in high speed mode. |
| * @max_pkt_fs: Maximum packet size in full speed mode. |
| */ |
| struct iudma_ch_cfg { |
| int ep_num; |
| int n_bds; |
| int ep_type; |
| int dir; |
| int n_fifo_slots; |
| int max_pkt_hs; |
| int max_pkt_fs; |
| }; |
| |
| static const struct iudma_ch_cfg iudma_defaults[] = { |
| |
| /* This controller was designed to support a CDC/RNDIS application. |
| It may be possible to reconfigure some of the endpoints, but |
| the hardware limitations (FIFO sizing and number of DMA channels) |
| may significantly impact flexibility and/or stability. Change |
| these values at your own risk. |
| |
| ep_num ep_type n_fifo_slots max_pkt_fs |
| idx | n_bds | dir | max_pkt_hs | |
| | | | | | | | | */ |
| [0] = { -1, 4, BCMEP_CTRL, BCMEP_OUT, 32, 64, 64 }, |
| [1] = { 0, 4, BCMEP_CTRL, BCMEP_OUT, 32, 64, 64 }, |
| [2] = { 2, 16, BCMEP_BULK, BCMEP_OUT, 128, 512, 64 }, |
| [3] = { 1, 16, BCMEP_BULK, BCMEP_IN, 128, 512, 64 }, |
| [4] = { 4, 4, BCMEP_INTR, BCMEP_OUT, 32, 64, 64 }, |
| [5] = { 3, 4, BCMEP_INTR, BCMEP_IN, 32, 64, 64 }, |
| }; |
| |
| struct bcm63xx_udc; |
| |
| /** |
| * struct iudma_ch - Represents the current state of a single IUDMA channel. |
| * @ch_idx: IUDMA channel index (0 to BCM63XX_NUM_IUDMA-1). |
| * @ep_num: USB endpoint number. -1 for ep0 RX. |
| * @enabled: Whether bcm63xx_ep_enable() has been called. |
| * @max_pkt: "Chunk size" on the USB interface. Based on interface speed. |
| * @is_tx: true for TX, false for RX. |
| * @bep: Pointer to the associated endpoint. NULL for ep0 RX. |
| * @udc: Reference to the device controller. |
| * @read_bd: Next buffer descriptor to reap from the hardware. |
| * @write_bd: Next BD available for a new packet. |
| * @end_bd: Points to the final BD in the ring. |
| * @n_bds_used: Number of BD entries currently occupied. |
| * @bd_ring: Base pointer to the BD ring. |
| * @bd_ring_dma: Physical (DMA) address of bd_ring. |
| * @n_bds: Total number of BDs in the ring. |
| * |
| * ep0 has two IUDMA channels (IUDMA_EP0_RXCHAN and IUDMA_EP0_TXCHAN), as it is |
| * bidirectional. The "struct usb_ep" associated with ep0 is for TX (IN) |
| * only. |
| * |
| * Each bulk/intr endpoint has a single IUDMA channel and a single |
| * struct usb_ep. |
| */ |
| struct iudma_ch { |
| unsigned int ch_idx; |
| int ep_num; |
| bool enabled; |
| int max_pkt; |
| bool is_tx; |
| struct bcm63xx_ep *bep; |
| struct bcm63xx_udc *udc; |
| |
| struct bcm_enet_desc *read_bd; |
| struct bcm_enet_desc *write_bd; |
| struct bcm_enet_desc *end_bd; |
| int n_bds_used; |
| |
| struct bcm_enet_desc *bd_ring; |
| dma_addr_t bd_ring_dma; |
| unsigned int n_bds; |
| }; |
| |
| /** |
| * struct bcm63xx_ep - Internal (driver) state of a single endpoint. |
| * @ep_num: USB endpoint number. |
| * @iudma: Pointer to IUDMA channel state. |
| * @ep: USB gadget layer representation of the EP. |
| * @udc: Reference to the device controller. |
| * @queue: Linked list of outstanding requests for this EP. |
| * @halted: 1 if the EP is stalled; 0 otherwise. |
| */ |
| struct bcm63xx_ep { |
| unsigned int ep_num; |
| struct iudma_ch *iudma; |
| struct usb_ep ep; |
| struct bcm63xx_udc *udc; |
| struct list_head queue; |
| unsigned halted:1; |
| }; |
| |
| /** |
| * struct bcm63xx_req - Internal (driver) state of a single request. |
| * @queue: Links back to the EP's request list. |
| * @req: USB gadget layer representation of the request. |
| * @offset: Current byte offset into the data buffer (next byte to queue). |
| * @bd_bytes: Number of data bytes in outstanding BD entries. |
| * @iudma: IUDMA channel used for the request. |
| */ |
| struct bcm63xx_req { |
| struct list_head queue; /* ep's requests */ |
| struct usb_request req; |
| unsigned int offset; |
| unsigned int bd_bytes; |
| struct iudma_ch *iudma; |
| }; |
| |
| /** |
| * struct bcm63xx_udc - Driver/hardware private context. |
| * @lock: Spinlock to mediate access to this struct, and (most) HW regs. |
| * @dev: Generic Linux device structure. |
| * @pd: Platform data (board/port info). |
| * @usbd_clk: Clock descriptor for the USB device block. |
| * @usbh_clk: Clock descriptor for the USB host block. |
| * @gadget: USB slave device. |
| * @driver: Driver for USB slave devices. |
| * @usbd_regs: Base address of the USBD/USB20D block. |
| * @iudma_regs: Base address of the USBD's associated IUDMA block. |
| * @bep: Array of endpoints, including ep0. |
| * @iudma: Array of all IUDMA channels used by this controller. |
| * @cfg: USB configuration number, from SET_CONFIGURATION wValue. |
| * @iface: USB interface number, from SET_INTERFACE wIndex. |
| * @alt_iface: USB alt interface number, from SET_INTERFACE wValue. |
| * @ep0_ctrl_req: Request object for bcm63xx_udc-initiated ep0 transactions. |
| * @ep0_ctrl_buf: Data buffer for ep0_ctrl_req. |
| * @ep0state: Current state of the ep0 state machine. |
| * @ep0_wq: Workqueue struct used to wake up the ep0 state machine. |
| * @wedgemap: Bitmap of wedged endpoints. |
| * @ep0_req_reset: USB reset is pending. |
| * @ep0_req_set_cfg: Need to spoof a SET_CONFIGURATION packet. |
| * @ep0_req_set_iface: Need to spoof a SET_INTERFACE packet. |
| * @ep0_req_shutdown: Driver is shutting down; requesting ep0 to halt activity. |
| * @ep0_req_completed: ep0 request has completed; worker has not seen it yet. |
| * @ep0_reply: Pending reply from gadget driver. |
| * @ep0_request: Outstanding ep0 request. |
| * @debugfs_root: debugfs directory: /sys/kernel/debug/<DRV_MODULE_NAME>. |
| */ |
| struct bcm63xx_udc { |
| spinlock_t lock; |
| |
| struct device *dev; |
| struct bcm63xx_usbd_platform_data *pd; |
| struct clk *usbd_clk; |
| struct clk *usbh_clk; |
| |
| struct usb_gadget gadget; |
| struct usb_gadget_driver *driver; |
| |
| void __iomem *usbd_regs; |
| void __iomem *iudma_regs; |
| |
| struct bcm63xx_ep bep[BCM63XX_NUM_EP]; |
| struct iudma_ch iudma[BCM63XX_NUM_IUDMA]; |
| |
| int cfg; |
| int iface; |
| int alt_iface; |
| |
| struct bcm63xx_req ep0_ctrl_req; |
| u8 *ep0_ctrl_buf; |
| |
| int ep0state; |
| struct work_struct ep0_wq; |
| |
| unsigned long wedgemap; |
| |
| unsigned ep0_req_reset:1; |
| unsigned ep0_req_set_cfg:1; |
| unsigned ep0_req_set_iface:1; |
| unsigned ep0_req_shutdown:1; |
| |
| unsigned ep0_req_completed:1; |
| struct usb_request *ep0_reply; |
| struct usb_request *ep0_request; |
| |
| struct dentry *debugfs_root; |
| }; |
| |
| static const struct usb_ep_ops bcm63xx_udc_ep_ops; |
| |
| /*********************************************************************** |
| * Convenience functions |
| ***********************************************************************/ |
| |
| static inline struct bcm63xx_udc *gadget_to_udc(struct usb_gadget *g) |
| { |
| return container_of(g, struct bcm63xx_udc, gadget); |
| } |
| |
| static inline struct bcm63xx_ep *our_ep(struct usb_ep *ep) |
| { |
| return container_of(ep, struct bcm63xx_ep, ep); |
| } |
| |
| static inline struct bcm63xx_req *our_req(struct usb_request *req) |
| { |
| return container_of(req, struct bcm63xx_req, req); |
| } |
| |
| static inline u32 usbd_readl(struct bcm63xx_udc *udc, u32 off) |
| { |
| return bcm_readl(udc->usbd_regs + off); |
| } |
| |
| static inline void usbd_writel(struct bcm63xx_udc *udc, u32 val, u32 off) |
| { |
| bcm_writel(val, udc->usbd_regs + off); |
| } |
| |
| static inline u32 usb_dma_readl(struct bcm63xx_udc *udc, u32 off) |
| { |
| return bcm_readl(udc->iudma_regs + off); |
| } |
| |
| static inline void usb_dma_writel(struct bcm63xx_udc *udc, u32 val, u32 off) |
| { |
| bcm_writel(val, udc->iudma_regs + off); |
| } |
| |
| static inline u32 usb_dmac_readl(struct bcm63xx_udc *udc, u32 off, int chan) |
| { |
| return bcm_readl(udc->iudma_regs + IUDMA_DMAC_OFFSET + off + |
| (ENETDMA_CHAN_WIDTH * chan)); |
| } |
| |
| static inline void usb_dmac_writel(struct bcm63xx_udc *udc, u32 val, u32 off, |
| int chan) |
| { |
| bcm_writel(val, udc->iudma_regs + IUDMA_DMAC_OFFSET + off + |
| (ENETDMA_CHAN_WIDTH * chan)); |
| } |
| |
| static inline u32 usb_dmas_readl(struct bcm63xx_udc *udc, u32 off, int chan) |
| { |
| return bcm_readl(udc->iudma_regs + IUDMA_DMAS_OFFSET + off + |
| (ENETDMA_CHAN_WIDTH * chan)); |
| } |
| |
| static inline void usb_dmas_writel(struct bcm63xx_udc *udc, u32 val, u32 off, |
| int chan) |
| { |
| bcm_writel(val, udc->iudma_regs + IUDMA_DMAS_OFFSET + off + |
| (ENETDMA_CHAN_WIDTH * chan)); |
| } |
| |
| static inline void set_clocks(struct bcm63xx_udc *udc, bool is_enabled) |
| { |
| if (is_enabled) { |
| clk_enable(udc->usbh_clk); |
| clk_enable(udc->usbd_clk); |
| udelay(10); |
| } else { |
| clk_disable(udc->usbd_clk); |
| clk_disable(udc->usbh_clk); |
| } |
| } |
| |
| /*********************************************************************** |
| * Low-level IUDMA / FIFO operations |
| ***********************************************************************/ |
| |
| /** |
| * bcm63xx_ep_dma_select - Helper function to set up the init_sel signal. |
| * @udc: Reference to the device controller. |
| * @idx: Desired init_sel value. |
| * |
| * The "init_sel" signal is used as a selection index for both endpoints |
| * and IUDMA channels. Since these do not map 1:1, the use of this signal |
| * depends on the context. |
| */ |
| static void bcm63xx_ep_dma_select(struct bcm63xx_udc *udc, int idx) |
| { |
| u32 val = usbd_readl(udc, USBD_CONTROL_REG); |
| |
| val &= ~USBD_CONTROL_INIT_SEL_MASK; |
| val |= idx << USBD_CONTROL_INIT_SEL_SHIFT; |
| usbd_writel(udc, val, USBD_CONTROL_REG); |
| } |
| |
| /** |
| * bcm63xx_set_stall - Enable/disable stall on one endpoint. |
| * @udc: Reference to the device controller. |
| * @bep: Endpoint on which to operate. |
| * @is_stalled: true to enable stall, false to disable. |
| * |
| * See notes in bcm63xx_update_wedge() regarding automatic clearing of |
| * halt/stall conditions. |
| */ |
| static void bcm63xx_set_stall(struct bcm63xx_udc *udc, struct bcm63xx_ep *bep, |
| bool is_stalled) |
| { |
| u32 val; |
| |
| val = USBD_STALL_UPDATE_MASK | |
| (is_stalled ? USBD_STALL_ENABLE_MASK : 0) | |
| (bep->ep_num << USBD_STALL_EPNUM_SHIFT); |
| usbd_writel(udc, val, USBD_STALL_REG); |
| } |
| |
| /** |
| * bcm63xx_fifo_setup - (Re)initialize FIFO boundaries and settings. |
| * @udc: Reference to the device controller. |
| * |
| * These parameters depend on the USB link speed. Settings are |
| * per-IUDMA-channel-pair. |
| */ |
| static void bcm63xx_fifo_setup(struct bcm63xx_udc *udc) |
| { |
| int is_hs = udc->gadget.speed == USB_SPEED_HIGH; |
| u32 i, val, rx_fifo_slot, tx_fifo_slot; |
| |
| /* set up FIFO boundaries and packet sizes; this is done in pairs */ |
| rx_fifo_slot = tx_fifo_slot = 0; |
| for (i = 0; i < BCM63XX_NUM_IUDMA; i += 2) { |
| const struct iudma_ch_cfg *rx_cfg = &iudma_defaults[i]; |
| const struct iudma_ch_cfg *tx_cfg = &iudma_defaults[i + 1]; |
| |
| bcm63xx_ep_dma_select(udc, i >> 1); |
| |
| val = (rx_fifo_slot << USBD_RXFIFO_CONFIG_START_SHIFT) | |
| ((rx_fifo_slot + rx_cfg->n_fifo_slots - 1) << |
| USBD_RXFIFO_CONFIG_END_SHIFT); |
| rx_fifo_slot += rx_cfg->n_fifo_slots; |
| usbd_writel(udc, val, USBD_RXFIFO_CONFIG_REG); |
| usbd_writel(udc, |
| is_hs ? rx_cfg->max_pkt_hs : rx_cfg->max_pkt_fs, |
| USBD_RXFIFO_EPSIZE_REG); |
| |
| val = (tx_fifo_slot << USBD_TXFIFO_CONFIG_START_SHIFT) | |
| ((tx_fifo_slot + tx_cfg->n_fifo_slots - 1) << |
| USBD_TXFIFO_CONFIG_END_SHIFT); |
| tx_fifo_slot += tx_cfg->n_fifo_slots; |
| usbd_writel(udc, val, USBD_TXFIFO_CONFIG_REG); |
| usbd_writel(udc, |
| is_hs ? tx_cfg->max_pkt_hs : tx_cfg->max_pkt_fs, |
| USBD_TXFIFO_EPSIZE_REG); |
| |
| usbd_readl(udc, USBD_TXFIFO_EPSIZE_REG); |
| } |
| } |
| |
| /** |
| * bcm63xx_fifo_reset_ep - Flush a single endpoint's FIFO. |
| * @udc: Reference to the device controller. |
| * @ep_num: Endpoint number. |
| */ |
| static void bcm63xx_fifo_reset_ep(struct bcm63xx_udc *udc, int ep_num) |
| { |
| u32 val; |
| |
| bcm63xx_ep_dma_select(udc, ep_num); |
| |
| val = usbd_readl(udc, USBD_CONTROL_REG); |
| val |= USBD_CONTROL_FIFO_RESET_MASK; |
| usbd_writel(udc, val, USBD_CONTROL_REG); |
| usbd_readl(udc, USBD_CONTROL_REG); |
| } |
| |
| /** |
| * bcm63xx_fifo_reset - Flush all hardware FIFOs. |
| * @udc: Reference to the device controller. |
| */ |
| static void bcm63xx_fifo_reset(struct bcm63xx_udc *udc) |
| { |
| int i; |
| |
| for (i = 0; i < BCM63XX_NUM_FIFO_PAIRS; i++) |
| bcm63xx_fifo_reset_ep(udc, i); |
| } |
| |
| /** |
| * bcm63xx_ep_init - Initial (one-time) endpoint initialization. |
| * @udc: Reference to the device controller. |
| */ |
| static void bcm63xx_ep_init(struct bcm63xx_udc *udc) |
| { |
| u32 i, val; |
| |
| for (i = 0; i < BCM63XX_NUM_IUDMA; i++) { |
| const struct iudma_ch_cfg *cfg = &iudma_defaults[i]; |
| |
| if (cfg->ep_num < 0) |
| continue; |
| |
| bcm63xx_ep_dma_select(udc, cfg->ep_num); |
| val = (cfg->ep_type << USBD_EPNUM_TYPEMAP_TYPE_SHIFT) | |
| ((i >> 1) << USBD_EPNUM_TYPEMAP_DMA_CH_SHIFT); |
| usbd_writel(udc, val, USBD_EPNUM_TYPEMAP_REG); |
| } |
| } |
| |
| /** |
| * bcm63xx_ep_setup - Configure per-endpoint settings. |
| * @udc: Reference to the device controller. |
| * |
| * This needs to be rerun if the speed/cfg/intf/altintf changes. |
| */ |
| static void bcm63xx_ep_setup(struct bcm63xx_udc *udc) |
| { |
| u32 val, i; |
| |
| usbd_writel(udc, USBD_CSR_SETUPADDR_DEF, USBD_CSR_SETUPADDR_REG); |
| |
| for (i = 0; i < BCM63XX_NUM_IUDMA; i++) { |
| const struct iudma_ch_cfg *cfg = &iudma_defaults[i]; |
| int max_pkt = udc->gadget.speed == USB_SPEED_HIGH ? |
| cfg->max_pkt_hs : cfg->max_pkt_fs; |
| int idx = cfg->ep_num; |
| |
| udc->iudma[i].max_pkt = max_pkt; |
| |
| if (idx < 0) |
| continue; |
| usb_ep_set_maxpacket_limit(&udc->bep[idx].ep, max_pkt); |
| |
| val = (idx << USBD_CSR_EP_LOG_SHIFT) | |
| (cfg->dir << USBD_CSR_EP_DIR_SHIFT) | |
| (cfg->ep_type << USBD_CSR_EP_TYPE_SHIFT) | |
| (udc->cfg << USBD_CSR_EP_CFG_SHIFT) | |
| (udc->iface << USBD_CSR_EP_IFACE_SHIFT) | |
| (udc->alt_iface << USBD_CSR_EP_ALTIFACE_SHIFT) | |
| (max_pkt << USBD_CSR_EP_MAXPKT_SHIFT); |
| usbd_writel(udc, val, USBD_CSR_EP_REG(idx)); |
| } |
| } |
| |
| /** |
| * iudma_write - Queue a single IUDMA transaction. |
| * @udc: Reference to the device controller. |
| * @iudma: IUDMA channel to use. |
| * @breq: Request containing the transaction data. |
| * |
| * For RX IUDMA, this will queue a single buffer descriptor, as RX IUDMA |
| * does not honor SOP/EOP so the handling of multiple buffers is ambiguous. |
| * So iudma_write() may be called several times to fulfill a single |
| * usb_request. |
| * |
| * For TX IUDMA, this can queue multiple buffer descriptors if needed. |
| */ |
| static void iudma_write(struct bcm63xx_udc *udc, struct iudma_ch *iudma, |
| struct bcm63xx_req *breq) |
| { |
| int first_bd = 1, last_bd = 0, extra_zero_pkt = 0; |
| unsigned int bytes_left = breq->req.length - breq->offset; |
| const int max_bd_bytes = !irq_coalesce && !iudma->is_tx ? |
| iudma->max_pkt : IUDMA_MAX_FRAGMENT; |
| |
| iudma->n_bds_used = 0; |
| breq->bd_bytes = 0; |
| breq->iudma = iudma; |
| |
| if ((bytes_left % iudma->max_pkt == 0) && bytes_left && breq->req.zero) |
| extra_zero_pkt = 1; |
| |
| do { |
| struct bcm_enet_desc *d = iudma->write_bd; |
| u32 dmaflags = 0; |
| unsigned int n_bytes; |
| |
| if (d == iudma->end_bd) { |
| dmaflags |= DMADESC_WRAP_MASK; |
| iudma->write_bd = iudma->bd_ring; |
| } else { |
| iudma->write_bd++; |
| } |
| iudma->n_bds_used++; |
| |
| n_bytes = min_t(int, bytes_left, max_bd_bytes); |
| if (n_bytes) |
| dmaflags |= n_bytes << DMADESC_LENGTH_SHIFT; |
| else |
| dmaflags |= (1 << DMADESC_LENGTH_SHIFT) | |
| DMADESC_USB_ZERO_MASK; |
| |
| dmaflags |= DMADESC_OWNER_MASK; |
| if (first_bd) { |
| dmaflags |= DMADESC_SOP_MASK; |
| first_bd = 0; |
| } |
| |
| /* |
| * extra_zero_pkt forces one more iteration through the loop |
| * after all data is queued up, to send the zero packet |
| */ |
| if (extra_zero_pkt && !bytes_left) |
| extra_zero_pkt = 0; |
| |
| if (!iudma->is_tx || iudma->n_bds_used == iudma->n_bds || |
| (n_bytes == bytes_left && !extra_zero_pkt)) { |
| last_bd = 1; |
| dmaflags |= DMADESC_EOP_MASK; |
| } |
| |
| d->address = breq->req.dma + breq->offset; |
| mb(); |
| d->len_stat = dmaflags; |
| |
| breq->offset += n_bytes; |
| breq->bd_bytes += n_bytes; |
| bytes_left -= n_bytes; |
| } while (!last_bd); |
| |
| usb_dmac_writel(udc, ENETDMAC_CHANCFG_EN_MASK, |
| ENETDMAC_CHANCFG_REG, iudma->ch_idx); |
| } |
| |
| /** |
| * iudma_read - Check for IUDMA buffer completion. |
| * @udc: Reference to the device controller. |
| * @iudma: IUDMA channel to use. |
| * |
| * This checks to see if ALL of the outstanding BDs on the DMA channel |
| * have been filled. If so, it returns the actual transfer length; |
| * otherwise it returns -EBUSY. |
| */ |
| static int iudma_read(struct bcm63xx_udc *udc, struct iudma_ch *iudma) |
| { |
| int i, actual_len = 0; |
| struct bcm_enet_desc *d = iudma->read_bd; |
| |
| if (!iudma->n_bds_used) |
| return -EINVAL; |
| |
| for (i = 0; i < iudma->n_bds_used; i++) { |
| u32 dmaflags; |
| |
| dmaflags = d->len_stat; |
| |
| if (dmaflags & DMADESC_OWNER_MASK) |
| return -EBUSY; |
| |
| actual_len += (dmaflags & DMADESC_LENGTH_MASK) >> |
| DMADESC_LENGTH_SHIFT; |
| if (d == iudma->end_bd) |
| d = iudma->bd_ring; |
| else |
| d++; |
| } |
| |
| iudma->read_bd = d; |
| iudma->n_bds_used = 0; |
| return actual_len; |
| } |
| |
| /** |
| * iudma_reset_channel - Stop DMA on a single channel. |
| * @udc: Reference to the device controller. |
| * @iudma: IUDMA channel to reset. |
| */ |
| static void iudma_reset_channel(struct bcm63xx_udc *udc, struct iudma_ch *iudma) |
| { |
| int timeout = IUDMA_RESET_TIMEOUT_US; |
| struct bcm_enet_desc *d; |
| int ch_idx = iudma->ch_idx; |
| |
| if (!iudma->is_tx) |
| bcm63xx_fifo_reset_ep(udc, max(0, iudma->ep_num)); |
| |
| /* stop DMA, then wait for the hardware to wrap up */ |
| usb_dmac_writel(udc, 0, ENETDMAC_CHANCFG_REG, ch_idx); |
| |
| while (usb_dmac_readl(udc, ENETDMAC_CHANCFG_REG, ch_idx) & |
| ENETDMAC_CHANCFG_EN_MASK) { |
| udelay(1); |
| |
| /* repeatedly flush the FIFO data until the BD completes */ |
| if (iudma->is_tx && iudma->ep_num >= 0) |
| bcm63xx_fifo_reset_ep(udc, iudma->ep_num); |
| |
| if (!timeout--) { |
| dev_err(udc->dev, "can't reset IUDMA channel %d\n", |
| ch_idx); |
| break; |
| } |
| if (timeout == IUDMA_RESET_TIMEOUT_US / 2) { |
| dev_warn(udc->dev, "forcibly halting IUDMA channel %d\n", |
| ch_idx); |
| usb_dmac_writel(udc, ENETDMAC_CHANCFG_BUFHALT_MASK, |
| ENETDMAC_CHANCFG_REG, ch_idx); |
| } |
| } |
| usb_dmac_writel(udc, ~0, ENETDMAC_IR_REG, ch_idx); |
| |
| /* don't leave "live" HW-owned entries for the next guy to step on */ |
| for (d = iudma->bd_ring; d <= iudma->end_bd; d++) |
| d->len_stat = 0; |
| mb(); |
| |
| iudma->read_bd = iudma->write_bd = iudma->bd_ring; |
| iudma->n_bds_used = 0; |
| |
| /* set up IRQs, UBUS burst size, and BD base for this channel */ |
| usb_dmac_writel(udc, ENETDMAC_IR_BUFDONE_MASK, |
| ENETDMAC_IRMASK_REG, ch_idx); |
| usb_dmac_writel(udc, 8, ENETDMAC_MAXBURST_REG, ch_idx); |
| |
| usb_dmas_writel(udc, iudma->bd_ring_dma, ENETDMAS_RSTART_REG, ch_idx); |
| usb_dmas_writel(udc, 0, ENETDMAS_SRAM2_REG, ch_idx); |
| } |
| |
| /** |
| * iudma_init_channel - One-time IUDMA channel initialization. |
| * @udc: Reference to the device controller. |
| * @ch_idx: Channel to initialize. |
| */ |
| static int iudma_init_channel(struct bcm63xx_udc *udc, unsigned int ch_idx) |
| { |
| struct iudma_ch *iudma = &udc->iudma[ch_idx]; |
| const struct iudma_ch_cfg *cfg = &iudma_defaults[ch_idx]; |
| unsigned int n_bds = cfg->n_bds; |
| struct bcm63xx_ep *bep = NULL; |
| |
| iudma->ep_num = cfg->ep_num; |
| iudma->ch_idx = ch_idx; |
| iudma->is_tx = !!(ch_idx & 0x01); |
| if (iudma->ep_num >= 0) { |
| bep = &udc->bep[iudma->ep_num]; |
| bep->iudma = iudma; |
| INIT_LIST_HEAD(&bep->queue); |
| } |
| |
| iudma->bep = bep; |
| iudma->udc = udc; |
| |
| /* ep0 is always active; others are controlled by the gadget driver */ |
| if (iudma->ep_num <= 0) |
| iudma->enabled = true; |
| |
| iudma->n_bds = n_bds; |
| iudma->bd_ring = dmam_alloc_coherent(udc->dev, |
| n_bds * sizeof(struct bcm_enet_desc), |
| &iudma->bd_ring_dma, GFP_KERNEL); |
| if (!iudma->bd_ring) |
| return -ENOMEM; |
| iudma->end_bd = &iudma->bd_ring[n_bds - 1]; |
| |
| return 0; |
| } |
| |
| /** |
| * iudma_init - One-time initialization of all IUDMA channels. |
| * @udc: Reference to the device controller. |
| * |
| * Enable DMA, flush channels, and enable global IUDMA IRQs. |
| */ |
| static int iudma_init(struct bcm63xx_udc *udc) |
| { |
| int i, rc; |
| |
| usb_dma_writel(udc, ENETDMA_CFG_EN_MASK, ENETDMA_CFG_REG); |
| |
| for (i = 0; i < BCM63XX_NUM_IUDMA; i++) { |
| rc = iudma_init_channel(udc, i); |
| if (rc) |
| return rc; |
| iudma_reset_channel(udc, &udc->iudma[i]); |
| } |
| |
| usb_dma_writel(udc, BIT(BCM63XX_NUM_IUDMA)-1, ENETDMA_GLB_IRQMASK_REG); |
| return 0; |
| } |
| |
| /** |
| * iudma_uninit - Uninitialize IUDMA channels. |
| * @udc: Reference to the device controller. |
| * |
| * Kill global IUDMA IRQs, flush channels, and kill DMA. |
| */ |
| static void iudma_uninit(struct bcm63xx_udc *udc) |
| { |
| int i; |
| |
| usb_dma_writel(udc, 0, ENETDMA_GLB_IRQMASK_REG); |
| |
| for (i = 0; i < BCM63XX_NUM_IUDMA; i++) |
| iudma_reset_channel(udc, &udc->iudma[i]); |
| |
| usb_dma_writel(udc, 0, ENETDMA_CFG_REG); |
| } |
| |
| /*********************************************************************** |
| * Other low-level USBD operations |
| ***********************************************************************/ |
| |
| /** |
| * bcm63xx_set_ctrl_irqs - Mask/unmask control path interrupts. |
| * @udc: Reference to the device controller. |
| * @enable_irqs: true to enable, false to disable. |
| */ |
| static void bcm63xx_set_ctrl_irqs(struct bcm63xx_udc *udc, bool enable_irqs) |
| { |
| u32 val; |
| |
| usbd_writel(udc, 0, USBD_STATUS_REG); |
| |
| val = BIT(USBD_EVENT_IRQ_USB_RESET) | |
| BIT(USBD_EVENT_IRQ_SETUP) | |
| BIT(USBD_EVENT_IRQ_SETCFG) | |
| BIT(USBD_EVENT_IRQ_SETINTF) | |
| BIT(USBD_EVENT_IRQ_USB_LINK); |
| usbd_writel(udc, enable_irqs ? val : 0, USBD_EVENT_IRQ_MASK_REG); |
| usbd_writel(udc, val, USBD_EVENT_IRQ_STATUS_REG); |
| } |
| |
| /** |
| * bcm63xx_select_phy_mode - Select between USB device and host mode. |
| * @udc: Reference to the device controller. |
| * @is_device: true for device, false for host. |
| * |
| * This should probably be reworked to use the drivers/usb/otg |
| * infrastructure. |
| * |
| * By default, the AFE/pullups are disabled in device mode, until |
| * bcm63xx_select_pullup() is called. |
| */ |
| static void bcm63xx_select_phy_mode(struct bcm63xx_udc *udc, bool is_device) |
| { |
| u32 val, portmask = BIT(udc->pd->port_no); |
| |
| if (BCMCPU_IS_6328()) { |
| /* configure pinmux to sense VBUS signal */ |
| val = bcm_gpio_readl(GPIO_PINMUX_OTHR_REG); |
| val &= ~GPIO_PINMUX_OTHR_6328_USB_MASK; |
| val |= is_device ? GPIO_PINMUX_OTHR_6328_USB_DEV : |
| GPIO_PINMUX_OTHR_6328_USB_HOST; |
| bcm_gpio_writel(val, GPIO_PINMUX_OTHR_REG); |
| } |
| |
| val = bcm_rset_readl(RSET_USBH_PRIV, USBH_PRIV_UTMI_CTL_6368_REG); |
| if (is_device) { |
| val |= (portmask << USBH_PRIV_UTMI_CTL_HOSTB_SHIFT); |
| val |= (portmask << USBH_PRIV_UTMI_CTL_NODRIV_SHIFT); |
| } else { |
| val &= ~(portmask << USBH_PRIV_UTMI_CTL_HOSTB_SHIFT); |
| val &= ~(portmask << USBH_PRIV_UTMI_CTL_NODRIV_SHIFT); |
| } |
| bcm_rset_writel(RSET_USBH_PRIV, val, USBH_PRIV_UTMI_CTL_6368_REG); |
| |
| val = bcm_rset_readl(RSET_USBH_PRIV, USBH_PRIV_SWAP_6368_REG); |
| if (is_device) |
| val |= USBH_PRIV_SWAP_USBD_MASK; |
| else |
| val &= ~USBH_PRIV_SWAP_USBD_MASK; |
| bcm_rset_writel(RSET_USBH_PRIV, val, USBH_PRIV_SWAP_6368_REG); |
| } |
| |
| /** |
| * bcm63xx_select_pullup - Enable/disable the pullup on D+ |
| * @udc: Reference to the device controller. |
| * @is_on: true to enable the pullup, false to disable. |
| * |
| * If the pullup is active, the host will sense a FS/HS device connected to |
| * the port. If the pullup is inactive, the host will think the USB |
| * device has been disconnected. |
| */ |
| static void bcm63xx_select_pullup(struct bcm63xx_udc *udc, bool is_on) |
| { |
| u32 val, portmask = BIT(udc->pd->port_no); |
| |
| val = bcm_rset_readl(RSET_USBH_PRIV, USBH_PRIV_UTMI_CTL_6368_REG); |
| if (is_on) |
| val &= ~(portmask << USBH_PRIV_UTMI_CTL_NODRIV_SHIFT); |
| else |
| val |= (portmask << USBH_PRIV_UTMI_CTL_NODRIV_SHIFT); |
| bcm_rset_writel(RSET_USBH_PRIV, val, USBH_PRIV_UTMI_CTL_6368_REG); |
| } |
| |
| /** |
| * bcm63xx_uninit_udc_hw - Shut down the hardware prior to driver removal. |
| * @udc: Reference to the device controller. |
| * |
| * This just masks the IUDMA IRQs and releases the clocks. It is assumed |
| * that bcm63xx_udc_stop() has already run, and the clocks are stopped. |
| */ |
| static void bcm63xx_uninit_udc_hw(struct bcm63xx_udc *udc) |
| { |
| set_clocks(udc, true); |
| iudma_uninit(udc); |
| set_clocks(udc, false); |
| |
| clk_put(udc->usbd_clk); |
| clk_put(udc->usbh_clk); |
| } |
| |
| /** |
| * bcm63xx_init_udc_hw - Initialize the controller hardware and data structures. |
| * @udc: Reference to the device controller. |
| */ |
| static int bcm63xx_init_udc_hw(struct bcm63xx_udc *udc) |
| { |
| int i, rc = 0; |
| u32 val; |
| |
| udc->ep0_ctrl_buf = devm_kzalloc(udc->dev, BCM63XX_MAX_CTRL_PKT, |
| GFP_KERNEL); |
| if (!udc->ep0_ctrl_buf) |
| return -ENOMEM; |
| |
| INIT_LIST_HEAD(&udc->gadget.ep_list); |
| for (i = 0; i < BCM63XX_NUM_EP; i++) { |
| struct bcm63xx_ep *bep = &udc->bep[i]; |
| |
| bep->ep.name = bcm63xx_ep_info[i].name; |
| bep->ep.caps = bcm63xx_ep_info[i].caps; |
| bep->ep_num = i; |
| bep->ep.ops = &bcm63xx_udc_ep_ops; |
| list_add_tail(&bep->ep.ep_list, &udc->gadget.ep_list); |
| bep->halted = 0; |
| usb_ep_set_maxpacket_limit(&bep->ep, BCM63XX_MAX_CTRL_PKT); |
| bep->udc = udc; |
| bep->ep.desc = NULL; |
| INIT_LIST_HEAD(&bep->queue); |
| } |
| |
| udc->gadget.ep0 = &udc->bep[0].ep; |
| list_del(&udc->bep[0].ep.ep_list); |
| |
| udc->gadget.speed = USB_SPEED_UNKNOWN; |
| udc->ep0state = EP0_SHUTDOWN; |
| |
| udc->usbh_clk = clk_get(udc->dev, "usbh"); |
| if (IS_ERR(udc->usbh_clk)) |
| return -EIO; |
| |
| udc->usbd_clk = clk_get(udc->dev, "usbd"); |
| if (IS_ERR(udc->usbd_clk)) { |
| clk_put(udc->usbh_clk); |
| return -EIO; |
| } |
| |
| set_clocks(udc, true); |
| |
| val = USBD_CONTROL_AUTO_CSRS_MASK | |
| USBD_CONTROL_DONE_CSRS_MASK | |
| (irq_coalesce ? USBD_CONTROL_RXZSCFG_MASK : 0); |
| usbd_writel(udc, val, USBD_CONTROL_REG); |
| |
| val = USBD_STRAPS_APP_SELF_PWR_MASK | |
| USBD_STRAPS_APP_RAM_IF_MASK | |
| USBD_STRAPS_APP_CSRPRGSUP_MASK | |
| USBD_STRAPS_APP_8BITPHY_MASK | |
| USBD_STRAPS_APP_RMTWKUP_MASK; |
| |
| if (udc->gadget.max_speed == USB_SPEED_HIGH) |
| val |= (BCM63XX_SPD_HIGH << USBD_STRAPS_SPEED_SHIFT); |
| else |
| val |= (BCM63XX_SPD_FULL << USBD_STRAPS_SPEED_SHIFT); |
| usbd_writel(udc, val, USBD_STRAPS_REG); |
| |
| bcm63xx_set_ctrl_irqs(udc, false); |
| |
| usbd_writel(udc, 0, USBD_EVENT_IRQ_CFG_LO_REG); |
| |
| val = USBD_EVENT_IRQ_CFG_FALLING(USBD_EVENT_IRQ_ENUM_ON) | |
| USBD_EVENT_IRQ_CFG_FALLING(USBD_EVENT_IRQ_SET_CSRS); |
| usbd_writel(udc, val, USBD_EVENT_IRQ_CFG_HI_REG); |
| |
| rc = iudma_init(udc); |
| set_clocks(udc, false); |
| if (rc) |
| bcm63xx_uninit_udc_hw(udc); |
| |
| return 0; |
| } |
| |
| /*********************************************************************** |
| * Standard EP gadget operations |
| ***********************************************************************/ |
| |
| /** |
| * bcm63xx_ep_enable - Enable one endpoint. |
| * @ep: Endpoint to enable. |
| * @desc: Contains max packet, direction, etc. |
| * |
| * Most of the endpoint parameters are fixed in this controller, so there |
| * isn't much for this function to do. |
| */ |
| static int bcm63xx_ep_enable(struct usb_ep *ep, |
| const struct usb_endpoint_descriptor *desc) |
| { |
| struct bcm63xx_ep *bep = our_ep(ep); |
| struct bcm63xx_udc *udc = bep->udc; |
| struct iudma_ch *iudma = bep->iudma; |
| unsigned long flags; |
| |
| if (!ep || !desc || ep->name == bcm63xx_ep0name) |
| return -EINVAL; |
| |
| if (!udc->driver) |
| return -ESHUTDOWN; |
| |
| spin_lock_irqsave(&udc->lock, flags); |
| if (iudma->enabled) { |
| spin_unlock_irqrestore(&udc->lock, flags); |
| return -EINVAL; |
| } |
| |
| iudma->enabled = true; |
| BUG_ON(!list_empty(&bep->queue)); |
| |
| iudma_reset_channel(udc, iudma); |
| |
| bep->halted = 0; |
| bcm63xx_set_stall(udc, bep, false); |
| clear_bit(bep->ep_num, &udc->wedgemap); |
| |
| ep->desc = desc; |
| ep->maxpacket = usb_endpoint_maxp(desc); |
| |
| spin_unlock_irqrestore(&udc->lock, flags); |
| return 0; |
| } |
| |
| /** |
| * bcm63xx_ep_disable - Disable one endpoint. |
| * @ep: Endpoint to disable. |
| */ |
| static int bcm63xx_ep_disable(struct usb_ep *ep) |
| { |
| struct bcm63xx_ep *bep = our_ep(ep); |
| struct bcm63xx_udc *udc = bep->udc; |
| struct iudma_ch *iudma = bep->iudma; |
| struct bcm63xx_req *breq, *n; |
| unsigned long flags; |
| |
| if (!ep || !ep->desc) |
| return -EINVAL; |
| |
| spin_lock_irqsave(&udc->lock, flags); |
| if (!iudma->enabled) { |
| spin_unlock_irqrestore(&udc->lock, flags); |
| return -EINVAL; |
| } |
| iudma->enabled = false; |
| |
| iudma_reset_channel(udc, iudma); |
| |
| if (!list_empty(&bep->queue)) { |
| list_for_each_entry_safe(breq, n, &bep->queue, queue) { |
| usb_gadget_unmap_request(&udc->gadget, &breq->req, |
| iudma->is_tx); |
| list_del(&breq->queue); |
| breq->req.status = -ESHUTDOWN; |
| |
| spin_unlock_irqrestore(&udc->lock, flags); |
| usb_gadget_giveback_request(&iudma->bep->ep, &breq->req); |
| spin_lock_irqsave(&udc->lock, flags); |
| } |
| } |
| ep->desc = NULL; |
| |
| spin_unlock_irqrestore(&udc->lock, flags); |
| return 0; |
| } |
| |
| /** |
| * bcm63xx_udc_alloc_request - Allocate a new request. |
| * @ep: Endpoint associated with the request. |
| * @mem_flags: Flags to pass to kzalloc(). |
| */ |
| static struct usb_request *bcm63xx_udc_alloc_request(struct usb_ep *ep, |
| gfp_t mem_flags) |
| { |
| struct bcm63xx_req *breq; |
| |
| breq = kzalloc(sizeof(*breq), mem_flags); |
| if (!breq) |
| return NULL; |
| return &breq->req; |
| } |
| |
| /** |
| * bcm63xx_udc_free_request - Free a request. |
| * @ep: Endpoint associated with the request. |
| * @req: Request to free. |
| */ |
| static void bcm63xx_udc_free_request(struct usb_ep *ep, |
| struct usb_request *req) |
| { |
| struct bcm63xx_req *breq = our_req(req); |
| kfree(breq); |
| } |
| |
| /** |
| * bcm63xx_udc_queue - Queue up a new request. |
| * @ep: Endpoint associated with the request. |
| * @req: Request to add. |
| * @mem_flags: Unused. |
| * |
| * If the queue is empty, start this request immediately. Otherwise, add |
| * it to the list. |
| * |
| * ep0 replies are sent through this function from the gadget driver, but |
| * they are treated differently because they need to be handled by the ep0 |
| * state machine. (Sometimes they are replies to control requests that |
| * were spoofed by this driver, and so they shouldn't be transmitted at all.) |
| */ |
| static int bcm63xx_udc_queue(struct usb_ep *ep, struct usb_request *req, |
| gfp_t mem_flags) |
| { |
| struct bcm63xx_ep *bep = our_ep(ep); |
| struct bcm63xx_udc *udc = bep->udc; |
| struct bcm63xx_req *breq = our_req(req); |
| unsigned long flags; |
| int rc = 0; |
| |
| if (unlikely(!req || !req->complete || !req->buf || !ep)) |
| return -EINVAL; |
| |
| req->actual = 0; |
| req->status = 0; |
| breq->offset = 0; |
| |
| if (bep == &udc->bep[0]) { |
| /* only one reply per request, please */ |
| if (udc->ep0_reply) |
| return -EINVAL; |
| |
| udc->ep0_reply = req; |
| schedule_work(&udc->ep0_wq); |
| return 0; |
| } |
| |
| spin_lock_irqsave(&udc->lock, flags); |
| if (!bep->iudma->enabled) { |
| rc = -ESHUTDOWN; |
| goto out; |
| } |
| |
| rc = usb_gadget_map_request(&udc->gadget, req, bep->iudma->is_tx); |
| if (rc == 0) { |
| list_add_tail(&breq->queue, &bep->queue); |
| if (list_is_singular(&bep->queue)) |
| iudma_write(udc, bep->iudma, breq); |
| } |
| |
| out: |
| spin_unlock_irqrestore(&udc->lock, flags); |
| return rc; |
| } |
| |
| /** |
| * bcm63xx_udc_dequeue - Remove a pending request from the queue. |
| * @ep: Endpoint associated with the request. |
| * @req: Request to remove. |
| * |
| * If the request is not at the head of the queue, this is easy - just nuke |
| * it. If the request is at the head of the queue, we'll need to stop the |
| * DMA transaction and then queue up the successor. |
| */ |
| static int bcm63xx_udc_dequeue(struct usb_ep *ep, struct usb_request *req) |
| { |
| struct bcm63xx_ep *bep = our_ep(ep); |
| struct bcm63xx_udc *udc = bep->udc; |
| struct bcm63xx_req *breq = our_req(req), *cur; |
| unsigned long flags; |
| int rc = 0; |
| |
| spin_lock_irqsave(&udc->lock, flags); |
| if (list_empty(&bep->queue)) { |
| rc = -EINVAL; |
| goto out; |
| } |
| |
| cur = list_first_entry(&bep->queue, struct bcm63xx_req, queue); |
| usb_gadget_unmap_request(&udc->gadget, &breq->req, bep->iudma->is_tx); |
| |
| if (breq == cur) { |
| iudma_reset_channel(udc, bep->iudma); |
| list_del(&breq->queue); |
| |
| if (!list_empty(&bep->queue)) { |
| struct bcm63xx_req *next; |
| |
| next = list_first_entry(&bep->queue, |
| struct bcm63xx_req, queue); |
| iudma_write(udc, bep->iudma, next); |
| } |
| } else { |
| list_del(&breq->queue); |
| } |
| |
| out: |
| spin_unlock_irqrestore(&udc->lock, flags); |
| |
| req->status = -ESHUTDOWN; |
| req->complete(ep, req); |
| |
| return rc; |
| } |
| |
| /** |
| * bcm63xx_udc_set_halt - Enable/disable STALL flag in the hardware. |
| * @ep: Endpoint to halt. |
| * @value: Zero to clear halt; nonzero to set halt. |
| * |
| * See comments in bcm63xx_update_wedge(). |
| */ |
| static int bcm63xx_udc_set_halt(struct usb_ep *ep, int value) |
| { |
| struct bcm63xx_ep *bep = our_ep(ep); |
| struct bcm63xx_udc *udc = bep->udc; |
| unsigned long flags; |
| |
| spin_lock_irqsave(&udc->lock, flags); |
| bcm63xx_set_stall(udc, bep, !!value); |
| bep->halted = value; |
| spin_unlock_irqrestore(&udc->lock, flags); |
| |
| return 0; |
| } |
| |
| /** |
| * bcm63xx_udc_set_wedge - Stall the endpoint until the next reset. |
| * @ep: Endpoint to wedge. |
| * |
| * See comments in bcm63xx_update_wedge(). |
| */ |
| static int bcm63xx_udc_set_wedge(struct usb_ep *ep) |
| { |
| struct bcm63xx_ep *bep = our_ep(ep); |
| struct bcm63xx_udc *udc = bep->udc; |
| unsigned long flags; |
| |
| spin_lock_irqsave(&udc->lock, flags); |
| set_bit(bep->ep_num, &udc->wedgemap); |
| bcm63xx_set_stall(udc, bep, true); |
| spin_unlock_irqrestore(&udc->lock, flags); |
| |
| return 0; |
| } |
| |
| static const struct usb_ep_ops bcm63xx_udc_ep_ops = { |
| .enable = bcm63xx_ep_enable, |
| .disable = bcm63xx_ep_disable, |
| |
| .alloc_request = bcm63xx_udc_alloc_request, |
| .free_request = bcm63xx_udc_free_request, |
| |
| .queue = bcm63xx_udc_queue, |
| .dequeue = bcm63xx_udc_dequeue, |
| |
| .set_halt = bcm63xx_udc_set_halt, |
| .set_wedge = bcm63xx_udc_set_wedge, |
| }; |
| |
| /*********************************************************************** |
| * EP0 handling |
| ***********************************************************************/ |
| |
| /** |
| * bcm63xx_ep0_setup_callback - Drop spinlock to invoke ->setup callback. |
| * @udc: Reference to the device controller. |
| * @ctrl: 8-byte SETUP request. |
| */ |
| static int bcm63xx_ep0_setup_callback(struct bcm63xx_udc *udc, |
| struct usb_ctrlrequest *ctrl) |
| { |
| int rc; |
| |
| spin_unlock_irq(&udc->lock); |
| rc = udc->driver->setup(&udc->gadget, ctrl); |
| spin_lock_irq(&udc->lock); |
| return rc; |
| } |
| |
| /** |
| * bcm63xx_ep0_spoof_set_cfg - Synthesize a SET_CONFIGURATION request. |
| * @udc: Reference to the device controller. |
| * |
| * Many standard requests are handled automatically in the hardware, but |
| * we still need to pass them to the gadget driver so that it can |
| * reconfigure the interfaces/endpoints if necessary. |
| * |
| * Unfortunately we are not able to send a STALL response if the host |
| * requests an invalid configuration. If this happens, we'll have to be |
| * content with printing a warning. |
| */ |
| static int bcm63xx_ep0_spoof_set_cfg(struct bcm63xx_udc *udc) |
| { |
| struct usb_ctrlrequest ctrl; |
| int rc; |
| |
| ctrl.bRequestType = USB_DIR_OUT | USB_RECIP_DEVICE; |
| ctrl.bRequest = USB_REQ_SET_CONFIGURATION; |
| ctrl.wValue = cpu_to_le16(udc->cfg); |
| ctrl.wIndex = 0; |
| ctrl.wLength = 0; |
| |
| rc = bcm63xx_ep0_setup_callback(udc, &ctrl); |
| if (rc < 0) { |
| dev_warn_ratelimited(udc->dev, |
| "hardware auto-acked bad SET_CONFIGURATION(%d) request\n", |
| udc->cfg); |
| } |
| return rc; |
| } |
| |
| /** |
| * bcm63xx_ep0_spoof_set_iface - Synthesize a SET_INTERFACE request. |
| * @udc: Reference to the device controller. |
| */ |
| static int bcm63xx_ep0_spoof_set_iface(struct bcm63xx_udc *udc) |
| { |
| struct usb_ctrlrequest ctrl; |
| int rc; |
| |
| ctrl.bRequestType = USB_DIR_OUT | USB_RECIP_INTERFACE; |
| ctrl.bRequest = USB_REQ_SET_INTERFACE; |
| ctrl.wValue = cpu_to_le16(udc->alt_iface); |
| ctrl.wIndex = cpu_to_le16(udc->iface); |
| ctrl.wLength = 0; |
| |
| rc = bcm63xx_ep0_setup_callback(udc, &ctrl); |
| if (rc < 0) { |
| dev_warn_ratelimited(udc->dev, |
| "hardware auto-acked bad SET_INTERFACE(%d,%d) request\n", |
| udc->iface, udc->alt_iface); |
| } |
| return rc; |
| } |
| |
| /** |
| * bcm63xx_ep0_map_write - dma_map and iudma_write a single request. |
| * @udc: Reference to the device controller. |
| * @ch_idx: IUDMA channel number. |
| * @req: USB gadget layer representation of the request. |
| */ |
| static void bcm63xx_ep0_map_write(struct bcm63xx_udc *udc, int ch_idx, |
| struct usb_request *req) |
| { |
| struct bcm63xx_req *breq = our_req(req); |
| struct iudma_ch *iudma = &udc->iudma[ch_idx]; |
| |
| BUG_ON(udc->ep0_request); |
| udc->ep0_request = req; |
| |
| req->actual = 0; |
| breq->offset = 0; |
| usb_gadget_map_request(&udc->gadget, req, iudma->is_tx); |
| iudma_write(udc, iudma, breq); |
| } |
| |
| /** |
| * bcm63xx_ep0_complete - Set completion status and "stage" the callback. |
| * @udc: Reference to the device controller. |
| * @req: USB gadget layer representation of the request. |
| * @status: Status to return to the gadget driver. |
| */ |
| static void bcm63xx_ep0_complete(struct bcm63xx_udc *udc, |
| struct usb_request *req, int status) |
| { |
| req->status = status; |
| if (status) |
| req->actual = 0; |
| if (req->complete) { |
| spin_unlock_irq(&udc->lock); |
| req->complete(&udc->bep[0].ep, req); |
| spin_lock_irq(&udc->lock); |
| } |
| } |
| |
| /** |
| * bcm63xx_ep0_nuke_reply - Abort request from the gadget driver due to |
| * reset/shutdown. |
| * @udc: Reference to the device controller. |
| * @is_tx: Nonzero for TX (IN), zero for RX (OUT). |
| */ |
| static void bcm63xx_ep0_nuke_reply(struct bcm63xx_udc *udc, int is_tx) |
| { |
| struct usb_request *req = udc->ep0_reply; |
| |
| udc->ep0_reply = NULL; |
| usb_gadget_unmap_request(&udc->gadget, req, is_tx); |
| if (udc->ep0_request == req) { |
| udc->ep0_req_completed = 0; |
| udc->ep0_request = NULL; |
| } |
| bcm63xx_ep0_complete(udc, req, -ESHUTDOWN); |
| } |
| |
| /** |
| * bcm63xx_ep0_read_complete - Close out the pending ep0 request; return |
| * transfer len. |
| * @udc: Reference to the device controller. |
| */ |
| static int bcm63xx_ep0_read_complete(struct bcm63xx_udc *udc) |
| { |
| struct usb_request *req = udc->ep0_request; |
| |
| udc->ep0_req_completed = 0; |
| udc->ep0_request = NULL; |
| |
| return req->actual; |
| } |
| |
| /** |
| * bcm63xx_ep0_internal_request - Helper function to submit an ep0 request. |
| * @udc: Reference to the device controller. |
| * @ch_idx: IUDMA channel number. |
| * @length: Number of bytes to TX/RX. |
| * |
| * Used for simple transfers performed by the ep0 worker. This will always |
| * use ep0_ctrl_req / ep0_ctrl_buf. |
| */ |
| static void bcm63xx_ep0_internal_request(struct bcm63xx_udc *udc, int ch_idx, |
| int length) |
| { |
| struct usb_request *req = &udc->ep0_ctrl_req.req; |
| |
| req->buf = udc->ep0_ctrl_buf; |
| req->length = length; |
| req->complete = NULL; |
| |
| bcm63xx_ep0_map_write(udc, ch_idx, req); |
| } |
| |
| /** |
| * bcm63xx_ep0_do_setup - Parse new SETUP packet and decide how to handle it. |
| * @udc: Reference to the device controller. |
| * |
| * EP0_IDLE probably shouldn't ever happen. EP0_REQUEUE means we're ready |
| * for the next packet. Anything else means the transaction requires multiple |
| * stages of handling. |
| */ |
| static enum bcm63xx_ep0_state bcm63xx_ep0_do_setup(struct bcm63xx_udc *udc) |
| { |
| int rc; |
| struct usb_ctrlrequest *ctrl = (void *)udc->ep0_ctrl_buf; |
| |
| rc = bcm63xx_ep0_read_complete(udc); |
| |
| if (rc < 0) { |
| dev_err(udc->dev, "missing SETUP packet\n"); |
| return EP0_IDLE; |
| } |
| |
| /* |
| * Handle 0-byte IN STATUS acknowledgement. The hardware doesn't |
| * ALWAYS deliver these 100% of the time, so if we happen to see one, |
| * just throw it away. |
| */ |
| if (rc == 0) |
| return EP0_REQUEUE; |
| |
| /* Drop malformed SETUP packets */ |
| if (rc != sizeof(*ctrl)) { |
| dev_warn_ratelimited(udc->dev, |
| "malformed SETUP packet (%d bytes)\n", rc); |
| return EP0_REQUEUE; |
| } |
| |
| /* Process new SETUP packet arriving on ep0 */ |
| rc = bcm63xx_ep0_setup_callback(udc, ctrl); |
| if (rc < 0) { |
| bcm63xx_set_stall(udc, &udc->bep[0], true); |
| return EP0_REQUEUE; |
| } |
| |
| if (!ctrl->wLength) |
| return EP0_REQUEUE; |
| else if (ctrl->bRequestType & USB_DIR_IN) |
| return EP0_IN_DATA_PHASE_SETUP; |
| else |
| return EP0_OUT_DATA_PHASE_SETUP; |
| } |
| |
| /** |
| * bcm63xx_ep0_do_idle - Check for outstanding requests if ep0 is idle. |
| * @udc: Reference to the device controller. |
| * |
| * In state EP0_IDLE, the RX descriptor is either pending, or has been |
| * filled with a SETUP packet from the host. This function handles new |
| * SETUP packets, control IRQ events (which can generate fake SETUP packets), |
| * and reset/shutdown events. |
| * |
| * Returns 0 if work was done; -EAGAIN if nothing to do. |
| */ |
| static int bcm63xx_ep0_do_idle(struct bcm63xx_udc *udc) |
| { |
| if (udc->ep0_req_reset) { |
| udc->ep0_req_reset = 0; |
| } else if (udc->ep0_req_set_cfg) { |
| udc->ep0_req_set_cfg = 0; |
| if (bcm63xx_ep0_spoof_set_cfg(udc) >= 0) |
| udc->ep0state = EP0_IN_FAKE_STATUS_PHASE; |
| } else if (udc->ep0_req_set_iface) { |
| udc->ep0_req_set_iface = 0; |
| if (bcm63xx_ep0_spoof_set_iface(udc) >= 0) |
| udc->ep0state = EP0_IN_FAKE_STATUS_PHASE; |
| } else if (udc->ep0_req_completed) { |
| udc->ep0state = bcm63xx_ep0_do_setup(udc); |
| return udc->ep0state == EP0_IDLE ? -EAGAIN : 0; |
| } else if (udc->ep0_req_shutdown) { |
| udc->ep0_req_shutdown = 0; |
| udc->ep0_req_completed = 0; |
| udc->ep0_request = NULL; |
| iudma_reset_channel(udc, &udc->iudma[IUDMA_EP0_RXCHAN]); |
| usb_gadget_unmap_request(&udc->gadget, |
| &udc->ep0_ctrl_req.req, 0); |
| |
| /* bcm63xx_udc_pullup() is waiting for this */ |
| mb(); |
| udc->ep0state = EP0_SHUTDOWN; |
| } else if (udc->ep0_reply) { |
| /* |
| * This could happen if a USB RESET shows up during an ep0 |
| * transaction (especially if a laggy driver like gadgetfs |
| * is in use). |
| */ |
| dev_warn(udc->dev, "nuking unexpected reply\n"); |
| bcm63xx_ep0_nuke_reply(udc, 0); |
| } else { |
| return -EAGAIN; |
| } |
| |
| return 0; |
| } |
| |
| /** |
| * bcm63xx_ep0_one_round - Handle the current ep0 state. |
| * @udc: Reference to the device controller. |
| * |
| * Returns 0 if work was done; -EAGAIN if nothing to do. |
| */ |
| static int bcm63xx_ep0_one_round(struct bcm63xx_udc *udc) |
| { |
| enum bcm63xx_ep0_state ep0state = udc->ep0state; |
| bool shutdown = udc->ep0_req_reset || udc->ep0_req_shutdown; |
| |
| switch (udc->ep0state) { |
| case EP0_REQUEUE: |
| /* set up descriptor to receive SETUP packet */ |
| bcm63xx_ep0_internal_request(udc, IUDMA_EP0_RXCHAN, |
| BCM63XX_MAX_CTRL_PKT); |
| ep0state = EP0_IDLE; |
| break; |
| case EP0_IDLE: |
| return bcm63xx_ep0_do_idle(udc); |
| case EP0_IN_DATA_PHASE_SETUP: |
| /* |
| * Normal case: TX request is in ep0_reply (queued by the |
| * callback), or will be queued shortly. When it's here, |
| * send it to the HW and go to EP0_IN_DATA_PHASE_COMPLETE. |
| * |
| * Shutdown case: Stop waiting for the reply. Just |
| * REQUEUE->IDLE. The gadget driver is NOT expected to |
| * queue anything else now. |
| */ |
| if (udc->ep0_reply) { |
| bcm63xx_ep0_map_write(udc, IUDMA_EP0_TXCHAN, |
| udc->ep0_reply); |
| ep0state = EP0_IN_DATA_PHASE_COMPLETE; |
| } else if (shutdown) { |
| ep0state = EP0_REQUEUE; |
| } |
| break; |
| case EP0_IN_DATA_PHASE_COMPLETE: { |
| /* |
| * Normal case: TX packet (ep0_reply) is in flight; wait for |
| * it to finish, then go back to REQUEUE->IDLE. |
| * |
| * Shutdown case: Reset the TX channel, send -ESHUTDOWN |
| * completion to the gadget driver, then REQUEUE->IDLE. |
| */ |
| if (udc->ep0_req_completed) { |
| udc->ep0_reply = NULL; |
| bcm63xx_ep0_read_complete(udc); |
| /* |
| * the "ack" sometimes gets eaten (see |
| * bcm63xx_ep0_do_idle) |
| */ |
| ep0state = EP0_REQUEUE; |
| } else if (shutdown) { |
| iudma_reset_channel(udc, &udc->iudma[IUDMA_EP0_TXCHAN]); |
| bcm63xx_ep0_nuke_reply(udc, 1); |
| ep0state = EP0_REQUEUE; |
| } |
| break; |
| } |
| case EP0_OUT_DATA_PHASE_SETUP: |
| /* Similar behavior to EP0_IN_DATA_PHASE_SETUP */ |
| if (udc->ep0_reply) { |
| bcm63xx_ep0_map_write(udc, IUDMA_EP0_RXCHAN, |
| udc->ep0_reply); |
| ep0state = EP0_OUT_DATA_PHASE_COMPLETE; |
| } else if (shutdown) { |
| ep0state = EP0_REQUEUE; |
| } |
| break; |
| case EP0_OUT_DATA_PHASE_COMPLETE: { |
| /* Similar behavior to EP0_IN_DATA_PHASE_COMPLETE */ |
| if (udc->ep0_req_completed) { |
| udc->ep0_reply = NULL; |
| bcm63xx_ep0_read_complete(udc); |
| |
| /* send 0-byte ack to host */ |
| bcm63xx_ep0_internal_request(udc, IUDMA_EP0_TXCHAN, 0); |
| ep0state = EP0_OUT_STATUS_PHASE; |
| } else if (shutdown) { |
| iudma_reset_channel(udc, &udc->iudma[IUDMA_EP0_RXCHAN]); |
| bcm63xx_ep0_nuke_reply(udc, 0); |
| ep0state = EP0_REQUEUE; |
| } |
| break; |
| } |
| case EP0_OUT_STATUS_PHASE: |
| /* |
| * Normal case: 0-byte OUT ack packet is in flight; wait |
| * for it to finish, then go back to REQUEUE->IDLE. |
| * |
| * Shutdown case: just cancel the transmission. Don't bother |
| * calling the completion, because it originated from this |
| * function anyway. Then go back to REQUEUE->IDLE. |
| */ |
| if (udc->ep0_req_completed) { |
| bcm63xx_ep0_read_complete(udc); |
| ep0state = EP0_REQUEUE; |
| } else if (shutdown) { |
| iudma_reset_channel(udc, &udc->iudma[IUDMA_EP0_TXCHAN]); |
| udc->ep0_request = NULL; |
| ep0state = EP0_REQUEUE; |
| } |
| break; |
| case EP0_IN_FAKE_STATUS_PHASE: { |
| /* |
| * Normal case: we spoofed a SETUP packet and are now |
| * waiting for the gadget driver to send a 0-byte reply. |
| * This doesn't actually get sent to the HW because the |
| * HW has already sent its own reply. Once we get the |
| * response, return to IDLE. |
| * |
| * Shutdown case: return to IDLE immediately. |
| * |
| * Note that the ep0 RX descriptor has remained queued |
| * (and possibly unfilled) during this entire transaction. |
| * The HW datapath (IUDMA) never even sees SET_CONFIGURATION |
| * or SET_INTERFACE transactions. |
| */ |
| struct usb_request *r = udc->ep0_reply; |
| |
| if (!r) { |
| if (shutdown) |
| ep0state = EP0_IDLE; |
| break; |
| } |
| |
| bcm63xx_ep0_complete(udc, r, 0); |
| udc->ep0_reply = NULL; |
| ep0state = EP0_IDLE; |
| break; |
| } |
| case EP0_SHUTDOWN: |
| break; |
| } |
| |
| if (udc->ep0state == ep0state) |
| return -EAGAIN; |
| |
| udc->ep0state = ep0state; |
| return 0; |
| } |
| |
| /** |
| * bcm63xx_ep0_process - ep0 worker thread / state machine. |
| * @w: Workqueue struct. |
| * |
| * bcm63xx_ep0_process is triggered any time an event occurs on ep0. It |
| * is used to synchronize ep0 events and ensure that both HW and SW events |
| * occur in a well-defined order. When the ep0 IUDMA queues are idle, it may |
| * synthesize SET_CONFIGURATION / SET_INTERFACE requests that were consumed |
| * by the USBD hardware. |
| * |
| * The worker function will continue iterating around the state machine |
| * until there is nothing left to do. Usually "nothing left to do" means |
| * that we're waiting for a new event from the hardware. |
| */ |
| static void bcm63xx_ep0_process(struct work_struct *w) |
| { |
| struct bcm63xx_udc *udc = container_of(w, struct bcm63xx_udc, ep0_wq); |
| spin_lock_irq(&udc->lock); |
| while (bcm63xx_ep0_one_round(udc) == 0) |
| ; |
| spin_unlock_irq(&udc->lock); |
| } |
| |
| /*********************************************************************** |
| * Standard UDC gadget operations |
| ***********************************************************************/ |
| |
| /** |
| * bcm63xx_udc_get_frame - Read current SOF frame number from the HW. |
| * @gadget: USB slave device. |
| */ |
| static int bcm63xx_udc_get_frame(struct usb_gadget *gadget) |
| { |
| struct bcm63xx_udc *udc = gadget_to_udc(gadget); |
| |
| return (usbd_readl(udc, USBD_STATUS_REG) & |
| USBD_STATUS_SOF_MASK) >> USBD_STATUS_SOF_SHIFT; |
| } |
| |
| /** |
| * bcm63xx_udc_pullup - Enable/disable pullup on D+ line. |
| * @gadget: USB slave device. |
| * @is_on: 0 to disable pullup, 1 to enable. |
| * |
| * See notes in bcm63xx_select_pullup(). |
| */ |
| static int bcm63xx_udc_pullup(struct usb_gadget *gadget, int is_on) |
| { |
| struct bcm63xx_udc *udc = gadget_to_udc(gadget); |
| unsigned long flags; |
| int i, rc = -EINVAL; |
| |
| spin_lock_irqsave(&udc->lock, flags); |
| if (is_on && udc->ep0state == EP0_SHUTDOWN) { |
| udc->gadget.speed = USB_SPEED_UNKNOWN; |
| udc->ep0state = EP0_REQUEUE; |
| bcm63xx_fifo_setup(udc); |
| bcm63xx_fifo_reset(udc); |
| bcm63xx_ep_setup(udc); |
| |
| bitmap_zero(&udc->wedgemap, BCM63XX_NUM_EP); |
| for (i = 0; i < BCM63XX_NUM_EP; i++) |
| bcm63xx_set_stall(udc, &udc->bep[i], false); |
| |
| bcm63xx_set_ctrl_irqs(udc, true); |
| bcm63xx_select_pullup(gadget_to_udc(gadget), true); |
| rc = 0; |
| } else if (!is_on && udc->ep0state != EP0_SHUTDOWN) { |
| bcm63xx_select_pullup(gadget_to_udc(gadget), false); |
| |
| udc->ep0_req_shutdown = 1; |
| spin_unlock_irqrestore(&udc->lock, flags); |
| |
| while (1) { |
| schedule_work(&udc->ep0_wq); |
| if (udc->ep0state == EP0_SHUTDOWN) |
| break; |
| msleep(50); |
| } |
| bcm63xx_set_ctrl_irqs(udc, false); |
| cancel_work_sync(&udc->ep0_wq); |
| return 0; |
| } |
| |
| spin_unlock_irqrestore(&udc->lock, flags); |
| return rc; |
| } |
| |
| /** |
| * bcm63xx_udc_start - Start the controller. |
| * @gadget: USB slave device. |
| * @driver: Driver for USB slave devices. |
| */ |
| static int bcm63xx_udc_start(struct usb_gadget *gadget, |
| struct usb_gadget_driver *driver) |
| { |
| struct bcm63xx_udc *udc = gadget_to_udc(gadget); |
| unsigned long flags; |
| |
| if (!driver || driver->max_speed < USB_SPEED_HIGH || |
| !driver->setup) |
| return -EINVAL; |
| if (!udc) |
| return -ENODEV; |
| if (udc->driver) |
| return -EBUSY; |
| |
| spin_lock_irqsave(&udc->lock, flags); |
| |
| set_clocks(udc, true); |
| bcm63xx_fifo_setup(udc); |
| bcm63xx_ep_init(udc); |
| bcm63xx_ep_setup(udc); |
| bcm63xx_fifo_reset(udc); |
| bcm63xx_select_phy_mode(udc, true); |
| |
| udc->driver = driver; |
| driver->driver.bus = NULL; |
| udc->gadget.dev.of_node = udc->dev->of_node; |
| |
| spin_unlock_irqrestore(&udc->lock, flags); |
| |
| return 0; |
| } |
| |
| /** |
| * bcm63xx_udc_stop - Shut down the controller. |
| * @gadget: USB slave device. |
| * @driver: Driver for USB slave devices. |
| */ |
| static int bcm63xx_udc_stop(struct usb_gadget *gadget) |
| { |
| struct bcm63xx_udc *udc = gadget_to_udc(gadget); |
| unsigned long flags; |
| |
| spin_lock_irqsave(&udc->lock, flags); |
| |
| udc->driver = NULL; |
| |
| /* |
| * If we switch the PHY too abruptly after dropping D+, the host |
| * will often complain: |
| * |
| * hub 1-0:1.0: port 1 disabled by hub (EMI?), re-enabling... |
| */ |
| msleep(100); |
| |
| bcm63xx_select_phy_mode(udc, false); |
| set_clocks(udc, false); |
| |
| spin_unlock_irqrestore(&udc->lock, flags); |
| |
| return 0; |
| } |
| |
| static const struct usb_gadget_ops bcm63xx_udc_ops = { |
| .get_frame = bcm63xx_udc_get_frame, |
| .pullup = bcm63xx_udc_pullup, |
| .udc_start = bcm63xx_udc_start, |
| .udc_stop = bcm63xx_udc_stop, |
| }; |
| |
| /*********************************************************************** |
| * IRQ handling |
| ***********************************************************************/ |
| |
| /** |
| * bcm63xx_update_cfg_iface - Read current configuration/interface settings. |
| * @udc: Reference to the device controller. |
| * |
| * This controller intercepts SET_CONFIGURATION and SET_INTERFACE messages. |
| * The driver never sees the raw control packets coming in on the ep0 |
| * IUDMA channel, but at least we get an interrupt event to tell us that |
| * new values are waiting in the USBD_STATUS register. |
| */ |
| static void bcm63xx_update_cfg_iface(struct bcm63xx_udc *udc) |
| { |
| u32 reg = usbd_readl(udc, USBD_STATUS_REG); |
| |
| udc->cfg = (reg & USBD_STATUS_CFG_MASK) >> USBD_STATUS_CFG_SHIFT; |
| udc->iface = (reg & USBD_STATUS_INTF_MASK) >> USBD_STATUS_INTF_SHIFT; |
| udc->alt_iface = (reg & USBD_STATUS_ALTINTF_MASK) >> |
| USBD_STATUS_ALTINTF_SHIFT; |
| bcm63xx_ep_setup(udc); |
| } |
| |
| /** |
| * bcm63xx_update_link_speed - Check to see if the link speed has changed. |
| * @udc: Reference to the device controller. |
| * |
| * The link speed update coincides with a SETUP IRQ. Returns 1 if the |
| * speed has changed, so that the caller can update the endpoint settings. |
| */ |
| static int bcm63xx_update_link_speed(struct bcm63xx_udc *udc) |
| { |
| u32 reg = usbd_readl(udc, USBD_STATUS_REG); |
| enum usb_device_speed oldspeed = udc->gadget.speed; |
| |
| switch ((reg & USBD_STATUS_SPD_MASK) >> USBD_STATUS_SPD_SHIFT) { |
| case BCM63XX_SPD_HIGH: |
| udc->gadget.speed = USB_SPEED_HIGH; |
| break; |
| case BCM63XX_SPD_FULL: |
| udc->gadget.speed = USB_SPEED_FULL; |
| break; |
| default: |
| /* this should never happen */ |
| udc->gadget.speed = USB_SPEED_UNKNOWN; |
| dev_err(udc->dev, |
| "received SETUP packet with invalid link speed\n"); |
| return 0; |
| } |
| |
| if (udc->gadget.speed != oldspeed) { |
| dev_info(udc->dev, "link up, %s-speed mode\n", |
| udc->gadget.speed == USB_SPEED_HIGH ? "high" : "full"); |
| return 1; |
| } else { |
| return 0; |
| } |
| } |
| |
| /** |
| * bcm63xx_update_wedge - Iterate through wedged endpoints. |
| * @udc: Reference to the device controller. |
| * @new_status: true to "refresh" wedge status; false to clear it. |
| * |
| * On a SETUP interrupt, we need to manually "refresh" the wedge status |
| * because the controller hardware is designed to automatically clear |
| * stalls in response to a CLEAR_FEATURE request from the host. |
| * |
| * On a RESET interrupt, we do want to restore all wedged endpoints. |
| */ |
| static void bcm63xx_update_wedge(struct bcm63xx_udc *udc, bool new_status) |
| { |
| int i; |
| |
| for_each_set_bit(i, &udc->wedgemap, BCM63XX_NUM_EP) { |
| bcm63xx_set_stall(udc, &udc->bep[i], new_status); |
| if (!new_status) |
| clear_bit(i, &udc->wedgemap); |
| } |
| } |
| |
| /** |
| * bcm63xx_udc_ctrl_isr - ISR for control path events (USBD). |
| * @irq: IRQ number (unused). |
| * @dev_id: Reference to the device controller. |
| * |
| * This is where we handle link (VBUS) down, USB reset, speed changes, |
| * SET_CONFIGURATION, and SET_INTERFACE events. |
| */ |
| static irqreturn_t bcm63xx_udc_ctrl_isr(int irq, void *dev_id) |
| { |
| struct bcm63xx_udc *udc = dev_id; |
| u32 stat; |
| bool disconnected = false, bus_reset = false; |
| |
| stat = usbd_readl(udc, USBD_EVENT_IRQ_STATUS_REG) & |
| usbd_readl(udc, USBD_EVENT_IRQ_MASK_REG); |
| |
| usbd_writel(udc, stat, USBD_EVENT_IRQ_STATUS_REG); |
| |
| spin_lock(&udc->lock); |
| if (stat & BIT(USBD_EVENT_IRQ_USB_LINK)) { |
| /* VBUS toggled */ |
| |
| if (!(usbd_readl(udc, USBD_EVENTS_REG) & |
| USBD_EVENTS_USB_LINK_MASK) && |
| udc->gadget.speed != USB_SPEED_UNKNOWN) |
| dev_info(udc->dev, "link down\n"); |
| |
| udc->gadget.speed = USB_SPEED_UNKNOWN; |
| disconnected = true; |
| } |
| if (stat & BIT(USBD_EVENT_IRQ_USB_RESET)) { |
| bcm63xx_fifo_setup(udc); |
| bcm63xx_fifo_reset(udc); |
| bcm63xx_ep_setup(udc); |
| |
| bcm63xx_update_wedge(udc, false); |
| |
| udc->ep0_req_reset = 1; |
| schedule_work(&udc->ep0_wq); |
| bus_reset = true; |
| } |
| if (stat & BIT(USBD_EVENT_IRQ_SETUP)) { |
| if (bcm63xx_update_link_speed(udc)) { |
| bcm63xx_fifo_setup(udc); |
| bcm63xx_ep_setup(udc); |
| } |
| bcm63xx_update_wedge(udc, true); |
| } |
| if (stat & BIT(USBD_EVENT_IRQ_SETCFG)) { |
| bcm63xx_update_cfg_iface(udc); |
| udc->ep0_req_set_cfg = 1; |
| schedule_work(&udc->ep0_wq); |
| } |
| if (stat & BIT(USBD_EVENT_IRQ_SETINTF)) { |
| bcm63xx_update_cfg_iface(udc); |
| udc->ep0_req_set_iface = 1; |
| schedule_work(&udc->ep0_wq); |
| } |
| spin_unlock(&udc->lock); |
| |
| if (disconnected && udc->driver) |
| udc->driver->disconnect(&udc->gadget); |
| else if (bus_reset && udc->driver) |
| usb_gadget_udc_reset(&udc->gadget, udc->driver); |
| |
| return IRQ_HANDLED; |
| } |
| |
| /** |
| * bcm63xx_udc_data_isr - ISR for data path events (IUDMA). |
| * @irq: IRQ number (unused). |
| * @dev_id: Reference to the IUDMA channel that generated the interrupt. |
| * |
| * For the two ep0 channels, we have special handling that triggers the |
| * ep0 worker thread. For normal bulk/intr channels, either queue up |
| * the next buffer descriptor for the transaction (incomplete transaction), |
| * or invoke the completion callback (complete transactions). |
| */ |
| static irqreturn_t bcm63xx_udc_data_isr(int irq, void *dev_id) |
| { |
| struct iudma_ch *iudma = dev_id; |
| struct bcm63xx_udc *udc = iudma->udc; |
| struct bcm63xx_ep *bep; |
| struct usb_request *req = NULL; |
| struct bcm63xx_req *breq = NULL; |
| int rc; |
| bool is_done = false; |
| |
| spin_lock(&udc->lock); |
| |
| usb_dmac_writel(udc, ENETDMAC_IR_BUFDONE_MASK, |
| ENETDMAC_IR_REG, iudma->ch_idx); |
| bep = iudma->bep; |
| rc = iudma_read(udc, iudma); |
| |
| /* special handling for EP0 RX (0) and TX (1) */ |
| if (iudma->ch_idx == IUDMA_EP0_RXCHAN || |
| iudma->ch_idx == IUDMA_EP0_TXCHAN) { |
| req = udc->ep0_request; |
| breq = our_req(req); |
| |
| /* a single request could require multiple submissions */ |
| if (rc >= 0) { |
| req->actual += rc; |
| |
| if (req->actual >= req->length || breq->bd_bytes > rc) { |
| udc->ep0_req_completed = 1; |
| is_done = true; |
| schedule_work(&udc->ep0_wq); |
| |
| /* "actual" on a ZLP is 1 byte */ |
| req->actual = min(req->actual, req->length); |
| } else { |
| /* queue up the next BD (same request) */ |
| iudma_write(udc, iudma, breq); |
| } |
| } |
| } else if (!list_empty(&bep->queue)) { |
| breq = list_first_entry(&bep->queue, struct bcm63xx_req, queue); |
| req = &breq->req; |
| |
| if (rc >= 0) { |
| req->actual += rc; |
| |
| if (req->actual >= req->length || breq->bd_bytes > rc) { |
| is_done = true; |
| list_del(&breq->queue); |
| |
| req->actual = min(req->actual, req->length); |
| |
| if (!list_empty(&bep->queue)) { |
| struct bcm63xx_req *next; |
| |
| next = list_first_entry(&bep->queue, |
| struct bcm63xx_req, queue); |
| iudma_write(udc, iudma, next); |
| } |
| } else { |
| iudma_write(udc, iudma, breq); |
| } |
| } |
| } |
| spin_unlock(&udc->lock); |
| |
| if (is_done) { |
| usb_gadget_unmap_request(&udc->gadget, req, iudma->is_tx); |
| if (req->complete) |
| req->complete(&bep->ep, req); |
| } |
| |
| return IRQ_HANDLED; |
| } |
| |
| /*********************************************************************** |
| * Debug filesystem |
| ***********************************************************************/ |
| |
| /* |
| * bcm63xx_usbd_dbg_show - Show USBD controller state. |
| * @s: seq_file to which the information will be written. |
| * @p: Unused. |
| * |
| * This file nominally shows up as /sys/kernel/debug/bcm63xx_udc/usbd |
| */ |
| static int bcm63xx_usbd_dbg_show(struct seq_file *s, void *p) |
| { |
| struct bcm63xx_udc *udc = s->private; |
| |
| if (!udc->driver) |
| return -ENODEV; |
| |
| seq_printf(s, "ep0 state: %s\n", |
| bcm63xx_ep0_state_names[udc->ep0state]); |
| seq_printf(s, " pending requests: %s%s%s%s%s%s%s\n", |
| udc->ep0_req_reset ? "reset " : "", |
| udc->ep0_req_set_cfg ? "set_cfg " : "", |
| udc->ep0_req_set_iface ? "set_iface " : "", |
| udc->ep0_req_shutdown ? "shutdown " : "", |
| udc->ep0_request ? "pending " : "", |
| udc->ep0_req_completed ? "completed " : "", |
| udc->ep0_reply ? "reply " : ""); |
| seq_printf(s, "cfg: %d; iface: %d; alt_iface: %d\n", |
| udc->cfg, udc->iface, udc->alt_iface); |
| seq_printf(s, "regs:\n"); |
| seq_printf(s, " control: %08x; straps: %08x; status: %08x\n", |
| usbd_readl(udc, USBD_CONTROL_REG), |
| usbd_readl(udc, USBD_STRAPS_REG), |
| usbd_readl(udc, USBD_STATUS_REG)); |
| seq_printf(s, " events: %08x; stall: %08x\n", |
| usbd_readl(udc, USBD_EVENTS_REG), |
| usbd_readl(udc, USBD_STALL_REG)); |
| |
| return 0; |
| } |
| DEFINE_SHOW_ATTRIBUTE(bcm63xx_usbd_dbg); |
| |
| /* |
| * bcm63xx_iudma_dbg_show - Show IUDMA status and descriptors. |
| * @s: seq_file to which the information will be written. |
| * @p: Unused. |
| * |
| * This file nominally shows up as /sys/kernel/debug/bcm63xx_udc/iudma |
| */ |
| static int bcm63xx_iudma_dbg_show(struct seq_file *s, void *p) |
| { |
| struct bcm63xx_udc *udc = s->private; |
| int ch_idx, i; |
| u32 sram2, sram3; |
| |
| if (!udc->driver) |
| return -ENODEV; |
| |
| for (ch_idx = 0; ch_idx < BCM63XX_NUM_IUDMA; ch_idx++) { |
| struct iudma_ch *iudma = &udc->iudma[ch_idx]; |
| struct list_head *pos; |
| |
| seq_printf(s, "IUDMA channel %d -- ", ch_idx); |
| switch (iudma_defaults[ch_idx].ep_type) { |
| case BCMEP_CTRL: |
| seq_printf(s, "control"); |
| break; |
| case BCMEP_BULK: |
| seq_printf(s, "bulk"); |
| break; |
| case BCMEP_INTR: |
| seq_printf(s, "interrupt"); |
| break; |
| } |
| seq_printf(s, ch_idx & 0x01 ? " tx" : " rx"); |
| seq_printf(s, " [ep%d]:\n", |
| max_t(int, iudma_defaults[ch_idx].ep_num, 0)); |
| seq_printf(s, " cfg: %08x; irqstat: %08x; irqmask: %08x; maxburst: %08x\n", |
| usb_dmac_readl(udc, ENETDMAC_CHANCFG_REG, ch_idx), |
| usb_dmac_readl(udc, ENETDMAC_IR_REG, ch_idx), |
| usb_dmac_readl(udc, ENETDMAC_IRMASK_REG, ch_idx), |
| usb_dmac_readl(udc, ENETDMAC_MAXBURST_REG, ch_idx)); |
| |
| sram2 = usb_dmas_readl(udc, ENETDMAS_SRAM2_REG, ch_idx); |
| sram3 = usb_dmas_readl(udc, ENETDMAS_SRAM3_REG, ch_idx); |
| seq_printf(s, " base: %08x; index: %04x_%04x; desc: %04x_%04x %08x\n", |
| usb_dmas_readl(udc, ENETDMAS_RSTART_REG, ch_idx), |
| sram2 >> 16, sram2 & 0xffff, |
| sram3 >> 16, sram3 & 0xffff, |
| usb_dmas_readl(udc, ENETDMAS_SRAM4_REG, ch_idx)); |
| seq_printf(s, " desc: %d/%d used", iudma->n_bds_used, |
| iudma->n_bds); |
| |
| if (iudma->bep) { |
| i = 0; |
| list_for_each(pos, &iudma->bep->queue) |
| i++; |
| seq_printf(s, "; %d queued\n", i); |
| } else { |
| seq_printf(s, "\n"); |
| } |
| |
| for (i = 0; i < iudma->n_bds; i++) { |
| struct bcm_enet_desc *d = &iudma->bd_ring[i]; |
| |
| seq_printf(s, " %03x (%02x): len_stat: %04x_%04x; pa %08x", |
| i * sizeof(*d), i, |
| d->len_stat >> 16, d->len_stat & 0xffff, |
| d->address); |
| if (d == iudma->read_bd) |
| seq_printf(s, " <<RD"); |
| if (d == iudma->write_bd) |
| seq_printf(s, " <<WR"); |
| seq_printf(s, "\n"); |
| } |
| |
| seq_printf(s, "\n"); |
| } |
| |
| return 0; |
| } |
| DEFINE_SHOW_ATTRIBUTE(bcm63xx_iudma_dbg); |
| |
| /** |
| * bcm63xx_udc_init_debugfs - Create debugfs entries. |
| * @udc: Reference to the device controller. |
| */ |
| static void bcm63xx_udc_init_debugfs(struct bcm63xx_udc *udc) |
| { |
| struct dentry *root; |
| |
| if (!IS_ENABLED(CONFIG_USB_GADGET_DEBUG_FS)) |
| return; |
| |
| root = debugfs_create_dir(udc->gadget.name, NULL); |
| udc->debugfs_root = root; |
| |
| debugfs_create_file("usbd", 0400, root, udc, &bcm63xx_usbd_dbg_fops); |
| debugfs_create_file("iudma", 0400, root, udc, &bcm63xx_iudma_dbg_fops); |
| } |
| |
| /** |
| * bcm63xx_udc_cleanup_debugfs - Remove debugfs entries. |
| * @udc: Reference to the device controller. |
| * |
| * debugfs_remove() is safe to call with a NULL argument. |
| */ |
| static void bcm63xx_udc_cleanup_debugfs(struct bcm63xx_udc *udc) |
| { |
| debugfs_remove_recursive(udc->debugfs_root); |
| } |
| |
| /*********************************************************************** |
| * Driver init/exit |
| ***********************************************************************/ |
| |
| /** |
| * bcm63xx_udc_probe - Initialize a new instance of the UDC. |
| * @pdev: Platform device struct from the bcm63xx BSP code. |
| * |
| * Note that platform data is required, because pd.port_no varies from chip |
| * to chip and is used to switch the correct USB port to device mode. |
| */ |
| static int bcm63xx_udc_probe(struct platform_device *pdev) |
| { |
| struct device *dev = &pdev->dev; |
| struct bcm63xx_usbd_platform_data *pd = dev_get_platdata(dev); |
| struct bcm63xx_udc *udc; |
| struct resource *res; |
| int rc = -ENOMEM, i, irq; |
| |
| udc = devm_kzalloc(dev, sizeof(*udc), GFP_KERNEL); |
| if (!udc) |
| return -ENOMEM; |
| |
| platform_set_drvdata(pdev, udc); |
| udc->dev = dev; |
| udc->pd = pd; |
| |
| if (!pd) { |
| dev_err(dev, "missing platform data\n"); |
| return -EINVAL; |
| } |
| |
| res = platform_get_resource(pdev, IORESOURCE_MEM, 0); |
| udc->usbd_regs = devm_ioremap_resource(dev, res); |
| if (IS_ERR(udc->usbd_regs)) |
| return PTR_ERR(udc->usbd_regs); |
| |
| res = platform_get_resource(pdev, IORESOURCE_MEM, 1); |
| udc->iudma_regs = devm_ioremap_resource(dev, res); |
| if (IS_ERR(udc->iudma_regs)) |
| return PTR_ERR(udc->iudma_regs); |
| |
| spin_lock_init(&udc->lock); |
| INIT_WORK(&udc->ep0_wq, bcm63xx_ep0_process); |
| |
| udc->gadget.ops = &bcm63xx_udc_ops; |
| udc->gadget.name = dev_name(dev); |
| |
| if (!pd->use_fullspeed && !use_fullspeed) |
| udc->gadget.max_speed = USB_SPEED_HIGH; |
| else |
| udc->gadget.max_speed = USB_SPEED_FULL; |
| |
| /* request clocks, allocate buffers, and clear any pending IRQs */ |
| rc = bcm63xx_init_udc_hw(udc); |
| if (rc) |
| return rc; |
| |
| rc = -ENXIO; |
| |
| /* IRQ resource #0: control interrupt (VBUS, speed, etc.) */ |
| irq = platform_get_irq(pdev, 0); |
| if (irq < 0) { |
| dev_err(dev, "missing IRQ resource #0\n"); |
| goto out_uninit; |
| } |
| if (devm_request_irq(dev, irq, &bcm63xx_udc_ctrl_isr, 0, |
| dev_name(dev), udc) < 0) |
| goto report_request_failure; |
| |
| /* IRQ resources #1-6: data interrupts for IUDMA channels 0-5 */ |
| for (i = 0; i < BCM63XX_NUM_IUDMA; i++) { |
| irq = platform_get_irq(pdev, i + 1); |
| if (irq < 0) { |
| dev_err(dev, "missing IRQ resource #%d\n", i + 1); |
| goto out_uninit; |
| } |
| if (devm_request_irq(dev, irq, &bcm63xx_udc_data_isr, 0, |
| dev_name(dev), &udc->iudma[i]) < 0) |
| goto report_request_failure; |
| } |
| |
| bcm63xx_udc_init_debugfs(udc); |
| rc = usb_add_gadget_udc(dev, &udc->gadget); |
| if (!rc) |
| return 0; |
| |
| bcm63xx_udc_cleanup_debugfs(udc); |
| out_uninit: |
| bcm63xx_uninit_udc_hw(udc); |
| return rc; |
| |
| report_request_failure: |
| dev_err(dev, "error requesting IRQ #%d\n", irq); |
| goto out_uninit; |
| } |
| |
| /** |
| * bcm63xx_udc_remove - Remove the device from the system. |
| * @pdev: Platform device struct from the bcm63xx BSP code. |
| */ |
| static int bcm63xx_udc_remove(struct platform_device *pdev) |
| { |
| struct bcm63xx_udc *udc = platform_get_drvdata(pdev); |
| |
| bcm63xx_udc_cleanup_debugfs(udc); |
| usb_del_gadget_udc(&udc->gadget); |
| BUG_ON(udc->driver); |
| |
| bcm63xx_uninit_udc_hw(udc); |
| |
| return 0; |
| } |
| |
| static struct platform_driver bcm63xx_udc_driver = { |
| .probe = bcm63xx_udc_probe, |
| .remove = bcm63xx_udc_remove, |
| .driver = { |
| .name = DRV_MODULE_NAME, |
| }, |
| }; |
| module_platform_driver(bcm63xx_udc_driver); |
| |
| MODULE_DESCRIPTION("BCM63xx USB Peripheral Controller"); |
| MODULE_AUTHOR("Kevin Cernekee <cernekee@gmail.com>"); |
| MODULE_LICENSE("GPL"); |
| MODULE_ALIAS("platform:" DRV_MODULE_NAME); |