// SPDX-License-Identifier: GPL-2.0+ /* * USB Gadget driver for LPC32xx * * Authors: * Kevin Wells * Mike James * Roland Stigge * * Copyright (C) 2006 Philips Semiconductors * Copyright (C) 2009 NXP Semiconductors * Copyright (C) 2012 Roland Stigge * * Note: This driver is based on original work done by Mike James for * the LPC3180. */ #include #include #include #include #include #include #include #include #include #include #include #include #include #include #ifdef CONFIG_USB_GADGET_DEBUG_FILES #include #include #endif #include /* * USB device configuration structure */ typedef void (*usc_chg_event)(int); struct lpc32xx_usbd_cfg { int vbus_drv_pol; /* 0=active low drive for VBUS via ISP1301 */ usc_chg_event conn_chgb; /* Connection change event (optional) */ usc_chg_event susp_chgb; /* Suspend/resume event (optional) */ usc_chg_event rmwk_chgb; /* Enable/disable remote wakeup */ }; /* * controller driver data structures */ /* 16 endpoints (not to be confused with 32 hardware endpoints) */ #define NUM_ENDPOINTS 16 /* * IRQ indices make reading the code a little easier */ #define IRQ_USB_LP 0 #define IRQ_USB_HP 1 #define IRQ_USB_DEVDMA 2 #define IRQ_USB_ATX 3 #define EP_OUT 0 /* RX (from host) */ #define EP_IN 1 /* TX (to host) */ /* Returns the interrupt mask for the selected hardware endpoint */ #define EP_MASK_SEL(ep, dir) (1 << (((ep) * 2) + dir)) #define EP_INT_TYPE 0 #define EP_ISO_TYPE 1 #define EP_BLK_TYPE 2 #define EP_CTL_TYPE 3 /* EP0 states */ #define WAIT_FOR_SETUP 0 /* Wait for setup packet */ #define DATA_IN 1 /* Expect dev->host transfer */ #define DATA_OUT 2 /* Expect host->dev transfer */ /* DD (DMA Descriptor) structure, requires word alignment, this is already * defined in the LPC32XX USB device header file, but this version is slightly * modified to tag some work data with each DMA descriptor. */ struct lpc32xx_usbd_dd_gad { u32 dd_next_phy; u32 dd_setup; u32 dd_buffer_addr; u32 dd_status; u32 dd_iso_ps_mem_addr; u32 this_dma; u32 iso_status[6]; /* 5 spare */ u32 dd_next_v; }; /* * Logical endpoint structure */ struct lpc32xx_ep { struct usb_ep ep; struct list_head queue; struct lpc32xx_udc *udc; u32 hwep_num_base; /* Physical hardware EP */ u32 hwep_num; /* Maps to hardware endpoint */ u32 maxpacket; u32 lep; bool is_in; bool req_pending; u32 eptype; u32 totalints; bool wedge; }; /* * Common UDC structure */ struct lpc32xx_udc { struct usb_gadget gadget; struct usb_gadget_driver *driver; struct platform_device *pdev; struct device *dev; struct dentry *pde; spinlock_t lock; struct i2c_client *isp1301_i2c_client; /* Board and device specific */ struct lpc32xx_usbd_cfg *board; u32 io_p_start; u32 io_p_size; void __iomem *udp_baseaddr; int udp_irq[4]; struct clk *usb_slv_clk; /* DMA support */ u32 *udca_v_base; u32 udca_p_base; struct dma_pool *dd_cache; /* Common EP and control data */ u32 enabled_devints; u32 enabled_hwepints; u32 dev_status; u32 realized_eps; /* VBUS detection, pullup, and power flags */ u8 vbus; u8 last_vbus; int pullup; int poweron; /* Work queues related to I2C support */ struct work_struct pullup_job; struct work_struct vbus_job; struct work_struct power_job; /* USB device peripheral - various */ struct lpc32xx_ep ep[NUM_ENDPOINTS]; bool enabled; bool clocked; bool suspended; int ep0state; atomic_t enabled_ep_cnt; wait_queue_head_t ep_disable_wait_queue; }; /* * Endpoint request */ struct lpc32xx_request { struct usb_request req; struct list_head queue; struct lpc32xx_usbd_dd_gad *dd_desc_ptr; bool mapped; bool send_zlp; }; static inline struct lpc32xx_udc *to_udc(struct usb_gadget *g) { return container_of(g, struct lpc32xx_udc, gadget); } #define ep_dbg(epp, fmt, arg...) \ dev_dbg(epp->udc->dev, "%s: " fmt, __func__, ## arg) #define ep_err(epp, fmt, arg...) \ dev_err(epp->udc->dev, "%s: " fmt, __func__, ## arg) #define ep_info(epp, fmt, arg...) \ dev_info(epp->udc->dev, "%s: " fmt, __func__, ## arg) #define ep_warn(epp, fmt, arg...) \ dev_warn(epp->udc->dev, "%s:" fmt, __func__, ## arg) #define UDCA_BUFF_SIZE (128) /********************************************************************** * USB device controller register offsets **********************************************************************/ #define USBD_DEVINTST(x) ((x) + 0x200) #define USBD_DEVINTEN(x) ((x) + 0x204) #define USBD_DEVINTCLR(x) ((x) + 0x208) #define USBD_DEVINTSET(x) ((x) + 0x20C) #define USBD_CMDCODE(x) ((x) + 0x210) #define USBD_CMDDATA(x) ((x) + 0x214) #define USBD_RXDATA(x) ((x) + 0x218) #define USBD_TXDATA(x) ((x) + 0x21C) #define USBD_RXPLEN(x) ((x) + 0x220) #define USBD_TXPLEN(x) ((x) + 0x224) #define USBD_CTRL(x) ((x) + 0x228) #define USBD_DEVINTPRI(x) ((x) + 0x22C) #define USBD_EPINTST(x) ((x) + 0x230) #define USBD_EPINTEN(x) ((x) + 0x234) #define USBD_EPINTCLR(x) ((x) + 0x238) #define USBD_EPINTSET(x) ((x) + 0x23C) #define USBD_EPINTPRI(x) ((x) + 0x240) #define USBD_REEP(x) ((x) + 0x244) #define USBD_EPIND(x) ((x) + 0x248) #define USBD_EPMAXPSIZE(x) ((x) + 0x24C) /* DMA support registers only below */ /* Set, clear, or get enabled state of the DMA request status. If * enabled, an IN or OUT token will start a DMA transfer for the EP */ #define USBD_DMARST(x) ((x) + 0x250) #define USBD_DMARCLR(x) ((x) + 0x254) #define USBD_DMARSET(x) ((x) + 0x258) /* DMA UDCA head pointer */ #define USBD_UDCAH(x) ((x) + 0x280) /* EP DMA status, enable, and disable. This is used to specifically * enabled or disable DMA for a specific EP */ #define USBD_EPDMAST(x) ((x) + 0x284) #define USBD_EPDMAEN(x) ((x) + 0x288) #define USBD_EPDMADIS(x) ((x) + 0x28C) /* DMA master interrupts enable and pending interrupts */ #define USBD_DMAINTST(x) ((x) + 0x290) #define USBD_DMAINTEN(x) ((x) + 0x294) /* DMA end of transfer interrupt enable, disable, status */ #define USBD_EOTINTST(x) ((x) + 0x2A0) #define USBD_EOTINTCLR(x) ((x) + 0x2A4) #define USBD_EOTINTSET(x) ((x) + 0x2A8) /* New DD request interrupt enable, disable, status */ #define USBD_NDDRTINTST(x) ((x) + 0x2AC) #define USBD_NDDRTINTCLR(x) ((x) + 0x2B0) #define USBD_NDDRTINTSET(x) ((x) + 0x2B4) /* DMA error interrupt enable, disable, status */ #define USBD_SYSERRTINTST(x) ((x) + 0x2B8) #define USBD_SYSERRTINTCLR(x) ((x) + 0x2BC) #define USBD_SYSERRTINTSET(x) ((x) + 0x2C0) /********************************************************************** * USBD_DEVINTST/USBD_DEVINTEN/USBD_DEVINTCLR/USBD_DEVINTSET/ * USBD_DEVINTPRI register definitions **********************************************************************/ #define USBD_ERR_INT (1 << 9) #define USBD_EP_RLZED (1 << 8) #define USBD_TXENDPKT (1 << 7) #define USBD_RXENDPKT (1 << 6) #define USBD_CDFULL (1 << 5) #define USBD_CCEMPTY (1 << 4) #define USBD_DEV_STAT (1 << 3) #define USBD_EP_SLOW (1 << 2) #define USBD_EP_FAST (1 << 1) #define USBD_FRAME (1 << 0) /********************************************************************** * USBD_EPINTST/USBD_EPINTEN/USBD_EPINTCLR/USBD_EPINTSET/ * USBD_EPINTPRI register definitions **********************************************************************/ /* End point selection macro (RX) */ #define USBD_RX_EP_SEL(e) (1 << ((e) << 1)) /* End point selection macro (TX) */ #define USBD_TX_EP_SEL(e) (1 << (((e) << 1) + 1)) /********************************************************************** * USBD_REEP/USBD_DMARST/USBD_DMARCLR/USBD_DMARSET/USBD_EPDMAST/ * USBD_EPDMAEN/USBD_EPDMADIS/ * USBD_NDDRTINTST/USBD_NDDRTINTCLR/USBD_NDDRTINTSET/ * USBD_EOTINTST/USBD_EOTINTCLR/USBD_EOTINTSET/ * USBD_SYSERRTINTST/USBD_SYSERRTINTCLR/USBD_SYSERRTINTSET * register definitions **********************************************************************/ /* Endpoint selection macro */ #define USBD_EP_SEL(e) (1 << (e)) /********************************************************************** * SBD_DMAINTST/USBD_DMAINTEN **********************************************************************/ #define USBD_SYS_ERR_INT (1 << 2) #define USBD_NEW_DD_INT (1 << 1) #define USBD_EOT_INT (1 << 0) /********************************************************************** * USBD_RXPLEN register definitions **********************************************************************/ #define USBD_PKT_RDY (1 << 11) #define USBD_DV (1 << 10) #define USBD_PK_LEN_MASK 0x3FF /********************************************************************** * USBD_CTRL register definitions **********************************************************************/ #define USBD_LOG_ENDPOINT(e) ((e) << 2) #define USBD_WR_EN (1 << 1) #define USBD_RD_EN (1 << 0) /********************************************************************** * USBD_CMDCODE register definitions **********************************************************************/ #define USBD_CMD_CODE(c) ((c) << 16) #define USBD_CMD_PHASE(p) ((p) << 8) /********************************************************************** * USBD_DMARST/USBD_DMARCLR/USBD_DMARSET register definitions **********************************************************************/ #define USBD_DMAEP(e) (1 << (e)) /* DD (DMA Descriptor) structure, requires word alignment */ struct lpc32xx_usbd_dd { u32 *dd_next; u32 dd_setup; u32 dd_buffer_addr; u32 dd_status; u32 dd_iso_ps_mem_addr; }; /* dd_setup bit defines */ #define DD_SETUP_ATLE_DMA_MODE 0x01 #define DD_SETUP_NEXT_DD_VALID 0x04 #define DD_SETUP_ISO_EP 0x10 #define DD_SETUP_PACKETLEN(n) (((n) & 0x7FF) << 5) #define DD_SETUP_DMALENBYTES(n) (((n) & 0xFFFF) << 16) /* dd_status bit defines */ #define DD_STATUS_DD_RETIRED 0x01 #define DD_STATUS_STS_MASK 0x1E #define DD_STATUS_STS_NS 0x00 /* Not serviced */ #define DD_STATUS_STS_BS 0x02 /* Being serviced */ #define DD_STATUS_STS_NC 0x04 /* Normal completion */ #define DD_STATUS_STS_DUR 0x06 /* Data underrun (short packet) */ #define DD_STATUS_STS_DOR 0x08 /* Data overrun */ #define DD_STATUS_STS_SE 0x12 /* System error */ #define DD_STATUS_PKT_VAL 0x20 /* Packet valid */ #define DD_STATUS_LSB_EX 0x40 /* LS byte extracted (ATLE) */ #define DD_STATUS_MSB_EX 0x80 /* MS byte extracted (ATLE) */ #define DD_STATUS_MLEN(n) (((n) >> 8) & 0x3F) #define DD_STATUS_CURDMACNT(n) (((n) >> 16) & 0xFFFF) /* * * Protocol engine bits below * */ /* Device Interrupt Bit Definitions */ #define FRAME_INT 0x00000001 #define EP_FAST_INT 0x00000002 #define EP_SLOW_INT 0x00000004 #define DEV_STAT_INT 0x00000008 #define CCEMTY_INT 0x00000010 #define CDFULL_INT 0x00000020 #define RxENDPKT_INT 0x00000040 #define TxENDPKT_INT 0x00000080 #define EP_RLZED_INT 0x00000100 #define ERR_INT 0x00000200 /* Rx & Tx Packet Length Definitions */ #define PKT_LNGTH_MASK 0x000003FF #define PKT_DV 0x00000400 #define PKT_RDY 0x00000800 /* USB Control Definitions */ #define CTRL_RD_EN 0x00000001 #define CTRL_WR_EN 0x00000002 /* Command Codes */ #define CMD_SET_ADDR 0x00D00500 #define CMD_CFG_DEV 0x00D80500 #define CMD_SET_MODE 0x00F30500 #define CMD_RD_FRAME 0x00F50500 #define DAT_RD_FRAME 0x00F50200 #define CMD_RD_TEST 0x00FD0500 #define DAT_RD_TEST 0x00FD0200 #define CMD_SET_DEV_STAT 0x00FE0500 #define CMD_GET_DEV_STAT 0x00FE0500 #define DAT_GET_DEV_STAT 0x00FE0200 #define CMD_GET_ERR_CODE 0x00FF0500 #define DAT_GET_ERR_CODE 0x00FF0200 #define CMD_RD_ERR_STAT 0x00FB0500 #define DAT_RD_ERR_STAT 0x00FB0200 #define DAT_WR_BYTE(x) (0x00000100 | ((x) << 16)) #define CMD_SEL_EP(x) (0x00000500 | ((x) << 16)) #define DAT_SEL_EP(x) (0x00000200 | ((x) << 16)) #define CMD_SEL_EP_CLRI(x) (0x00400500 | ((x) << 16)) #define DAT_SEL_EP_CLRI(x) (0x00400200 | ((x) << 16)) #define CMD_SET_EP_STAT(x) (0x00400500 | ((x) << 16)) #define CMD_CLR_BUF 0x00F20500 #define DAT_CLR_BUF 0x00F20200 #define CMD_VALID_BUF 0x00FA0500 /* Device Address Register Definitions */ #define DEV_ADDR_MASK 0x7F #define DEV_EN 0x80 /* Device Configure Register Definitions */ #define CONF_DVICE 0x01 /* Device Mode Register Definitions */ #define AP_CLK 0x01 #define INAK_CI 0x02 #define INAK_CO 0x04 #define INAK_II 0x08 #define INAK_IO 0x10 #define INAK_BI 0x20 #define INAK_BO 0x40 /* Device Status Register Definitions */ #define DEV_CON 0x01 #define DEV_CON_CH 0x02 #define DEV_SUS 0x04 #define DEV_SUS_CH 0x08 #define DEV_RST 0x10 /* Error Code Register Definitions */ #define ERR_EC_MASK 0x0F #define ERR_EA 0x10 /* Error Status Register Definitions */ #define ERR_PID 0x01 #define ERR_UEPKT 0x02 #define ERR_DCRC 0x04 #define ERR_TIMOUT 0x08 #define ERR_EOP 0x10 #define ERR_B_OVRN 0x20 #define ERR_BTSTF 0x40 #define ERR_TGL 0x80 /* Endpoint Select Register Definitions */ #define EP_SEL_F 0x01 #define EP_SEL_ST 0x02 #define EP_SEL_STP 0x04 #define EP_SEL_PO 0x08 #define EP_SEL_EPN 0x10 #define EP_SEL_B_1_FULL 0x20 #define EP_SEL_B_2_FULL 0x40 /* Endpoint Status Register Definitions */ #define EP_STAT_ST 0x01 #define EP_STAT_DA 0x20 #define EP_STAT_RF_MO 0x40 #define EP_STAT_CND_ST 0x80 /* Clear Buffer Register Definitions */ #define CLR_BUF_PO 0x01 /* DMA Interrupt Bit Definitions */ #define EOT_INT 0x01 #define NDD_REQ_INT 0x02 #define SYS_ERR_INT 0x04 #define DRIVER_VERSION "1.03" static const char driver_name[] = "lpc32xx_udc"; /* * * proc interface support * */ #ifdef CONFIG_USB_GADGET_DEBUG_FILES static char *epnames[] = {"INT", "ISO", "BULK", "CTRL"}; static const char debug_filename[] = "driver/udc"; static void proc_ep_show(struct seq_file *s, struct lpc32xx_ep *ep) { struct lpc32xx_request *req; seq_printf(s, "\n"); seq_printf(s, "%12s, maxpacket %4d %3s", ep->ep.name, ep->ep.maxpacket, ep->is_in ? "in" : "out"); seq_printf(s, " type %4s", epnames[ep->eptype]); seq_printf(s, " ints: %12d", ep->totalints); if (list_empty(&ep->queue)) seq_printf(s, "\t(queue empty)\n"); else { list_for_each_entry(req, &ep->queue, queue) { u32 length = req->req.actual; seq_printf(s, "\treq %p len %d/%d buf %p\n", &req->req, length, req->req.length, req->req.buf); } } } static int proc_udc_show(struct seq_file *s, void *unused) { struct lpc32xx_udc *udc = s->private; struct lpc32xx_ep *ep; unsigned long flags; seq_printf(s, "%s: version %s\n", driver_name, DRIVER_VERSION); spin_lock_irqsave(&udc->lock, flags); seq_printf(s, "vbus %s, pullup %s, %s powered%s, gadget %s\n\n", udc->vbus ? "present" : "off", udc->enabled ? (udc->vbus ? "active" : "enabled") : "disabled", udc->gadget.is_selfpowered ? "self" : "VBUS", udc->suspended ? ", suspended" : "", udc->driver ? udc->driver->driver.name : "(none)"); if (udc->enabled && udc->vbus) { proc_ep_show(s, &udc->ep[0]); list_for_each_entry(ep, &udc->gadget.ep_list, ep.ep_list) proc_ep_show(s, ep); } spin_unlock_irqrestore(&udc->lock, flags); return 0; } static int proc_udc_open(struct inode *inode, struct file *file) { return single_open(file, proc_udc_show, PDE_DATA(inode)); } static const struct file_operations proc_ops = { .owner = THIS_MODULE, .open = proc_udc_open, .read = seq_read, .llseek = seq_lseek, .release = single_release, }; static void create_debug_file(struct lpc32xx_udc *udc) { udc->pde = debugfs_create_file(debug_filename, 0, NULL, udc, &proc_ops); } static void remove_debug_file(struct lpc32xx_udc *udc) { debugfs_remove(udc->pde); } #else static inline void create_debug_file(struct lpc32xx_udc *udc) {} static inline void remove_debug_file(struct lpc32xx_udc *udc) {} #endif /* Primary initialization sequence for the ISP1301 transceiver */ static void isp1301_udc_configure(struct lpc32xx_udc *udc) { /* LPC32XX only supports DAT_SE0 USB mode */ /* This sequence is important */ /* Disable transparent UART mode first */ i2c_smbus_write_byte_data(udc->isp1301_i2c_client, (ISP1301_I2C_MODE_CONTROL_1 | ISP1301_I2C_REG_CLEAR_ADDR), MC1_UART_EN); /* Set full speed and SE0 mode */ i2c_smbus_write_byte_data(udc->isp1301_i2c_client, (ISP1301_I2C_MODE_CONTROL_1 | ISP1301_I2C_REG_CLEAR_ADDR), ~0); i2c_smbus_write_byte_data(udc->isp1301_i2c_client, ISP1301_I2C_MODE_CONTROL_1, (MC1_SPEED_REG | MC1_DAT_SE0)); /* * The PSW_OE enable bit state is reversed in the ISP1301 User's Guide */ i2c_smbus_write_byte_data(udc->isp1301_i2c_client, (ISP1301_I2C_MODE_CONTROL_2 | ISP1301_I2C_REG_CLEAR_ADDR), ~0); i2c_smbus_write_byte_data(udc->isp1301_i2c_client, ISP1301_I2C_MODE_CONTROL_2, (MC2_BI_DI | MC2_SPD_SUSP_CTRL)); /* Driver VBUS_DRV high or low depending on board setup */ if (udc->board->vbus_drv_pol != 0) i2c_smbus_write_byte_data(udc->isp1301_i2c_client, ISP1301_I2C_OTG_CONTROL_1, OTG1_VBUS_DRV); else i2c_smbus_write_byte_data(udc->isp1301_i2c_client, ISP1301_I2C_OTG_CONTROL_1 | ISP1301_I2C_REG_CLEAR_ADDR, OTG1_VBUS_DRV); /* Bi-directional mode with suspend control * Enable both pulldowns for now - the pullup will be enable when VBUS * is detected */ i2c_smbus_write_byte_data(udc->isp1301_i2c_client, (ISP1301_I2C_OTG_CONTROL_1 | ISP1301_I2C_REG_CLEAR_ADDR), ~0); i2c_smbus_write_byte_data(udc->isp1301_i2c_client, ISP1301_I2C_OTG_CONTROL_1, (0 | OTG1_DM_PULLDOWN | OTG1_DP_PULLDOWN)); /* Discharge VBUS (just in case) */ i2c_smbus_write_byte_data(udc->isp1301_i2c_client, ISP1301_I2C_OTG_CONTROL_1, OTG1_VBUS_DISCHRG); msleep(1); i2c_smbus_write_byte_data(udc->isp1301_i2c_client, (ISP1301_I2C_OTG_CONTROL_1 | ISP1301_I2C_REG_CLEAR_ADDR), OTG1_VBUS_DISCHRG); /* Clear and enable VBUS high edge interrupt */ i2c_smbus_write_byte_data(udc->isp1301_i2c_client, ISP1301_I2C_INTERRUPT_LATCH | ISP1301_I2C_REG_CLEAR_ADDR, ~0); i2c_smbus_write_byte_data(udc->isp1301_i2c_client, ISP1301_I2C_INTERRUPT_FALLING | ISP1301_I2C_REG_CLEAR_ADDR, ~0); i2c_smbus_write_byte_data(udc->isp1301_i2c_client, ISP1301_I2C_INTERRUPT_FALLING, INT_VBUS_VLD); i2c_smbus_write_byte_data(udc->isp1301_i2c_client, ISP1301_I2C_INTERRUPT_RISING | ISP1301_I2C_REG_CLEAR_ADDR, ~0); i2c_smbus_write_byte_data(udc->isp1301_i2c_client, ISP1301_I2C_INTERRUPT_RISING, INT_VBUS_VLD); dev_info(udc->dev, "ISP1301 Vendor ID : 0x%04x\n", i2c_smbus_read_word_data(udc->isp1301_i2c_client, 0x00)); dev_info(udc->dev, "ISP1301 Product ID : 0x%04x\n", i2c_smbus_read_word_data(udc->isp1301_i2c_client, 0x02)); dev_info(udc->dev, "ISP1301 Version ID : 0x%04x\n", i2c_smbus_read_word_data(udc->isp1301_i2c_client, 0x14)); } /* Enables or disables the USB device pullup via the ISP1301 transceiver */ static void isp1301_pullup_set(struct lpc32xx_udc *udc) { if (udc->pullup) /* Enable pullup for bus signalling */ i2c_smbus_write_byte_data(udc->isp1301_i2c_client, ISP1301_I2C_OTG_CONTROL_1, OTG1_DP_PULLUP); else /* Enable pullup for bus signalling */ i2c_smbus_write_byte_data(udc->isp1301_i2c_client, ISP1301_I2C_OTG_CONTROL_1 | ISP1301_I2C_REG_CLEAR_ADDR, OTG1_DP_PULLUP); } static void pullup_work(struct work_struct *work) { struct lpc32xx_udc *udc = container_of(work, struct lpc32xx_udc, pullup_job); isp1301_pullup_set(udc); } static void isp1301_pullup_enable(struct lpc32xx_udc *udc, int en_pullup, int block) { if (en_pullup == udc->pullup) return; udc->pullup = en_pullup; if (block) isp1301_pullup_set(udc); else /* defer slow i2c pull up setting */ schedule_work(&udc->pullup_job); } #ifdef CONFIG_PM /* Powers up or down the ISP1301 transceiver */ static void isp1301_set_powerstate(struct lpc32xx_udc *udc, int enable) { if (enable != 0) /* Power up ISP1301 - this ISP1301 will automatically wakeup when VBUS is detected */ i2c_smbus_write_byte_data(udc->isp1301_i2c_client, ISP1301_I2C_MODE_CONTROL_2 | ISP1301_I2C_REG_CLEAR_ADDR, MC2_GLOBAL_PWR_DN); else /* Power down ISP1301 */ i2c_smbus_write_byte_data(udc->isp1301_i2c_client, ISP1301_I2C_MODE_CONTROL_2, MC2_GLOBAL_PWR_DN); } static void power_work(struct work_struct *work) { struct lpc32xx_udc *udc = container_of(work, struct lpc32xx_udc, power_job); isp1301_set_powerstate(udc, udc->poweron); } #endif /* * * USB protocol engine command/data read/write helper functions * */ /* Issues a single command to the USB device state machine */ static void udc_protocol_cmd_w(struct lpc32xx_udc *udc, u32 cmd) { u32 pass = 0; int to; /* EP may lock on CLRI if this read isn't done */ u32 tmp = readl(USBD_DEVINTST(udc->udp_baseaddr)); (void) tmp; while (pass == 0) { writel(USBD_CCEMPTY, USBD_DEVINTCLR(udc->udp_baseaddr)); /* Write command code */ writel(cmd, USBD_CMDCODE(udc->udp_baseaddr)); to = 10000; while (((readl(USBD_DEVINTST(udc->udp_baseaddr)) & USBD_CCEMPTY) == 0) && (to > 0)) { to--; } if (to > 0) pass = 1; cpu_relax(); } } /* Issues 2 commands (or command and data) to the USB device state machine */ static inline void udc_protocol_cmd_data_w(struct lpc32xx_udc *udc, u32 cmd, u32 data) { udc_protocol_cmd_w(udc, cmd); udc_protocol_cmd_w(udc, data); } /* Issues a single command to the USB device state machine and reads * response data */ static u32 udc_protocol_cmd_r(struct lpc32xx_udc *udc, u32 cmd) { u32 tmp; int to = 1000; /* Write a command and read data from the protocol engine */ writel((USBD_CDFULL | USBD_CCEMPTY), USBD_DEVINTCLR(udc->udp_baseaddr)); /* Write command code */ udc_protocol_cmd_w(udc, cmd); tmp = readl(USBD_DEVINTST(udc->udp_baseaddr)); while ((!(readl(USBD_DEVINTST(udc->udp_baseaddr)) & USBD_CDFULL)) && (to > 0)) to--; if (!to) dev_dbg(udc->dev, "Protocol engine didn't receive response (CDFULL)\n"); return readl(USBD_CMDDATA(udc->udp_baseaddr)); } /* * * USB device interrupt mask support functions * */ /* Enable one or more USB device interrupts */ static inline void uda_enable_devint(struct lpc32xx_udc *udc, u32 devmask) { udc->enabled_devints |= devmask; writel(udc->enabled_devints, USBD_DEVINTEN(udc->udp_baseaddr)); } /* Disable one or more USB device interrupts */ static inline void uda_disable_devint(struct lpc32xx_udc *udc, u32 mask) { udc->enabled_devints &= ~mask; writel(udc->enabled_devints, USBD_DEVINTEN(udc->udp_baseaddr)); } /* Clear one or more USB device interrupts */ static inline void uda_clear_devint(struct lpc32xx_udc *udc, u32 mask) { writel(mask, USBD_DEVINTCLR(udc->udp_baseaddr)); } /* * * Endpoint interrupt disable/enable functions * */ /* Enable one or more USB endpoint interrupts */ static void uda_enable_hwepint(struct lpc32xx_udc *udc, u32 hwep) { udc->enabled_hwepints |= (1 << hwep); writel(udc->enabled_hwepints, USBD_EPINTEN(udc->udp_baseaddr)); } /* Disable one or more USB endpoint interrupts */ static void uda_disable_hwepint(struct lpc32xx_udc *udc, u32 hwep) { udc->enabled_hwepints &= ~(1 << hwep); writel(udc->enabled_hwepints, USBD_EPINTEN(udc->udp_baseaddr)); } /* Clear one or more USB endpoint interrupts */ static inline void uda_clear_hwepint(struct lpc32xx_udc *udc, u32 hwep) { writel((1 << hwep), USBD_EPINTCLR(udc->udp_baseaddr)); } /* Enable DMA for the HW channel */ static inline void udc_ep_dma_enable(struct lpc32xx_udc *udc, u32 hwep) { writel((1 << hwep), USBD_EPDMAEN(udc->udp_baseaddr)); } /* Disable DMA for the HW channel */ static inline void udc_ep_dma_disable(struct lpc32xx_udc *udc, u32 hwep) { writel((1 << hwep), USBD_EPDMADIS(udc->udp_baseaddr)); } /* * * Endpoint realize/unrealize functions * */ /* Before an endpoint can be used, it needs to be realized * in the USB protocol engine - this realizes the endpoint. * The interrupt (FIFO or DMA) is not enabled with this function */ static void udc_realize_hwep(struct lpc32xx_udc *udc, u32 hwep, u32 maxpacket) { int to = 1000; writel(USBD_EP_RLZED, USBD_DEVINTCLR(udc->udp_baseaddr)); writel(hwep, USBD_EPIND(udc->udp_baseaddr)); udc->realized_eps |= (1 << hwep); writel(udc->realized_eps, USBD_REEP(udc->udp_baseaddr)); writel(maxpacket, USBD_EPMAXPSIZE(udc->udp_baseaddr)); /* Wait until endpoint is realized in hardware */ while ((!(readl(USBD_DEVINTST(udc->udp_baseaddr)) & USBD_EP_RLZED)) && (to > 0)) to--; if (!to) dev_dbg(udc->dev, "EP not correctly realized in hardware\n"); writel(USBD_EP_RLZED, USBD_DEVINTCLR(udc->udp_baseaddr)); } /* Unrealize an EP */ static void udc_unrealize_hwep(struct lpc32xx_udc *udc, u32 hwep) { udc->realized_eps &= ~(1 << hwep); writel(udc->realized_eps, USBD_REEP(udc->udp_baseaddr)); } /* * * Endpoint support functions * */ /* Select and clear endpoint interrupt */ static u32 udc_selep_clrint(struct lpc32xx_udc *udc, u32 hwep) { udc_protocol_cmd_w(udc, CMD_SEL_EP_CLRI(hwep)); return udc_protocol_cmd_r(udc, DAT_SEL_EP_CLRI(hwep)); } /* Disables the endpoint in the USB protocol engine */ static void udc_disable_hwep(struct lpc32xx_udc *udc, u32 hwep) { udc_protocol_cmd_data_w(udc, CMD_SET_EP_STAT(hwep), DAT_WR_BYTE(EP_STAT_DA)); } /* Stalls the endpoint - endpoint will return STALL */ static void udc_stall_hwep(struct lpc32xx_udc *udc, u32 hwep) { udc_protocol_cmd_data_w(udc, CMD_SET_EP_STAT(hwep), DAT_WR_BYTE(EP_STAT_ST)); } /* Clear stall or reset endpoint */ static void udc_clrstall_hwep(struct lpc32xx_udc *udc, u32 hwep) { udc_protocol_cmd_data_w(udc, CMD_SET_EP_STAT(hwep), DAT_WR_BYTE(0)); } /* Select an endpoint for endpoint status, clear, validate */ static void udc_select_hwep(struct lpc32xx_udc *udc, u32 hwep) { udc_protocol_cmd_w(udc, CMD_SEL_EP(hwep)); } /* * * Endpoint buffer management functions * */ /* Clear the current endpoint's buffer */ static void udc_clr_buffer_hwep(struct lpc32xx_udc *udc, u32 hwep) { udc_select_hwep(udc, hwep); udc_protocol_cmd_w(udc, CMD_CLR_BUF); } /* Validate the current endpoint's buffer */ static void udc_val_buffer_hwep(struct lpc32xx_udc *udc, u32 hwep) { udc_select_hwep(udc, hwep); udc_protocol_cmd_w(udc, CMD_VALID_BUF); } static inline u32 udc_clearep_getsts(struct lpc32xx_udc *udc, u32 hwep) { /* Clear EP interrupt */ uda_clear_hwepint(udc, hwep); return udc_selep_clrint(udc, hwep); } /* * * USB EP DMA support * */ /* Allocate a DMA Descriptor */ static struct lpc32xx_usbd_dd_gad *udc_dd_alloc(struct lpc32xx_udc *udc) { dma_addr_t dma; struct lpc32xx_usbd_dd_gad *dd; dd = dma_pool_alloc(udc->dd_cache, GFP_ATOMIC | GFP_DMA, &dma); if (dd) dd->this_dma = dma; return dd; } /* Free a DMA Descriptor */ static void udc_dd_free(struct lpc32xx_udc *udc, struct lpc32xx_usbd_dd_gad *dd) { dma_pool_free(udc->dd_cache, dd, dd->this_dma); } /* * * USB setup and shutdown functions * */ /* Enables or disables most of the USB system clocks when low power mode is * needed. Clocks are typically started on a connection event, and disabled * when a cable is disconnected */ static void udc_clk_set(struct lpc32xx_udc *udc, int enable) { if (enable != 0) { if (udc->clocked) return; udc->clocked = 1; clk_prepare_enable(udc->usb_slv_clk); } else { if (!udc->clocked) return; udc->clocked = 0; clk_disable_unprepare(udc->usb_slv_clk); } } /* Set/reset USB device address */ static void udc_set_address(struct lpc32xx_udc *udc, u32 addr) { /* Address will be latched at the end of the status phase, or latched immediately if function is called twice */ udc_protocol_cmd_data_w(udc, CMD_SET_ADDR, DAT_WR_BYTE(DEV_EN | addr)); } /* Setup up a IN request for DMA transfer - this consists of determining the * list of DMA addresses for the transfer, allocating DMA Descriptors, * installing the DD into the UDCA, and then enabling the DMA for that EP */ static int udc_ep_in_req_dma(struct lpc32xx_udc *udc, struct lpc32xx_ep *ep) { struct lpc32xx_request *req; u32 hwep = ep->hwep_num; ep->req_pending = 1; /* There will always be a request waiting here */ req = list_entry(ep->queue.next, struct lpc32xx_request, queue); /* Place the DD Descriptor into the UDCA */ udc->udca_v_base[hwep] = req->dd_desc_ptr->this_dma; /* Enable DMA and interrupt for the HW EP */ udc_ep_dma_enable(udc, hwep); /* Clear ZLP if last packet is not of MAXP size */ if (req->req.length % ep->ep.maxpacket) req->send_zlp = 0; return 0; } /* Setup up a OUT request for DMA transfer - this consists of determining the * list of DMA addresses for the transfer, allocating DMA Descriptors, * installing the DD into the UDCA, and then enabling the DMA for that EP */ static int udc_ep_out_req_dma(struct lpc32xx_udc *udc, struct lpc32xx_ep *ep) { struct lpc32xx_request *req; u32 hwep = ep->hwep_num; ep->req_pending = 1; /* There will always be a request waiting here */ req = list_entry(ep->queue.next, struct lpc32xx_request, queue); /* Place the DD Descriptor into the UDCA */ udc->udca_v_base[hwep] = req->dd_desc_ptr->this_dma; /* Enable DMA and interrupt for the HW EP */ udc_ep_dma_enable(udc, hwep); return 0; } static void udc_disable(struct lpc32xx_udc *udc) { u32 i; /* Disable device */ udc_protocol_cmd_data_w(udc, CMD_CFG_DEV, DAT_WR_BYTE(0)); udc_protocol_cmd_data_w(udc, CMD_SET_DEV_STAT, DAT_WR_BYTE(0)); /* Disable all device interrupts (including EP0) */ uda_disable_devint(udc, 0x3FF); /* Disable and reset all endpoint interrupts */ for (i = 0; i < 32; i++) { uda_disable_hwepint(udc, i); uda_clear_hwepint(udc, i); udc_disable_hwep(udc, i); udc_unrealize_hwep(udc, i); udc->udca_v_base[i] = 0; /* Disable and clear all interrupts and DMA */ udc_ep_dma_disable(udc, i); writel((1 << i), USBD_EOTINTCLR(udc->udp_baseaddr)); writel((1 << i), USBD_NDDRTINTCLR(udc->udp_baseaddr)); writel((1 << i), USBD_SYSERRTINTCLR(udc->udp_baseaddr)); writel((1 << i), USBD_DMARCLR(udc->udp_baseaddr)); } /* Disable DMA interrupts */ writel(0, USBD_DMAINTEN(udc->udp_baseaddr)); writel(0, USBD_UDCAH(udc->udp_baseaddr)); } static void udc_enable(struct lpc32xx_udc *udc) { u32 i; struct lpc32xx_ep *ep = &udc->ep[0]; /* Start with known state */ udc_disable(udc); /* Enable device */ udc_protocol_cmd_data_w(udc, CMD_SET_DEV_STAT, DAT_WR_BYTE(DEV_CON)); /* EP interrupts on high priority, FRAME interrupt on low priority */ writel(USBD_EP_FAST, USBD_DEVINTPRI(udc->udp_baseaddr)); writel(0xFFFF, USBD_EPINTPRI(udc->udp_baseaddr)); /* Clear any pending device interrupts */ writel(0x3FF, USBD_DEVINTCLR(udc->udp_baseaddr)); /* Setup UDCA - not yet used (DMA) */ writel(udc->udca_p_base, USBD_UDCAH(udc->udp_baseaddr)); /* Only enable EP0 in and out for now, EP0 only works in FIFO mode */ for (i = 0; i <= 1; i++) { udc_realize_hwep(udc, i, ep->ep.maxpacket); uda_enable_hwepint(udc, i); udc_select_hwep(udc, i); udc_clrstall_hwep(udc, i); udc_clr_buffer_hwep(udc, i); } /* Device interrupt setup */ uda_clear_devint(udc, (USBD_ERR_INT | USBD_DEV_STAT | USBD_EP_SLOW | USBD_EP_FAST)); uda_enable_devint(udc, (USBD_ERR_INT | USBD_DEV_STAT | USBD_EP_SLOW | USBD_EP_FAST)); /* Set device address to 0 - called twice to force a latch in the USB engine without the need of a setup packet status closure */ udc_set_address(udc, 0); udc_set_address(udc, 0); /* Enable master DMA interrupts */ writel((USBD_SYS_ERR_INT | USBD_EOT_INT), USBD_DMAINTEN(udc->udp_baseaddr)); udc->dev_status = 0; } /* * * USB device board specific events handled via callbacks * */ /* Connection change event - notify board function of change */ static void uda_power_event(struct lpc32xx_udc *udc, u32 conn) { /* Just notify of a connection change event (optional) */ if (udc->board->conn_chgb != NULL) udc->board->conn_chgb(conn); } /* Suspend/resume event - notify board function of change */ static void uda_resm_susp_event(struct lpc32xx_udc *udc, u32 conn) { /* Just notify of a Suspend/resume change event (optional) */ if (udc->board->susp_chgb != NULL) udc->board->susp_chgb(conn); if (conn) udc->suspended = 0; else udc->suspended = 1; } /* Remote wakeup enable/disable - notify board function of change */ static void uda_remwkp_cgh(struct lpc32xx_udc *udc) { if (udc->board->rmwk_chgb != NULL) udc->board->rmwk_chgb(udc->dev_status & (1 << USB_DEVICE_REMOTE_WAKEUP)); } /* Reads data from FIFO, adjusts for alignment and data size */ static void udc_pop_fifo(struct lpc32xx_udc *udc, u8 *data, u32 bytes) { int n, i, bl; u16 *p16; u32 *p32, tmp, cbytes; /* Use optimal data transfer method based on source address and size */ switch (((u32) data) & 0x3) { case 0: /* 32-bit aligned */ p32 = (u32 *) data; cbytes = (bytes & ~0x3); /* Copy 32-bit aligned data first */ for (n = 0; n < cbytes; n += 4) *p32++ = readl(USBD_RXDATA(udc->udp_baseaddr)); /* Handle any remaining bytes */ bl = bytes - cbytes; if (bl) { tmp = readl(USBD_RXDATA(udc->udp_baseaddr)); for (n = 0; n < bl; n++) data[cbytes + n] = ((tmp >> (n * 8)) & 0xFF); } break; case 1: /* 8-bit aligned */ case 3: /* Each byte has to be handled independently */ for (n = 0; n < bytes; n += 4) { tmp = readl(USBD_RXDATA(udc->udp_baseaddr)); bl = bytes - n; if (bl > 4) bl = 4; for (i = 0; i < bl; i++) data[n + i] = (u8) ((tmp >> (i * 8)) & 0xFF); } break; case 2: /* 16-bit aligned */ p16 = (u16 *) data; cbytes = (bytes & ~0x3); /* Copy 32-bit sized objects first with 16-bit alignment */ for (n = 0; n < cbytes; n += 4) { tmp = readl(USBD_RXDATA(udc->udp_baseaddr)); *p16++ = (u16)(tmp & 0xFFFF); *p16++ = (u16)((tmp >> 16) & 0xFFFF); } /* Handle any remaining bytes */ bl = bytes - cbytes; if (bl) { tmp = readl(USBD_RXDATA(udc->udp_baseaddr)); for (n = 0; n < bl; n++) data[cbytes + n] = ((tmp >> (n * 8)) & 0xFF); } break; } } /* Read data from the FIFO for an endpoint. This function is for endpoints (such * as EP0) that don't use DMA. This function should only be called if a packet * is known to be ready to read for the endpoint. Note that the endpoint must * be selected in the protocol engine prior to this call. */ static u32 udc_read_hwep(struct lpc32xx_udc *udc, u32 hwep, u32 *data, u32 bytes) { u32 tmpv; int to = 1000; u32 tmp, hwrep = ((hwep & 0x1E) << 1) | CTRL_RD_EN; /* Setup read of endpoint */ writel(hwrep, USBD_CTRL(udc->udp_baseaddr)); /* Wait until packet is ready */ while ((((tmpv = readl(USBD_RXPLEN(udc->udp_baseaddr))) & PKT_RDY) == 0) && (to > 0)) to--; if (!to) dev_dbg(udc->dev, "No packet ready on FIFO EP read\n"); /* Mask out count */ tmp = tmpv & PKT_LNGTH_MASK; if (bytes < tmp) tmp = bytes; if ((tmp > 0) && (data != NULL)) udc_pop_fifo(udc, (u8 *) data, tmp); writel(((hwep & 0x1E) << 1), USBD_CTRL(udc->udp_baseaddr)); /* Clear the buffer */ udc_clr_buffer_hwep(udc, hwep); return tmp; } /* Stuffs data into the FIFO, adjusts for alignment and data size */ static void udc_stuff_fifo(struct lpc32xx_udc *udc, u8 *data, u32 bytes) { int n, i, bl; u16 *p16; u32 *p32, tmp, cbytes; /* Use optimal data transfer method based on source address and size */ switch (((u32) data) & 0x3) { case 0: /* 32-bit aligned */ p32 = (u32 *) data; cbytes = (bytes & ~0x3); /* Copy 32-bit aligned data first */ for (n = 0; n < cbytes; n += 4) writel(*p32++, USBD_TXDATA(udc->udp_baseaddr)); /* Handle any remaining bytes */ bl = bytes - cbytes; if (bl) { tmp = 0; for (n = 0; n < bl; n++) tmp |= data[cbytes + n] << (n * 8); writel(tmp, USBD_TXDATA(udc->udp_baseaddr)); } break; case 1: /* 8-bit aligned */ case 3: /* Each byte has to be handled independently */ for (n = 0; n < bytes; n += 4) { bl = bytes - n; if (bl > 4) bl = 4; tmp = 0; for (i = 0; i < bl; i++) tmp |= data[n + i] << (i * 8); writel(tmp, USBD_TXDATA(udc->udp_baseaddr)); } break; case 2: /* 16-bit aligned */ p16 = (u16 *) data; cbytes = (bytes & ~0x3); /* Copy 32-bit aligned data first */ for (n = 0; n < cbytes; n += 4) { tmp = *p16++ & 0xFFFF; tmp |= (*p16++ & 0xFFFF) << 16; writel(tmp, USBD_TXDATA(udc->udp_baseaddr)); } /* Handle any remaining bytes */ bl = bytes - cbytes; if (bl) { tmp = 0; for (n = 0; n < bl; n++) tmp |= data[cbytes + n] << (n * 8); writel(tmp, USBD_TXDATA(udc->udp_baseaddr)); } break; } } /* Write data to the FIFO for an endpoint. This function is for endpoints (such * as EP0) that don't use DMA. Note that the endpoint must be selected in the * protocol engine prior to this call. */ static void udc_write_hwep(struct lpc32xx_udc *udc, u32 hwep, u32 *data, u32 bytes) { u32 hwwep = ((hwep & 0x1E) << 1) | CTRL_WR_EN; if ((bytes > 0) && (data == NULL)) return; /* Setup write of endpoint */ writel(hwwep, USBD_CTRL(udc->udp_baseaddr)); writel(bytes, USBD_TXPLEN(udc->udp_baseaddr)); /* Need at least 1 byte to trigger TX */ if (bytes == 0) writel(0, USBD_TXDATA(udc->udp_baseaddr)); else udc_stuff_fifo(udc, (u8 *) data, bytes); writel(((hwep & 0x1E) << 1), USBD_CTRL(udc->udp_baseaddr)); udc_val_buffer_hwep(udc, hwep); } /* USB device reset - resets USB to a default state with just EP0 enabled */ static void uda_usb_reset(struct lpc32xx_udc *udc) { u32 i = 0; /* Re-init device controller and EP0 */ udc_enable(udc); udc->gadget.speed = USB_SPEED_FULL; for (i = 1; i < NUM_ENDPOINTS; i++) { struct lpc32xx_ep *ep = &udc->ep[i]; ep->req_pending = 0; } } /* Send a ZLP on EP0 */ static void udc_ep0_send_zlp(struct lpc32xx_udc *udc) { udc_write_hwep(udc, EP_IN, NULL, 0); } /* Get current frame number */ static u16 udc_get_current_frame(struct lpc32xx_udc *udc) { u16 flo, fhi; udc_protocol_cmd_w(udc, CMD_RD_FRAME); flo = (u16) udc_protocol_cmd_r(udc, DAT_RD_FRAME); fhi = (u16) udc_protocol_cmd_r(udc, DAT_RD_FRAME); return (fhi << 8) | flo; } /* Set the device as configured - enables all endpoints */ static inline void udc_set_device_configured(struct lpc32xx_udc *udc) { udc_protocol_cmd_data_w(udc, CMD_CFG_DEV, DAT_WR_BYTE(CONF_DVICE)); } /* Set the device as unconfigured - disables all endpoints */ static inline void udc_set_device_unconfigured(struct lpc32xx_udc *udc) { udc_protocol_cmd_data_w(udc, CMD_CFG_DEV, DAT_WR_BYTE(0)); } /* reinit == restore initial software state */ static void udc_reinit(struct lpc32xx_udc *udc) { u32 i; INIT_LIST_HEAD(&udc->gadget.ep_list); INIT_LIST_HEAD(&udc->gadget.ep0->ep_list); for (i = 0; i < NUM_ENDPOINTS; i++) { struct lpc32xx_ep *ep = &udc->ep[i]; if (i != 0) list_add_tail(&ep->ep.ep_list, &udc->gadget.ep_list); usb_ep_set_maxpacket_limit(&ep->ep, ep->maxpacket); INIT_LIST_HEAD(&ep->queue); ep->req_pending = 0; } udc->ep0state = WAIT_FOR_SETUP; } /* Must be called with lock */ static void done(struct lpc32xx_ep *ep, struct lpc32xx_request *req, int status) { struct lpc32xx_udc *udc = ep->udc; list_del_init(&req->queue); if (req->req.status == -EINPROGRESS) req->req.status = status; else status = req->req.status; if (ep->lep) { usb_gadget_unmap_request(&udc->gadget, &req->req, ep->is_in); /* Free DDs */ udc_dd_free(udc, req->dd_desc_ptr); } if (status && status != -ESHUTDOWN) ep_dbg(ep, "%s done %p, status %d\n", ep->ep.name, req, status); ep->req_pending = 0; spin_unlock(&udc->lock); usb_gadget_giveback_request(&ep->ep, &req->req); spin_lock(&udc->lock); } /* Must be called with lock */ static void nuke(struct lpc32xx_ep *ep, int status) { struct lpc32xx_request *req; while (!list_empty(&ep->queue)) { req = list_entry(ep->queue.next, struct lpc32xx_request, queue); done(ep, req, status); } if (status == -ESHUTDOWN) { uda_disable_hwepint(ep->udc, ep->hwep_num); udc_disable_hwep(ep->udc, ep->hwep_num); } } /* IN endpoint 0 transfer */ static int udc_ep0_in_req(struct lpc32xx_udc *udc) { struct lpc32xx_request *req; struct lpc32xx_ep *ep0 = &udc->ep[0]; u32 tsend, ts = 0; if (list_empty(&ep0->queue)) /* Nothing to send */ return 0; else req = list_entry(ep0->queue.next, struct lpc32xx_request, queue); tsend = ts = req->req.length - req->req.actual; if (ts == 0) { /* Send a ZLP */ udc_ep0_send_zlp(udc); done(ep0, req, 0); return 1; } else if (ts > ep0->ep.maxpacket) ts = ep0->ep.maxpacket; /* Just send what we can */ /* Write data to the EP0 FIFO and start transfer */ udc_write_hwep(udc, EP_IN, (req->req.buf + req->req.actual), ts); /* Increment data pointer */ req->req.actual += ts; if (tsend >= ep0->ep.maxpacket) return 0; /* Stay in data transfer state */ /* Transfer request is complete */ udc->ep0state = WAIT_FOR_SETUP; done(ep0, req, 0); return 1; } /* OUT endpoint 0 transfer */ static int udc_ep0_out_req(struct lpc32xx_udc *udc) { struct lpc32xx_request *req; struct lpc32xx_ep *ep0 = &udc->ep[0]; u32 tr, bufferspace; if (list_empty(&ep0->queue)) return 0; else req = list_entry(ep0->queue.next, struct lpc32xx_request, queue); if (req) { if (req->req.length == 0) { /* Just dequeue request */ done(ep0, req, 0); udc->ep0state = WAIT_FOR_SETUP; return 1; } /* Get data from FIFO */ bufferspace = req->req.length - req->req.actual; if (bufferspace > ep0->ep.maxpacket) bufferspace = ep0->ep.maxpacket; /* Copy data to buffer */ prefetchw(req->req.buf + req->req.actual); tr = udc_read_hwep(udc, EP_OUT, req->req.buf + req->req.actual, bufferspace); req->req.actual += bufferspace; if (tr < ep0->ep.maxpacket) { /* This is the last packet */ done(ep0, req, 0); udc->ep0state = WAIT_FOR_SETUP; return 1; } } return 0; } /* Must be called with lock */ static void stop_activity(struct lpc32xx_udc *udc) { struct usb_gadget_driver *driver = udc->driver; int i; if (udc->gadget.speed == USB_SPEED_UNKNOWN) driver = NULL; udc->gadget.speed = USB_SPEED_UNKNOWN; udc->suspended = 0; for (i = 0; i < NUM_ENDPOINTS; i++) { struct lpc32xx_ep *ep = &udc->ep[i]; nuke(ep, -ESHUTDOWN); } if (driver) { spin_unlock(&udc->lock); driver->disconnect(&udc->gadget); spin_lock(&udc->lock); } isp1301_pullup_enable(udc, 0, 0); udc_disable(udc); udc_reinit(udc); } /* * Activate or kill host pullup * Can be called with or without lock */ static void pullup(struct lpc32xx_udc *udc, int is_on) { if (!udc->clocked) return; if (!udc->enabled || !udc->vbus) is_on = 0; if (is_on != udc->pullup) isp1301_pullup_enable(udc, is_on, 0); } /* Must be called without lock */ static int lpc32xx_ep_disable(struct usb_ep *_ep) { struct lpc32xx_ep *ep = container_of(_ep, struct lpc32xx_ep, ep); struct lpc32xx_udc *udc = ep->udc; unsigned long flags; if ((ep->hwep_num_base == 0) || (ep->hwep_num == 0)) return -EINVAL; spin_lock_irqsave(&udc->lock, flags); nuke(ep, -ESHUTDOWN); /* Clear all DMA statuses for this EP */ udc_ep_dma_disable(udc, ep->hwep_num); writel(1 << ep->hwep_num, USBD_EOTINTCLR(udc->udp_baseaddr)); writel(1 << ep->hwep_num, USBD_NDDRTINTCLR(udc->udp_baseaddr)); writel(1 << ep->hwep_num, USBD_SYSERRTINTCLR(udc->udp_baseaddr)); writel(1 << ep->hwep_num, USBD_DMARCLR(udc->udp_baseaddr)); /* Remove the DD pointer in the UDCA */ udc->udca_v_base[ep->hwep_num] = 0; /* Disable and reset endpoint and interrupt */ uda_clear_hwepint(udc, ep->hwep_num); udc_unrealize_hwep(udc, ep->hwep_num); ep->hwep_num = 0; spin_unlock_irqrestore(&udc->lock, flags); atomic_dec(&udc->enabled_ep_cnt); wake_up(&udc->ep_disable_wait_queue); return 0; } /* Must be called without lock */ static int lpc32xx_ep_enable(struct usb_ep *_ep, const struct usb_endpoint_descriptor *desc) { struct lpc32xx_ep *ep = container_of(_ep, struct lpc32xx_ep, ep); struct lpc32xx_udc *udc; u16 maxpacket; u32 tmp; unsigned long flags; /* Verify EP data */ if ((!_ep) || (!ep) || (!desc) || (desc->bDescriptorType != USB_DT_ENDPOINT)) return -EINVAL; udc = ep->udc; maxpacket = usb_endpoint_maxp(desc); if ((maxpacket == 0) || (maxpacket > ep->maxpacket)) { dev_dbg(udc->dev, "bad ep descriptor's packet size\n"); return -EINVAL; } /* Don't touch EP0 */ if (ep->hwep_num_base == 0) { dev_dbg(udc->dev, "Can't re-enable EP0!!!\n"); return -EINVAL; } /* Is driver ready? */ if ((!udc->driver) || (udc->gadget.speed == USB_SPEED_UNKNOWN)) { dev_dbg(udc->dev, "bogus device state\n"); return -ESHUTDOWN; } tmp = desc->bmAttributes & USB_ENDPOINT_XFERTYPE_MASK; switch (tmp) { case USB_ENDPOINT_XFER_CONTROL: return -EINVAL; case USB_ENDPOINT_XFER_INT: if (maxpacket > ep->maxpacket) { dev_dbg(udc->dev, "Bad INT endpoint maxpacket %d\n", maxpacket); return -EINVAL; } break; case USB_ENDPOINT_XFER_BULK: switch (maxpacket) { case 8: case 16: case 32: case 64: break; default: dev_dbg(udc->dev, "Bad BULK endpoint maxpacket %d\n", maxpacket); return -EINVAL; } break; case USB_ENDPOINT_XFER_ISOC: break; } spin_lock_irqsave(&udc->lock, flags); /* Initialize endpoint to match the selected descriptor */ ep->is_in = (desc->bEndpointAddress & USB_DIR_IN) != 0; ep->ep.maxpacket = maxpacket; /* Map hardware endpoint from base and direction */ if (ep->is_in) /* IN endpoints are offset 1 from the OUT endpoint */ ep->hwep_num = ep->hwep_num_base + EP_IN; else ep->hwep_num = ep->hwep_num_base; ep_dbg(ep, "EP enabled: %s, HW:%d, MP:%d IN:%d\n", ep->ep.name, ep->hwep_num, maxpacket, (ep->is_in == 1)); /* Realize the endpoint, interrupt is enabled later when * buffers are queued, IN EPs will NAK until buffers are ready */ udc_realize_hwep(udc, ep->hwep_num, ep->ep.maxpacket); udc_clr_buffer_hwep(udc, ep->hwep_num); uda_disable_hwepint(udc, ep->hwep_num); udc_clrstall_hwep(udc, ep->hwep_num); /* Clear all DMA statuses for this EP */ udc_ep_dma_disable(udc, ep->hwep_num); writel(1 << ep->hwep_num, USBD_EOTINTCLR(udc->udp_baseaddr)); writel(1 << ep->hwep_num, USBD_NDDRTINTCLR(udc->udp_baseaddr)); writel(1 << ep->hwep_num, USBD_SYSERRTINTCLR(udc->udp_baseaddr)); writel(1 << ep->hwep_num, USBD_DMARCLR(udc->udp_baseaddr)); spin_unlock_irqrestore(&udc->lock, flags); atomic_inc(&udc->enabled_ep_cnt); return 0; } /* * Allocate a USB request list * Can be called with or without lock */ static struct usb_request *lpc32xx_ep_alloc_request(struct usb_ep *_ep, gfp_t gfp_flags) { struct lpc32xx_request *req; req = kzalloc(sizeof(struct lpc32xx_request), gfp_flags); if (!req) return NULL; INIT_LIST_HEAD(&req->queue); return &req->req; } /* * De-allocate a USB request list * Can be called with or without lock */ static void lpc32xx_ep_free_request(struct usb_ep *_ep, struct usb_request *_req) { struct lpc32xx_request *req; req = container_of(_req, struct lpc32xx_request, req); BUG_ON(!list_empty(&req->queue)); kfree(req); } /* Must be called without lock */ static int lpc32xx_ep_queue(struct usb_ep *_ep, struct usb_request *_req, gfp_t gfp_flags) { struct lpc32xx_request *req; struct lpc32xx_ep *ep; struct lpc32xx_udc *udc; unsigned long flags; int status = 0; req = container_of(_req, struct lpc32xx_request, req); ep = container_of(_ep, struct lpc32xx_ep, ep); if (!_ep || !_req || !_req->complete || !_req->buf || !list_empty(&req->queue)) return -EINVAL; udc = ep->udc; if (udc->gadget.speed == USB_SPEED_UNKNOWN) return -EPIPE; if (ep->lep) { struct lpc32xx_usbd_dd_gad *dd; status = usb_gadget_map_request(&udc->gadget, _req, ep->is_in); if (status) return status; /* For the request, build a list of DDs */ dd = udc_dd_alloc(udc); if (!dd) { /* Error allocating DD */ return -ENOMEM; } req->dd_desc_ptr = dd; /* Setup the DMA descriptor */ dd->dd_next_phy = dd->dd_next_v = 0; dd->dd_buffer_addr = req->req.dma; dd->dd_status = 0; /* Special handling for ISO EPs */ if (ep->eptype == EP_ISO_TYPE) { dd->dd_setup = DD_SETUP_ISO_EP | DD_SETUP_PACKETLEN(0) | DD_SETUP_DMALENBYTES(1); dd->dd_iso_ps_mem_addr = dd->this_dma + 24; if (ep->is_in) dd->iso_status[0] = req->req.length; else dd->iso_status[0] = 0; } else dd->dd_setup = DD_SETUP_PACKETLEN(ep->ep.maxpacket) | DD_SETUP_DMALENBYTES(req->req.length); } ep_dbg(ep, "%s queue req %p len %d buf %p (in=%d) z=%d\n", _ep->name, _req, _req->length, _req->buf, ep->is_in, _req->zero); spin_lock_irqsave(&udc->lock, flags); _req->status = -EINPROGRESS; _req->actual = 0; req->send_zlp = _req->zero; /* Kickstart empty queues */ if (list_empty(&ep->queue)) { list_add_tail(&req->queue, &ep->queue); if (ep->hwep_num_base == 0) { /* Handle expected data direction */ if (ep->is_in) { /* IN packet to host */ udc->ep0state = DATA_IN; status = udc_ep0_in_req(udc); } else { /* OUT packet from host */ udc->ep0state = DATA_OUT; status = udc_ep0_out_req(udc); } } else if (ep->is_in) { /* IN packet to host and kick off transfer */ if (!ep->req_pending) udc_ep_in_req_dma(udc, ep); } else /* OUT packet from host and kick off list */ if (!ep->req_pending) udc_ep_out_req_dma(udc, ep); } else list_add_tail(&req->queue, &ep->queue); spin_unlock_irqrestore(&udc->lock, flags); return (status < 0) ? status : 0; } /* Must be called without lock */ static int lpc32xx_ep_dequeue(struct usb_ep *_ep, struct usb_request *_req) { struct lpc32xx_ep *ep; struct lpc32xx_request *req; unsigned long flags; ep = container_of(_ep, struct lpc32xx_ep, ep); if (!_ep || ep->hwep_num_base == 0) return -EINVAL; spin_lock_irqsave(&ep->udc->lock, flags); /* make sure it's actually queued on this endpoint */ list_for_each_entry(req, &ep->queue, queue) { if (&req->req == _req) break; } if (&req->req != _req) { spin_unlock_irqrestore(&ep->udc->lock, flags); return -EINVAL; } done(ep, req, -ECONNRESET); spin_unlock_irqrestore(&ep->udc->lock, flags); return 0; } /* Must be called without lock */ static int lpc32xx_ep_set_halt(struct usb_ep *_ep, int value) { struct lpc32xx_ep *ep = container_of(_ep, struct lpc32xx_ep, ep); struct lpc32xx_udc *udc; unsigned long flags; if ((!ep) || (ep->hwep_num <= 1)) return -EINVAL; /* Don't halt an IN EP */ if (ep->is_in) return -EAGAIN; udc = ep->udc; spin_lock_irqsave(&udc->lock, flags); if (value == 1) { /* stall */ udc_protocol_cmd_data_w(udc, CMD_SET_EP_STAT(ep->hwep_num), DAT_WR_BYTE(EP_STAT_ST)); } else { /* End stall */ ep->wedge = 0; udc_protocol_cmd_data_w(udc, CMD_SET_EP_STAT(ep->hwep_num), DAT_WR_BYTE(0)); } spin_unlock_irqrestore(&udc->lock, flags); return 0; } /* set the halt feature and ignores clear requests */ static int lpc32xx_ep_set_wedge(struct usb_ep *_ep) { struct lpc32xx_ep *ep = container_of(_ep, struct lpc32xx_ep, ep); if (!_ep || !ep->udc) return -EINVAL; ep->wedge = 1; return usb_ep_set_halt(_ep); } static const struct usb_ep_ops lpc32xx_ep_ops = { .enable = lpc32xx_ep_enable, .disable = lpc32xx_ep_disable, .alloc_request = lpc32xx_ep_alloc_request, .free_request = lpc32xx_ep_free_request, .queue = lpc32xx_ep_queue, .dequeue = lpc32xx_ep_dequeue, .set_halt = lpc32xx_ep_set_halt, .set_wedge = lpc32xx_ep_set_wedge, }; /* Send a ZLP on a non-0 IN EP */ void udc_send_in_zlp(struct lpc32xx_udc *udc, struct lpc32xx_ep *ep) { /* Clear EP status */ udc_clearep_getsts(udc, ep->hwep_num); /* Send ZLP via FIFO mechanism */ udc_write_hwep(udc, ep->hwep_num, NULL, 0); } /* * Handle EP completion for ZLP * This function will only be called when a delayed ZLP needs to be sent out * after a DMA transfer has filled both buffers. */ void udc_handle_eps(struct lpc32xx_udc *udc, struct lpc32xx_ep *ep) { u32 epstatus; struct lpc32xx_request *req; if (ep->hwep_num <= 0) return; uda_clear_hwepint(udc, ep->hwep_num); /* If this interrupt isn't enabled, return now */ if (!(udc->enabled_hwepints & (1 << ep->hwep_num))) return; /* Get endpoint status */ epstatus = udc_clearep_getsts(udc, ep->hwep_num); /* * This should never happen, but protect against writing to the * buffer when full. */ if (epstatus & EP_SEL_F) return; if (ep->is_in) { udc_send_in_zlp(udc, ep); uda_disable_hwepint(udc, ep->hwep_num); } else return; /* If there isn't a request waiting, something went wrong */ req = list_entry(ep->queue.next, struct lpc32xx_request, queue); if (req) { done(ep, req, 0); /* Start another request if ready */ if (!list_empty(&ep->queue)) { if (ep->is_in) udc_ep_in_req_dma(udc, ep); else udc_ep_out_req_dma(udc, ep); } else ep->req_pending = 0; } } /* DMA end of transfer completion */ static void udc_handle_dma_ep(struct lpc32xx_udc *udc, struct lpc32xx_ep *ep) { u32 status, epstatus; struct lpc32xx_request *req; struct lpc32xx_usbd_dd_gad *dd; #ifdef CONFIG_USB_GADGET_DEBUG_FILES ep->totalints++; #endif req = list_entry(ep->queue.next, struct lpc32xx_request, queue); if (!req) { ep_err(ep, "DMA interrupt on no req!\n"); return; } dd = req->dd_desc_ptr; /* DMA descriptor should always be retired for this call */ if (!(dd->dd_status & DD_STATUS_DD_RETIRED)) ep_warn(ep, "DMA descriptor did not retire\n"); /* Disable DMA */ udc_ep_dma_disable(udc, ep->hwep_num); writel((1 << ep->hwep_num), USBD_EOTINTCLR(udc->udp_baseaddr)); writel((1 << ep->hwep_num), USBD_NDDRTINTCLR(udc->udp_baseaddr)); /* System error? */ if (readl(USBD_SYSERRTINTST(udc->udp_baseaddr)) & (1 << ep->hwep_num)) { writel((1 << ep->hwep_num), USBD_SYSERRTINTCLR(udc->udp_baseaddr)); ep_err(ep, "AHB critical error!\n"); ep->req_pending = 0; /* The error could have occurred on a packet of a multipacket * transfer, so recovering the transfer is not possible. Close * the request with an error */ done(ep, req, -ECONNABORTED); return; } /* Handle the current DD's status */ status = dd->dd_status; switch (status & DD_STATUS_STS_MASK) { case DD_STATUS_STS_NS: /* DD not serviced? This shouldn't happen! */ ep->req_pending = 0; ep_err(ep, "DMA critical EP error: DD not serviced (0x%x)!\n", status); done(ep, req, -ECONNABORTED); return; case DD_STATUS_STS_BS: /* Interrupt only fires on EOT - This shouldn't happen! */ ep->req_pending = 0; ep_err(ep, "DMA critical EP error: EOT prior to service completion (0x%x)!\n", status); done(ep, req, -ECONNABORTED); return; case DD_STATUS_STS_NC: case DD_STATUS_STS_DUR: /* Really just a short packet, not an underrun */ /* This is a good status and what we expect */ break; default: /* Data overrun, system error, or unknown */ ep->req_pending = 0; ep_err(ep, "DMA critical EP error: System error (0x%x)!\n", status); done(ep, req, -ECONNABORTED); return; } /* ISO endpoints are handled differently */ if (ep->eptype == EP_ISO_TYPE) { if (ep->is_in) req->req.actual = req->req.length; else req->req.actual = dd->iso_status[0] & 0xFFFF; } else req->req.actual += DD_STATUS_CURDMACNT(status); /* Send a ZLP if necessary. This will be done for non-int * packets which have a size that is a divisor of MAXP */ if (req->send_zlp) { /* * If at least 1 buffer is available, send the ZLP now. * Otherwise, the ZLP send needs to be deferred until a * buffer is available. */ if (udc_clearep_getsts(udc, ep->hwep_num) & EP_SEL_F) { udc_clearep_getsts(udc, ep->hwep_num); uda_enable_hwepint(udc, ep->hwep_num); epstatus = udc_clearep_getsts(udc, ep->hwep_num); /* Let the EP interrupt handle the ZLP */ return; } else udc_send_in_zlp(udc, ep); } /* Transfer request is complete */ done(ep, req, 0); /* Start another request if ready */ udc_clearep_getsts(udc, ep->hwep_num); if (!list_empty((&ep->queue))) { if (ep->is_in) udc_ep_in_req_dma(udc, ep); else udc_ep_out_req_dma(udc, ep); } else ep->req_pending = 0; } /* * * Endpoint 0 functions * */ static void udc_handle_dev(struct lpc32xx_udc *udc) { u32 tmp; udc_protocol_cmd_w(udc, CMD_GET_DEV_STAT); tmp = udc_protocol_cmd_r(udc, DAT_GET_DEV_STAT); if (tmp & DEV_RST) uda_usb_reset(udc); else if (tmp & DEV_CON_CH) uda_power_event(udc, (tmp & DEV_CON)); else if (tmp & DEV_SUS_CH) { if (tmp & DEV_SUS) { if (udc->vbus == 0) stop_activity(udc); else if ((udc->gadget.speed != USB_SPEED_UNKNOWN) && udc->driver) { /* Power down transceiver */ udc->poweron = 0; schedule_work(&udc->pullup_job); uda_resm_susp_event(udc, 1); } } else if ((udc->gadget.speed != USB_SPEED_UNKNOWN) && udc->driver && udc->vbus) { uda_resm_susp_event(udc, 0); /* Power up transceiver */ udc->poweron = 1; schedule_work(&udc->pullup_job); } } } static int udc_get_status(struct lpc32xx_udc *udc, u16 reqtype, u16 wIndex) { struct lpc32xx_ep *ep; u32 ep0buff = 0, tmp; switch (reqtype & USB_RECIP_MASK) { case USB_RECIP_INTERFACE: break; /* Not supported */ case USB_RECIP_DEVICE: ep0buff = udc->gadget.is_selfpowered; if (udc->dev_status & (1 << USB_DEVICE_REMOTE_WAKEUP)) ep0buff |= (1 << USB_DEVICE_REMOTE_WAKEUP); break; case USB_RECIP_ENDPOINT: tmp = wIndex & USB_ENDPOINT_NUMBER_MASK; ep = &udc->ep[tmp]; if ((tmp == 0) || (tmp >= NUM_ENDPOINTS)) return -EOPNOTSUPP; if (wIndex & USB_DIR_IN) { if (!ep->is_in) return -EOPNOTSUPP; /* Something's wrong */ } else if (ep->is_in) return -EOPNOTSUPP; /* Not an IN endpoint */ /* Get status of the endpoint */ udc_protocol_cmd_w(udc, CMD_SEL_EP(ep->hwep_num)); tmp = udc_protocol_cmd_r(udc, DAT_SEL_EP(ep->hwep_num)); if (tmp & EP_SEL_ST) ep0buff = (1 << USB_ENDPOINT_HALT); else ep0buff = 0; break; default: break; } /* Return data */ udc_write_hwep(udc, EP_IN, &ep0buff, 2); return 0; } static void udc_handle_ep0_setup(struct lpc32xx_udc *udc) { struct lpc32xx_ep *ep, *ep0 = &udc->ep[0]; struct usb_ctrlrequest ctrlpkt; int i, bytes; u16 wIndex, wValue, wLength, reqtype, req, tmp; /* Nuke previous transfers */ nuke(ep0, -EPROTO); /* Get setup packet */ bytes = udc_read_hwep(udc, EP_OUT, (u32 *) &ctrlpkt, 8); if (bytes != 8) { ep_warn(ep0, "Incorrectly sized setup packet (s/b 8, is %d)!\n", bytes); return; } /* Native endianness */ wIndex = le16_to_cpu(ctrlpkt.wIndex); wValue = le16_to_cpu(ctrlpkt.wValue); wLength = le16_to_cpu(ctrlpkt.wLength); reqtype = le16_to_cpu(ctrlpkt.bRequestType); /* Set direction of EP0 */ if (likely(reqtype & USB_DIR_IN)) ep0->is_in = 1; else ep0->is_in = 0; /* Handle SETUP packet */ req = le16_to_cpu(ctrlpkt.bRequest); switch (req) { case USB_REQ_CLEAR_FEATURE: case USB_REQ_SET_FEATURE: switch (reqtype) { case (USB_TYPE_STANDARD | USB_RECIP_DEVICE): if (wValue != USB_DEVICE_REMOTE_WAKEUP) goto stall; /* Nothing else handled */ /* Tell board about event */ if (req == USB_REQ_CLEAR_FEATURE) udc->dev_status &= ~(1 << USB_DEVICE_REMOTE_WAKEUP); else udc->dev_status |= (1 << USB_DEVICE_REMOTE_WAKEUP); uda_remwkp_cgh(udc); goto zlp_send; case (USB_TYPE_STANDARD | USB_RECIP_ENDPOINT): tmp = wIndex & USB_ENDPOINT_NUMBER_MASK; if ((wValue != USB_ENDPOINT_HALT) || (tmp >= NUM_ENDPOINTS)) break; /* Find hardware endpoint from logical endpoint */ ep = &udc->ep[tmp]; tmp = ep->hwep_num; if (tmp == 0) break; if (req == USB_REQ_SET_FEATURE) udc_stall_hwep(udc, tmp); else if (!ep->wedge) udc_clrstall_hwep(udc, tmp); goto zlp_send; default: break; } case USB_REQ_SET_ADDRESS: if (reqtype == (USB_TYPE_STANDARD | USB_RECIP_DEVICE)) { udc_set_address(udc, wValue); goto zlp_send; } break; case USB_REQ_GET_STATUS: udc_get_status(udc, reqtype, wIndex); return; default: break; /* Let GadgetFS handle the descriptor instead */ } if (likely(udc->driver)) { /* device-2-host (IN) or no data setup command, process * immediately */ spin_unlock(&udc->lock); i = udc->driver->setup(&udc->gadget, &ctrlpkt); spin_lock(&udc->lock); if (req == USB_REQ_SET_CONFIGURATION) { /* Configuration is set after endpoints are realized */ if (wValue) { /* Set configuration */ udc_set_device_configured(udc); udc_protocol_cmd_data_w(udc, CMD_SET_MODE, DAT_WR_BYTE(AP_CLK | INAK_BI | INAK_II)); } else { /* Clear configuration */ udc_set_device_unconfigured(udc); /* Disable NAK interrupts */ udc_protocol_cmd_data_w(udc, CMD_SET_MODE, DAT_WR_BYTE(AP_CLK)); } } if (i < 0) { /* setup processing failed, force stall */ dev_dbg(udc->dev, "req %02x.%02x protocol STALL; stat %d\n", reqtype, req, i); udc->ep0state = WAIT_FOR_SETUP; goto stall; } } if (!ep0->is_in) udc_ep0_send_zlp(udc); /* ZLP IN packet on data phase */ return; stall: udc_stall_hwep(udc, EP_IN); return; zlp_send: udc_ep0_send_zlp(udc); return; } /* IN endpoint 0 transfer */ static void udc_handle_ep0_in(struct lpc32xx_udc *udc) { struct lpc32xx_ep *ep0 = &udc->ep[0]; u32 epstatus; /* Clear EP interrupt */ epstatus = udc_clearep_getsts(udc, EP_IN); #ifdef CONFIG_USB_GADGET_DEBUG_FILES ep0->totalints++; #endif /* Stalled? Clear stall and reset buffers */ if (epstatus & EP_SEL_ST) { udc_clrstall_hwep(udc, EP_IN); nuke(ep0, -ECONNABORTED); udc->ep0state = WAIT_FOR_SETUP; return; } /* Is a buffer available? */ if (!(epstatus & EP_SEL_F)) { /* Handle based on current state */ if (udc->ep0state == DATA_IN) udc_ep0_in_req(udc); else { /* Unknown state for EP0 oe end of DATA IN phase */ nuke(ep0, -ECONNABORTED); udc->ep0state = WAIT_FOR_SETUP; } } } /* OUT endpoint 0 transfer */ static void udc_handle_ep0_out(struct lpc32xx_udc *udc) { struct lpc32xx_ep *ep0 = &udc->ep[0]; u32 epstatus; /* Clear EP interrupt */ epstatus = udc_clearep_getsts(udc, EP_OUT); #ifdef CONFIG_USB_GADGET_DEBUG_FILES ep0->totalints++; #endif /* Stalled? */ if (epstatus & EP_SEL_ST) { udc_clrstall_hwep(udc, EP_OUT); nuke(ep0, -ECONNABORTED); udc->ep0state = WAIT_FOR_SETUP; return; } /* A NAK may occur if a packet couldn't be received yet */ if (epstatus & EP_SEL_EPN) return; /* Setup packet incoming? */ if (epstatus & EP_SEL_STP) { nuke(ep0, 0); udc->ep0state = WAIT_FOR_SETUP; } /* Data available? */ if (epstatus & EP_SEL_F) /* Handle based on current state */ switch (udc->ep0state) { case WAIT_FOR_SETUP: udc_handle_ep0_setup(udc); break; case DATA_OUT: udc_ep0_out_req(udc); break; default: /* Unknown state for EP0 */ nuke(ep0, -ECONNABORTED); udc->ep0state = WAIT_FOR_SETUP; } } /* Must be called without lock */ static int lpc32xx_get_frame(struct usb_gadget *gadget) { int frame; unsigned long flags; struct lpc32xx_udc *udc = to_udc(gadget); if (!udc->clocked) return -EINVAL; spin_lock_irqsave(&udc->lock, flags); frame = (int) udc_get_current_frame(udc); spin_unlock_irqrestore(&udc->lock, flags); return frame; } static int lpc32xx_wakeup(struct usb_gadget *gadget) { return -ENOTSUPP; } static int lpc32xx_set_selfpowered(struct usb_gadget *gadget, int is_on) { gadget->is_selfpowered = (is_on != 0); return 0; } /* * vbus is here! turn everything on that's ready * Must be called without lock */ static int lpc32xx_vbus_session(struct usb_gadget *gadget, int is_active) { unsigned long flags; struct lpc32xx_udc *udc = to_udc(gadget); spin_lock_irqsave(&udc->lock, flags); /* Doesn't need lock */ if (udc->driver) { udc_clk_set(udc, 1); udc_enable(udc); pullup(udc, is_active); } else { stop_activity(udc); pullup(udc, 0); spin_unlock_irqrestore(&udc->lock, flags); /* * Wait for all the endpoints to disable, * before disabling clocks. Don't wait if * endpoints are not enabled. */ if (atomic_read(&udc->enabled_ep_cnt)) wait_event_interruptible(udc->ep_disable_wait_queue, (atomic_read(&udc->enabled_ep_cnt) == 0)); spin_lock_irqsave(&udc->lock, flags); udc_clk_set(udc, 0); } spin_unlock_irqrestore(&udc->lock, flags); return 0; } /* Can be called with or without lock */ static int lpc32xx_pullup(struct usb_gadget *gadget, int is_on) { struct lpc32xx_udc *udc = to_udc(gadget); /* Doesn't need lock */ pullup(udc, is_on); return 0; } static int lpc32xx_start(struct usb_gadget *, struct usb_gadget_driver *); static int lpc32xx_stop(struct usb_gadget *); static const struct usb_gadget_ops lpc32xx_udc_ops = { .get_frame = lpc32xx_get_frame, .wakeup = lpc32xx_wakeup, .set_selfpowered = lpc32xx_set_selfpowered, .vbus_session = lpc32xx_vbus_session, .pullup = lpc32xx_pullup, .udc_start = lpc32xx_start, .udc_stop = lpc32xx_stop, }; static void nop_release(struct device *dev) { /* nothing to free */ } static const struct lpc32xx_udc controller_template = { .gadget = { .ops = &lpc32xx_udc_ops, .name = driver_name, .dev = { .init_name = "gadget", .release = nop_release, } }, .ep[0] = { .ep = { .name = "ep0", .ops = &lpc32xx_ep_ops, .caps = USB_EP_CAPS(USB_EP_CAPS_TYPE_CONTROL, USB_EP_CAPS_DIR_ALL), }, .maxpacket = 64, .hwep_num_base = 0, .hwep_num = 0, /* Can be 0 or 1, has special handling */ .lep = 0, .eptype = EP_CTL_TYPE, }, .ep[1] = { .ep = { .name = "ep1-int", .ops = &lpc32xx_ep_ops, .caps = USB_EP_CAPS(USB_EP_CAPS_TYPE_INT, USB_EP_CAPS_DIR_ALL), }, .maxpacket = 64, .hwep_num_base = 2, .hwep_num = 0, /* 2 or 3, will be set later */ .lep = 1, .eptype = EP_INT_TYPE, }, .ep[2] = { .ep = { .name = "ep2-bulk", .ops = &lpc32xx_ep_ops, .caps = USB_EP_CAPS(USB_EP_CAPS_TYPE_BULK, USB_EP_CAPS_DIR_ALL), }, .maxpacket = 64, .hwep_num_base = 4, .hwep_num = 0, /* 4 or 5, will be set later */ .lep = 2, .eptype = EP_BLK_TYPE, }, .ep[3] = { .ep = { .name = "ep3-iso", .ops = &lpc32xx_ep_ops, .caps = USB_EP_CAPS(USB_EP_CAPS_TYPE_ISO, USB_EP_CAPS_DIR_ALL), }, .maxpacket = 1023, .hwep_num_base = 6, .hwep_num = 0, /* 6 or 7, will be set later */ .lep = 3, .eptype = EP_ISO_TYPE, }, .ep[4] = { .ep = { .name = "ep4-int", .ops = &lpc32xx_ep_ops, .caps = USB_EP_CAPS(USB_EP_CAPS_TYPE_INT, USB_EP_CAPS_DIR_ALL), }, .maxpacket = 64, .hwep_num_base = 8, .hwep_num = 0, /* 8 or 9, will be set later */ .lep = 4, .eptype = EP_INT_TYPE, }, .ep[5] = { .ep = { .name = "ep5-bulk", .ops = &lpc32xx_ep_ops, .caps = USB_EP_CAPS(USB_EP_CAPS_TYPE_BULK, USB_EP_CAPS_DIR_ALL), }, .maxpacket = 64, .hwep_num_base = 10, .hwep_num = 0, /* 10 or 11, will be set later */ .lep = 5, .eptype = EP_BLK_TYPE, }, .ep[6] = { .ep = { .name = "ep6-iso", .ops = &lpc32xx_ep_ops, .caps = USB_EP_CAPS(USB_EP_CAPS_TYPE_ISO, USB_EP_CAPS_DIR_ALL), }, .maxpacket = 1023, .hwep_num_base = 12, .hwep_num = 0, /* 12 or 13, will be set later */ .lep = 6, .eptype = EP_ISO_TYPE, }, .ep[7] = { .ep = { .name = "ep7-int", .ops = &lpc32xx_ep_ops, .caps = USB_EP_CAPS(USB_EP_CAPS_TYPE_INT, USB_EP_CAPS_DIR_ALL), }, .maxpacket = 64, .hwep_num_base = 14, .hwep_num = 0, .lep = 7, .eptype = EP_INT_TYPE, }, .ep[8] = { .ep = { .name = "ep8-bulk", .ops = &lpc32xx_ep_ops, .caps = USB_EP_CAPS(USB_EP_CAPS_TYPE_BULK, USB_EP_CAPS_DIR_ALL), }, .maxpacket = 64, .hwep_num_base = 16, .hwep_num = 0, .lep = 8, .eptype = EP_BLK_TYPE, }, .ep[9] = { .ep = { .name = "ep9-iso", .ops = &lpc32xx_ep_ops, .caps = USB_EP_CAPS(USB_EP_CAPS_TYPE_ISO, USB_EP_CAPS_DIR_ALL), }, .maxpacket = 1023, .hwep_num_base = 18, .hwep_num = 0, .lep = 9, .eptype = EP_ISO_TYPE, }, .ep[10] = { .ep = { .name = "ep10-int", .ops = &lpc32xx_ep_ops, .caps = USB_EP_CAPS(USB_EP_CAPS_TYPE_INT, USB_EP_CAPS_DIR_ALL), }, .maxpacket = 64, .hwep_num_base = 20, .hwep_num = 0, .lep = 10, .eptype = EP_INT_TYPE, }, .ep[11] = { .ep = { .name = "ep11-bulk", .ops = &lpc32xx_ep_ops, .caps = USB_EP_CAPS(USB_EP_CAPS_TYPE_BULK, USB_EP_CAPS_DIR_ALL), }, .maxpacket = 64, .hwep_num_base = 22, .hwep_num = 0, .lep = 11, .eptype = EP_BLK_TYPE, }, .ep[12] = { .ep = { .name = "ep12-iso", .ops = &lpc32xx_ep_ops, .caps = USB_EP_CAPS(USB_EP_CAPS_TYPE_ISO, USB_EP_CAPS_DIR_ALL), }, .maxpacket = 1023, .hwep_num_base = 24, .hwep_num = 0, .lep = 12, .eptype = EP_ISO_TYPE, }, .ep[13] = { .ep = { .name = "ep13-int", .ops = &lpc32xx_ep_ops, .caps = USB_EP_CAPS(USB_EP_CAPS_TYPE_INT, USB_EP_CAPS_DIR_ALL), }, .maxpacket = 64, .hwep_num_base = 26, .hwep_num = 0, .lep = 13, .eptype = EP_INT_TYPE, }, .ep[14] = { .ep = { .name = "ep14-bulk", .ops = &lpc32xx_ep_ops, .caps = USB_EP_CAPS(USB_EP_CAPS_TYPE_BULK, USB_EP_CAPS_DIR_ALL), }, .maxpacket = 64, .hwep_num_base = 28, .hwep_num = 0, .lep = 14, .eptype = EP_BLK_TYPE, }, .ep[15] = { .ep = { .name = "ep15-bulk", .ops = &lpc32xx_ep_ops, .caps = USB_EP_CAPS(USB_EP_CAPS_TYPE_BULK, USB_EP_CAPS_DIR_ALL), }, .maxpacket = 1023, .hwep_num_base = 30, .hwep_num = 0, .lep = 15, .eptype = EP_BLK_TYPE, }, }; /* ISO and status interrupts */ static irqreturn_t lpc32xx_usb_lp_irq(int irq, void *_udc) { u32 tmp, devstat; struct lpc32xx_udc *udc = _udc; spin_lock(&udc->lock); /* Read the device status register */ devstat = readl(USBD_DEVINTST(udc->udp_baseaddr)); devstat &= ~USBD_EP_FAST; writel(devstat, USBD_DEVINTCLR(udc->udp_baseaddr)); devstat = devstat & udc->enabled_devints; /* Device specific handling needed? */ if (devstat & USBD_DEV_STAT) udc_handle_dev(udc); /* Start of frame? (devstat & FRAME_INT): * The frame interrupt isn't really needed for ISO support, * as the driver will queue the necessary packets */ /* Error? */ if (devstat & ERR_INT) { /* All types of errors, from cable removal during transfer to * misc protocol and bit errors. These are mostly for just info, * as the USB hardware will work around these. If these errors * happen alot, something is wrong. */ udc_protocol_cmd_w(udc, CMD_RD_ERR_STAT); tmp = udc_protocol_cmd_r(udc, DAT_RD_ERR_STAT); dev_dbg(udc->dev, "Device error (0x%x)!\n", tmp); } spin_unlock(&udc->lock); return IRQ_HANDLED; } /* EP interrupts */ static irqreturn_t lpc32xx_usb_hp_irq(int irq, void *_udc) { u32 tmp; struct lpc32xx_udc *udc = _udc; spin_lock(&udc->lock); /* Read the device status register */ writel(USBD_EP_FAST, USBD_DEVINTCLR(udc->udp_baseaddr)); /* Endpoints */ tmp = readl(USBD_EPINTST(udc->udp_baseaddr)); /* Special handling for EP0 */ if (tmp & (EP_MASK_SEL(0, EP_OUT) | EP_MASK_SEL(0, EP_IN))) { /* Handle EP0 IN */ if (tmp & (EP_MASK_SEL(0, EP_IN))) udc_handle_ep0_in(udc); /* Handle EP0 OUT */ if (tmp & (EP_MASK_SEL(0, EP_OUT))) udc_handle_ep0_out(udc); } /* All other EPs */ if (tmp & ~(EP_MASK_SEL(0, EP_OUT) | EP_MASK_SEL(0, EP_IN))) { int i; /* Handle other EP interrupts */ for (i = 1; i < NUM_ENDPOINTS; i++) { if (tmp & (1 << udc->ep[i].hwep_num)) udc_handle_eps(udc, &udc->ep[i]); } } spin_unlock(&udc->lock); return IRQ_HANDLED; } static irqreturn_t lpc32xx_usb_devdma_irq(int irq, void *_udc) { struct lpc32xx_udc *udc = _udc; int i; u32 tmp; spin_lock(&udc->lock); /* Handle EP DMA EOT interrupts */ tmp = readl(USBD_EOTINTST(udc->udp_baseaddr)) | (readl(USBD_EPDMAST(udc->udp_baseaddr)) & readl(USBD_NDDRTINTST(udc->udp_baseaddr))) | readl(USBD_SYSERRTINTST(udc->udp_baseaddr)); for (i = 1; i < NUM_ENDPOINTS; i++) { if (tmp & (1 << udc->ep[i].hwep_num)) udc_handle_dma_ep(udc, &udc->ep[i]); } spin_unlock(&udc->lock); return IRQ_HANDLED; } /* * * VBUS detection, pullup handler, and Gadget cable state notification * */ static void vbus_work(struct work_struct *work) { u8 value; struct lpc32xx_udc *udc = container_of(work, struct lpc32xx_udc, vbus_job); if (udc->enabled != 0) { /* Discharge VBUS real quick */ i2c_smbus_write_byte_data(udc->isp1301_i2c_client, ISP1301_I2C_OTG_CONTROL_1, OTG1_VBUS_DISCHRG); /* Give VBUS some time (100mS) to discharge */ msleep(100); /* Disable VBUS discharge resistor */ i2c_smbus_write_byte_data(udc->isp1301_i2c_client, ISP1301_I2C_OTG_CONTROL_1 | ISP1301_I2C_REG_CLEAR_ADDR, OTG1_VBUS_DISCHRG); /* Clear interrupt */ i2c_smbus_write_byte_data(udc->isp1301_i2c_client, ISP1301_I2C_INTERRUPT_LATCH | ISP1301_I2C_REG_CLEAR_ADDR, ~0); /* Get the VBUS status from the transceiver */ value = i2c_smbus_read_byte_data(udc->isp1301_i2c_client, ISP1301_I2C_INTERRUPT_SOURCE); /* VBUS on or off? */ if (value & INT_SESS_VLD) udc->vbus = 1; else udc->vbus = 0; /* VBUS changed? */ if (udc->last_vbus != udc->vbus) { udc->last_vbus = udc->vbus; lpc32xx_vbus_session(&udc->gadget, udc->vbus); } } /* Re-enable after completion */ enable_irq(udc->udp_irq[IRQ_USB_ATX]); } static irqreturn_t lpc32xx_usb_vbus_irq(int irq, void *_udc) { struct lpc32xx_udc *udc = _udc; /* Defer handling of VBUS IRQ to work queue */ disable_irq_nosync(udc->udp_irq[IRQ_USB_ATX]); schedule_work(&udc->vbus_job); return IRQ_HANDLED; } static int lpc32xx_start(struct usb_gadget *gadget, struct usb_gadget_driver *driver) { struct lpc32xx_udc *udc = to_udc(gadget); int i; if (!driver || driver->max_speed < USB_SPEED_FULL || !driver->setup) { dev_err(udc->dev, "bad parameter.\n"); return -EINVAL; } if (udc->driver) { dev_err(udc->dev, "UDC already has a gadget driver\n"); return -EBUSY; } udc->driver = driver; udc->gadget.dev.of_node = udc->dev->of_node; udc->enabled = 1; udc->gadget.is_selfpowered = 1; udc->vbus = 0; /* Force VBUS process once to check for cable insertion */ udc->last_vbus = udc->vbus = 0; schedule_work(&udc->vbus_job); /* Do not re-enable ATX IRQ (3) */ for (i = IRQ_USB_LP; i < IRQ_USB_ATX; i++) enable_irq(udc->udp_irq[i]); return 0; } static int lpc32xx_stop(struct usb_gadget *gadget) { int i; struct lpc32xx_udc *udc = to_udc(gadget); for (i = IRQ_USB_LP; i <= IRQ_USB_ATX; i++) disable_irq(udc->udp_irq[i]); if (udc->clocked) { spin_lock(&udc->lock); stop_activity(udc); spin_unlock(&udc->lock); /* * Wait for all the endpoints to disable, * before disabling clocks. Don't wait if * endpoints are not enabled. */ if (atomic_read(&udc->enabled_ep_cnt)) wait_event_interruptible(udc->ep_disable_wait_queue, (atomic_read(&udc->enabled_ep_cnt) == 0)); spin_lock(&udc->lock); udc_clk_set(udc, 0); spin_unlock(&udc->lock); } udc->enabled = 0; udc->driver = NULL; return 0; } static void lpc32xx_udc_shutdown(struct platform_device *dev) { /* Force disconnect on reboot */ struct lpc32xx_udc *udc = platform_get_drvdata(dev); pullup(udc, 0); } /* * Callbacks to be overridden by options passed via OF (TODO) */ static void lpc32xx_usbd_conn_chg(int conn) { /* Do nothing, it might be nice to enable an LED * based on conn state being !0 */ } static void lpc32xx_usbd_susp_chg(int susp) { /* Device suspend if susp != 0 */ } static void lpc32xx_rmwkup_chg(int remote_wakup_enable) { /* Enable or disable USB remote wakeup */ } struct lpc32xx_usbd_cfg lpc32xx_usbddata = { .vbus_drv_pol = 0, .conn_chgb = &lpc32xx_usbd_conn_chg, .susp_chgb = &lpc32xx_usbd_susp_chg, .rmwk_chgb = &lpc32xx_rmwkup_chg, }; static u64 lpc32xx_usbd_dmamask = ~(u32) 0x7F; static int lpc32xx_udc_probe(struct platform_device *pdev) { struct device *dev = &pdev->dev; struct lpc32xx_udc *udc; int retval, i; struct resource *res; dma_addr_t dma_handle; struct device_node *isp1301_node; udc = kmemdup(&controller_template, sizeof(*udc), GFP_KERNEL); if (!udc) return -ENOMEM; for (i = 0; i <= 15; i++) udc->ep[i].udc = udc; udc->gadget.ep0 = &udc->ep[0].ep; /* init software state */ udc->gadget.dev.parent = dev; udc->pdev = pdev; udc->dev = &pdev->dev; udc->enabled = 0; if (pdev->dev.of_node) { isp1301_node = of_parse_phandle(pdev->dev.of_node, "transceiver", 0); } else { isp1301_node = NULL; } udc->isp1301_i2c_client = isp1301_get_client(isp1301_node); of_node_put(isp1301_node); if (!udc->isp1301_i2c_client) { retval = -EPROBE_DEFER; goto phy_fail; } dev_info(udc->dev, "ISP1301 I2C device at address 0x%x\n", udc->isp1301_i2c_client->addr); pdev->dev.dma_mask = &lpc32xx_usbd_dmamask; retval = dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(32)); if (retval) goto resource_fail; udc->board = &lpc32xx_usbddata; /* * Resources are mapped as follows: * IORESOURCE_MEM, base address and size of USB space * IORESOURCE_IRQ, USB device low priority interrupt number * IORESOURCE_IRQ, USB device high priority interrupt number * IORESOURCE_IRQ, USB device interrupt number * IORESOURCE_IRQ, USB transceiver interrupt number */ res = platform_get_resource(pdev, IORESOURCE_MEM, 0); if (!res) { retval = -ENXIO; goto resource_fail; } spin_lock_init(&udc->lock); /* Get IRQs */ for (i = 0; i < 4; i++) { udc->udp_irq[i] = platform_get_irq(pdev, i); if (udc->udp_irq[i] < 0) { dev_err(udc->dev, "irq resource %d not available!\n", i); retval = udc->udp_irq[i]; goto irq_fail; } } udc->io_p_start = res->start; udc->io_p_size = resource_size(res); if (!request_mem_region(udc->io_p_start, udc->io_p_size, driver_name)) { dev_err(udc->dev, "someone's using UDC memory\n"); retval = -EBUSY; goto request_mem_region_fail; } udc->udp_baseaddr = ioremap(udc->io_p_start, udc->io_p_size); if (!udc->udp_baseaddr) { retval = -ENOMEM; dev_err(udc->dev, "IO map failure\n"); goto io_map_fail; } /* Get USB device clock */ udc->usb_slv_clk = clk_get(&pdev->dev, NULL); if (IS_ERR(udc->usb_slv_clk)) { dev_err(udc->dev, "failed to acquire USB device clock\n"); retval = PTR_ERR(udc->usb_slv_clk); goto usb_clk_get_fail; } /* Enable USB device clock */ retval = clk_prepare_enable(udc->usb_slv_clk); if (retval < 0) { dev_err(udc->dev, "failed to start USB device clock\n"); goto usb_clk_enable_fail; } /* Setup deferred workqueue data */ udc->poweron = udc->pullup = 0; INIT_WORK(&udc->pullup_job, pullup_work); INIT_WORK(&udc->vbus_job, vbus_work); #ifdef CONFIG_PM INIT_WORK(&udc->power_job, power_work); #endif /* All clocks are now on */ udc->clocked = 1; isp1301_udc_configure(udc); /* Allocate memory for the UDCA */ udc->udca_v_base = dma_alloc_coherent(&pdev->dev, UDCA_BUFF_SIZE, &dma_handle, (GFP_KERNEL | GFP_DMA)); if (!udc->udca_v_base) { dev_err(udc->dev, "error getting UDCA region\n"); retval = -ENOMEM; goto i2c_fail; } udc->udca_p_base = dma_handle; dev_dbg(udc->dev, "DMA buffer(0x%x bytes), P:0x%08x, V:0x%p\n", UDCA_BUFF_SIZE, udc->udca_p_base, udc->udca_v_base); /* Setup the DD DMA memory pool */ udc->dd_cache = dma_pool_create("udc_dd", udc->dev, sizeof(struct lpc32xx_usbd_dd_gad), sizeof(u32), 0); if (!udc->dd_cache) { dev_err(udc->dev, "error getting DD DMA region\n"); retval = -ENOMEM; goto dma_alloc_fail; } /* Clear USB peripheral and initialize gadget endpoints */ udc_disable(udc); udc_reinit(udc); /* Request IRQs - low and high priority USB device IRQs are routed to * the same handler, while the DMA interrupt is routed elsewhere */ retval = request_irq(udc->udp_irq[IRQ_USB_LP], lpc32xx_usb_lp_irq, 0, "udc_lp", udc); if (retval < 0) { dev_err(udc->dev, "LP request irq %d failed\n", udc->udp_irq[IRQ_USB_LP]); goto irq_lp_fail; } retval = request_irq(udc->udp_irq[IRQ_USB_HP], lpc32xx_usb_hp_irq, 0, "udc_hp", udc); if (retval < 0) { dev_err(udc->dev, "HP request irq %d failed\n", udc->udp_irq[IRQ_USB_HP]); goto irq_hp_fail; } retval = request_irq(udc->udp_irq[IRQ_USB_DEVDMA], lpc32xx_usb_devdma_irq, 0, "udc_dma", udc); if (retval < 0) { dev_err(udc->dev, "DEV request irq %d failed\n", udc->udp_irq[IRQ_USB_DEVDMA]); goto irq_dev_fail; } /* The transceiver interrupt is used for VBUS detection and will kick off the VBUS handler function */ retval = request_irq(udc->udp_irq[IRQ_USB_ATX], lpc32xx_usb_vbus_irq, 0, "udc_otg", udc); if (retval < 0) { dev_err(udc->dev, "VBUS request irq %d failed\n", udc->udp_irq[IRQ_USB_ATX]); goto irq_xcvr_fail; } /* Initialize wait queue */ init_waitqueue_head(&udc->ep_disable_wait_queue); atomic_set(&udc->enabled_ep_cnt, 0); /* Keep all IRQs disabled until GadgetFS starts up */ for (i = IRQ_USB_LP; i <= IRQ_USB_ATX; i++) disable_irq(udc->udp_irq[i]); retval = usb_add_gadget_udc(dev, &udc->gadget); if (retval < 0) goto add_gadget_fail; dev_set_drvdata(dev, udc); device_init_wakeup(dev, 1); create_debug_file(udc); /* Disable clocks for now */ udc_clk_set(udc, 0); dev_info(udc->dev, "%s version %s\n", driver_name, DRIVER_VERSION); return 0; add_gadget_fail: free_irq(udc->udp_irq[IRQ_USB_ATX], udc); irq_xcvr_fail: free_irq(udc->udp_irq[IRQ_USB_DEVDMA], udc); irq_dev_fail: free_irq(udc->udp_irq[IRQ_USB_HP], udc); irq_hp_fail: free_irq(udc->udp_irq[IRQ_USB_LP], udc); irq_lp_fail: dma_pool_destroy(udc->dd_cache); dma_alloc_fail: dma_free_coherent(&pdev->dev, UDCA_BUFF_SIZE, udc->udca_v_base, udc->udca_p_base); i2c_fail: clk_disable_unprepare(udc->usb_slv_clk); usb_clk_enable_fail: clk_put(udc->usb_slv_clk); usb_clk_get_fail: iounmap(udc->udp_baseaddr); io_map_fail: release_mem_region(udc->io_p_start, udc->io_p_size); dev_err(udc->dev, "%s probe failed, %d\n", driver_name, retval); request_mem_region_fail: irq_fail: resource_fail: phy_fail: kfree(udc); return retval; } static int lpc32xx_udc_remove(struct platform_device *pdev) { struct lpc32xx_udc *udc = platform_get_drvdata(pdev); usb_del_gadget_udc(&udc->gadget); if (udc->driver) return -EBUSY; udc_clk_set(udc, 1); udc_disable(udc); pullup(udc, 0); free_irq(udc->udp_irq[IRQ_USB_ATX], udc); device_init_wakeup(&pdev->dev, 0); remove_debug_file(udc); dma_pool_destroy(udc->dd_cache); dma_free_coherent(&pdev->dev, UDCA_BUFF_SIZE, udc->udca_v_base, udc->udca_p_base); free_irq(udc->udp_irq[IRQ_USB_DEVDMA], udc); free_irq(udc->udp_irq[IRQ_USB_HP], udc); free_irq(udc->udp_irq[IRQ_USB_LP], udc); clk_disable_unprepare(udc->usb_slv_clk); clk_put(udc->usb_slv_clk); iounmap(udc->udp_baseaddr); release_mem_region(udc->io_p_start, udc->io_p_size); kfree(udc); return 0; } #ifdef CONFIG_PM static int lpc32xx_udc_suspend(struct platform_device *pdev, pm_message_t mesg) { struct lpc32xx_udc *udc = platform_get_drvdata(pdev); if (udc->clocked) { /* Power down ISP */ udc->poweron = 0; isp1301_set_powerstate(udc, 0); /* Disable clocking */ udc_clk_set(udc, 0); /* Keep clock flag on, so we know to re-enable clocks on resume */ udc->clocked = 1; /* Kill global USB clock */ clk_disable_unprepare(udc->usb_slv_clk); } return 0; } static int lpc32xx_udc_resume(struct platform_device *pdev) { struct lpc32xx_udc *udc = platform_get_drvdata(pdev); if (udc->clocked) { /* Enable global USB clock */ clk_prepare_enable(udc->usb_slv_clk); /* Enable clocking */ udc_clk_set(udc, 1); /* ISP back to normal power mode */ udc->poweron = 1; isp1301_set_powerstate(udc, 1); } return 0; } #else #define lpc32xx_udc_suspend NULL #define lpc32xx_udc_resume NULL #endif #ifdef CONFIG_OF static const struct of_device_id lpc32xx_udc_of_match[] = { { .compatible = "nxp,lpc3220-udc", }, { }, }; MODULE_DEVICE_TABLE(of, lpc32xx_udc_of_match); #endif static struct platform_driver lpc32xx_udc_driver = { .remove = lpc32xx_udc_remove, .shutdown = lpc32xx_udc_shutdown, .suspend = lpc32xx_udc_suspend, .resume = lpc32xx_udc_resume, .driver = { .name = (char *) driver_name, .of_match_table = of_match_ptr(lpc32xx_udc_of_match), }, }; module_platform_driver_probe(lpc32xx_udc_driver, lpc32xx_udc_probe); MODULE_DESCRIPTION("LPC32XX udc driver"); MODULE_AUTHOR("Kevin Wells "); MODULE_AUTHOR("Roland Stigge "); MODULE_LICENSE("GPL"); MODULE_ALIAS("platform:lpc32xx_udc");